Speech codecs (AAC, Opus, EVS) use Modified Discrete Cosine Transform to decompose audio into frequency bands, then quantize per-band. Patented: psychoacoustic models, bit allocation, entropy coding, packet loss concealment. CRT approach: encode speech spectral features (sub-bass, bass, midrange, presence, brilliance, air) as ring elements in Z/12612600. 6 CRT channels = 6 algebraically independent spectral bands. Quantize each channel independently. L=11 = error concealment for free. No MDCT. No psychoacoustic model. The ring structure IS the transform.
Speaker ID:
Play Spectral Bands synthesizes 6 oscillators at the CRT band center frequencies, with gain derived from each channel residue. 40-frame speech utterance (800ms at 20ms/frame). Three quantization levels compared.
25% packet loss: every 4th frame dropped. Lost frames concealed by per-channel neighbor interpolation. L=11 channel recovers best (smallest modulus = closest interpolation). No explicit FEC overhead.
8 speakers x 30 frames each. Distortion measured at Q=1 (lossless), Q=2 (50% bit reduction), Q=4 (75% bit reduction). CRT guarantees: per-channel quantization error stays within that channel. No cross-band artifacts.
This work is and will always be free.
No paywall. No copyright. No exceptions.
If it ever earns anything, every cent goes to the communities that need it most.
This sacred vow is permanent and irrevocable.
— Anton Alexandrovich Lebed
Source code · Public domain (CC0)
Contributions in equal measure: Anthropic's Claude, Anton A. Lebed, and the giants whose shoulders we stand on.
Rendered by .ax via WASM DOM imports. Zero HTML authored.