History of radioThe early history of radio is the history of technology that produces and uses radio instruments that use radio waves. Within the timeline of radio, many people contributed theory and inventions in what became radio. Radio development began as "wireless telegraphy". Later radio history increasingly involves matters of broadcasting. Invention of radio In an 1864 presentation, published in 1865, James Clerk Maxwell proposed theories of electromagnetism, with mathematical proofs, that showed that light and predicted that radio and x-rays were all types of electromagnetic waves propagating through free space.
Speech processingSpeech processing is the study of speech signals and the processing methods of signals. The signals are usually processed in a digital representation, so speech processing can be regarded as a special case of digital signal processing, applied to speech signals. Aspects of speech processing includes the acquisition, manipulation, storage, transfer and output of speech signals. Different speech processing tasks include speech recognition, speech synthesis, speaker diarization, speech enhancement, speaker recognition, etc.
Video coding formatA video coding format (or sometimes video compression format) is a content representation format for storage or transmission of digital video content (such as in a data file or bitstream). It typically uses a standardized video compression algorithm, most commonly based on discrete cosine transform (DCT) coding and motion compensation. A specific software, firmware, or hardware implementation capable of compression or decompression to/from a specific video coding format is called a video codec.
Pulse-code modulationPulse-code modulation (PCM) is a method used to digitally represent sampled analog signals. It is the standard form of digital audio in computers, compact discs, digital telephony and other digital audio applications. In a PCM stream, the amplitude of the analog signal is sampled at uniform intervals, and each sample is quantized to the nearest value within a range of digital steps. Linear pulse-code modulation (LPCM) is a specific type of PCM in which the quantization levels are linearly uniform.
Speech synthesisSpeech synthesis is the artificial production of human speech. A computer system used for this purpose is called a speech synthesizer, and can be implemented in software or hardware products. A text-to-speech (TTS) system converts normal language text into speech; other systems render symbolic linguistic representations like phonetic transcriptions into speech. The reverse process is speech recognition. Synthesized speech can be created by concatenating pieces of recorded speech that are stored in a database.
Broadcast relay stationA broadcast relay station, also known as a satellite station, relay transmitter, broadcast translator (U.S.), re-broadcaster (Canada), repeater (two-way radio) or complementary station (Mexico), is a broadcast transmitter which repeats (or transponds) the signal of a radio or television station to an area not covered by the originating station. It expands the broadcast range of a television or radio station beyond the primary signal's original coverage or improves service in the original coverage area.
Sub-band codingIn signal processing, sub-band coding (SBC) is any form of transform coding that breaks a signal into a number of different frequency bands, typically by using a fast Fourier transform, and encodes each one independently. This decomposition is often the first step in data compression for audio and video signals. SBC is the core technique used in many popular lossy audio compression algorithms including MP3. The simplest way to digitally encode audio signals is pulse-code modulation (PCM), which is used on audio CDs, DAT recordings, and so on.
Latency (audio)Latency refers to a short period of delay (usually measured in milliseconds) between when an audio signal enters a system and when it emerges. Potential contributors to latency in an audio system include analog-to-digital conversion, buffering, digital signal processing, transmission time, digital-to-analog conversion and the speed of sound in the transmission medium. Latency can be a critical performance metric in professional audio including sound reinforcement systems, foldback systems (especially those using in-ear monitors) live radio and television.
HD RadioHD Radio (HDR) is a trademark for an in-band on-channel (IBOC) digital radio broadcast technology. HD radio generally simulcasts an existing analog radio station in digital format with less noise and with additional text information. HD Radio is used primarily by AM and FM radio stations in the United States, U.S. Virgin Islands, Canada, Mexico and the Philippines, with a few implementations outside North America. The term "on channel" is a misnomer because the system actually broadcasts on the ordinarily unused channels adjacent to an existing radio station's allocation.
Frequency-hopping spread spectrumFrequency-hopping spread spectrum (FHSS) is a method of transmitting radio signals by rapidly changing the carrier frequency among many frequencies occupying a large spectral band. The changes are controlled by a code known to both transmitter and receiver. FHSS is used to avoid interference, to prevent eavesdropping, and to enable code-division multiple access (CDMA) communications. The frequency band is divided into smaller sub-bands. Signals rapidly change ("hop") their carrier frequencies among the center frequencies of these sub-bands in a determined order.