Professional Documents
Culture Documents
Chapter 6
Basics of Digital Audio
Audio information is crucial for multimedia presentations. It is one of the simplest types of
multimedia data. The element of sound can be important to achieve the goals of a multimedia
product. Sound adds another dimension. If used well, sound is an extremely powerful element
which can stimulate emotional responses that would never be activated from text and graphics
alone. Sound helps convey the intended message or complement the purpose of multimedia
applications. To help ensure effectiveness, it is important to incorporate sound files that are of
high quality, appropriate, and consistent with the goals of the multimedia product.
What is Sound?
Sound is a continuous wave that travels through the air.
Without air there is no sound. The wave is made up of pressure
differences. Sound is detected by measuring the pressure level
at a location. Sound waves have normal wave properties
(reflection, refraction, diffraction, etc.). When we speak,
vibrations, called sound waves, are created. Sound waves have
a recurring pattern or an analog wave pattern called a waveform. Humans can hear frequencies
in the range 20 - 20,000 Hz.
Amplitude (loudness/intensity)
Frequency (pitch)
Envelope (waveform)
E Amplitude: Distance between the valley and the peak of a waveform; determines volume.
Volume is measured in decibels (dB) Decibel (dB) is a logarithmic unit used to describe a
ratio. One dB is close to Just Noticeable Difference (JND) for sound level.
E Frequency: Number of peaks that occur in one second measured by the distance between
the peaks; it determines pitch.
1|Page
Chapter 6 [Department of IT]
Why to Digitize?
Sound is required input into a computer. Therefore it needs to sampled or digitized. For
example microphones, video cameras produce analog signals (continuous-valued voltages) as
illustrated in Fig below:
Digitization
To get audio or video into a computer, we have to digitize it (convert it into a stream of
numbers).We need to special hardware Analog-to-Digital Converter for the purpose. Digitization
process involves discrete sampling and quantization methods.
Quantization: The process of converting continuous sample values into discrete values. It
divides the vertical axis (signal strength) into pieces. Sometimes, a non-linear function is applied.
8-bit quantization divides the vertical axis into 256 levels. 16-bit gives you 65536 levels. Size of
quantization interval is called quantization step. The number of quantization levels is related to
the quality of digital audio. The higher the quantization, the better the resulting sound quality. If
more quantization levels are allowed, the difference between the original value and the quantified
2|Page
Chapter 6 [Department of IT]
value will be smaller and we will get a better quality of the digital representation. However, this
would also mean a higher cost for storage and processing of these values inside a computer
(disks of larger capacity and more powerful CPUs are required).
Amplitude Amplitude
Coding: The process of representing quantized values digitally. In fig below, Sound sampling converts
analog sound to digital audio is shown below. Here 4-bit quantization used.
4|Page
Chapter 6 [Department of IT]
We must pass input through a Low Pass Filter before sampling. Otherwise strange
artifacts from high frequency signals appear and are audible.
5|Page
Chapter 6 [Department of IT]
Stereo sounds: much more dynamic and lifelike Music almost always must be recorded and
saved in stereo.
Digital Audio File Size
Audio Compression
Similar to image compression, mathematical algorithms are used to reduce audio file sizes.
File size is the primary consideration when using audio files on the Web. Compression is
beneficial for storing and transferring audio files. Different compression schemes are available
for different file formats, and some file formats include the compression scheme within the
format itself. The compression scheme applied impacts both the quality of audio file and the file
size. A simple and widely used audio compression method is Adaptive Delta Pulse Code
Modulation (ADPCM).
6|Page
Chapter 6 [Department of IT]
computer receives streaming audio data more quickly than required, the excess data will be
stored in a buffer. If the users computer receives streaming audio data slower than required, the
data stored in the buffer will be used. If the buffer becomes empty, the user will experience a
break.
Two types of streamed audio:
On demand: Streamed audio is stored on a server for a long period of time, and is
available for transmission at a users request.
Live: Live streams are only available at one particular time.
Example: live radio program
Multiple tracks editing: Edit and combine multiple tracks and then merge the tracks and
export them in a final mix to a single audio file.
Trimming: Remove dead air or blank space from the front of a recording and any
unnecessary extra time off the end.
7|Page
Chapter 6 [Department of IT]
Splicing and assembly: Remove the extraneous noises that inevitably creep into a recording.
Fade-ins and fade-outs: Smooth out the very beginning and the very end of a sound file
8|Page