Jump to content

Sound synthesis: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Line 8: Line 8:


== Introduction ==
== Introduction ==
When any mechanical collision occurs sound is produced. The energy from the collision is transferred through the [[air]] as sound waves, which are perceived by the human auditory system. Sound waves are the aggregate of one or many periodic vibrations, described mathematically by [[sine waves]]. The characteristics of a sound, known as [[pitch]] and [[timbre]], are defined by the [[amplitude]] and [[pitch]] of each individual sine wave that makes up what is perceived as sound.
When any mechanical collision occurs sound is produced. The energy from the collision is transferred through the [[air]] as sound waves, which are perceived by the human auditory system. Sound waves are the aggregate of one or many periodic vibrations, described mathematically by [[sine waves]]. The characteristics of a sound, known as [[pitch]] and [[timbre]], are defined by the [[amplitude]] and [[pitch]] of each individual sine wave, collectively know as the partials or [[harmonics]]. Generally, a sound that does not change over time will include a [[fundamental frequency|fundamental partial]] or harmonic, and any number of partials.

Generally, a single "sound" will include a [[fundamental frequency]], and any number of [[overtone]]s. The frequencies of these overtones are either integer multiples of the fundamental frequency, or integer fractions thereof ([[subharmonics]]). This study of how complex waveforms can be alternately represented is covered in [[Laplace Transform|Laplace]] and [[Fourier transform|Fourier]] transforms.


When natural tonal instruments' sounds are analyzed in the [[frequency domain]] (as on a spectrum analyzer), the [[frequency spectrum|spectra]] of their sounds will exhibit [[amplitude]] spikes at each of the fundamental tone's [[harmonic series (music)|harmonics]]. Some harmonics may have higher amplitudes than others. The specific set of harmonic-vs-amplitude pairs is known as a sound's [[harmonic content]].
When natural tonal instruments' sounds are analyzed in the [[frequency domain]] (as on a spectrum analyzer), the [[frequency spectrum|spectra]] of their sounds will exhibit [[amplitude]] spikes at each of the fundamental tone's [[harmonic series (music)|harmonics]]. Some harmonics may have higher amplitudes than others. The specific set of harmonic-vs-amplitude pairs is known as a sound's [[harmonic content]].

Revision as of 11:51, 26 November 2008

In music technology, sound synthesis is the process of generating sound from analogue and digital electronic equipment, often for musical, artistic or entertainment purposes. In particular, it refers to the process of generating, combining or mixing sounds from a set of fundamental building blocks or routines in order to create sounds of a greater complexity and richness. Sound synthesis can be used to mimic acoustic sound sources or generate sound that may be impossible to realise naturally. Since its development in the first half of the 20th Century, it has provided applications to music, computer science, film, acoustics and even biology.

Introduction

When any mechanical collision occurs sound is produced. The energy from the collision is transferred through the air as sound waves, which are perceived by the human auditory system. Sound waves are the aggregate of one or many periodic vibrations, described mathematically by sine waves. The characteristics of a sound, known as pitch and timbre, are defined by the amplitude and pitch of each individual sine wave, collectively know as the partials or harmonics. Generally, a sound that does not change over time will include a fundamental partial or harmonic, and any number of partials.

When natural tonal instruments' sounds are analyzed in the frequency domain (as on a spectrum analyzer), the spectra of their sounds will exhibit amplitude spikes at each of the fundamental tone's harmonics. Some harmonics may have higher amplitudes than others. The specific set of harmonic-vs-amplitude pairs is known as a sound's harmonic content.

When analyzed in the time domain, a sound does not necessarily have the same harmonic content throughout the duration of the sound. Typically, high-frequency harmonics will die out more quickly than the lower harmonics. For a synthesized sound to "sound" right, it requires accurate reproduction of the original sound in both the frequency domain and the time domain.

Percussion instruments and rasps have very low harmonic content, and exhibit spectra that are comprised mainly of noise shaped by the resonant frequencies of the structures that produce the sounds. However, the resonant properties of the instruments (the spectral peaks of which are also referred to as formants) also shape an instrument's spectrum (esp. in string, wind, voice and other natural instruments).

In most conventional synthesizers, for purposes of re-synthesis, recordings of real instruments are composed of several components.

These component sounds represent the acoustic responses of different parts of the instrument, the sounds produced by the instrument during different parts of a performance, or the behavior of the instrument under different playing conditions (pitch, intensity of playing, fingering, etc.) The distinctive timbre, intonation and attack of a real instrument can therefore be created by mixing together these components in such a way as resembles the natural behavior of the real instrument. Nomenclature varies by synthesizer methodology and manufacturer, but the components are often referred to as oscillators or partials. A higher fidelity reproduction of a natural instrument can typically be achieved using more oscillators, but increased computational power and human programming is required, and most synthesizers use between one and four oscillators by default.

Schematic of ADSR

One of the most important parts of any sound is its amplitude envelope. This envelope determines whether the sound is percussive, like a snare drum, or persistent, like a violin string. Most often, this shaping of the sound's amplitude profile is realized with an "ADSR" (Attack Decay Sustain Release) envelope model applied to control oscillator volumes. Apart from Sustain, each of these stages is modeled by a change in volume (typically exponential).

Although the oscillations in real instruments also change frequency, most instruments can be modeled well without this refinement. This refinement is necessary to generate a vibrato.

Attack time

is the time taken for initial run-up of the sound level from nil (or blank sustainment) to 100% or what ever designated percentage.

Decay time

is the time taken for the subsequent run down from the attack level to the designated Sustain level.

Sustain level

is the steady volume produced when a key is held down.

Release time

is the time taken for the sound to decay from the Sustain level to nil when the key is released. If a key is released during the Attack or Decay stage, the Sustain phase is usually skipped. Similarly, a Sustain level of zero will produce a more-or-less piano-like (or percussive) envelope, with no continuous steady level, even when a key is held. Exponential rates are commonly used because they closely model real physical vibrations, which usually rise or decay exponentially.

Overview of popular synthesis methods

Subtractive synthesizers use a simple acoustic model that assumes an instrument can be approximated by a simple signal generator (producing sawtooth waves, square waves, etc...) followed by a filter which represents the frequency-dependent losses and resonances in the instrument body. For reasons of simplicity and economy, these filters are typically low-order lowpass filters. The combination of simple modulation routings (such as pulse width modulation and oscillator sync), along with the physically unrealistic lowpass filters, is responsible for the "classic synthesizer" sound commonly associated with "analog synthesis" and often mistakenly used when referring to software synthesizers using subtractive synthesis. Although physical modeling synthesis, synthesis wherein the sound is generated according to the physics of the instrument, has superseded subtractive synthesis for accurately reproducing natural instrument timbres, the subtractive synthesis paradigm is still ubiquitous in synthesizers with most modern designs still offering low-order lowpass or bandpass filters following the oscillator stage.

One of the newest systems to evolve inside music synthesis is physical modeling. This involves taking up models of components of musical objects and creating systems which define action, filters, envelopes and other parameters over time. The definition of such instruments is virtually limitless, as one can combine any given models available with any amount of sources of modulation in terms of pitch, frequency and contour. For example, the model of a violin with characteristics of a pedal steel guitar and perhaps the action of piano hammer ... physical modeling on computers gets better and faster with higher processing.

One of the easiest synthesis systems is to record a real instrument as a digitized waveform, and then play back its recordings at different speeds to produce different tones. This is the technique used in "sampling". Most samplers designate a part of the sample for each component of the ADSR envelope, and then repeat that section while changing the volume for that segment of the envelope. This lets the sampler have a persuasively different envelope using the same note..

Synthesizer basics

There are three major kinds of synthesizers, analog, digital and software. In addition there are synthesizers that rely upon combinations of those three kinds, known as hybrid synthesizers.

There are also many different kinds of synthesis methods, each applicable to both analog and digital synthesizers. These techniques tend to be mathematically related, especially frequency modulation and phase modulation.