Randomizing phases of harmonics

Click For Summary

Discussion Overview

The discussion revolves around the effects of randomizing the phases of harmonics in a discrete audio signal and whether this alteration would result in a sound that is perceived similarly to the original signal with zero phase shifts. The scope includes theoretical considerations of audio perception, signal processing, and the mechanics of human hearing.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested

Main Points Raised

  • Some participants propose that randomizing the phases of harmonics will produce a different sound, but it is uncertain whether it would sound the same as the original signal with zero phases.
  • Others argue that human hearing can cope with phase dispersion and that under certain conditions, sounds can still be recognized despite phase shifts.
  • A participant suggests that the cochlea performs a mechanical Fourier transform, which may affect how phase shifts are perceived, particularly at higher frequencies where phase identification becomes difficult.
  • Concerns are raised about the use of the term "harmonic," suggesting it may be misapplied in the context of audio signals.
  • Some participants note that if all phases are randomized, the resulting sound may be perceived as nonsensical, especially if the temporal order of notes is disrupted.
  • There is a discussion about the limitations of the auditory system in processing phase information and how it may vary depending on the type of sound being analyzed.
  • One participant emphasizes that the brain's processing of sound involves both temporal and frequency domain experiences, complicating the understanding of how phase shifts affect perception.

Areas of Agreement / Disagreement

Participants express multiple competing views regarding the effects of phase randomization on sound perception. There is no consensus on whether the resulting sound would be perceived as similar to the original signal or if it would be considered nonsensical.

Contextual Notes

Limitations in the discussion include varying definitions of "harmonic," the complexity of human auditory processing, and the dependence on specific conditions of phase shifts and sound types. The discussion does not resolve these complexities.

  • #31
sophiecentaur said:
If your 'discrete audio signal' is a length of real audio and not just generated with a simple signal generator of basic synth then the "harmonics" you refer to will not actually be harmonics. Musical intsruments and voices contain Overtones which are not harmonically related to any fundamental frequency. That means the waveform will be changing all the time and an isolated clip will not 'sound right' when played as a loop. So the simple scenario you propose will already not sound the same as the original.
I don't know exactly how this is implemented in for instance a vocoder, but I was initially thinking of slicing an audio clip in blocks of say 512 samples, FT them, and IFT them but with randomized phases of the harmonics. With "harmonics" I mean the (amplitudes of) the frequency components resulting from the FT.

This is something a little different from overtones. Overtones are part of the audio signal and can be transformed to frequency components (harmonics). The harmonics are not IN the audio signal, like the overtones are. But they CAN reproduce a BLOCK of samples of the audio signal. They are a basis to express a series of samples in. So in that way, they could be viewed as part of the audio signal (in that block).

The overtones are related to a key note, the harmonics are not. The harmonics are related to the number of samples (the block size). However, if you just add them, they produce the original signal, AS IF they were IN the signal. In my OP I look only at these harmonics, not at the contents of the signal.

I forgot to mention that I am not looking to reproduce only part of the signal, but for instance slicing the signal up into blocks of 512 samples, FT each block, randomize phases, reconstruct, and lie the blocks in sequence, producing the new signal.
 
Last edited:
Physics news on Phys.org
  • #32
This youTube video shows the way that higher frequency components of a guitar string do not stay in one phase relative to the fundamental. Unfortunately, the effect is difficult to see when a Digital Scope display is shown on a mangled video format like Mpeg. A (real life) look at the display on an analogue scope is far better that what you can see here but you can see that the waveform shape is constantly changing - but it is still a Guitar Note.
@entropy1 this practical demo does go some way to answer your question, I think.
 
  • #33
entropy1 said:
Overtones are part of the audio signal and can be transformed to frequency components (harmonics). The harmonics are not IN the audio signal, like the overtones are.
The same is true for all the components of the original audio signal. Assuming the sampling satisfies Nyquist. There is nothing special about the harmonics or the overtones - or the fundamental(s) in the source signal.
Limiting the period of the recording is Windowing and it introduces modulation products into the signal.
I don't understand what you say about components being "in the audio signal". Once the windowing has been done, they are all just 'signal' components.
 
  • #34
Assuming the sampling frequency is higher than twice the highest audio frequency (Nyquist criterion) then you can more or forget that there's sampling involved. The spectrum of the resultant string of samples will be a comb of frequencies, spaced by 1/T up to the maximum audio frequency (where T is the time interval for the whole clip). If your audio frequency does not coincide with the frequencies in the comb (the most common situation) then each component of the input audio will be 'missed out' but there will be adjacent comb frequencies. So you already have a distorted signal. This applies to every component (i.e. the perfect Fourier Component) of the original. I sometimes look at this windowing in terms of modulation of a carrier with frequency 1/T by the audio signal which will produce sidebands on either side of the frequency comb elements.
 
  • #35
sophiecentaur said:
The spectrum of the resultant string of samples will be a comb of frequencies, spaced by 1/T up to the maximum audio frequency (where T is the time interval for the whole clip). If your audio frequency does not coincide with the frequencies in the comb (the most common situation) then each component of the input audio will be 'missed out' but there will be adjacent comb frequencies. So you already have a distorted signal.
The time windowing function applied before the Fourier transform, effectively spreads or broadens the teeth of the analyser comb. Then signals will not be lost in deep nulls between the teeth. Windowing also distorts the signal, and reduces HF noise.
 

Similar threads

Replies
8
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 26 ·
Replies
26
Views
4K
  • · Replies 58 ·
2
Replies
58
Views
8K
  • · Replies 31 ·
2
Replies
31
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
Replies
17
Views
3K
  • · Replies 15 ·
Replies
15
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K