Duration of sound in our brain (0.1s)

  • Context: Undergrad 
  • Thread starter Thread starter fog37
  • Start date Start date
  • Tags Tags
    Brain Sound
Click For Summary
SUMMARY

The duration of sound retention in the human brain is approximately 0.1 seconds, which influences how we perceive echoes and reverberations. Echoes occur when sound travels a distance of at least 17 meters, while shorter distances result in reverberation. The brain processes sounds not linearly like a digital computer but through complex pattern recognition, allowing for the distinction between echoes and overlapping sounds. This processing is akin to how analogue TV signals are affected by multipath conditions, where reflections can distort sound perception.

PREREQUISITES
  • Understanding of auditory perception and sound processing
  • Familiarity with the concepts of echo and reverberation
  • Knowledge of pattern recognition in cognitive science
  • Basic principles of signal processing and its comparison to digital systems
NEXT STEPS
  • Research the mechanisms of auditory perception in cognitive neuroscience
  • Explore the differences between echo and reverberation in acoustics
  • Study the principles of pattern recognition in human cognition
  • Investigate the effects of multipath conditions on sound and signal processing
USEFUL FOR

Cognitive scientists, audio engineers, psychologists, and anyone interested in the intricacies of sound perception and processing in the human brain.

fog37
Messages
1,566
Reaction score
108
Hello,
a certain sound remains in our brain for 1/10 of a second. The phenomenon of echo can happen only if the distance is at least 17 meters: we say the letter A, which remains in our brain for 1/10s, and the reflection (echo) of A arrives to our ear after that 0.1s...If the distance is shorter than 17 meters we get reverberation.

Aside from echo and reverberation, when we speak a whole word, like book, we clearly hear each single letter (b-o-o-k). Why doesn't the 0.1s memory of each letter play a factor? that would make each letter overlap in our brain...
 
Physics news on Phys.org
A brain is not like a digital computer which manipulates data according to a linear set of instructions which execute at 'n' amount of CPU ticks per second.
It is a system which is good at recognizing patterns, and the more a pattern is re-inforced the faster a brain responds to it.
 
Last edited:
This is a very complicated subject. Our hearing does both frequency and time analysis on the sounds we receive. When a sound arrives at our ear via two different paths, There are two possible effects. A long relative delay can be distinguished as two separate sounds (an echo) and a short relative delay will result in a distinguishable change in the frequency spectrum of the sound which you can't resolve as a time difference. There is a good analogy with reception of analogue TV signals under 'multipath conditions'. Reflected signals from a distant building will produce identifiable low level delayed images, superimposed on the main picture which we actually call echos. (Football players running across a green field will each have faint a shadow, following them). Reflections from a nearby obstruction (short relative delay) are invisible (not resolved) but can cause the sound or colour information to be distorted because they sit in a part of the TV signal spectrum that's different from the coarse detail (main shapes in the picture).
Your brain evolved to get the most information out of the sounds it receives for the 'cost' of running your hearing system. Your 'memory' of a situation is formed by a complicated signal processing system (different from the DSP that we can do with computers- see rootone's comments). Was your '1/10s' figure just a rough idea?
Your example of the way the word 'book' is recognised is a bit too simplistic because that word is very familiar and you associate that sound pattern with all your other memories of books. The response to it is very different from hearing the (made up) word 'foons'. People say that 'we hear what we want to hear', and I agree. The total input of information into our brains is far too high to deal with so we have to condense it into understandable and familiar concepts. That process starts very early on in our processing systems. (The same thing happens with vision, with a lot of processing taking part on the actual surface of the retina.) No one 'designed' these systems.
 

Similar threads

  • · Replies 10 ·
Replies
10
Views
13K
  • · Replies 6 ·
Replies
6
Views
5K
Replies
2
Views
3K
  • · Replies 30 ·
2
Replies
30
Views
7K
  • · Replies 58 ·
2
Replies
58
Views
7K
  • · Replies 15 ·
Replies
15
Views
6K
  • · Replies 2 ·
Replies
2
Views
4K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 62 ·
3
Replies
62
Views
11K