[sdiy] harmonics & vibrato - drums

Chris Stecker cstecker at umich.edu
Fri Dec 20 21:02:04 CET 2002


Grant (and interested list),

Although "masking" is not the term I would necessarily use (it has a more 
specific, but related, meaning), there are many examples of adaptation in the 
nervous system, including (as Mr. Gravenhorst noted) fatigue of sensory 
receptors, and also lateral inhibition (stimulated regions of the retina 
suppress activity in neighboring cells, producing "Mach bands" or bright 
lines next to black/white boundaries). There are higher-order adaptive 
phenomena at many later stages of neural processing, that adapt to more 
complex or abstract aspects of stimuli. Some of these produce observable 
effects like afterimages, and some are more subtle or immediate. In general, 
adaptation helps to reduce the response to steady-state (continuous) 
stimulation, so that dynamic or transient stimuli are easier to detect. It's 
a bit like high-pass filtering or taking the derivative of any parameter 
you're interested in (across time, frequency, space, or whatever); changes 
become more noticeable, and the system's "dynamic range" is increased.

Lateral inhibition in the auditory system occurs between neighboring 
frequencies (as evidenced by "two-tone suppression" as well as various 
simultaneous masking phenomena), and also over time, as you suggested. 
Forward-masking (the reduced detectability of sounds following exposure to a 
masking--generally more intense--sound) is one example, and so is the 
precedence or Haas effect (reduced ability to localize or spatially segregate 
sounds following onsets). These phenomena and others are characterized by a 
limited duration of suppression (usually they have some measurable "recovery 
function").  [My specific research interest in this area is adaptation to the 
acoustics of rooms.]

What you're suggesting for drumming, however, is difficult for me to 
understand, primarily because we (I, especially) don't yet understand how 
rhythm, timing, and repetition are represented or processed in the brain. 
This theory would require a representation of intervals, whereby "units" 
representing particular intervals would fatique if repetitively stimulated. 
Jittering the intervals would spread activation around a small group of such 
units, causing less fatigue. It's not a totally unreasonable idea, and in 
fact at much shorter time scales (intervals around 5-10 ms) randomly varying 
the intervals (by a very large "jitter") reduces the strength of the 
precedence effect...but that's really a totally different time scale, a 
different perceptual dimension (space), and presumably a totally different 
type of mechanism (there is no precedence effect for intervals beyond 10 ms 
[clicks] to 50-100 ms [speech or music]). I don't know of any data suggesting 
that the perceptibility of sound is actually diminished by highly regular 
repetition, but it's possible that part of the "appeal" of less-quantized 
rhythm is in the "spreading around" of stimulation to more than one 
exact-interval processing "unit."

Here are two other possibilities I have in mind:

a) slight randomization sounds better because it sounds "human," and it sounds 
human because that's what we've been exposed to, listening to human drummers. 
(not that children raised by robots would think hard-quantization sounds 
"human," but...) Or maybe its a bit like the difference between 
photo-realistic painting and more expressive styles. (Or is it more like the 
difference between photorealistic painting and photography?)

b) slight randomization reduces the predictability of the beat. we like 
surprises, (because we pay attention to the derivatives?) so totally 
predictable rhythms sound "flat."  This is really just another way to frame 
your "crackpot" adaptation theory, but it does not depend on a low-level 
process that suppresses sound perception, per se. Rather, it hypothesizes the 
existence of fairly high-level perceptual adaptative mechanisms that track 
regularity in an analytic-synthetic, predictive, or change-oriented manner.

My money's on "a," as it seems simpler and in keeping with variability between 
listeners; some prefer "tight" (versus "sloppy") drummers, while others 
prefer "expressive" (versus "flat" or "robotic") ones. "b" appeals more to my 
views of neural computation, however, and is more aesthetically 
appealing/interesting, but that doesn't make it right (it's probably more 
testable though). 

-Chris


On Friday 20 December 2002 01:43 pm, Grant Richter wrote:
> > Ok, this makes more sense.  But the question remains, is there a random
> > factor or not?  If not, then why does doing a *slight* randomization
> > (weighted in my case) sound better (um more "human") than a quantized
> > track?
>
> VERY Crackpot theory,
>
> Neural systems have a masking factor. This prevents a stuck input from
> tying up the whole net. Like when you stare at something and it leaves a
> negative image in your eye. The negative image is the mask generated by the
> neural system.
>
> That may apply to time related items also, like when you are in a noisy
> room and your ears adapt by "tuning" out the noise so you can hear better.
>
> It may be that drummers instinctively play around the outside of the neural
> "mask" time. If they hit perfectly on the beat every time, they eventually
> couldn't hear themselves.



More information about the Synth-diy mailing list