1. This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
    By continuing to use this site, you are consenting to our use of cookies.

    Dismiss Notice

DIY Stereo Smyth Realiser

  1. dude_500
    A quick divergence from DIY electrostatic projects... working on making a DIY stereo smyth realiser (couldn't care less about any more than 2 channels, only listen to music - can get by with a lot less DSP than the 8-channel realiser needs).
     
    IMG_20140515_173642.jpg
     
    Currently my hardware is a Digilent Nexys 4 dev board with an Artix 7 100T FPGA. Hard to beat for the price, 240 18x24 multipliers and 135 36kb RAM blocks. I built an add-on board which receives and transmits SPDIF and also has a stereo mic input amplifier and 24 bit ADC. 
     
    So far I've got all the IO systems setup for SPDIF and the ADC, and designed/tested 16384 sample 4-way stereo true-convolution (104gbit/s memory bandwidth, 2.9 billion 24x24 multiplies/s... aren't FPGA's awesome?)
     
    Next up will be designing the room capture algorithm. This is actually not all that difficult in practice. Do a log sine sweep, record it, and perform ifft(fft(measurement)/fft(test_signal)) to get the convolution impulse. Actually implementing will be a bit of work since I'll need to manually code the FFT library using an external RAM chip to handle a 524288 sample FFT. Fortunately the dev board comes with a 128mbit RAM chip. I'll store the captured impulses on an SD card.
     
    Ultimately I'll add a head-tracker too, but that's still a ways down the road. The hardware should be simple enough, just an accelerator on a wireless card. The software is not that bad either, I can stream a new kernal into the RAM buffer every few milliseconds in real-time while the convolution is operating.
     
     
    More to come...
     
  2. jgazal
    [​IMG]
     
    Which microphone capsules are you going to use? What is the plug in power voltage of your microphone input?
     
    Knowles FB-AE-30209-000 microphone with MB6022ASC-1 capsule seems to work with 10Vdc. 
     
    Do you know any other microphone/capsule that can be inserted deeply in the ear canal?
     
    I wish you all the best with this project!
     
  3. dude_500
    I'm using these: http://www.soundprofessionals.com/cgi-bin/gold/item/SP-TFB-2
     
    I don't know exactly what capsule the use, but they seem to have decent reviews and they at least report a full-range frequency response. Although I'm a bit suspicious of their frequency response since basically all the responses on the sound professionals site is a copy of the WM-61 response, which is discontinued. 
     
    It doesn't insert deep into the ear canal, but it should get all the head acoustics which I believe is all that really matters for this project.
     
    I have them biased at about 3V in this case. I could go up to 10V on the board, but I believe most electret capsules would be unhappy with 10V (not sure, documentation is sparse on exactly what the realistic max is, and I don't even know what capsule I have).
     
  4. jgazal
    There is also Primo EM-158. You can buy it here. Just cut a cheap headphone and diy.
     
    I have tried this one, with matched capsules. I record people and street ambience, with such microphones placed in my own ears, but it simple does not work when I play back the file: all sounds stay behind my head, like every other generic binaural recording. However, it apparently worked for other people recording with my gear. They say to have better 3d rendering playback.
     
    Perhaps the microphone must be inserted deeply in the ear-canal? For tapers point of view in this subject, see here.
    Perhaps my recorder does not have a good fidelity? It has 3v pip power (maybe less when loaded), it records at 16 bit, 44 kHz, but I do not know the quality of the internal preamp. 
    Would Head-tracking work better for me?
     
    I really do not know.
     
    My reasoning for higher diaphragm voltage: the higher pip dc voltage, the less distortion when exposed to a higher volume sine sweep, then better signal to noise ratio reaches your processor.
     
    There must be a reason for Smyth Research choosing a microphone that enters the ear canal.
     
  5. dude_500
    I don't think the smyth mics really enter the canal, here's a picture of them. And since we know the mic is facing outwards, it's not at the inner tip (which barely enters the canal anyways). 
     
    Actually, I think you want them right at the surface of the ear canal. The reason being that the canal is still there for the headphones, so you actually shouldn't measure the canal response. If you did, you'd be convolving it in but it's also there in reality so the response of the canal would be "double counted". Instead, you just want to measure the head transfer function which is not present when wearing headphones.
     
    At least, that's my logic. I could be totally off.
     
    Smyth-A8_03.jpg
     
  6. jgazal
    Why SOS, when you can read directly from the source:
     
     
  7. jgazal
     
    Just one more thought about stereo music vs. multichannel video content. 
     
    I am sure you are going to read every paragraph of Dr. Smyth paper, but reading it again this is now screaming to me:
     
     
    People usually underestimate how our brain relies on visual cues to give sense to spatial hearing. Multichannel content is usually linked with visual content, which may help you too better understand the auditory cues. 
     
    I thought the remembrance of the moment when I was recording my binaural files would help me to render the position of each source, but perhaps the neural networks that I use to recall visual memory are not related to the neural networks of my visual cortex… 
     
  8. Sopranino
    I just want to say bravo, this looks amazing.  Also you got me interested in learning how to code FPGA.  Should I learn Verilog or VHDL?  
     
  9. dude_500
     
    I've only worked with verilog.
     
    I believe VHDL is the standard for most of the world, but the United States is almost entirely Verilog centered in industry. Probably not relevant to a hobbyist but very much relevant if you plan to ever use the skills at work. 
     
    Some people say it is harder to learn VHDL after having learned Verilog, but ultimately people generally agree that it doesn't really matter which one you learn first, and if you get really serious about FPGA's you'll surely end up learning both eventually.
     

Share This Page