Animata / eyesweb / MAC
Hi kinolab, here interesting news for you:
As a gift for the beginning of 2011, there's a new "unstable" version
of OpenNI & Sensor modules with MacOSX support!
Other new features:
* Added support for multiple sensors.
* Improved README! (much more detailed instructions...)
* Some bug fixes...
Previous version features:
* A brand new .NET wrapper and .NET samples (OpenNI.net).
NITE binaries for all supported platforms (including MacOSX) will also
be released very soon.
The source code is located at:
https://github.com/OpenNI/OpenNI/tree/unstable
https://github.com/PrimeSense/Sensor/tree/unstable
GitHub:
git clone https://github.com/OpenNI/OpenNI.git -b unstable
git clone https://github.com/PrimeSense/Sensor.git -b unstable
Binaries:
Windows:
http://openni.org/sites/default/files/unstable/OpenNI-Bin-Win32-v1.0....
http://downloads.primesense.com/unstable/Sensor-Bin-Win32-v5.0.0.25.exe
Ubuntu 10.10 x86:
http://openni.org/sites/default/files/unstable/OpenNI-Bin-Linux32-v1....
http://downloads.primesense.com/unstable/Sensor-Bin-Linux32-v5.0.0.25...
Ubuntu 10.10 x64:
http://openni.org/sites/default/files/unstable/OpenNI-Bin-Linux64-v1....
http://downloads.primesense.com/unstable/Sensor-Bin-Linux64-v5.0.0.25...
MacOSX 10.6:
http://openni.org/sites/default/files/unstable/OpenNI-Bin-MacOSX-v1.0...
http://downloads.primesense.com/unstable/Sensor-Bin-MacOSX-v5.0.0.25....
We're really amazed from the awesome response of the community!
And thanks for everyone that's helped.
Ziv.
Extracting partial information
I was actually thinking of using [sigmund~] in peaks mode and routing the frequency and amplitude data of the partials to [osc~]. But looking at it now, I see that it orders them in amplitude, not harmonic number, so it might be difficult to get the specific harmonic you're after. And I don't really know how to get the correct phase from that, either.
I know that [bp~] and [vcf~] can ring a bit if you try to narrow the Q too much, so they're not ideal for extracting a single frequency from a complex signal.
FFT has its own complications. The frequencies of each bin in the analysis frame range from 0 to the samplerate, and they are evenly spaced at a distance of SR/blocksize. In other words, if your block size (which is also the number of bins) is 512 and sample rate is 44100Hz, then you have a bin for frequencies 0, 86.13, 172.27, etc. That's fine if you want a frequency that falls right on one bin; you can just resynthesize that bin. But if it falls between two bins, then it actually gets represented across several bins. I don't really know how to determine which bins should be used to resynthesize one frequency. I'm sure it can be done, though (I believe Spear does it).
After writing all that, I'm still kind of leaning toward [sigmund~].
Interaction Design Student Patches Available
Greetings all,
I have just posted a collection of student patches for an interaction design course I was teaching at Emily Carr University of Art and Design. I hope that the patches will be useful to people playing around with Pure Data in a learning environment, installation artwork and other uses.
The link is: http://bit.ly/8OtDAq
or: http://www.sfu.ca/~leonardp/VideoGameAudio/main.htm#patches
The patches include multi-area motion detection, colour tracking, live audio looping, live video looping, collision detection, real-time video effects, real-time audio effects, 3D object manipulation and more...
Cheers,
Leonard
Pure Data Interaction Design Patches
These are projects from the Emily Carr University of Art and Design DIVA 202 Interaction Design course for Spring 2010 term. All projects use Pure Data Extended and run on Mac OS X. They could likely be modified with small changes to run on other platforms as well. The focus was on education so the patches are sometimes "works in progress" technically but should be quite useful for others learning about PD and interaction design.
NOTE: This page may move, please link from: http://www.VideoGameAudio.com for correct location.
Instructor: Leonard J. Paul
Students: Ben, Christine, Collin, Euginia, Gabriel K, Gabriel P, Gokce, Huan, Jing, Katy, Nasrin, Quinton, Tony and Sandy
GabrielK-AsteroidTracker - An entire game based on motion tracking. This is a simple arcade-style game in which the user must navigate the spaceship through a field of oncoming asteroids. The user controls the spaceship by moving a specifically coloured object in front of the camera.
Features: Motion tracking, collision detection, texture mapping, real-time music synthesis, game logic
GabrielP-DogHead - Maps your face from the webcam onto different dog's bodies in real-time with an interactive audio loop jammer. Fun!
Features: Colour tracking, audio loop jammer, real-time webcam texture mapping
Euginia-DanceMix - Live audio loop playback of four separate channels. Loop selection is random for first two channels and sequenced for last two channels. Slow volume muting of channels allows for crossfading. Tempo-based video crossfading.
Features: Four channel live loop jammer (extended from Hardoff's ma4u patch), beat-based video cross-cutting
Huan-CarDance - Rotates 3D object based on the audio output level so that it looks like it's dancing to the music.
Features: 3D object display, 3d line synthesis, live audio looper
Ben-VideoGameWiiMix - Randomly remixes classic video game footage and music together. Uses the wiimote to trigger new video by DarwiinRemote and OSC messages.
Features: Wiimote control, OSC, tempo-based video crossmixing, music loop remixing and effects
Christine-eMotionAudio - Mixes together video with recorded sounds and music depending on the amount of motion in the webcam. Intensity level of music increases and speed of video playback increases with more motion.
Features: Adaptive music branching, motion blur, blob size motion detection, video mixing
Collin-LouderCars - Videos of cars respond to audio input level.
Features: Video switching, audio input level detection.
Gokce-AVmixer - Live remixing of video and audio loops.
Features: video remixing, live audio looper
Jing-LadyGaga-ing - Remixes video from Lady Gaga's videos with video effects and music effects.
Features: Video warping, video stuttering, live audio looper, audio effects
KatyC_Bunnies - Triggers video and audio using multi-area motion detection. There are three areas on each side to control the video and audio loop selections. Video and audio loops are loaded from directories.
Features: Multi-area motion detection, audio loop directory loader, video loop directory loader
Nasrin-AnimationMixer - Hand animation videos are superimposed over the webcam image and chosen by multi-area motion sensing. Audio loop playback is randomly chosen with each new video.
Features: Multi-area motion sensing, audio loop directory loader
Quintons-AmericaRedux - Videos are remixed in response to live audio loop playback. Some audio effects are mirrored with corresponding video effects.
Features: Real-time video effects, live audio looper
Tony-MusicGame - A music game where the player needs to find how to piece together the music segments triggered by multi-area motion detection on a webcam.
Features: Multi-area motion detection, audio loop directory loader
Sandy-Exerciser - An exercise game where you move to the motions of the video above the webcam video. Stutter effects on video and live audio looper.
Features: Video stutter effect, real-time webcam video effects
\[peak~\] peak picking
Hello every one!!!!!
So, a small intro... what i try to do it to compute the harmonic peaks of my
input (an "aaaaa" vowel)....
What i do is : my input-> hanning -> fft() -> abs() -> 20*log10()
and from that result i want to see where i have peaks....
I found the object peak~ witch takes a signal input and gives me a signal output...
really what i need...but the big big problem is that it only gives me 5-12 peaks in each processed block, and also not all the peaks that exist between its results,
(I hope i am clear). I looked for the help of the object, or the source code, but i did
not find anything.
What i want to do is to find around the first 160 peaks of my input signal and had
a correct output as in a format like the output of [peak~]....
Is there any ideas????
Pack - cold inlets and lists
The three messages will never actually appear at the same time. Everything is processed serially, and so each process will hit [pack]'s inlets one after the other. From the way it sounds, you're attempting to send three values to [pack] at the same time, but they're not hitting in the order you'd like them to. If so, there might be simpler options than what domien posted (though, I must say, it's pretty clever). One option is to ensure that the values are processed in the correct order, maybe using [trigger]. Another (perhaps simpler one as it might require less repatching) is to find out what order they are being processed and make sure the last one goes into [pack]'s hot inlet. Then you can reorder them with a message like this:
[pack f f f]
|
[$3 $1 $2(
Assuming the one you want to be the first in the list goes into the third inlet. This, of course, will only work if they are always being sent to [pack] in the same order. If not, domien's solution or the [trigger] method might be best.
Formant vocoder algorithm
Does any of this help? I'm not sure if a phase vocoder is the same thing as a formant vocoder, but googling formant always brings up phase, so here you are!
http://www.panix.com/~jens/pvoc-dolson.par
http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=1094078
Help in project
http://www.busim.ee.boun.edu.tr/~speech/research.html
these guys are actively working in speech processing domain. Few years back I also studied some of their papers to extract information. Have a look over their research work and even make a request for guidance , may be some one guide you as well.
Help in project
Thank you so much ahmad
I've followed your steps
and it works and I've manged to write the result in file
but how can I compute the 12 parameters??
-- FN low-frequency formant (fixed at 250 Hz)
-- ALF amplitude of low-frequency region
-- F1 frequency of 1st formant
-- A1 amplitude of 1st formant
-- F2 frequency of 2nd formant
-- A2 amplitude of 2nd formant
-- F3 frequency of 3rd formant
-- A3 amplitude of 3rd formant
-- AHF amplitude of 4th formant (fixed at 3500 Hz)
-- V degree of voicing
-- F0 fundamental frequency
-- MS glottal pulse mark/space ratio (fixed)
Help in project
The 'Holmes' parallel formant synthesiser Capable of producing speech from12 parameters (updated every 10 msecs):
-- FN low-frequency formant (fixed at 250 Hz)
-- ALF amplitude of low-frequency region
-- F1 frequency of 1st formant
-- A1 amplitude of 1st formant
-- F2 frequency of 2nd formant
-- A2 amplitude of 2nd formant
-- F3 frequency of 3rd formant
-- A3 amplitude of 3rd formant
-- AHF amplitude of 4th formant (fixed at 3500 Hz)
-- V degree of voicing
-- F0 fundamental frequency
-- MS glottal pulse mark/space ratio (fixed)
the required patch should accept wave file and produce pfs file that contains these parameters
how can I do it?
please ...........
Speech Formant Synthesizer (With bonus Turkish vowels pack :P)
Thanks, I'm quite amazed how clear the vowels are - excellent! Sorry I can't help you with further links about speech synthesis beyond the ones you would find with google yourself.
Also, formants for consonants? I was living with the impression that formants exclusively describe vowels. I also thought so far that in speech synthesis they use filtered noise for the various consonants.