I am looking for reading suggestions (papers, dissertation etc) related to the use of the human voice's morphophysiology as a mapping model for gestural controllers.
You might have a look at Peter Birkholz' Diploma thesis (German) and the description of his articulatory speech synthesizer (http://www.vocaltractlab.de). His software includes a gestural score to control the synthesis. As far as I know both air stream and voicing can be controlled to allow for precise VOT timing and e.g. h-productions.