Spatial Sound Scene Synthesis and Manipulation for Virtual Reality and Audio Effects
PubDate: June 2018
Teams: Ville Pulkki; Symeon Delikaris-Manias; Archontis Politis
Writers: Ville Pulkki; Symeon Delikaris-Manias; Archontis Politis
This chapter covers DirAC‐based parametric time‐frequency domain (TF domain) audio techniques where the virtual sound scene is synthesized based on a geometric description of it around the avatar. Time‐frequency domain parametric spatial audio techniques also have applications in audio engineering, in addition to applications in virtual reality. The basic processing block of DirAC‐based spatialization in virtual worlds is the DirAC monosynth. When producing audio for virtual reality, often some recorded speech, music, and/or environmental sounds are used as signals for virtual sources. The chapter presents the application of parametric analysis and synthesis to spatial audio effects, using the DirAC framework. The ambience extraction method attempts to make foreground sounds with a clear direction softer, while preserving background sounds or ambience. Spatialization of a monophonic audio channel, either to a multichannel audio reproduction setup or as background sound in virtual reality, is a useful task in many audio engineering applications.