Loading…
AES Show 2024 NY has ended
Exhibits+ badges provide access to the ADAM Audio Immersive Room, the Genelec Immersive Room, Tech Tours, and the presentations on the Main Stage.

All Access badges provide access to all content in the Program (Tech Tours still require registration)

View the Exhibit Floor Plan.
Monday, October 7
 

10:30am EDT

AI tools for sound design: emotional music and sound effects in visual media
Monday October 7, 2024 10:30am - 12:00pm EDT
Join us for an engaging workshop exploring deep learning tools for sound design. You'll receive two 30-second video clips, each intended to evoke different emotional responses, and be tasked with creating a musical accompaniment (including sound design) using the AI tools introduced. Gain insights from professionals who offer diverse perspectives on these technologies and learn how to integrate commercial and open-source AI effectively into your creative process. We’ll discuss the future of AI in sound design, focusing on its potential benefits, challenges, and areas for improvement. Come explore how AI can support rather than replace audio professionals.
Monday October 7, 2024 10:30am - 12:00pm EDT
1E03

11:00am EDT

Exhibitor Registration
Monday October 7, 2024 11:00am - 5:00pm EDT
Monday October 7, 2024 11:00am - 5:00pm EDT
Crystal Palalce South

12:15pm EDT

Dolby Atmos Music Essentials: From Stereo to Atmos
Monday October 7, 2024 12:15pm - 1:45pm EDT
In this workshop you’ll gain insights into Dolby Atmos audio with a particular emphasis on using your original stereo mix as a starting point to create an Atmos mix. You will learn how to set up your stereo mix session in a way that will make the transition into Atmos easy as well as create an Atmos mix using stems.
Speakers
avatar for Oscar Zambrano

Oscar Zambrano

Chief Engineer / Partner, Zampol Productions
Oscar Zambrano was born and raised in Mexico City. After graduating from Berklee College of Music in 2003, he moved to New York where he started working at Sound on Sound. In 2004 he and Jorge Castellanos founded Zampol Productions, a recording studio that focuses on mastering, mixing... Read More →
Monday October 7, 2024 12:15pm - 1:45pm EDT
1E03

2:00pm EDT

Creative sample-based music making in Ableton Live 12
Monday October 7, 2024 2:00pm - 3:30pm EDT
Join us for an in-depth workshop where you will explore the advanced sampling capabilities of Ableton Live 12 with expert presenter Ben Casey. This session will dive into the enhanced features and tools that make Live 12 a powerhouse for creative sample-based music production. This workshop will equip you with practical skills and techniques to elevate your music-making process. Learn how to discover, organize, and manipulate samples, generate new ideas, and integrate them seamlessly into your projects, all within the intuitive workflow of Ableton Live 12.
Speakers
avatar for Ben Casey

Ben Casey

Certified Trainer, Ableton
Ben Casey is a Brooklyn-based electronic musician, Ableton Certified Trainer, and overall music tech nerd. When he’s not surrounded by wires and drum machines or tinkering with Max for Live, Ben teaches Ableton Live to musicians across all genres, from avant-garde to zydeco.
Monday October 7, 2024 2:00pm - 3:30pm EDT
1E03

3:45pm EDT

A practical introduction to Remote ADR and Music Overdub workflows
Monday October 7, 2024 3:45pm - 5:15pm EDT
Outcome:
A deeper understanding of the voice ADR and music overdub process between the engineer, talent and client, advanced remote collaboration techniques for audio engineers, and general knowledge of the challenges of remote recording and how to overcome them to achieve flawless remote sessions.

Summary:
In this workshop you will learn in-depth the critical concepts you need to know for stress-free Remote ADR and Music Overdub sessions, when the talent (e.g. voice actor or music recording artist), audio engineer and perhaps client are all remote from each other. The primary challenge in recording remotely is that the talent will not have the engineer alongside them, and this requires additional technical setup on the talent’s side. Guiding the talent through this process is as important as knowing how to configure your own setup. In this workshop you will learn ADR and Overdub terms and concepts, how to set up a DAW for remote recording, using remote timeline synchronization techniques, and how to manage remote talent and client sessions using professional remote collaboration tools.
Bio:
Robert Marshall is musician, engineer, and producer, he is the visionary behind the idea of leveraging Internet connectivity to make post-production faster and more productive. He is a crucial force behind the solutions that have become synonymous with Source Elements such as Remote Transport Sync and Auto-Restore & Replace. Vincent dePierro is a sound engineer and musician. As Head of Support for Source Elements, as well as being a core member of their Innovations and Solutions team, Vincent has in-depth knowledge of how engineers and talent work together online.

Prior experience:
Assumes you already have knowledge in basic acoustics and digital audio workstations, existing knowledge of ADR workflows is not required.
Monday October 7, 2024 3:45pm - 5:15pm EDT
1E03
 
Tuesday, October 8
 

8:00am EDT

Attendee Registration
Tuesday October 8, 2024 8:00am - 6:00pm EDT
Tuesday October 8, 2024 8:00am - 6:00pm EDT
Crystal Palace South

9:00am EDT

Student Welcome Meeting
Tuesday October 8, 2024 9:00am - 9:30am EDT
Students! Joins us so we can find out where you are from, and tell you about all the exciting things happening at the convention.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Angela Piva

Angela Piva

Angela Piva, Audio Pro/Audio Professor, highly skilled in all aspects of music & audio production, recording, mixing and mastering with over 35 years of professional audio engineering experience and accolades. Known as an innovator in sound technology, and for contributing to the... Read More →
Tuesday October 8, 2024 9:00am - 9:30am EDT
1E08

9:00am EDT

NPR's Tiny Desk: A look back and a look forward
Tuesday October 8, 2024 9:00am - 10:00am EDT
NPR's Tiny Desk Concert Series has seen and increase in its popularity post-pandemic and is now a must for up and coming as well as established artists. Neil Tevault started the concert series when Bob Boilen came to NPR's music studio with an idea to record an artist at his desk. Josh Newell is the current technical director of NPR Music who is working to keep the standards set at the highest level and move the concert series into the future.

Neil will present a couple of concerts from early in the series looking at the small recording footprint that was employed. The very first concert with Laura Gibson, which was a single microphone and a DI and Bela Fleck, Edgar Meyer, and Zakir Hussain who played around a single stereo shotgun microphone.

Josh will present a couple of recent concerts to show how everything has grown and we have moved all the microphones closer and increased the number of microphones to capture the sound that the Tiny Desk is known for. Josh will also look at what we're doing to make things work better and make the recording process more efficient on concert day. We'll discuss Josh's vision for the sound of the Tiny Desk moving into the future.

We'll also leave ample time for Q&A

Tuesday October 8, 2024 9:00am - 10:00am EDT
1E09

9:00am EDT

Personalized Spatial Audio for Accessibility in Xbox Gaming
Tuesday October 8, 2024 9:00am - 10:00am EDT
In this enlightening panel, technologists from the Xbox platform and creatives from Xbox studios will join us to discuss how they are driving audio innovation and game sound design towards the vision of gaming for everyone. The discussion will focus on how personalized spatial audio can foster inclusivity by accommodating the unique auditory profiles of different ethnicities, genders, and age groups of gamers. By integrating these personalized Head-Related Transfer Functions (HRTFs) into audio middleware, we aim to enhance the Xbox gaming experience for all gamers. This approach not only enriches the auditory landscape but also breaks down barriers, making immersive gaming a truly inclusive experience. Join us as we explore the future of spatial audio on Xbox, where every gamer is heard and can fully engage with the immersive worlds we create.
Speakers
KS

Kaushik Sunder

VP of Engineering, Embody
avatar for Kapil Jain

Kapil Jain

CEO, Embody
We are Embody, and we believe in the power of AI to push the boundaries of immersive sound experience in gaming and entertainment.We are a team of data scientists, audio engineers, musicians and gamers who are building an ecosystem of tools and technologies for Immersive entertainment... Read More →
avatar for Robert Ridihalgh

Robert Ridihalgh

Technical Audio Specialist, Microsoft
A 33-year veteran of the games industry, Robert is an audio designer, composer, integrator, voice expert, and programmer with a passion for future audio technologies and audio accessibility.
Tuesday October 8, 2024 9:00am - 10:00am EDT
1E16

9:00am EDT

Adapting Immersive Microphone Techniques for different Acoustics Spaces
Tuesday October 8, 2024 9:00am - 10:00am EDT
By now we know a lot about how immersive microphone arrays work, from fully coincident, to semi-coincident, to fully spaced etc., and how they can be used to make spectacular 3D productions, But what happens when we use the same recording techniques in very different acoustic spaces with different size ensembles etc.? Various hi-resolution immersive recordings of acoustic music from around the globe made in extraordinary acoustic spaces including Skywalker Scoring Stage, Zlin Concert Hall, Notre Dame University, NYU's Paulson Center, CNSO Studios, Smecky Studios, and other exceptional churches and concert halls will be presented. Along with critical listening exercises, we will share microphone set-ups, multi-track recordings, videos and photo documentation of the spaces with the participants.
Speakers
PG

Paul Geluso

Director of the Music Technolo, New York University
avatar for David Bowles

David Bowles

Owner, Swineshead Productions, LLC
David v.R Bowles formed Swineshead Productions, LLC as a classical recording production company in 1995. His recordings have been GRAMMY- and JUNO-nominated and critically acclaimed worldwide. His releases in 3D Dolby Atmos can be found on Avie, OutHere Music (Delos) and Navona labels.Mr... Read More →
Tuesday October 8, 2024 9:00am - 10:00am EDT
1E06

9:00am EDT

Best practices for wireless audio in live productions
Tuesday October 8, 2024 9:00am - 10:00am EDT
Wireless audio, both mics and in-ear-monitors, has become essential in many live productions of music and theatre, but it is often fraught with uneasiness and uncertainty. The panel of presenters will draw on their varied experience and knowledge to show how practitioners can use best engineering practices to ensure reliability and performance of their wireless mic and in-ear-monitor systems.
Speakers
avatar for Bob Lee

Bob Lee

Applications Engineer / Trainer, RF Venue, Inc.
I'm a fellow of the AES, an RF and electronics geek, and live audio specialist, especially in both amateur and professional theater. My résumé includes Senhheiser, ARRL, and a 27-year-long tenure at QSC. Now I help live audio practitioners up their wireless mic and IEM game.I play... Read More →
HC

Henry Cohen

ASR Co-Chair, AES 151
Tuesday October 8, 2024 9:00am - 10:00am EDT
1E07

9:00am EDT

dLive Certification
Tuesday October 8, 2024 9:00am - 12:00pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.
 
The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Sign up link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Tuesday October 8, 2024 9:00am - 12:00pm EDT
1E05

9:30am EDT

An Industry Focused Investigation into Immersive Commercial Melodic Rap Production - Part Two
Tuesday October 8, 2024 9:30am - 9:50am EDT
In part one of this study, five professional mixing engineers were asked to create a Dolby Atmos 7.1.4 mix of the same melodic rap song adhering to the following commercial music industry specifications: follow the framework of the stereo reference, implement binaural distance settings, and conform to –18LKFS, -1dBTP loudness levels. An analysis of the mix sessions and post-mix interviews with the engineers revealed that they felt creatively limited in their approaches due to the imposed industry specifications. The restricted approaches were evident through the minimal applications of mix processing, automation, and traditional positioning of key elements in the completed mixes.
In part two of this study, the same mix engineers were asked to complete a second mix of the same song without any imposed limitations and were encouraged to approach the mix creatively. Intra-subject comparisons between the restricted and unrestricted mixes were explored to identify differences in element positioning, mix processing techniques, panning automation, loudness levels, and binaural distance settings. Analysis of the mix sessions and interviews showed that when no restrictions were imposed on their work, the mix engineers emphasized the musical narrative through more diverse element positioning, increased use of automation, and applications of additional reverb with characteristics that differed from the reverb in the source material.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers
avatar for Christal Jerez

Christal Jerez

Engineer, Christal's Sonic Lab
Christal Jerez is an audio engineer with experience recording, mixing and mastering music. After studying audio production at American University for her B.A. in Audio Production and at New York University for her Masters degree in Music Technology, she started working professionally... Read More →
Authors
avatar for Christal Jerez

Christal Jerez

Engineer, Christal's Sonic Lab
Christal Jerez is an audio engineer with experience recording, mixing and mastering music. After studying audio production at American University for her B.A. in Audio Production and at New York University for her Masters degree in Music Technology, she started working professionally... Read More →
avatar for Andrew Scheps

Andrew Scheps

Owner, Tonequake Records
Andrew Scheps has worked with some of the biggest bands in the world: Green Day, Red Hot Chili Peppers, Weezer, Audioslave, Black Sabbath, Metallica, Linkin Park, Hozier, Kaleo and U2. He’s worked with legends such as Johnny Cash, Neil Diamond and Iggy Pop, as well as indie artists... Read More →
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
Tuesday October 8, 2024 9:30am - 9:50am EDT
1E03

9:30am EDT

Physiological measurement of the arousing effect of bass amplification in music
Tuesday October 8, 2024 9:30am - 10:00am EDT
Music's amazing ability to evoke emotions has been the focus of various scientific studies, with researchers testing how different musical structures or interpretations impacted the emotions induced in the listener. However, in the context of amplified music, little is known about the influence of the sound reinforcement system. In this study, we investigate whether the amount of low-frequency amplification produced by a sound system impacts the listener's arousal. We organized two listening experiments whereby we measured the skin conductance of the participants while they were listening to music excerpts with different levels of low-frequency amplification. Our results indicate that an increase in the level of bass is correlated with a small but measurable rise in electrodermal activity, which is correlated with arousal. In addition this effect seems to depend on the nature of the music.
Moderators
avatar for Brett Leonard

Brett Leonard

Director of Music Technology Programs, University of Indianapolis
Speakers
avatar for Nicolas Epain

Nicolas Epain

Application Research Engineer, L-Acoustics
Authors
avatar for Nicolas Epain

Nicolas Epain

Application Research Engineer, L-Acoustics
avatar for Thomas Mouterde

Thomas Mouterde

Field application research engineer, L-Acoustics
Thomas Mouterde is a field application research engineer at L-Acoustics, a French manufacturer of loudspeakers, amplifiers, and signal processing devices. He is a member of the “Education and Scientific Outreach” department that aims at developing the education program of the... Read More →
Tuesday October 8, 2024 9:30am - 10:00am EDT
1E04

9:45am EDT

Introducing the 2025 AES AI and ML for Audio Conference and its New Format
Tuesday October 8, 2024 9:45am - 10:30am EDT
The upcoming 2025 AES International Conference on Artificial Intelligence and Machine Learning for Audio (AIMLA) aims to foster a collaborative environment where researchers and practitioners from academia and industry can converge to share their latest work in Artificial Intelligence (AI) and Machine Learning (ML) for Audio.

We want to advertise the upcoming AIMLA at the AES Show, to encourage early involvement and awareness from the AES community. To better accommodate the central themes of the conference, we propose new additions to the typical AES proceedings, such as challenges and long workshops, that can more appropriately showcase the rapidly growing state of the art. In this presentation, we plan to give an overview and a discussion space about the upcoming conference, and the changes we want to bring into play, tailored for AI/ML research communities, with references to successfully organized cases outside of AES. Finally, we propose a standardized template with guidelines for hosting crowdsourced challenges and presenting long workshops.

Challenges are a staple in the ML/AI community, providing a platform where specific problems are tackled by multiple teams who develop and submit models to address the given issue. These events not only spur competition but also encourage collaboration and knowledge sharing, ultimately driving forward the collective understanding and capabilities of the community.

Complementing the challenges, we introduce long-format workshops to exchange knowledge about emerging AI approaches in audio. These workshops can help develop novel approaches from the ground up and produce high-quality material for diffusion among participants. Both additions could help the conference become an exciting and beneficial event at the forefront of AI/ML for audio, as they intend to cultivate a setting where ideas can be exchanged effectively, drawing inspiration from established conferences such as ISMIR, DCASE, and ICASSP, which have successfully fostered AI/ML communities.

As evidenced by the recent AES International Symposium on AI and the Musician, we believe AI and ML will play an increasingly important role in audio and music engineering. To facilitate and standardize the procedures for featuring and conducting challenges and long-form workshops, we will present a complete guideline for hosting long-form workshops and challenges at AES conferences.

Our final goal is to promote the upcoming 2025 International Conference on AI and Machine Learning for Audio, generate a space to discuss the new additions and ideas, connect with interested parties, advertise and provide guidelines regarding the calls for crowd-sourced challenges and workshops, and ultimately get feedback from the AES as a whole to tailor the new conference to the requirements of both our AES and the AI/ML communities.
Speakers
avatar for Soumya Sai Vanka

Soumya Sai Vanka

PhD Researcher, Queen Mary University of London
I am a doctoral researcher at the Centre for Digital MusicQueen Mary University of London under the AI and Music Centre for Doctoral Training Program. My research focuses on the design of user-centric context-aware AI-based tools for music production. As a hobbyist musician and producer myself, I am interested in developing tools that can support creativity and collaboration resulting in emergence and novelty. I am also interested... Read More →
avatar for Franco Caspe

Franco Caspe

Student, Queen Mary University of London
I’m an electronic engineer, a maker, hobbyist musician and a PhD Student at the Artificial Intelligence and Music CDT at Queen Mary University of London. I have experience in development of real-time systems for applications such as communication, neural network inference, and DSP... Read More →
avatar for Brecht De Man

Brecht De Man

Head of Research, PXL University of Applied Sciences and Arts
Brecht is an audio engineer with a broad background comprising research, software development, management and creative practice. He holds a PhD from the Centre for Digital Music at Queen Mary University of London on the topic of intelligent software tools for music production, and... Read More →
Tuesday October 8, 2024 9:45am - 10:30am EDT
1E08

9:50am EDT

Investigation of spatial resolution of first and high order ambisonics microphones as capturing tool for auralization of real spaces in recording studios equipped with virtual acoustics systems
Tuesday October 8, 2024 9:50am - 10:10am EDT
This paper proposes a methodology for studying the spatial resolution of a collection of first-order and high-order ambisonic microphones when employed as a capturing tool of Spatial Room Impulse Responses (SRIRs) for virtual acoustics applications. In this study, the spatial resolution is defined as the maximum number of mono statistically independent impulse responses that can be extracted through beamforming technique and used in multichannel convolution reverbs. The correlation of the responses is assessed as a function of the beam angle and frequency bands, adapted to the frequency response of the loudspeakers in use, with the aim to be used in recording studios equipped with virtual acoustics systems that operate in the creation of the spatial impression of reverberation of real environments. The study examines the differences introduced by the physical characteristics of the microphones, the normalization methodologies of the spherical harmonics, and the number of spherical harmonics introduced in the encoding (ambisonic order). Preliminary results show that the correlation is inversely proportional to frequency as a function of wavelength. 
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers
avatar for Gianluca Grazioli

Gianluca Grazioli

Montreal, Canada, McGill University
Authors
Tuesday October 8, 2024 9:50am - 10:10am EDT
1E03

10:00am EDT

Exploring trends in audio mixes and masters: Insights from a dataset analysis
Tuesday October 8, 2024 10:00am - 10:30am EDT
We present an analysis of a dataset of audio metrics and aesthetic considerations about mixes and masters provided by the web platform MixCheck studio.The platform is designed for educational purposes, primarily targeting amateur music producers, and aimed at analysing their recordings prior to them being released. The analysis focuses on the following data points: integrated loudness, mono compatibility, presence of clipping and phase issues, compression and tonal profile across 30 user-specified genres. Both mixed (mixes) and mastered audio (masters)are included in the analysis, where mixes refer to the initial combination and balance of individual tracks, and masters refer to the final refined version optimized for distribution. Results show that loudness-related issues along with dynamics issues are the most prevalent, particularly in mastered audio. However mastered audio presents better results in compression than just mixed audio. Additionally, results show that mastered audio has a lower percentage of stereo field and phase issues.
Moderators
avatar for Brett Leonard

Brett Leonard

Director of Music Technology Programs, University of Indianapolis
Speakers
avatar for David Ronan

David Ronan

CEO, RoEx
CEO
Authors
Tuesday October 8, 2024 10:00am - 10:30am EDT
1E04

10:10am EDT

A comparative study of volumetric microphone techniques and methods in a classical recording context
Tuesday October 8, 2024 10:10am - 10:30am EDT
This paper studies volumetric microphone techniques (i.e. using configurations of multiple Ambisonic microphones) in a classical recording context. A pilot study with expert opinions was designed to show its feasibility. Based on the findings from the pilot study, a trio recording of piano, violin, and cello was conducted where 6 Ambisonic microphones established a hexagon. Such a volumetric approach is believed to improve the sound characteristics where the recordings were processed with the SoundField by RØDE Ambisonic decoder and were produced into a 7.0.4 loudspeaker system. A blinded subject experiment was designed where the participants were asked to evaluate the volumetric hexagonal configuration, comparing it to a more traditional 5.0 immersive configuration and a single Ambisonic microphone, all of which were mixed with spot microphones. These results were quantitatively analyzed, and revealed that the volumetric configuration is the most localized amongst all, but less immersive than the single Ambisonic microphone. No significant difference occurred in focus, naturalness, and preference. The analyses are generalized because the demographic backgrounds of the participants have no effect on the sound characteristics.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers Authors
avatar for Parichat Songmuang

Parichat Songmuang

Studio Manager/PhD Student, New York University
Parichat Songmuang graduated from New York University with her Master of Music degree in Music Technology at New York University and Advanced Certificate in Tonmeister Studies. As an undergraduate, she studied for her Bachelor of Science in Electronics Media and Film with a concentration... Read More →
PG

Paul Geluso

Director of the Music Technolo, New York University
Tuesday October 8, 2024 10:10am - 10:30am EDT
1E03

10:15am EDT

For Future Use - Surround and Immersive from Familiar Sources
Tuesday October 8, 2024 10:15am - 11:15am EDT
In a perfect world, one would record projects in the perfect space with musicians performing in perfect balance, with perfect results, but for us, the perfect world doesn’t exist and we’re presented with projects that were never intended for surround or immersive presentation, In this illustrated talk, we will play projects that were recorded (even in the last century!) and have been turned into successful immersive and surround recordings..
Speakers
avatar for Jim Anderson

Jim Anderson

Producer/Engineer, Anderson Audio New York
Producer/Engineer
avatar for Ulrike Schwarz

Ulrike Schwarz

Engineer/Producer, Co-Founder, Anderson Audio New York
Engineer/Producer, Co-Founder
Tuesday October 8, 2024 10:15am - 11:15am EDT
1E06

10:15am EDT

Adventures in Livestreaming
Tuesday October 8, 2024 10:15am - 11:45am EDT
During the Covid pandemic, streaming was the only way for musicians and artists to connect. As the pandemic subsided, many organizations continued to stream and livestream, and oftentimes it was the audio engineering team who were left to figure out how to produce video. Join this panel of experienced engineers and educators who found themselves learning how to produce compelling live concert video. The special challenges of livestreaming (i.e. streaming without a net) will be discussed, along with lessons learned. It is the intention of the panelists to provide encouragement to fellow engineers who are learning this new skillset themselves.
Speakers
avatar for Scott Burgess

Scott Burgess

Director, Audio and Media Production, Aspen Music Festival and School
Scott Burgess has worked on all facets of symphonic production. A graduate of Interlochen and the Cleveland Institute of Music, he has played bassoon and contrabassoon, sung with the Cleveland Orchestra Chorus, and produced, engineered, or edited numerous orchestral and chamber music... Read More →
avatar for Mary Mazurek

Mary Mazurek

Audio Educator/ Recording Engineer, University of Lethbridge
Audio Educator/ Recording Engineer
Tuesday October 8, 2024 10:15am - 11:45am EDT
1E09

10:15am EDT

The Plugin Kitchen: Coding a custom effect and putting it in the mix
Tuesday October 8, 2024 10:15am - 11:45am EDT
We are excited to present an educational workshop centered around the creation of custom audio plugins in Matlab. This hands-on session is designed for both students and educators on the endless possibilities of plugin development, while also encouraging participation in AES student competitions and fostering increased membership.

The workshop, titled “The Plugin Kitchen: Coding a custom effect and putting it in the mix” will offer a unique blend of real-time coding and practical application. Participants will be guided through the process of developing a comprehensive channel strip plugin, incorporating an EQ, a compressor, and a reverb effect.

Our session will begin with an introduction to audio plugins and their importance in modern music production. We will then dive into the hands-on coding segments:

1. Multi-band Parametric EQ:
Participants will learn to implement a parametric EQ, implementing multiple bands and adjusting parameters such as frequency and gain.

2. Compressor:
We will use Matlab’s built-in functions to create a compressor, covering essential concepts like threshold, ratio, attack, and release times. Attendees will see the immediate impact of these parameters on audio dynamics.

3. Reverb:
The final coding segment will focus on adding a reverb effect, including experimenting with decay time and pre-delay.

The workshop will culminate in a live demonstration by renowned engineer Paul Womack, showcasing the practical application of the developed plugin in a mixing session. This will illustrate the real-world benefits and versatility of the plugin, inspiring participants to explore further and engage with AES competitions.

Throughout the session, we emphasize interaction and practical application, ensuring that participants leave with both theoretical knowledge and a functional plugin they can continue to develop. Join us to unlock the potential of Matlab for audio plugin development and take a step towards innovative audio engineering.
Speakers
avatar for Christoph Thompson

Christoph Thompson

Director of Music Media Production, Ball State University
Christoph Thompson is vice-chair of the AES audio education committee. He is the chair of the AES Student Design Competition and the Matlab Plugin Design Competition. He is the director of the music media production program at Ball State University. His research topics include audio... Read More →
avatar for Chris Bennett

Chris Bennett

Professor, University of Miami
avatar for Paul Womack

Paul Womack

Record Producer/Recording Engineer
A producer, engineer and sonic artist, Paul "Willie Green" Womack has built a discography boasting names such as Armand Hammer, Wiz Khalifa, The Alchemist, The Roots, Billy Woods, ELUCID and many more, and established himself as one of the top names in independent Hip-Hop & R&B. Currently... Read More →
Tuesday October 8, 2024 10:15am - 11:45am EDT
1E07

10:15am EDT

The Anatomy of a Recording Session: Where Audio Technology and Musical Creativity Intersect (Part III)
Tuesday October 8, 2024 10:15am - 11:45am EDT
Abstract: Humans and machines interact to create new and interesting music content. Here we look at video excerpts from a particular recording session where the behind-the-scenes action comes to the forefront. Artists working with each other, artists working with the producer and engineer, and the influence (good or bad) of the technology with which they work will all be discussed during the workshop.

Summary:
The workshop will center on a discussion of using recording studio sessions to study creativity as a collaborative, but often complex and subtle practice. Our interdisciplinary team of recording engineers/producers and musicologists aims to understand how the interactions of musicians, engineers, recording technology, and musical instruments shape a recording’s outcome. Statements by participant-observers and the analysis of video footage from recording sessions will provide the starting point for discussions. In addition to first-hand recollections by members of our team, we are interviewing musicians who participated in the sessions. The workshop will focus on both musical interactions and on the interpersonal dynamics that affect the flow of various contributions and ideas during the recording process. Technology used also plays a role and will be analyzed as part of the workshop. The first workshop of this kind was a huge success in Helsinki at the 154th AES Convention. The room was packed, and we had an engaging discussion between panelists and audience members. Part II took place last week in Madrid (156th Convention), with many attendees saying it was “one of the highlights of the convention”. The workshop room was quite full, even though it was in the last timeslot of the last day. For this third workshop we plan on looking under the hood of a recording session involving a funk band with full horn section, and three lead singers. We plan to dig deeply into analyzing underlying events that percolate over time; the “quiet voices” that subtly influence the outcome of a recording session.
Along with our regular team of experts we are very excited to invite a guest panelist this time around – a venerable expert in music production, collaboration, and perception, Dr. Susan Rogers. This workshop was also proposed for AES 155th New York last fall but was declined due to lack of presentation space. Our requirements are simply a PowerPoint video presentation with stereo audio playback.
Speakers
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
avatar for Lisa Barg

Lisa Barg

Associate Professor, McGill University
Lisa Barg is Associate Professor of Music History and Musicology at the Schulich School of Music at McGill University and Associate Dean of Graduate Studies. She has published articles on race and modernist opera, Duke Ellington, Billy Strayhorn, Melba Liston and Paul Robeson. She... Read More →
avatar for David Brackett

David Brackett

Professor, McGill University
David Brackett is Professor of Music History and Musicology at the Schulich School of Music of McGill University, and Canada Research Chair in Popular Music Studies. His publications include Interpreting Popular Music (2000), The Pop, Rock, and Soul Reader: Histories and Debates... Read More →
avatar for Susan Rogers

Susan Rogers

Professor, Berklee Online
Susan Rogers holds a doctoral degree in experimental psychology from McGill University (2010). Prior to her science career, Susan was a multiplatinum-earning record producer, engineer, mixer and audio technician. She is best known for her work with Prince during his peak creative... Read More →
avatar for George Massenburg

George Massenburg

Associate Professor of Sound Recording, Massenburg Design Works
George Y. Massenburg is a Grammy award-winning recording engineer and inventor. Working principally in Baltimore, Los Angeles, Nashville, and Macon, Georgia, Massenburg is widely known for submitting a paper to the Audio Engineering Society in 1972 regarding the parametric equali... Read More →
Tuesday October 8, 2024 10:15am - 11:45am EDT
1E16

10:15am EDT

SEIDS guide to Building Sustainable Businesses for Music Creators
Tuesday October 8, 2024 10:15am - 11:45am EDT
In this workshop, you'll learn how to turn your love for music into a successful business. Acclaimed music producer, Sabrina Seidman aka SEIDS' tutorials have garnered thousands of views online. She’ll talk about how to decide what you want to do, find the right clients, and make products or services they need. You’ll also learn how to get clients and create your own chances to succeed. By the end of the workshop, you'll have the tools to start building a music career that makes you happy and earns you money.
Speakers
Tuesday October 8, 2024 10:15am - 11:45am EDT
1E15

10:30am EDT

Bestiari: a hypnagogic experience created by combining complementary state-of-the-art spatial sound technologies, Catalan Pavilion, Venice Art Biennale 2024
Tuesday October 8, 2024 10:30am - 10:50am EDT
Bestiari, by artist Carlos Casas1, is a spatial audio installation created as the Catalan pavilion for the 2024 Venice Art Biennale. The installation was designed for ambulant visitors and the use of informal seating arrangements distributed throughout the reproduction space, so the technical installation design did not focus on listeners’ presence in a single “sweet-spot”. While high-quality conventional spatial loudspeaker arrays typically provide excellent surround-sound experiences, the particular challenge of this installation was to reach into the proximate space of individual, dispersed and mobile listeners, rather than providing an experience that was only peripherally enveloping. To that end, novel spatial audio workflows and combinations of reproduction technologies were employed, including: High-order Ambisonic (HoA), Wavefield Synthesis (WFS), beamforming icosahedral (IKO), directional/parametric ultrasound, and infrasound. The work features sound recordings made for each reproduction technology, e.g., ambient Ambisonic soundfields recorded in Catalan national parks combined with mono and stereo recordings of specific insects in that habitat simultaneously projected via the WFS system. In-situ production provided an opportunity to explore the differing attributes of the reproduction devices and their interactions with the acoustical characteristics of the space – a concrete and brick structure with a trussed wooden roof, built in the late 1800s for the Venetian shipping industry. The practitioners’ reflections on this exploration, including their perception of the capabilities of this unusual combination of spatial technologies, are presented. Design, workflows and implementation are detailed.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers
avatar for Craig Cieciura

Craig Cieciura

Research Fellow, University of Surrey
Craig graduated from the Music and Sound Recording (Tonmeister) course at The University of Surrey in 2016. He then completed his PhD at the same institution in 2022. His PhD topic concerned reproduction of object-based audio in the domestic environment using combinations of installed... Read More →
Authors
avatar for Craig Cieciura

Craig Cieciura

Research Fellow, University of Surrey
Craig graduated from the Music and Sound Recording (Tonmeister) course at The University of Surrey in 2016. He then completed his PhD at the same institution in 2022. His PhD topic concerned reproduction of object-based audio in the domestic environment using combinations of installed... Read More →
Tuesday October 8, 2024 10:30am - 10:50am EDT
1E03

10:30am EDT

Audience Effect in the Low-Frequency Range, Part 2: Impact on Time Alignment of Loudspeaker Systems
Tuesday October 8, 2024 10:30am - 11:00am EDT
A sound reinforcement system typically combines a full-range system with a subwoofer system to deliver a consistent frequency bandwidth. The two systems must be time-aligned, which is usually done without an audience. This paper investigates the impact of the audience on the time alignment of loudspeaker systems at low frequencies. The study demonstrates, through on-site measurements and simulations, that the audience significantly affects sound propagation. The research highlights the greater phase shift observed with ground-stacked subwoofers compared to flown systems due to the audience’s presence, requiring adjustments of the system time alignment with the audience when flown and ground-stacked sources are used together. Moreover, in this case, the results demonstrate the lower quality of the summation with the audience even with the alignment adjustment. Lastly, recommendations for system design and calibration are proposed.
Moderators
avatar for Brett Leonard

Brett Leonard

Director of Music Technology Programs, University of Indianapolis
Speakers
avatar for Thomas Mouterde

Thomas Mouterde

Field application research engineer, L-Acoustics
Thomas Mouterde is a field application research engineer at L-Acoustics, a French manufacturer of loudspeakers, amplifiers, and signal processing devices. He is a member of the “Education and Scientific Outreach” department that aims at developing the education program of the... Read More →
Authors
avatar for Thomas Mouterde

Thomas Mouterde

Field application research engineer, L-Acoustics
Thomas Mouterde is a field application research engineer at L-Acoustics, a French manufacturer of loudspeakers, amplifiers, and signal processing devices. He is a member of the “Education and Scientific Outreach” department that aims at developing the education program of the... Read More →
avatar for Nicolas Epain

Nicolas Epain

Application Research Engineer, L-Acoustics
Tuesday October 8, 2024 10:30am - 11:00am EDT
1E04

10:45am EDT

Generative AI For Novel Audio Content Creation
Tuesday October 8, 2024 10:45am - 11:45am EDT
The presence and hype associated with generative AI across most forms of recorded media have become undeniable realities. Generative AI tools are becoming increasingly more prevalent, with applications ranging from conversational chatbots to text-to-image generation. More recently, we have witnessed an influx of generative audio models which have the potential of disrupting how music may be created in the very near future. In this talk, we will highlight some of the core technologies that enable novel audio content creation for music production, reviewing some seminal text-to-music works from the past year. We will then delve deeper into common research themes and subsequent works which intend to map these technologies closer to musicians’ needs.


We will begin the talk by outlining a common framework underlying the generative audio models that we will touch on, consisting of an audio synthesizer “back-end” paired with a latent representation modeling “front-end.” Accordingly, we will overview two primary forms of back-ends in the forms of neural audio codecs and variational auto-encoders (with examples), and illustrate how they pair naturally with transformer language model (LM) and latent diffusion model (LDM) front-ends, respectively. Furthermore, we will briefly touch on CLAP and T5 embeddings as conditioning signals that enable text as an input interface, and explain the means by which they are integrated into modern text-to-audio systems.

Next, we will review some seminal works that have been released within the past year(s) (primarily in the field of text-to-music generation), and roughly categorize them according to the common framework that we have built up thus far. At the time of writing this proposal, we would naturally consider MusicLM/FX (LM), MusicGen (LM), Stable Audio (LDM), etc. as exemplary candidates for review. We will contextualize these new capabilities in terms of what they can enable for music production and opportunities for future improvements. Accordingly, we will draw on some subsequent works that intend on meeting musicians a bit closer to the creative process. At the time of writing this proposal, this may include but is not limited to ControlNet (LDM), SingSong (LM), StemGen (LM), VampNet (LM), as well as our own previous work, as approved time permits. We will cap off our talk by providing some perspectives on what AI researchers could stand to understand about music creators, and what musicians could stand to understand about scientific research. Time permitting, we may allow ourselves to conduct a live coding demonstration whereby we exemplify constructing, training, and inferring audio examples from a generative audio model on a toy data example leveraging several prevalent open source libraries.


We hope that such a talk would be both accessible and fruitful for technologists and musicians alike. It would assume no background knowledge in generative modeling, and may perhaps assume only the most notional conception as to how machine learning works. The goal of this talk would be for the audience at large to walk out with a rough understanding of the underlying technologies and challenges associated with novel audio content creation using generative AI.
Tuesday October 8, 2024 10:45am - 11:45am EDT
1E08

10:50am EDT

Influence of Dolby Atmos versus Stereo Formats on Narrative Engagement: A Comparative Study Using Physiological and Self-Report Measures
Tuesday October 8, 2024 10:50am - 11:10am EDT
As spatial audio technology rapidly evolves, the conversation around immersion becomes ever more relevant, particularly in how these advancements enhance the creation of compelling sonic experiences. However, immersion is a complex, multidimensional construct, making it challenging to study in its entirety. This paper narrows the focus to one particular dimension—narrative engagement—to explore how it shapes the immersive experience. Specifically, we investigate whether the multichannel audio format, here 7.1.4, enhances narrative engagement compared to traditional stereo storytelling. Participants were exposed to two storytelling examples: one in an immersive format and another in a stereo fold-down. Physiological responses were recorded during listening sessions, followed by a self-report survey adapted from the Narrative Engagement Scale. The lack of significant differences between two formats in both subjective and objective measures is discussed in the context of existing studies.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
Authors
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
Tuesday October 8, 2024 10:50am - 11:10am EDT
1E03

11:00am EDT

Part of the Band: Virtual Acoustic Space as a Participant in Musical Performance
Tuesday October 8, 2024 11:00am - 11:30am EDT
We detail a real-time application of active acoustics used to create a shared virtual environment over a closed audio network as a research-creation project exploring the concept of room participation in musical performance. As part of a concert given in the Immersive Media Lab at McGill University, musicians and audience members were located in a virtual acoustic environment while a second audience was located in an adjacent but acoustically isolated space on the same audio network. Overall, the blending of computer-generated and acoustic sources created a specific use case for virtual acoustics while the immersive capture and distribution method examined an avenue for producing a real-time shared experience. Future work in this area includes audio networks with multiple virtual acoustic environments.
Moderators
avatar for Brett Leonard

Brett Leonard

Director of Music Technology Programs, University of Indianapolis
Speakers
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Authors
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Tuesday October 8, 2024 11:00am - 11:30am EDT
1E04

11:10am EDT

Creation of representative head-related impulse responses for binaural rendering of moving audio objects
Tuesday October 8, 2024 11:10am - 11:30am EDT
To achieve highly realistic 3D audio reproduction in virtual reality (VR) or augmented reality (AR) through binaural rendering, we must address the considerable computational complexity involved in convoluting head-related impulse responses (HRIRs). To reduce this complexity, an algorithm is proposed where audio signals are distributed to pre-defined representative directions through panning. Only the distributed signals are then convoluted with the corresponding HRIRs. In this study, we explored a method for generating representative HRIRs through learning, utilizing a full-sphere HRIR set. This approach takes into account smooth transitions and minimal degradation introduced during rendering, for both moving and static audio objects. Compared with conventional panning, the proposed method reduces average distortion by approximately 47% while maintaining the runtime complexity of the rendering.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers Authors
avatar for Masayuki Nishiguchi

Masayuki Nishiguchi

Professor, Akita Prefectural University
Masayuki Nishiguchi received his B.E., M.S., and Ph.D. degrees from Tokyo Institute of Technology, University of California Santa Barbara, and Tokyo Institute of Technology, in 1981, 1989, and 2006 respectively.  He was with Sony corporation from 1981 to 2015, where he was involved... Read More →
Tuesday October 8, 2024 11:10am - 11:30am EDT
1E03

11:30am EDT

Quantifying the Impact of Head-Tracked Spatial Audio on Common User Auditory Experiences using Facial Microexpressions
Tuesday October 8, 2024 11:30am - 11:50am EDT
The study aims to enhance the understanding of how Head Tracked Spatial Audio technology influences both emotional responses and immersion levels among listeners. By employing micro facial gesture recognition technology, it quantifies the depth of immersion and the intensity of emotional responses elicited by various types of binaural content, measuring categories such as Neutral, Happy, Sad, Angry, Surprised, Scared, Disgusted, Contempt, Valence, and Arousal. Subjects were presented with a randomized set of audio stimuli consisting of stereo music, stereo speech, and 5.1 movie content. Each audio piece lasted 15 seconds, and the Spatial Audio processing was On or Off randomly throughout the experiment. The FaceReader software was detecting the facial microexpressions of the subjects constantly. Statistical analysis was conducted using R software, applying Granger causality tests in time series, T-tests, and the P-value criterion for hypothesis validation. After consolidating the records of 78 participants, the final database consisted of 212,862 unique data points. With a 95% confidence, it was determined that the average level of "Arousal" is significantly higher when Head Tracked Spatial Audio is activated compared to when it is deactivated, suggesting that HT technology increases the emotional arousal of audio listeners. Regarding the happiness reaction, the highest levels were recorded in mode 5 (HT on and Voice) with an average of 0.038, while the lowest levels were detected in mode 6 (HT off and Voice). Preliminary conclusions indicate that surprise effectively causes a decrease in neutrality, supporting the dynamic interaction between these emotional variables.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers Authors
Tuesday October 8, 2024 11:30am - 11:50am EDT
1E03

11:30am EDT

INFLUENCE OF RECORDING TECHNIQUE AND ENSEMBLE SIZE ON APPARENT SOURCE WIDTH
Tuesday October 8, 2024 11:30am - 12:00pm EDT
The impression of listeners to aurally “see” the size of a performing entity is crucial to the success of both a concert hall and a reproduced sound field. Previous studies have looked at how different concert halls with different lateral reflections affect apparent source width. Yet, the perceptual effects of different source distributions with different recording techniques on apparent source width are not well understood. This study explores how listeners perceive the width of an orchestra by using four stereo and one binaural recording techniques and three wave field synthesis ensemble settings. Subjective experiments were conducted using stereo loudspeakers and headphone to play back the recording clips asking the listeners to rate the perceived wideness of the sound source. Results show that recording techniques greatly influence how wide an orchestra is perceived. The primary mechanism used to judge auditory spatial impression differs between stereo loudspeaker and headphone listening. When western classical symphony is recorded and reproduced by two-channel stereophony, the changes in instrument positions in terms of increasing or reducing the physical source width do not lead to an obvious increase or reduction on the spatial impression of the performing entity.
Moderators
avatar for Brett Leonard

Brett Leonard

Director of Music Technology Programs, University of Indianapolis
Speakers Authors
Tuesday October 8, 2024 11:30am - 12:00pm EDT
1E04

11:50am EDT

Investigating the Role of Customized Interaural Time Differences on First-Person Shooter Gaming Performance
Tuesday October 8, 2024 11:50am - 12:10pm EDT
Binaural listening with personalized Head-Related Transfer Functions (HRTFs) is known to enhance a listener's auditory localization in virtual environments, including gaming. However, the methods for achieving personalized HRTFs are often inaccessible for average game players due to measurement complexity and cost. This study explores a simplified approach to improving game performance, particularly in First-Person Shooter (FPS) games by optimizing Interaural Time Difference (ITD). Recognizing that horizontal localization is particularly important for identifying opponent positions in FPS games, this study hypothesizes that optimizing ITD alone may be sufficient for better game performances, potentially alleviating the need for full HRTF personalization. To test this hypothesis, a simplified FPS game environment was developed in Unity. Participants performed tasks to detect sound positions under three HRTF conditions: MIT-KEMAR, Steam Audio’s default HRTF, and the proposed ITD optimization method. The results indicated that our proposed method significantly reduced players' response times compared to other HRTF conditions. These findings allow players to improve their gaming performance within FPS games through simplified HRTF optimization, broadening accessibility to optimized HRTFs for a wider range of game users.
Moderators
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
Speakers
avatar for Sungjoon Kim

Sungjoon Kim

Research Intern, Korea Advanced Institute of Science and Technology
Authors
avatar for Sungjoon Kim

Sungjoon Kim

Research Intern, Korea Advanced Institute of Science and Technology
avatar for Rai Sato

Rai Sato

Ph.D. Student, Korea Advanced Institute of Science and Technology
Rai Sato (佐藤 来) is currently pursuing a PhD at the Graduate School of Culture Technology at the Korea Advanced Institute of Science and Technology. He holds a Bachelor of Music from Tokyo University of the Arts, where he specialized in immersive audio recording and psychoacoustics... Read More →
Tuesday October 8, 2024 11:50am - 12:10pm EDT
1E03

12:00pm EDT

Opening & Awards Ceremony
Tuesday October 8, 2024 12:00pm - 12:45pm EDT
Join the AES Committee and Chairs as we celebrate award recipients from the past year. The Opening and Awards will be followed by the keynote speech by Ebonie Smith.
Speakers
avatar for Leslie Gaston-Bird

Leslie Gaston-Bird

Owner, Mix Messiah Productions
Leslie Gaston-Bird (AMPS, MPSE) is author of the book "Women in Audio", part of the AES Presents series and published by Focal Press (Routledge). She is a voting member of the Recording Academy (The Grammys®) and its P&E (Producers and Engineers) Wing. Currently, she is a freelance... Read More →
avatar for Gary Gottlieb

Gary Gottlieb

AES President-Elect, Mendocino College
President-Elect, Co-Chair of the Events Coordination Committee, Chair of the Conference Policy Committee, and former Vice President of the Eastern Region, US and Canada; AES Fellow, Engineer, Author, Educator and Guest Speaker Gary Gottlieb refers to himself as a music generalist... Read More →
avatar for Michael Hagen

Michael Hagen

System Administrator and Studio Manager, New York University - Clive Davis Institute
Michael Hagen is the driving force behind the technical operations at the Clive Davis Institute at New York University, where he oversees the maintenance and continual improvement of the institute’s advanced production facilities. With extensive hands-on experience in recording... Read More →
avatar for Jeanne Montalvo

Jeanne Montalvo

Engineer/Producer, Self
Jeanne Montalvo is a Grammy-nominated audio engineer and award-winning radio producer. She was selected amongst thousands of applicants as the 2018 EQL resident at Spotify Studios and Electric Lady Studios in New York City, assisting in the recording process for artists like John... Read More →
Tuesday October 8, 2024 12:00pm - 12:45pm EDT
Stage

12:45pm EDT

Keynote: Ebonie Smith
Tuesday October 8, 2024 12:45pm - 1:45pm EDT
Ebonie Smith is a celebrated music producer, audio engineer, and singer-songwriter, based in the vibrant hub of Los Angeles. As a prominent figure in the industry, she currently holds the esteemed roles of senior audio engineer and producer at Atlantic Records. Ebonie's remarkable portfolio features notable credits, including the Broadway cast album of Hamilton, Janelle Monae's groundbreaking Dirty Computer, and Cardi B's chart-topping Invasion Of Privacy.

Notably, Ebonie serves as the Co-Chair of the Producers & Engineers Wing of The Recording Academy, underscoring her dedication to advancing excellence in music production. Beyond her professional achievements, she is the visionary founder and president of Gender Amplified, Inc., a nonprofit organization committed to celebrating and empowering women music producers.

Ebonie's educational foundation includes a master's degree in music technology from New York University and an undergraduate degree from Barnard College, Columbia University, solidifying her position as a distinguished leader in the music industry.
Speakers
avatar for Ebonie Smith

Ebonie Smith

Ebonie Smith is a celebrated music producer, audio engineer, and singer-songwriter, based in the vibrant hub of Los Angeles. As a prominent figure in the industry, she currently holds the esteemed roles of senior audio engineer and producer at Atlantic Records. Ebonie's remarkable... Read More →
Tuesday October 8, 2024 12:45pm - 1:45pm EDT
Stage

1:00pm EDT

Mixing Monitors from FOH
Tuesday October 8, 2024 1:00pm - 2:00pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.
 
The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Sign up link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Tuesday October 8, 2024 1:00pm - 2:00pm EDT
1E05

1:00pm EDT

Exhibit Hall
Tuesday October 8, 2024 1:00pm - 6:00pm EDT
Step into the heart of innovation at the Audio Engineering Society’s Annual Conference Exhibit Hall. This dynamic space brings together leading companies and cutting-edge technologies from across the audio engineering industry. Attendees will have the opportunity to explore the latest advancements in audio equipment, software, and services, engage with industry experts, and discover new solutions to enhance their projects. Whether you’re looking to network with professionals, gain insights from live demonstrations, or simply stay ahead of the curve, the Exhibit Hall is the place to be. Don’t miss this chance to immerse yourself in the future of audio engineering!
Tuesday October 8, 2024 1:00pm - 6:00pm EDT
Exhibit Hall

1:45pm EDT

AES Ice Cream Social
Tuesday October 8, 2024 1:45pm - 2:45pm EDT
Tuesday October 8, 2024 1:45pm - 2:45pm EDT
AES Membership Booth

2:00pm EDT

Towards prediction of high-fidelity earplug subjective ratings using acoustic metrics
Tuesday October 8, 2024 2:00pm - 2:30pm EDT
High-fidelity earplugs are used by musicians and live sound engineers to prevent hearing damage while allowing musical sounds to reach the eardrum without distortion. To determine objective methods for judging earplug fidelity in a similar way to headphones or loudspeakers, a small sample of trained listeners were asked to judge the attenuation level and clarity of music through seven commercially available passive earplugs. These scores were then compared to acoustic/musical metrics measured in a laboratory. It was found that NRR is strongly predictive of both attenuation and clarity scores, and that insertion loss flatness provides no advantage over NRR. A different metric measuring spectral flatness distortion seems to predict clarity independently from attenuation and will be subject to further study.
Moderators Speakers
avatar for David Anderson

David Anderson

Assistant Professor, University of Minnesota Duluth
Authors
avatar for David Anderson

David Anderson

Assistant Professor, University of Minnesota Duluth
Tuesday October 8, 2024 2:00pm - 2:30pm EDT
1E03

2:00pm EDT

Fourier Paradoxes
Tuesday October 8, 2024 2:00pm - 2:30pm EDT
Fourier theory is quite ubiquitous in modern audio signal processing. However, this framework is often at odds with our intuitions behind audio signals. Strictly speaking, Fourier theory is ideal to analyze periodic behaviors but when periodicities change across time it is easy to misinterpret its results. Of course, we have developed strategies around it like the Short Time Fourier Transform, yet again many times our interpretations of it falls beyond what the theory really says. This paper pushes the exact theoretical description showing examples where our interpretation of the data is incorrect. Furthermore, it shows specific instances where we incorrectly take decisions based on such paradoxical framework.
Moderators
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Speakers
avatar for Juan Sierra

Juan Sierra

NYU
Currently, I am a PhD Candidate in Music Technology at NYU and am currently based in NYUAD as part of the Global Fellowship program. As a professional musician, my expertise lies in Audio Engineering, and I hold a master's degree in Music, Science, and Technology from the prestigious... Read More →
Authors
avatar for Juan Sierra

Juan Sierra

NYU
Currently, I am a PhD Candidate in Music Technology at NYU and am currently based in NYUAD as part of the Global Fellowship program. As a professional musician, my expertise lies in Audio Engineering, and I hold a master's degree in Music, Science, and Technology from the prestigious... Read More →
Tuesday October 8, 2024 2:00pm - 2:30pm EDT
1E04

2:00pm EDT

Archiving multi-track and multi-channel: challeges and recommendations
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
Where are the files of my immersive production and why don't they open correctly on another DAW?
This workshop will outline some of the major challenges when working with archival materials and it will discuss ongoing activities in AES Standards Group SC-03-06 relating to multi-track and multi-channel audio. The workshop will give an overview of the status of the document, "Implementation of a Recommended Workflow for the Creation and Archiving of Digital Archival Materials from Professional Audio Production Formats," as well as why the interaction with related partner institutions (NARAS P&E Wing, SMPTE) and internal working groups is so important. The topic is relevant not only for preservation in an archival setting, but far beyond, for keeping audio productions safely stored and accessible in general.
Speakers
avatar for Nadja Wallaszkovits

Nadja Wallaszkovits

Stuttgart State Academy of fine Arts
avatar for Brad McCoy

Brad McCoy

Audio Engineer, Retired, Library of Congress
Audio Engineer (Archiving/Preservation)
avatar for Ulrike Schwarz

Ulrike Schwarz

Engineer/Producer, Co-Founder, Anderson Audio New York
Engineer/Producer, Co-Founder
avatar for Jim Anderson

Jim Anderson

Producer/Engineer, Anderson Audio New York
Producer/Engineer
avatar for Jeff Willens

Jeff Willens

Media Preservation Engineer, New York Public Library
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
1E09

2:00pm EDT

Designing With Constraints in an Era of Abundance
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
Every year, the availability and capabilites of processors and sensors expand greatly while their cost decreases correspondingly. As an instrument designer, the temptation to include every possible advance is an obvious one, but one that comes at the cost of additional complexity and probably more concerningly, reduced *character*.

Looking back on electronic instruments from the past, note that what we love about the classics are the quirks and idiosyncracies that come from the technical limitations of the time and how the designers ended up leveraging those limitations creatively. This panel will bring together designers to discuss how we are self-imposing constraints on our designs in the face of that temptation to include everything that's possible.

As Brian Eno said: "Whatever you now find weird, ugly, uncomfortable and nasty about a new medium will surely become its signature. CD distortion, the jitteriness of digital video, the crap sound of 8-bit - all of these will be cherished and emulated as soon as they can be avoided."
Speakers
avatar for Brett Porter

Brett Porter

Lead Software Engineer, Artiphon
Brett g Porter is a software developer and engineering manager with 3 decades of experience in the pro audio/music instrument industry; currently Lead Software Engineer at Artiphon, he leads the team that develops companion applications for the company's family of instruments. Previously... Read More →
AF

Alexandra Fierra

Eternal Research
AM

Adam McHeffey

CMO, Artiphon
avatar for Ben Neill

Ben Neill

Former Professor, Ramapo College
Composer/performer Ben Neill is the inventor of the mutantrumpet, a hybrid electro-acoustic instrument, and is widely recognized as a musical innovator through his recordings, performances and installations. Neill has recorded ten CDs of his music on the Universal/Verve, Thirsty Ear... Read More →
avatar for Nick Yulman

Nick Yulman

Kickstarter
Nick Yulman has worked with Kickstarter’s community of creators for the last ten years and currently leads the company’s Outreach team, helping designers, technologists, and artists of all kinds bring their ideas to life through crowdfunding. He was previously Kickstarter’s... Read More →
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
1E08

2:00pm EDT

Recent Advances in Volumetric Recording Techniques for Music Production
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
Volumetric sound recording techniques , where multiple 3D sound signals are captured throughout the recording area and mapped later for playback, have shown great promise in the world of positional tracking immersive experiences. But is volumetric sound capture an effective and plausible technique for music production ? What advantages does volumetric have over more traditional recording techniques and single position Ambisonic recordings and how the signals are combined for a stationary immersive listening experience ? These questions will be addressed by sharing immersive recordings, documentation, and discussing the outcomes of recent and ongoing work conducted at New York University and McGill University.
Speakers
PG

Paul Geluso

Director of the Music Technolo, New York University
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
avatar for Parichat Songmuang

Parichat Songmuang

Studio Manager/PhD Student, New York University
Parichat Songmuang graduated from New York University with her Master of Music degree in Music Technology at New York University and Advanced Certificate in Tonmeister Studies. As an undergraduate, she studied for her Bachelor of Science in Electronics Media and Film with a concentration... Read More →
avatar for Michael Ikonomidis

Michael Ikonomidis

Doctoral student, McGill University
Michael Ikonomidis (Michail Oikonomidis) is an accomplished audio engineer and PhD student in Sound Recording at McGill University, specializing in immersive audio, high-channel count orchestral recordings and scoring sessions.With a diverse background in music production, live sound... Read More →
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
1E06

2:00pm EDT

Richard King: Enveloping Masterclass
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
Richard King plays high resolution 7.1.4 music recordings of Yo-Yo Ma and Eric Clapton, and from the excellent Chevalier sound track album, describing also some of the techniques used.

This masterclass series, featuring remarkable recording artists, is a chance to hear 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the important qualia, auditory envelopment (AE); and we evaluate how well AE latent in the content - from intimate to grand - comes across in this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
3D06

2:00pm EDT

Bridging the Gap: Lessons for Live Media Networking from IT
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
The rapid evolution of live media networking has brought it closer to converged networking, where robust and efficient communication is paramount. While protocols such as MILAN/AVB, Dante and AES67 are staples, significant opportunities exist to enhance live media networking by adopting architectural blueprints, tools, and widely used protocols from the Information Technology (IT) sector. This workshop explores the specific requirements of live media networking, identifies potential learnings from IT workflows, and examines how other industries, particularly broadcast and video markets, have successfully integrated IT principles to propose technical recommendations.
Live media networking, encompassing audio, video, and control signals, demands high precision, low latency, and synchronization. Unlike traditional IT networks, which prioritize data integrity and security, live media networks must ensure seamless real-time transmission without compromising quality. The workshop will delve into these specificities, highlighting the challenges unique to live media and how they differ from typical IT networking scenarios and the use of Time Sensitive Networking (TSN)..
A significant challenge in this transition is the learning curve faced by sound technicians. Traditionally focused on audio-specific knowledge, these professionals now need to acquire IT networking skills to manage complex media networks effectively. This gap in expertise necessitates a new role emerging in the industry: the "Live Media Network Manager," a specialist who bridges the knowledge gap between traditional sound engineering and advanced IT networking.
A key focus area will be examining IT architectural blueprints and their applicability to live media networking. IT networks often leverage scalable, redundant, and resilient architectures to ensure uninterrupted service delivery. By adopting similar principles, live media networks can achieve greater reliability and scalability. The workshop will discuss how concepts such as network segmentation, redundancy, and failover mechanisms from IT can be tailored to meet the stringent requirements of live media.
Additionally, we will explore the tools and protocols widely used in IT that can benefit live media networking. Network monitoring and management tools, such as SNMP and Syslog, offer comprehensive insights into network performance and can aid in proactive maintenance and troubleshooting. Furthermore, protocols like QoS can be adapted to prioritize media traffic, ensuring that critical audio and video streams are delivered with minimal delay and jitter.
The workshop will also draw parallels from the broadcast and video markets, which have already embraced IT-based solutions to enhance their networking capabilities. These industries have developed technical recommendations and standards, such as SMPTE ST 2110 for professional media over managed IP networks, which can serve as valuable references for the live media domain. By examining these examples, participants will gain a broader perspective on how cross-industry learnings can drive innovation in live media networking.
This workshop will provide a comprehensive overview of the specific needs of live media networking and present actionable insights from IT workflows and other industries. Participants will leave with a deeper understanding of how to leverage IT principles to enhance the efficiency, reliability, and scalability of live media networks, paving the way for a more integrated and future-proof approach.
Speakers
avatar for Nicolas Sturmel

Nicolas Sturmel

Directout GmbH
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
1E16

2:00pm EDT

HELA Certification: Elevating standards in live event sound management
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
The Healthy Ears, Limited Annoyance (HELA) certification scheme, which originated within the AES Technical Committee on Acoustics and Sound Reinforcement, launched in summer 2024. Tailored for event organizers, sound engineers, venue managers, musicians, and all other key live event industry stakeholders, HELA offers a comprehensive framework for delivering live music experiences that protect audience hearing and minimize neighborhood disturbances. This session will delve into the balance between sound quality, hearing health and community harmony. Attendees will gain practical insights into HELA's guidance on sound level management and effective communication strategies, fostering a community dedicated to sustainable live event production. Join us to discover how HELA Certification can set a new industry standard, creating memorable yet safe and respectful experiences for everyone involved.
Speakers
avatar for Adam Hill

Adam Hill

Associate Professor of Electroacoustics, University of Derby
Adam Hill is an Associate Professor of Electroacoustics at the University of Derby where he leads the Electro-Acoustics Research Lab (EARLab) and runs the MSc Audio Engineering program. He received a Ph.D. from the University of Essex, an M.Sc. in Acoustics and Music Technology from... Read More →
avatar for Jon Burton

Jon Burton

Senior Lecturer, Derby University
A live sound engineer with over 40 years of concert touring experience. Jon has toured internationally with artists such as Bryan Ferry, Stereophonics, Biffy Clyro and The Prodigy. Jon is also a partner in a five-studio recording complex in Sheffield, UK. Involved in education for... Read More →
avatar for Laura Sinnott

Laura Sinnott

Owner, Sound Culture
A long time audio engineer for film, Laura career-expanded into hearing health as an audiologist. She ran the hearing clinic at Sensaphonics, a Chicago-based institution that has served musicians for over 30 years. Now based in Central New York, she sees patients in her Utica, NY... Read More →
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
1E07

2:00pm EDT

WMAS – The way forward for multichannel wireless audio
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
Join the industry experts on this exciting panel to learn more about Wireless Multichannel Audio Systems, the wide scope of approaches that WMAS facilitates, and how the various technologies can help you with multi-channel wireless set ups in challenging RF environments.
Speakers
avatar for Joe Ciaudelli

Joe Ciaudelli

Sennheiser
Joe Ciaudelli was hired by Sennheiser in 1987 upon graduating from Columbia University with an electrical engineering degree.  He provided frequency coordination for large multi-channel wireless microphone systems used by Broadway productions, major theme parks, and broadcast networks... Read More →
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
1E15

2:00pm EDT

Jack Antonoff & Laura Sisk, Up Close and Personal
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
GRAMMY® Award-winning producer Jack Antonoff and GRAMMY® Award-winning recording and mix engineer Laura Sisk will be the focal point of a special event session titled “Up Close and Personal.” Jack Antonoff is an eleven-time GRAMMY® Award-winning producer, artist, songwriter, and musician, as well as the creative force behind Bleachers. In February 2024, Antonoff won Producer of the Year at the GRAMMY Awards® for an incredible third consecutive year, becoming only the second producer in history to win three years running. He will be joined by five-time GRAMMY® Award-winning recording and mix engineer and long-time collaborator Laura Sisk in a session revealing their studio work with such leading artists as Taylor Swift, Lana Del Rey, Sabrina Carpenter, Kendrick Lamar, Nick Cave, St. Vincent, Diana Ross and more. The event will be moderated by producer/engineer/musician Glenn Lorbecki.
Speakers
avatar for Jack Antonoff

Jack Antonoff

Credited by the BBC for having “redefined pop music,” the globally celebrated, eleven-time Grammy Award-winning singer, songwriter, musician, and producer, Jack Antonoff has collaborated with the likes of Taylor Swift, Kendrick Lamar, Lana Del Rey, The 1975, Diana Ross, Lorde... Read More →
avatar for Laura Sisk

Laura Sisk

Laura Sisk is a five-time GRAMMY® Award-winning recording and mix engineer, widely recognized for her work with producer Jack Antonoff on Taylor Swift, as well as working with renowned artists like Lana Del Rey, Jon Batiste, Florence + The Machine, Diana Ross, Lorde, St.Vincent... Read More →
Tuesday October 8, 2024 2:00pm - 3:00pm EDT
Stage

2:00pm EDT

Bitrate adaptation in object-based audio coding in communication immersive voice and audio systems
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
The object-based audio is one of the spatial audio representations providing an immersive audio experience. While it can be found in a wide variety of audio reproduction systems, its use in communication systems is very limited as it faces many constraints like the complexity of the system, short delay, or limited available bitrate for coding and transmission. This paper presents a new bitrate adaptation method to be used in object-based audio coding systems that overcomes these constraints and enables their use in 5G voice and audio communication systems. The presented method distributes an available codec bit budget to encode waveforms of the individual audio objects based on a classification of the objects’ subjective importance in particular frames. The presented method has been used in the Immersive Voice and Audio Services (IVAS) codec, recently standardized by 3GPP, but it can be employed in other codecs as well. Test results show the performance advantage of the bitrate adaptation method over the conventional uniformly distributed bitrate method. The paper also presents IVAS selection test results for object-based audio with four audio objects, rendered to binaural headphone representation, in which the presented method plays a substantial role.
Speakers Authors
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Enhancing Realism for Digital Piano Players: A Perceptual Evaluation of Head-Tracked Binaural Audio
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
This paper outlines a process for achieving and perceptually evaluating a head-tracked binaural audio system designed to enhance realism for players of digital pianos. Using an Ambisonic microphone to sample an acoustic piano, followed by leveraging off-the-shelf equipment, the system allows players to experience changes in the sound field in real-time as they rotate their heads while wearing headphones under three degrees of freedom (3DoF). The evaluation criteria included spatial clarity, spectral clarity, envelopment, and preference. These criteria were assessed across three different listening systems: stereo speakers, stereo headphones, and head-tracked binaural audio. Results showed a strong preference for the head-tracked binaural audio system, with players noting significantly greater realism and immersion.
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Exploring Immersive Opera: Recording and Post-Production with Spatial Multi-Microphone System and Volumetric Microphone Array
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Traditional opera recording techniques using large microphone systems are typically less flexible towards experimental singer choreographies, which have the potential of being adapted to immersive and interactive representations such as Virtual Reality (VR) applications. The authors present an engineering report on implementing two microphone systems for recording an experimental opera production in a medium-sized theatre: a 7.0.4 hybrid array of Lindberg’s 2L and the Bowles spatial arrays and a volumetric array consisting of three higher-order Ambisonic microphones in Left/Center/Right (LCR) formation. Details of both microphone setups are first described, followed by post-production techniques for multichannel loudspeaker playback and 6 degrees-of-freedom (6DoF) binaural rendering for VR experiences. Finally, the authors conclude with observations from informal listening critique sessions and discuss the technical challenges and aesthetic choices involved during the recording and post-production stages in the hope of inspiring future projects on a larger scale.
Speakers
JM

Jiawen Mao

PhD student, McGill University
Authors
JM

Jiawen Mao

PhD student, McGill University
avatar for Michael Ikonomidis

Michael Ikonomidis

Doctoral student, McGill University
Michael Ikonomidis (Michail Oikonomidis) is an accomplished audio engineer and PhD student in Sound Recording at McGill University, specializing in immersive audio, high-channel count orchestral recordings and scoring sessions.With a diverse background in music production, live sound... Read More →
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Exploring the Directivity of the Lute, Lavta, and Oud Plucked String Instruments
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
This study investigates the spherical directivity and radiation patterns of the Lute, Lavta, and Oud, pear-shaped traditional plucked-string instruments from the Middle East, Turkey, Greece, and the surrounding areas, providing insights into the acoustic qualities of their propagated sound in a three-dimensional space. Data was recorded in an acoustically controlled environment with a 29-microphone array, using multiple instruments of each type, performed by several professional musicians. Directivity is investigated in terms of sound projection and radiation patterns. Instruments were categorized according to string material. The analysis revealed that all instruments, regardless of their variations in geometry and material, exhibit similar radiation patterns across all frequency bands, justifying their intuitive classification within the “Lute family”. Nevertheless, variations in sound projection across all directions are evident between instrument types, which can be attributed to differences in construction details and string material. The impact of the musician's body on directivity is also observed. Practical implications of this study include the development of guidelines for the proper recording of these instruments, as well as the simulation of their directivity properties for use in spatial auralizations and acoustic simulations with direct applications in extended reality environments and remote collaborative music performances.
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Generate acoustic responses of virtual microphone arrays from a single set of measured FOA responses. - Apply to multiple sound sources.
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
V2MA (VSVerb Virtual Microphone Array)
Demos and related docs are available at https://bit.ly/3BmDBbL .
Once we have measured a set of four impulse responses (IRs) with an A-format microphone in a hall, we can make a virtual recording using a virtual microphone array built in the hall at will. The measurement does not require an A-format microphone and a loudspeaker to be placed at specific positions in the hall. Typical positions, such as at an audience seat and on a stage, are recommended, but you can place them anywhere you like. We will generate any type of virtual microphone response in a target room from an easy one-time IR measurement.
-------------------------------
We propose a method, V2MA, that virtually generates acoustic responses of any type of microphone array from a single set of FOA responses measured in a target room. An A-format microphone is used for the measurement, but no Ambisonics operation is included in the processing. V2MA is a method based on geometrical acoustics. We calculate sound intensities in the x, y, and z directions from a measured FOA response, then the virtual sound sources of the room are detected from them. Although it is desirable to have an A-format microphone place close to the attempted position of the virtual microphone array in the room, it is not a mandatory requirement. Since our method allows to generate SRIRs at arbitrary receiver positions in the room by updating the acoustic properties of the virtual sound sources detected at a certain position of the room, an A-format microphone can be placed anywhere you like. On the other hand, a loudspeaker must be placed at the source position where a player is assumed to be. Since the positions of virtual sound sources change when a real sound source moves, we used to measure the responses for each assumed real source position. To improve this inconvenient restriction, we developed the technique of updating the positions of the virtual sound sources when a real sound source moves from its original position. Although the technique requires some approximations, it is ascertained that the generated SRIRs provide fine acoustic properties in both physical and auditory aspect.
Speakers
avatar for Masataka Nakahara

Masataka Nakahara

Acoustic Designer / Acoustician, SONA Corp. / ONFUTURE Ltd.
Masataka Nakahra is an acoustician specializing in studio acoustic design and R&D work on room acoustics, as well as an educator.After studying acoustics at the Kyushu Institute of Design, he joined SONA Corporation and began his career as an acoustic designer.In 2005, he received... Read More →
Authors
avatar for Masataka Nakahara

Masataka Nakahara

Acoustic Designer / Acoustician, SONA Corp. / ONFUTURE Ltd.
Masataka Nakahra is an acoustician specializing in studio acoustic design and R&D work on room acoustics, as well as an educator.After studying acoustics at the Kyushu Institute of Design, he joined SONA Corporation and began his career as an acoustic designer.In 2005, he received... Read More →

Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Measurement and Applications of Directional Room Impulse Responses (DRIRs) for Immersive Sound Reproduction
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Traditional methods for characterizing Room Impulse Responses (RIRs) employing omnidirectional microphones do not fully capture the spatial properties of sound in an acoustic space. In this paper we explore a method for the characterization of room acoustics employing Directional Room Impulse Responses (DRIRs), which include the direction of arrival of the reflected sound waves in an acoustic space in addition to their time of arrival and strength. We measured DRIRs using a commercial 3D sound intensity probe (Weles Acoustics WA301) containing x, y, z acoustic velocity channels in addition to a scalar pressure channel. We then employed the measured DRIR’s to predict the binaural signals that would be measured by binaural dummy head microphones placed at the same location in the room where the DRIR was measured. The predictions can then be compared to the actual measured binaural signals. Successful implementation of DRIRs could significantly enhance applications in AR/VR and immersive sound reproduction by providing listeners with room-specific directional cues for early room reflections in addition to the diffuse reverberant impulse response tail.
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Quantitative Assessment of Acoustical Attributes and Listener Preferences in Binaural Renderers with Head-tracking Function
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
The rapid advancement of immersive audio technologies has popularized binaural renderers that create 3D auditory experiences using head-related transfer functions (HRTFs). Various renderers with unique algorithms have emerged, offering head-tracking functionality for real-time adjustments to spatial audio perception. Building on our previous study, we compared binauralized music from five renderers with the dynamic head-tracking function enabled, focusing on how differences in HRTFs and algorithms affect listener perceptions. Participants assessed overall preference, spatial fidelity, and timbral fidelity by comparing paired stimuli. Consistent with our earlier findings, one renderer received the highest ratings for overall preference and spatial fidelity, while others rated lower in these attributes. Physical analysis showed that interaural time differences (ITD), interaural level differences (ILD), and frequency response variations contributed to these outcomes. Notably, hierarchical cluster analysis of participants' timbral fidelity evaluations revealed two distinct groups, suggesting variability in individual sensitivities to timbral nuances. While spatial cues, enhanced by head tracking, were generally found to be more influential in determining overall preference, the results also highlight that timbral fidelity plays a significant role for certain listener groups, indicating that both spatial and timbral factors should be considered in future developments.
Speakers
avatar for Rai Sato

Rai Sato

Ph.D. Student, Korea Advanced Institute of Science and Technology
Rai Sato (佐藤 来) is currently pursuing a PhD at the Graduate School of Culture Technology at the Korea Advanced Institute of Science and Technology. He holds a Bachelor of Music from Tokyo University of the Arts, where he specialized in immersive audio recording and psychoacoustics... Read More →
Authors
avatar for Rai Sato

Rai Sato

Ph.D. Student, Korea Advanced Institute of Science and Technology
Rai Sato (佐藤 来) is currently pursuing a PhD at the Graduate School of Culture Technology at the Korea Advanced Institute of Science and Technology. He holds a Bachelor of Music from Tokyo University of the Arts, where he specialized in immersive audio recording and psychoacoustics... Read More →
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Review: Head-Related Impulse Response Measurement Methods
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
This review paper discusses the advancements in Head-Related Impulse Response measurement methods. HRIR (Head-Related Impulse Response) measurement methods, often referred to as HRTF (Head-Related Transfer Function) measurement methods, have undergone significant changes over the last few decades [1]. A frequently employed method is the Discrete stop-and-go method [1][2]. It involves changing the location of a single speaker, used as the sound source, and recording the impulse response at each location. [2]. Since the measurement is for 1 location of the sound source at a time, using the discrete stop-and-go method is time-consuming [1]. Hence improvements are required to enhance the efficiency of the measurement process such as using more sound sources (speakers) [1][3]. A typical HRTF measurement is usually conducted in an anechoic chamber to achieve a simulated free-field measurement condition without room reverberation. It measures the transfer function between the source and the ears to perceive localisation cues such as inter-aural time differences (ITDs), inter-aural level differences (ILDs), as well as monaural spectral cues [4]. Newer techniques such as the Multiple Exponential Sweep Method (MESM) and the reciprocal method offer alternatives. These methods enhance measurement efficiency and address challenges like inter-reflections and low-frequency response [5][6]. Individualised HRTF measurement techniques can be categorised into acoustical measurement, anthropometric data, and perceptual feedback [7]. Interpolation methods and non-anechoic environment measurements have expanded the practical application and feasibility of HRTF measurements [8][9][10][7].
Speakers
avatar for Jeremy Tsuaye

Jeremy Tsuaye

New York University
Authors
avatar for Jeremy Tsuaye

Jeremy Tsuaye

New York University
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

The effects of interaural time difference and interaural level difference on sound source localization on the horizontal plane
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Interaural Time Difference (ITD) and Interaural Level Difference (ILD) are the main cues used by the human auditory system to localize sound sources on the horizontal plane. To explore the relationship between ITD, ILD, and the perceived azimuth, a study was conducted to measure and analyze the localization effects on the horizontal plane by combining ITD and ILD. Pure tones were used as sound sources in the experiment. For each of the three different frequency bands, 25 combinations of ITD and ILD test values were selected. These combinations were used to process the perceived sound from directly in front of the listener (pure tone signals collected using an artificial head in an anechoic chamber). The tests were conducted using the 1up/2down and 2AFC (two-alternative forced-choice) psychophysical testing methods. The results showed that the perceived azimuth at 350 Hz and 570 Hz was generally higher than at 1000 Hz. Additionally, the perceived azimuth at 350 Hz and 570 Hz was similar under certain combinations. The experimental data and conclusions can provide foundational data and theoretical support for efficient compression of multi-channel audio.
Tuesday October 8, 2024 2:00pm - 4:00pm EDT
Poster

2:30pm EDT

Decoding Emotions: Lexical and Acoustical Cues in Vocal Affects
Tuesday October 8, 2024 2:30pm - 3:00pm EDT
This study investigates listeners’ ability to detect emotion from a diverse set of speech samples, including both spontaneous conversations and actor-posed speech. It explores the contributions of lexical content and acoustic properties when native listeners rate seven pairs of affective attributes. Two experimental conditions were employed: a text condition, where participants evaluated emotional attributes from written transcripts without vocal information, and a voice condition, where participants listened to audio recordings to assess emotions. Results showed that the importance of lexical and vocal cues varies across 14 affective states for posed and spontaneous speech. Vocal cues enhanced the expression of sadness and anger in posed speech, while they had less impact on conveying happiness. Notably, vocal cues tended to mitigate negative emotions conveyed by the lexical content in spontaneous speech. Further analysis on correlations between emotion ratings in text and voice conditions indicated that lexical meanings suggesting anger or hostility could be interpreted as positive affective states like intimacy or confidence. Linear regression analyses indicated that emotional ratings by native listeners could be predicted up to 59% by lexical content and up to 26% by vocal cues. Listeners relied more on vocal cues to perceive emotional tone when the lexical content was ambiguous in terms of feeling and attitude. Finally, the analysis identified statistically significant basic acoustical parameters and other non/para-linguistic information, after controlling for the effect of lexical content.
Moderators Speakers
EO

Eunmi Oh

Research Professor, Yonsei University
Authors
EO

Eunmi Oh

Research Professor, Yonsei University
Tuesday October 8, 2024 2:30pm - 3:00pm EDT
1E03

2:30pm EDT

Nonlinear distortion in analog modeled DSP plugins in consequence of recording levels
Tuesday October 8, 2024 2:30pm - 3:00pm EDT
The nominal audio level is where developers of professional analog equipment design their units to have an optimal performance. Audio levels above the nominal level will at some point lead to increased harmonic distortion and eventually clipping. DSP plugins emulating such nonlinear behavior must – in the same manner as analog equipment – align to a nominal level that is simulated within the digital environment. A listening test was tailored to investigate if, or to which extent, misalignments in the audio levels compared to the simulated nominal level in analog-modelled DSP plugins are audible, thus affecting the outcome, depending on which level you choose to record at. The results of this study indicate that harmonic distortion in analog-modeled DSP plugins may become audible as the recording level increases. However, for the plugins included in this study, the immediate consequence of the harmonics added is not critical and, in most cases, not noticed by the listener.
Moderators
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Speakers
avatar for Tore Teigland

Tore Teigland

Professor, Kristiania University College
Authors
avatar for Tore Teigland

Tore Teigland

Professor, Kristiania University College
Tuesday October 8, 2024 2:30pm - 3:00pm EDT
1E04

2:30pm EDT

Vocal Processing
Tuesday October 8, 2024 2:30pm - 3:30pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Tuesday October 8, 2024 2:30pm - 3:30pm EDT
1E05

3:00pm EDT

A comparison of in-ear headphone target curves for the Brüel & Kjær Head & Torso Simulator Type 5128
Tuesday October 8, 2024 3:00pm - 3:30pm EDT
Controlled listening tests were conducted on five different in-ear (IE) headphone target curves measured on the latest ITU-T Type 4.3 ear simulator (e.g. Bruel & Kjaer Head & Torso Simulator Type 5128). A total of 32 listeners rated each target on a 100-point scale based on preference for three different music programs with two observations each. When averaged across all listeners, two target curves were found to be equally preferred over the other choices. Agglomerative hierarchical clustering analysis further revealed two classes of listeners based on dissimilarities in their preferred target curves. Class One (72% of listeners) preferred the top two rated targets. Class two (28% of listeners) preferred targets with 2 dB less bass and 2 dB more treble than the target curves preferred by Class 1. Among the demographic factors examined, age was the best predictor of membership in each class.
Moderators Speakers Authors
Tuesday October 8, 2024 3:00pm - 3:30pm EDT
1E03

3:00pm EDT

A Survey of Methods for the Discretization of Phonograph Record Playback Filters
Tuesday October 8, 2024 3:00pm - 3:30pm EDT
Since the inception of electrical recording for phonograph records in 1924, records have been intentionally cut with a non-uniform frequency response to maximize the information density on a disc and to improve the signal-to-noise ratio. To reproduce a nominally flat signal within the available bandwidth, the effects of this cutting curve must be undone by applying an inverse curve on playback. Until 1953, with the introduction of what has become known as the RIAA curve, the playback curve required for any particular disc could vary by record company and over time. As a consequence, anyone seeking to hear or restore the information on a disc must have access to equipment that is capable of implementing multiple playback equalizations. This correction may be accomplished with either analog hardware or digital processing. The digital approach has the advantages of reduced cost and expanded versatility, but requires a transformation from continuous time, where the original curves are defined, to discrete time. This transformation inevitably comes with some deviations from the continuous-time response near the Nyquist frequency. There are many established methods for discretizing continuous-time filters, and these vary in performance, computational cost, and inherent latency. In this work, several methods for performing this transformation are explored in the context of phonograph playback equalization, and the performance of each approach is quantified. This work is intended as a resource for anyone developing systems for digital playback equalization or similar applications that require approximating the response of a continuous-time filter digitally.
Moderators
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Speakers
avatar for Benjamin Thompson

Benjamin Thompson

PhD Student, University of Rochester
Authors
Tuesday October 8, 2024 3:00pm - 3:30pm EDT
1E04

3:00pm EDT

Immersive and interactive audio management and production for new entertainment venues.
Tuesday October 8, 2024 3:00pm - 3:45pm EDT
A wide range of new immersive audio applications like The Sphere, the COSM entertainment domes, theatres, live, museum and art Installations came up. These venues need new audio management as well as production tools. But the way to create content has not changed. Music and audio producers’ home is a regular studio, also because it’s not possible to get time on the sites for extensive mixing sessions. Based on close to 30 years immersive and binaural audio production experience, New Audio Technology provide those tools and will give an insight of mixing, management, and playback strategies for these applications.
Speakers
avatar for Tom Ammermann

Tom Ammermann

New Audio Technology
Grammy-nominated music producer, Tom Ammermann, began his journey as a musician and music producer in the 1980s.At the turn of the 21st Century, Tom produced unique surround audio productions for music and film projects as well as pioneering the very first surround mixes for headphones... Read More →
Tuesday October 8, 2024 3:00pm - 3:45pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

3:15pm EDT

The Devil in the Details
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
What happens when you realize in the middle of a mass digitization project that most of your video assets have multi-track production audio instead of finished mixed audio, and your vendor doesn't offer a service to address the issue? Digitizing the Carlton Pearson Collection for the Harvard Divinity School produced just such a conundrum. This workshop will walk through a case study of the process of identifying problems from vendor work, QC and production workflows that had to be put into place to correct the issues that were surfaced as the project progressed, including a look at the technology stack that was developed internally in response to these issues and the necessary solutions, including a full GUI video editor that was developed for QC of audio and video and for implementing mass top/tail editing of assets while offering individual edit decision points. From problem identification to audio mix to video trimming to close captioning using AI solutions and project deposit to preservation repositories, the project team had only 3 months to complete the work on just shy of 4000 assets.
Speakers
avatar for Kaylie Ackerman

Kaylie Ackerman

Head of Media Preservation, Harvard Library
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
1E16

3:15pm EDT

Audio Design Roundtable
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Join us at the AES NY 2024 convention for an exciting panel featuring top-tier audio equipment designers who have set new standards in the industry! Geared toward aspiring audio designers, students, and educators, this session will dive into the creative and technical processes behind groundbreaking audio gear used in studios, live sound, and beyond. Hear from industry leaders as they share career highlights and insights on staying ahead in a rapidly evolving field. In addition, students will learn about the various student design competitions presented by AES, offering unique opportunities to showcase their skills and gain recognition. Stick around for a Q&A session, where you'll have the chance to ask these leading experts your burning questions and gain valuable knowledge to power your future in audio design!
Speakers
avatar for Christoph Thompson

Christoph Thompson

Director of Music Media Production, Ball State University
Christoph Thompson is vice-chair of the AES audio education committee. He is the chair of the AES Student Design Competition and the Matlab Plugin Design Competition. He is the director of the music media production program at Ball State University. His research topics include audio... Read More →
avatar for Brecht De Man

Brecht De Man

Head of Research, PXL University of Applied Sciences and Arts
Brecht is an audio engineer with a broad background comprising research, software development, management and creative practice. He holds a PhD from the Centre for Digital Music at Queen Mary University of London on the topic of intelligent software tools for music production, and... Read More →
avatar for George Massenburg

George Massenburg

Associate Professor of Sound Recording, Massenburg Design Works
George Y. Massenburg is a Grammy award-winning recording engineer and inventor. Working principally in Baltimore, Los Angeles, Nashville, and Macon, Georgia, Massenburg is widely known for submitting a paper to the Audio Engineering Society in 1972 regarding the parametric equali... Read More →
avatar for Anthony Agnello

Anthony Agnello

Managing Director, Eventide
Tony Agnello was born in Brooklyn, graduated from Brooklyn Technical High School in 1966, received the BSEE from City College of NY in 1971, the MSEE from the City University of NY in 1974 followed by post graduate studies in Digital Signal Processing at Brooklyn’s Polytechnical... Read More →
MW

Marek Walaszek

General, Addicted To Music
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Stage

3:15pm EDT

Mary Campbell In Memoriam
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Mary Campbell – In Memoriam
Manager Electric Lady Studios & Unique Recording, Sony Studios Booking Admin.

Mary Culum-Campbell was a legend of the NY studio scene for decades. She came from Montana in 1982 at age 21, and with a forged backstage pass and some other tales of delinquency talked her way into a job in the shop at Electric Lady Studios with tech extraordinaire Sal Greco, who went on to build Paisley Park for Prince. Sal’s comment recalling her job interview at her memorial was ‘She’s trouble. She’ll fit right in’. Within six months Mary became the studio manager. A trailblazer with attitude who also loved the music around her as a genuine fan, Mary was one of the few women in a power position at a major recording studio in the 80’s. Working in an intense and occasionally cutthroat environment during a tumultuous period in both music and music technology, she kept the studios booked and partied with the best (and worst) of them while managing both Electric Lady and the equally electric Unique Recording. She also had a stint working at the massive Sony Music Studios. The panel discussion will cover some of Mary’s history as well some of the history of the console and technology changes she presided over. Speakers will include former chief technicians Jim Gillis and Brian Macaluso, as well as associate Tony Drootin and engineer extraordinaire Ron Saint Germain.

Held together by Elliot Kissileff , Co-Moderator Roey Shamir

https://helenafuneralhome.com/obituaries/mary-c-campbell-age-61/
Speakers
avatar for Angela Piva

Angela Piva

Angela Piva, Audio Pro/Audio Professor, highly skilled in all aspects of music & audio production, recording, mixing and mastering with over 35 years of professional audio engineering experience and accolades. Known as an innovator in sound technology, and for contributing to the... Read More →
avatar for Eliot Kissileff

Eliot Kissileff

Owner, Music and Media Preservation Services
After graduating from New York University’s Tisch School of the Arts with a BFA in film production, Eliot Kissileff attended the Institute of Audio Research before landing a job at Electric Lady Studios in 1996. With the goal of learning how to make vintage-sounding recordings for... Read More →
avatar for Roey Shamir

Roey Shamir

Mixmaster, INFX PRODUCTIONS INC.
• Freelance Audio Engineer, Mixer and Producer • Audio Engineer Mixer, Producer, Remixer for extensive major label records and filmproductions (Mix to Picture) • Mixed and Recorded music / albums for numerous RIAA multi platinum and Grammynominated... Read More →
avatar for Ron St. Germain

Ron St. Germain

Owner, Saint Sounds Inc & 'Saint's Place' Studio
Ron's career in the music business began in 1970. He learned the art of recording at two of America's busiest and best recording studios, Record Plant Studios and Media Sound Studios, both in NYC. Some of Ron's ‘colleagues’ during those formative years were Tony Bongiovi, Bob... Read More →
avatar for Brian Macaluso

Brian Macaluso

Owner, Clandestine Recording
A lifelong music technology nerd, musician and songwriter, recording engineer, studio owner and former Chief Tech for Electric Lady Studios, JSM Music, and Dreamhire Professional Audio Rentals in NYC.Owner of Clandestine Recording, a private project studio in Kingston NY.
avatar for Tony Drootin

Tony Drootin

Manager, Sound on Sound Studios
Tony Graduated with a degree in Music Performance Percussion in 1984. Upon graduating college he took a position as receptionist at Unique Recording Studios in Times Square. Tony worked at Unique for 13 years, 10 of which he managed the facility. While at Unique Recording he formed... Read More →

Tuesday October 8, 2024 3:15pm - 4:15pm EDT
1E07

3:15pm EDT

Mastery in Ambisonic Recording
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Ambisonic microphones offer a dynamic and multifaceted approach to capturing and creating immersive audio experiences. This workshop delves into the process of transforming multichannel A-format raw recordings into versatile Ambisonic B-format files. These files can then be decoded into various audio formats, including stereo, binaural, surround, and spatial 3D audio. This makes it an ideal technique for crafting head-tracking binaural audio for 360/3D videos and VR gaming environments.
Participants will gain hands-on experience with the Zoom H3-VR portable recorder and learn how to create Ambisonic recordings. The session will guide you through processing raw recordings using Reaper and the dearVR AMBI MICRO plugin and illustrating the workflow for integrating Ambisonic audio into 360/3D video and VR games for head-tracking binaural sound. Attendees are encouraged to bring their laptops and headphones, install the necessary software, and download the provided demo projects for an interactive experience. Additionally, the workshop will offer insights into capturing immersive audio with various Ambisonic microphones, including the new em64 Eigenmike Spherical Microphone Array.
Speakers
avatar for Ming-Lun Lee

Ming-Lun Lee

Professor of Electrical and Computer Engineering, University of Rochester
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
1E09

3:15pm EDT

Susan Rogers: Enveloping Masterclass
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Susan talks about her incredible carrier in science and music, working with artists such as Julia Darling, David Byrne, Prince, Laurie Anderson, Tevin Campbell and many more; all garnished with high resolution listening examples.

This masterclass series, featuring remarkable recording artists, is a chance to hear stereo and 3D music at its best; as we discuss important factors of production, distribution and reproduction. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly intimacy and AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Susan Rogers

Susan Rogers

Professor, Berklee Online
Susan Rogers holds a doctoral degree in experimental psychology from McGill University (2010). Prior to her science career, Susan was a multiplatinum-earning record producer, engineer, mixer and audio technician. She is best known for her work with Prince during his peak creative... Read More →
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
3D06

3:15pm EDT

The Art of Mixing in Stereo and ATMOS Simultaneously
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Join veteran audio engineer Matt Boudreau, who has mixed in ATMOS for Alanis Morissette, Green Day, and Deafheaven, as he delves into immersive audio mixing. In this session, Matt will discuss his approach to mixing in ATMOS and stereo simultaneously and provide a look into his Pro Tools mixing template. Drawing from his extensive experience and his Mixing in Stereo and ATMOS course, Matt will share practical tips for creating dynamic mixes that excel in both formats. Whether you're new to ATMOS or refining your skills, this presentation offers valuable insights for all audio professionals.
Speakers
MB

Matt Boudreau

An audio engineer at Matt Boudreau MIXING|MASTERING + Producer/Host/Author Working Class Audio Podcast.
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
1E06

3:15pm EDT

Copying and attributing training data in audio generative models
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
MOTIVATION: Generative AI for audio is quickly proliferating in both the commercial and open source software communities, However, there is still no technical consensus on how novel the generated audio data is, and what (if any) characteristics from the training data are commonly replicated. This workshop will explore existing technical approaches for how to detect memorized audio training data, how often it happens, and what characteristics of the audio are memorized. Additionally, we plan to share recent research on how audio similarity algorithms can be used to attribute audio samples produced by a generative model to specific audio samples in the training set.

FORMAT: this will be in the form of a panel discussion with three parts to it (listed below) – the purpose is to inform the AES community about the technical feasibility of detecting memorization and attributing training data in audio generative models, give audio examples of the results, an outlook for future developments, and solicit feedback from the audience.

• Memorization and generalization in deep learning
• Searching for memorized training data in audio generative models
• Training data attribution using audio similarity measures
Speakers
avatar for Gordon Wichern

Gordon Wichern

Principal Research Scientist - Speech and Audio Team, MERL, Sr. Principal Research Scientist
Audio signal processing and machine learning resarcher
avatar for Yuki Mitsufuji

Yuki Mitsufuji

Lead Research Scientist/VP of AI Research, Sony AI
avatar for Philipp Lengeling

Philipp Lengeling

Senior Counsel, RafterMarsh
Philipp G. Lengeling, Mag. iur., LL.M. (New York), Esq. is an attorney based in New York (U.S.A.) and Hamburg (Germany), who is heading the New York and Hamburg based offices for RafterMarsh, a transatlantic boutique law firm (California, New York, U.K., Germany).
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
1E15

3:15pm EDT

Reporting from the Frontlines of the Recording Industry: Hear from Studio Owners and Chief Engineers on Present / Future Trends, Best Practices, and the State of the Recording Industry
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
Panel Description:
Join us for an in-depth, candid discussion with prominent NYC studio owners on the state recording industry of today and where it might go tomorrow. This panel will explore trends in studio bookings, predictions for the future, and the evolving landscape of recording studios. Our panelists will share their experiences of owning and managing studios, offering valuable insights for engineers on essential skills and best practices for success in today’s recording environment. We'll also discuss the importance of high-quality recording, and how these practices are crucial in maintaining the soul of music against the backdrop of AI and digital homogenization.

Topics also include:
What are studios looking for from aspiring engineers, and what does it take to work in a major market studio?
Studio best practices from deliverables to client relations
To Atmos or not to Atmos?
Proper crediting for engineers and staff
Working with record labels and film studios
Can recording and mix engineers find the same labor protections as the film industry?

Panelists (in alphabetical order)

Amon Drum, Owner / Engineer, The Bridge Studio (https://www.bridgerecordingstudio.com/)
Amon Drum is a recordist, producer, acoustic designer, owner, Chief Engineer, and designer of The Bridge Studio in Williamsburg, Brooklyn. He has an expertise in both analog and digital recording techniques. Amon specializes in the recording of non-western acoustic instruments, large ensembles, and live music for video. Amon is also a percussionist, having trained with master musicians including M’Bemba Bangora and Mamady Keita in Guinea West Africa, and brings this folkloric training to his recording & productions. Working with a wide variety of artists and genres from Jason Moran to Adi Oaisis, Run The Jewels, to Ill Divo as well as many Afro-Cuban ensembles from the diaspora.

Ben Kane, Owner, Electric Garden (https://electricgarden.com/)
Ben Kane is a Grammy Award winning recording and mix engineer, producer, and owner of Electric Garden, a prominent recording studio known for its focus on excellence in technical and engineering standards, and a unique handcrafted ambiance. Kane is known for his work with D'Angelo, Emily King, Chris Dave, and PJ Morton.

Shahzad Ismaily, Owner / Engineer, Figure 8 Recording (https://www.figure8recording.com/)
Shahzad Ismaily is a Grammy-nominated multi-instrumentalist, composer, owner and engineer at Figure 8 Recording in Brooklyn. Renowned for his collaborative spirit and eclectic musical range, Shahzad has contributed his unique sonic vision to projects spanning various genres and artists worldwide.

Zukye Ardella, Partner / Engineer, s5studio (https://www.s5studiony.com/)
Zukye is a New York City based certified gold audio engineer, music producer and commercial studio owner. In 2015, she began her professional career at the original s5studio located in Brooklyn, New York. Zukye teamed up with s5studio’s founder (Sonny Carson) to move s5studio to its current location in Chelsea, Manhattan. Over the years, Ardella has been an avid spokesperson for female empowerment organizations such as Women’s Audio Mission and She is the Music. Talents she’s worked with include NeYo, WizKid, Wale, Conway The Machine, A$AP Ferg, Lil Tecca, AZ, Dave East, Phillip Lawrence, Tay Keith, Lola Brooke, Princess Nokia, Vory, RMR, Yeat, DJ KaySlay, Kyrie Irving, Maliibu Mitch, Flipp Dinero, Fred The Godson, Jerry Wonda, ASAP 12vy and more.


--

Moderator:

Mona Kayhan, Owner, The Bridge Studio (https://www.bridgerecordingstudio.com/)
Mona is a talent manager for Grammy award winning artists, and consults for music and media companies in marketing and operations. She has experience as a tour manager, an international festival producer, and got her start in NYC working at Putumayo World Music. As an owner of The Bridge Studio, Mona focuses on client relationships, studio operations, and strategic partnerships. She also has an insatiable drive to support and advocate for the recording arts industry.
Tuesday October 8, 2024 3:15pm - 4:15pm EDT
1E08

3:30pm EDT

Leveraging TSN Protocols to Support AES67: Achieving AVB Quality with Layer 3 Benefits
Tuesday October 8, 2024 3:30pm - 3:50pm EDT
This paper investigates using Time-Sensitive Networking (TSN) protocols, particularly from Audio Video Bridging (AVB), to support AES67 audio transport. By leveraging the IEEE 1588 Level 3 Precision Time Protocol (PTP) Media Profile, packet scheduling, and bandwidth reservation, we demonstrate that AES67 can be transported with AVB-equivalent quality guarantees while benefiting from Layer 3 networking advantages. The evolution of professional audio networking has increased the demand for high-quality, interoperable, and efficiently managed networks. AVB provides robust Layer 2 delivery guarantees but is limited by Layer 2 constraints. AES67 offers Layer 3 interoperability but lacks strict quality of service (QoS) guarantees. This paper proposes combining the strengths of both approaches by using TSN protocols to support AES67, ensuring precise audio transmission with Layer 3 flexibility. TSN extends AVB standards for time synchronization, traffic shaping, and resource reservation, ensuring low latency, low jitter, and minimal packet loss. AES67, a standard for high-performance audio over IP, leverages ubiquitous IP infrastructure for scalability and flexibility but lacks the QoS needed for professional audio. Integrating TSN protocols with AES67 achieves AVB's QoS guarantees in a Layer 3 environment. IEEE 1588 Level 3 PTP Media Profile ensures precise synchronization, packet scheduling reduces latency and jitter, and bandwidth reservation prevents congestion. Experiments show that TSN protocols enable AES67 to achieve latency, jitter, and packet loss performance on par with AVB, providing reliable audio transmission suitable for professional applications in modern, scalable networks.
Moderators
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Speakers
avatar for Nicolas Sturmel

Nicolas Sturmel

Directout GmbH
Authors
Tuesday October 8, 2024 3:30pm - 3:50pm EDT
1E04

3:30pm EDT

A cepstrum analysis approach to perceptual modelling of the precedence effect
Tuesday October 8, 2024 3:30pm - 4:00pm EDT
The precedence effect describes our ability to perceive the spatial characteristics of lead and lag sound signals. When the time delay between the lead and lag is sufficiently small we will cease to hear two distinct sounds, instead perceiving the lead and lag as a single fused sound with its own spatial characteristics. Historically, precedence effect models have had difficulty differentiating between lead/lag signals and their fusions. The likelihood of fusion occurring is increased when the signal contains periodicity, such as in the case of music. In this work we present a cepstral analysis based perceptual model of the precedence effect, CEPBIMO, which is more resilient to the presence of fusions than its predecessors. To evaluate our model we employ four datasets of various signal types, each containing 10,000 synthetically generated room impulse responses. The results of the CEPBIMO model are then compared against results of the BICAM. Our results show that the CEPBIMO model is more resilient to the presence of fusions and signal periodicity than previous precedence effect models.
Moderators Speakers
avatar for Jeramey Tyler

Jeramey Tyler

Samtec
Jeramey is in the 3rd person. So it goes.
Authors
avatar for Jeramey Tyler

Jeramey Tyler

Samtec
Jeramey is in the 3rd person. So it goes.
Tuesday October 8, 2024 3:30pm - 4:00pm EDT
1E03

3:50pm EDT

Harnessing Diffuse Signal Processing (DiSP) to Mitigate Coherent Interference
Tuesday October 8, 2024 3:50pm - 4:10pm EDT
Coherent sound wave interference is a persistent challenge in live sound reinforcement, where phase differences between multiple loudspeakers lead to destructive interference, resulting in inconsistent audio coverage. This review paper presents a modern solution: Diffuse Signal Processing (DiSP), which utilizes Temporally Diffuse Impulses (TDIs) to mitigate phase cancellation. Unlike traditional methods focused on phase alignment, DiSP manipulates the temporal and spectral characteristics of sound, effectively diffusing coherent wavefronts. TDIs, designed to spread acoustic energy over time, are synthesized and convolved with audio signals to reduce the likelihood of interference. This process maintains the original sound’s perceptual integrity while enhancing spatial consistency, particularly in large-scale sound reinforcement systems. Practical implementation methods are demonstrated, including a MATLAB-based workflow for generating TDIs and optimizing them for specific frequency ranges or acoustic environments. Furthermore, dynamic DiSP is introduced as a method for addressing interference caused by early reflections in small-to-medium sized rooms. This technique adapts TDIs in real-time, ensuring ongoing decorrelation in complex environments. The potential for future developments, such as integrating DiSP with immersive audio systems or creating dedicated hardware for real-time signal processing, is also discussed.
Moderators
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Speakers
TS

Tommy Spurgeon

Physics Student & Undergraduate Researcher, University of South Carolina
Authors
TS

Tommy Spurgeon

Physics Student & Undergraduate Researcher, University of South Carolina
Tuesday October 8, 2024 3:50pm - 4:10pm EDT
1E04

4:00pm EDT

Categorical Perception of Neutral Thirds Within the Musical Context
Tuesday October 8, 2024 4:00pm - 4:30pm EDT
This paper investigates the contextual recognition of neutral thirds in music by integrating real-world musical context into the study of categorical perception. Traditionally, categorical perception has been studied using isolated auditory stimuli in controlled laboratory settings. However, music is typically experienced within a circumstantial framework, significantly influencing its reception. Our study involved musicians from various specializations who listened to precomposed musical fragments, each concluding with a 350-cent interval preceded by different harmonic contexts. The fragments included a monophonic synthesizer and orchestral mockups, with contexts such as major chords, minor chords, a single pitch, neutral thirds, and natural fifths. The results indicate that musical context remarkably affects the recognition of pseudotonal chords. Participants' accuracy in judging interval size varied based on the preceding harmonic context. A statistical analysis was conducted to determine if there were significant differences in the neutral third perception across the different harmonic contexts. The test led to the rejection of the null hypothesis: the findings underscore the need to consider real-world listening experiences in research on auditory processing and cognition.
Moderators Speakers Authors
Tuesday October 8, 2024 4:00pm - 4:30pm EDT
1E03

4:00pm EDT

Revolutionizing Loudspeaker Design with Ultra Thin Glass Technology
Tuesday October 8, 2024 4:00pm - 4:45pm EDT
Discover the groundbreaking potential of our latest Ultra Thin Glass (UTG) technology, engineered for speaker diaphragm applications. This innovation marries toughness, flexibility, and a distinctive finish with full ESG compliance. With thicknesses as fine as 25μm, UTG seamlessly integrates into earphones, headphones, micro-speakers, and a wide range of loudspeakers. Join us to explore how this cutting-edge material can elevate your audio designs to new heights.
Speakers
KC

Kwunkit Chan

Glass Acoustic Innovations Co., LTD
Tuesday October 8, 2024 4:00pm - 4:45pm EDT
3D04 (AES and AIMS Alliance Program Room)

4:30pm EDT

Mastering for Vinyl
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
Women's Audio Mission presents a panel discussion with award-winning mastering engineers and vinyl cutters who will talk about the intricacies of the vinyl format and the strategies used in mastering modern music to vinyl including how to address factors such as sibilance, managing low frequency stereo information, sequencing strategies, program length and volume, etc.

Moderator: Terri Winston

Panelists:

Kim Rosen - Kim Rosen is a GRAMMY- nominated mastering engineer who works out of her own “Knack Mastering” located in Ringwood, NJ, USA. Recordings Kim has mastered have garnered twenty-six Grammy nominations with five winning across multiple categories including Bonnie Raitt’s Song of the Year. She has mastered projects for Wynonna Judd, Johnny Cash, Aimee Mann, The Milk Carton Kids, Allison Russell, Superdrag, and Flogging Molly.

Margaret Luthar - Margaret Luthar is a Grammy-nominated Mastering Engineer who has worked at Chicago Mastering Service (where she learned to cut vinyl) and at Welcome to 1979 in Nashville. Currently a Broadcast Recording Technician at NPR and a freelance mastering engineer based in Los Angeles. She has mastered thousands of projects including The Lumineers, Tinashe, Spiritualized, Soccer Mommy and has significant experience in audio restoration, from her work at the Norwegian Institute of Recorded Sound.

Piper Payne - Piper Payne is a Mastering Engineer based in Nashville, TN. Piper has mastered a wide variety of music including nationally renowned artists such as Janis Ian (Best Folk Album Grammy Nomination 2023), Dolly Parton, Third Eye Blind, LeAnn Rimes, The Go-Go’s, Madame Gandhi, and many more. After moving her mastering studio to Nashville from the San Francisco Bay Area, she ventured into vinyl manufacturing, opening Physical Music Products.

Moderated by Terri Winston, Executive Director of Women's Audio Mission. Women's Audio Mission (WAM) is dedicated to closing the chronic gender gap in the music, recording and technology industries. Less than 5% of the people creating and shaping the sounds that make up the soundtrack of our lives are women and gender-expansive people. WAM is “changing the face of sound” by providing training, mentoring and career advancement that inspires the next generation of women and gender-expansive music producers, recording engineers and audio technologists and radically changes the narrative in these industries.
Speakers
avatar for Terri Winston

Terri Winston

Executive Director, Women's Audio Mission
Women's Audio Mission (WAM) is dedicated to closing the chronic gender gap in the music, recording and technology industries. Less than 5% of the people creating and shaping the sounds that make up the soundtrack of our lives are women and gender-expansive people. WAM is “changing... Read More →
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
1E07

4:30pm EDT

ADM-OSC 1.0
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
ADM-OSC is an industry initiative aimed at standardising Object-Based Audio (OBA) positioning data by implementing the Audio Definition Model (ADM) over Open Sound Control (OSC). As immersive audio gains traction across various industries, from music streaming to gaming, and from live sound to broadcasting, the Audio Definition Model (ADM) is becoming a popular standard for metadata. This includes Serial ADM for broadcast and ADM BWF or XML files for studio use.

The ADM-OSC workgroup was formed four years ago to bridge the gap between immersive live and studio ecosystems. It now includes leading developers and manufacturers who aim to facilitate the sharing of audio object metadata across different environments, from studios to broadcasts to live performances.

Since its initial draft implementation, ADM-OSC has been supported by various audio industry tools, including live rendering engines, digital audio workstations (DAWs), controllers, live tracking systems, and media server solutions. It is currently being deployed in both live and studio productions, with increasing interest from technology developers wanting to join and implement this standard.

ADM-OSC 1.0 is now the published specification, aiming to provide a basic interoperability layer between Object Editors and Object Renderers.

This presentation and workshop will take a dive on ADM-OSC and will cover:
- The origins of ADM-OSC
- Presentation of ADM-OSC 1.0 specification
- Use case/application demonstration
- DAW object positional data to external rendering engine (s)
- Controllers’ data to object panner in DAW for automation recording
- Live tracking (actors, artist) positional data to live rendering engine (s)
- Plugin fest 2023 and 2024 report
- Future considerations
- Application-specific subgroups such as broadcast, VR/Gaming, live rendering, show control
Speakers
avatar for Michael Zbyszynski

Michael Zbyszynski

Software Development Engineer, L-Acoustics
Michael Zbyszyński is musician, researcher, teacher and developer in the field of contemporary electroacoustic music. He is currently part of the Creative Technologies R&D group at L-Acoustics. As a musician, his work spans from brass bands to symphony orchestras, including composition... Read More →
avatar for Hugo Larin

Hugo Larin

Senior Mgr. Business Development | FLUX:: GPLM, Harman International
Hugo Larin is a key collaborator to the FLUX: SPAT Revolution project and has deep roots in audio mixing, design and operation, as well as in networked control and data distribution. He leads the FLUX:: business development at HARMAN. His recent involvements and interests include object-based spatial audio mixing workflows, interoperability... Read More →
avatar for Mathieu Delquignies

Mathieu Delquignies

Education & Application Support France, d&b audiotechnik
Mathieu holds a Bachelors's degree in applied physics from Paris 7 University and Master's degree in sound engineering from ENS Louis Lumière in 2003. He has years of diverse freelance mixing and system designer experiences internationally, as well as loudspeakers, amplifiers, dsp... Read More →
avatar for Lucas Zwicker

Lucas Zwicker

Senior Director, Workflow and Integration, CTO Office, Lawo AG
Lucas joined Lawo in 2014, having previously worked as a freelancer in the live sound and entertainment industry for several years. He holds a degree in event technology and a Bachelor of Engineering in electrical engineering and information technology from the University of Applied... Read More →
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
1E15

4:30pm EDT

Upmix and Format Conversion of Multichannel Audio: An Opportunity for AI-Based Breakthroughs?
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
The demand for the production, distribution and playback of audio content in surround or immersive multichannel formats and flexible playback setups has continued to grow in the cinema, broadcast and music industries. This trend motivates the continued development of multichannel audio signal processing methods for the conversion of recordings between different multichannel and spatial audio formats or layouts, including down- or up-mixing. These operations leverage a well researched frequency-domain processing framework nowadays commonly referred to as parametric spatial audio signal processing, to solve challenges including the primary-ambient decomposition of audio recordings. Bringing ML/AI techniques to bear in the development of novel spatial audio analysis/synthesis methods can unlock practical applications with transformative impact. In this workshop, we will review and discuss the fundamentals and recent progress in these methods.
Speakers
avatar for Jean-Marc Jot

Jean-Marc Jot

Founder and Principal, Virtuel Works LLC
Spatial audio and music technology expert and innovator. Virtuel Works provides audio technology strategy, IP creation and licensing services to help accelerate the development of audio and music spatial computing technology and interoperability solutions.
avatar for Gordon Wichern

Gordon Wichern

Principal Research Scientist - Speech and Audio Team, MERL, Sr. Principal Research Scientist
Audio signal processing and machine learning resarcher
avatar for Sunil Bharitkar

Sunil Bharitkar

Samsung Research America
avatar for Carlos Freitas

Carlos Freitas

CMO - Mastering Engineer, Spatial9
9 Latin Grammy nominees in the Best Engineering category, and has been the mastering engineer in 34 Grammy award winning records, with more than 100 nominations.The albums include artists such as Nathan East, Andres Cepeda, Ines Gaviria, João Gilberto, Caetano Veloso, Gilberto Gil... Read More →
avatar for Alan Silva

Alan Silva

CTO, Spatial9
Alan Silva is a well-seasoned Researcher and Engineer who focuses on leveraging machine learning algorithms and distributed computing to create innovative solutions. With a solid commitment to open-source development, Alan actively contributes to collaborative projects that drive... Read More →
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
1E06

4:30pm EDT

Psychoacoustics for Immersive Productions
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
3D audio has enormous potential to emotionally touch the audience: The potent effect occurs when the auditory system is given the illusion of being in a natural environment. When this is the case with impressive music, everyone gets goosebumps. Psychoacoustics forms the basis for remarkable results in music productions.
In the first part, Konrad Strauss explains the basics of psychoacoustics in the context of music production:
• How immersive differs from stereo
• Sound localization and perception
• The eye/ear/brain link
• Implications for recording and mixing in immersive
• Transitioning from stereo to immersive: Center speaker, LFE, working with the diffuse
surround field and height channels.
In the second part, Lasse Nipkow introduces the quasi-binaural spot miking technique he uses to capture the beautiful sound of acoustic instruments during his recordings. He explains the strategy for microphone placement and shows, using stereo and 3D audio sound examples, the potential of these signals for immersive productions.
This first contribution is linked to a second, subsequent presentation by Lasse Nipkow and Ronald Prent: ‘Tools for Impressive Immersive Productions’.
Speakers
avatar for Lasse Nipkow

Lasse Nipkow

CEO, Silent Work LLC
Since 2010, Lasse Nipkow has been a renowned keynote speaker in the field of 3D audio music production. His expertise spans from seminars to conferences, both online and offline, and has gained significant popularity. As one of the leading experts in Europe, he provides comprehensive... Read More →
avatar for Konrad Strauss

Konrad Strauss

Professor, Indiana University Jacobs School of Music
Konrad Strauss is a Professor of Music in the Department of Audio Engineering and Sound Production at Indiana University’s Jacobs School of Music. He served as department chair and director of Recording Arts from 2001 to 2022. Prior to joining the faculty of IU, he worked as an... Read More →
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
1E16

4:30pm EDT

Morten Lindberg: Enveloping Masterclass
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
Morten details 3D music recordings from his 2L catalogue, including high resolution listening examples from new releases.

This masterclass series, featuring remarkable recording artists, is a chance to hear 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Morten Lindberg

Morten Lindberg

Producer and Engineer, 2L (Lindberg Lyd, Norway)
Recording Producer and Balance Engineer with 43 GRAMMY-nominations, 35 of these in craft categories Best Engineered Album, Best Surround Sound Album, Best Immersive Audio Album and Producer of the Year. Founder and CEO of the record label 2L. Grammy Award-winner 2020.
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
3D06

4:30pm EDT

How Do We Embrace AI and Make It Work For Us in Audio? Hey AI, Get Yer Stinkin’ Hands Offa My Job! AES and SMPTE Joint Exploration
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
In this timely panel let’s discuss some of the pressing matters AI confronts us with in audio, and how we can turn a perceived foe into an ally.
We’ll discuss issues including:
During production, AI cannot deal with any artistic issues related to production and engineering, most of which depend on personal interaction - as well as perception.

In post-production, AI could be of use in repetitive tasks: making a track conform to a click-track and maintain proper pitch, perform pitch correction on a track, deal with extraneous clicks (without removing important vocal consonants), perform ambience-matching (particularly on live recordings), to name a few. Can AI running in the background on our DAW build macros for us?

The more we can use it as a tool for creativity and enhance our revenue streams the more it becomes a practical, positive approach. Many composers are using it to create or enhance their musical ideals almost instantaneously. The key here is that it is they are our ideas that AI adds to.

How do we embrace AI, adapt to it, and help it to adapt to us? Can we get to the point where we incorporate it as we have past innovations rather than fear it? How do we take control of AI instead of AI taking control of us?

What should we, the audio community, be asking AI to do?
Speakers
avatar for Gary Gottlieb

Gary Gottlieb

AES President-Elect, Mendocino College
President-Elect, Co-Chair of the Events Coordination Committee, Chair of the Conference Policy Committee, and former Vice President of the Eastern Region, US and Canada; AES Fellow, Engineer, Author, Educator and Guest Speaker Gary Gottlieb refers to himself as a music generalist... Read More →
avatar for Lenise Bent

Lenise Bent

Producer/Engineer/Editor/AES Governor, Soundflo Productions
Audio Recording HistoryWomen and Diversity in AudioAnalog Tape RecordingPost Production/Sound Design/FoleyVinyl RecordsAudio Recording Archiving, Repair and PreservationBasic and Essential Recording TechniquesOpportunities in the Audio IndustryAudio Adventurers
avatar for Franco Caspe

Franco Caspe

Student, Queen Mary University of London
I’m an electronic engineer, a maker, hobbyist musician and a PhD Student at the Artificial Intelligence and Music CDT at Queen Mary University of London. I have experience in development of real-time systems for applications such as communication, neural network inference, and DSP... Read More →
avatar for Soumya Sai Vanka

Soumya Sai Vanka

PhD Researcher, Queen Mary University of London
I am a doctoral researcher at the Centre for Digital MusicQueen Mary University of London under the AI and Music Centre for Doctoral Training Program. My research focuses on the design of user-centric context-aware AI-based tools for music production. As a hobbyist musician and producer myself, I am interested in developing tools that can support creativity and collaboration resulting in emergence and novelty. I am also interested... Read More →
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
1E08

4:30pm EDT

Sonic Mastery - A Room of One's Own
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
This panel brings together a group of audio engineers to explore their experiences of navigating and shaping the audio industry as Black women. As they delve into the creation and significance of safe spaces, the panelists will discuss the importance of community in a field where representation is often lacking. Through sharing stories of unique and impactful projects, the conversation will highlight the creativity and resilience required to thrive in a male-dominated industry. Additionally, we will celebrate the legacy of unsung women producers and their groundbreaking contributions to music and production. This session is a discussion by and tribute to the talent and influence of Black women in audio engineering, both past and present.
Speakers
avatar for Eve Horne

Eve Horne

Peak Music
avatar for Leslie Gaston-Bird

Leslie Gaston-Bird

Owner, Mix Messiah Productions
Leslie Gaston-Bird (AMPS, MPSE) is author of the book "Women in Audio", part of the AES Presents series and published by Focal Press (Routledge). She is a voting member of the Recording Academy (The Grammys®) and its P&E (Producers and Engineers) Wing. Currently, she is a freelance... Read More →
avatar for Gloria Kaba

Gloria Kaba

Studio Manager, Power Station at BerkleeNYC
Gloria Kaba is a Ghanaian-American sound engineer, producer, mixer, and writer with over a decade of experience in the studio, often operating under the moniker Redsoul. She’s worked on A Tribe Called Quest’s final album We Got It From Here...Thank You For Your Service and Solange’s... Read More →
avatar for Ebonie Smith

Ebonie Smith

Ebonie Smith is a celebrated music producer, audio engineer, and singer-songwriter, based in the vibrant hub of Los Angeles. As a prominent figure in the industry, she currently holds the esteemed roles of senior audio engineer and producer at Atlantic Records. Ebonie's remarkable... Read More →
Tuesday October 8, 2024 4:30pm - 5:30pm EDT
Stage

5:00pm EDT

Student Mixer at Telefunken Booth
Tuesday October 8, 2024 5:00pm - 6:00pm EDT
Open to anyone with a student badge.
Tuesday October 8, 2024 5:00pm - 6:00pm EDT
Booth 335

6:00pm EDT

Tech Tour: The Sonic Room at Amazon Studios
Tuesday October 8, 2024 6:00pm - 8:00pm EDT
All Tech Tours are full. Know Before You Go emails were sent to accepted Tech Tour registrants with information about the tour.  

Duration: 2 hours
Capacity: 50 (Once capacity is reached, registrants will be added to a wait list)

Tour Description
Come and check out the new Sonic Studio126 at Amazon Music in Brooklyn and get to meet multi-Grammy Award Winning producer and engineer, Mr Sonic. Take a tour of the various studios, writer's rooms and event spaces at Amazon Music's Brooklyn location and finish the evening with a mixer to meet and network with other creatives and professionals.

Arrival Instructions:  Arrive at the South Lobby of 25 Kent Avenue, Brooklyn, NY, and check in by presenting your full name and ID at the front desk of security. You will be heading to the 7th floor (Amazon Music) for the event.
Tuesday October 8, 2024 6:00pm - 8:00pm EDT
Offsite
 
Wednesday, October 9
 

8:00am EDT

Attendee Registration
Wednesday October 9, 2024 8:00am - 6:00pm EDT
Wednesday October 9, 2024 8:00am - 6:00pm EDT
Crystal Palace Hall

9:00am EDT

Student Recording Competition Category 1
Wednesday October 9, 2024 9:00am - 10:00am EDT
Traditional Acoustic Recording finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
avatar for Martha DeFrancisco

Martha DeFrancisco

Associate Professor, McGill University
Martha De Francisco is Associate Professor of Sound Recording at McGill University in Montreal, Canada. A native of Colombia, she has established a prominent international career for over thirty years as a record producer/engineer for classical music, frequently invited to produce... Read More →
avatar for Kseniya Kawko

Kseniya Kawko

Tonmeister, msm-studios & Jazz Club Unterfahrt
Kseniya Kawko is a producer and recording engineer specialized in classical music and jazz. She holds Master of Music degrees from two world-renowned audio programs: Sound Recording, McGill University (Montréal, Canada) and Musikregie / Tonmeister, Hochschule für Musik Detmold (Germany... Read More →
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E07

9:00am EDT

New Directions in Modular Synthesizer Design
Wednesday October 9, 2024 9:00am - 10:00am EDT
With the explosive growth of modular instruments in the Eurorack format, synthesizer modules have become more sophisticated pointing to new directions in synthesis capabilities as well as real time performance control. This panel will consider some new developments in the field from leading manufacturers.
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
avatar for Leon Gruenbaum

Leon Gruenbaum

Consultant, Eventide
Hello. I am here to present the Misha, Eventide’s Misha Eurorack instrument and sequencer, for which I served as a consultant
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E09

9:00am EDT

Sound Design for Multi-Participant Extended Reality Experiences
Wednesday October 9, 2024 9:00am - 10:00am EDT
This workshop explores the unique challenges and opportunities in sound design and technical implementation for multi-participant extended reality (XR) experiences. Participants will learn techniques for creating immersive and interactive spatial audio to enhance presence and facilitate communication and collaboration in multi-participant XR environments. Through demonstrations and an expert-led panel discussion, attendees will gain insights into creating auditory experiences in XR.
Speakers
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
avatar for Yi Wu

Yi Wu

PhD Student, New York University
avatar for Parichat Songmuang

Parichat Songmuang

Studio Manager/PhD Student, New York University
Parichat Songmuang graduated from New York University with her Master of Music degree in Music Technology at New York University and Advanced Certificate in Tonmeister Studies. As an undergraduate, she studied for her Bachelor of Science in Electronics Media and Film with a concentration... Read More →
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E08

9:00am EDT

Unlocking New Dimensions: Producing Music in Immersive Audio
Wednesday October 9, 2024 9:00am - 10:00am EDT
The field of audio production is always evolving. Now with immersive audio formats becoming more and more prominent, we should have a closer look at what possibilities come with it from a technical but most importantly from an artistic and musical standpoint.
In our Workshop, "Unlocking New Dimensions: Producing Music in Immersive Audio," we demonstrate how immersive audio formats can bring an artist's vision to life and how the storytelling in the music benefits from them.
In order to truly change the way people listen to music and provide an immersive experience, we must transform how we write and produce music, using immersive formats not just as a technical advancement but as a medium to create new art.
In this session, we will explore the entire production process, from recording to the final mix, and master with a focus on how one can create a dynamic and engaging listening experience with immersive formats like Dolby Atmos. We believe that immersive audio is more than just a technical upgrade—it's a new creative canvas. Our goal is to show how, by fully leveraging a format like Dolby Atmos, artists and producers can create soundscapes that envelop the listener and add new dimensions to the storytelling of music.

Philosophy

Artists often feel disconnected from the immersive production process. They rarely can give input on how their music is mixed in this format, leading to results that may not fully align with their artistic vision. At High Tide, we prioritize artist involvement, ensuring they are an integral part of the process. We believe that their input is crucial for creating an immersive experience that truly represents their vision. We will share insights and examples from our collaborations with artists like Amistat, an acoustic folk duo, and Tinush, an electronic music producer known for his attention to detail. These case studies will illustrate how our method fosters creativity and produces superior immersive audio experiences.

New workflows need new tools

A significant pain point in current immersive productions is the tendency to use only a few stems, which often limits the immersive potential. This often happens because the process of exporting individual tracks and preparing a mixing session can be time-consuming and labor-intensive. We will address these challenges in our presentation. We have developed innovative scripts and workflows that streamline this process, allowing us to work with all available tracks without the typical hassle. This approach not only enhances the quality of the final mix but also retains the intricate details and nuances of the original recordings.
Our workshop is designed to be interactive, with opportunities for attendees to ask questions throughout. We will provide real-world insights into our ProTools sessions, giving participants a detailed look at our Dolby Atmos mixing process. By walking through the entire workflow, from recording with Dolby Atmos in mind to the final mix, attendees will gain a comprehensive understanding of the steps involved and the benefits of this approach to create an engaging and immersive listening experience.
Speakers
avatar for Lennart Damann

Lennart Damann

Founder / Engineer, High Tide - Immersive Audio
avatar for Benedikt Ernst

Benedikt Ernst

High Tide - Immersive Audio
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E06

9:00am EDT

Groups, Matrices, and DCAS
Wednesday October 9, 2024 9:00am - 10:00am EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E05

9:00am EDT

Digital archiving and file management 101
Wednesday October 9, 2024 9:00am - 10:30am EDT
This hands-on tutorial will be an overview of digital archiving and file management concepts and methodologies. It will focus on best practices for file naming, folder structure and DAW session creation, preparation of DAW sessions and associated assets for long-term archiving, an overview of strategies and software solutions for backup and long-term archiving, and an survey of industry standards for digital archiving and deliverables. It will include demonstrations of archiving and backup software and asset preparation.

While this tutorial will be student-focused, anyone who wants to learn more about digital archiving and file management is welcome to attend.
Speakers
avatar for Konrad Strauss

Konrad Strauss

Professor, Indiana University Jacobs School of Music
Konrad Strauss is a Professor of Music in the Department of Audio Engineering and Sound Production at Indiana University’s Jacobs School of Music. He served as department chair and director of Recording Arts from 2001 to 2022. Prior to joining the faculty of IU, he worked as an... Read More →
Wednesday October 9, 2024 9:00am - 10:30am EDT
1E15

9:00am EDT

Applications of Artificial Intelligence and Machine Learning in Audio Quality Models, Part II
Wednesday October 9, 2024 9:00am - 10:30am EDT
The aim of this workshop is to provide participants with hands-on expertise in utilizing machine learning for audio quality modeling.

This workshop is a follow-up from the initial edition presented at the 156th AES Convention. The workshop expands on its first edition by including the participation of new panelists working at cutting-edge technology with multimedia industry leaders. Particularly, very recent advancements on deep learning for auditory perception will be presented, as well as a new open audio quality dataset for informing auditory models in audio quality metrics.

Machine learning techniques have been instrumental in understanding audio quality perception, revealing hidden relationships in subject response data to enhance device and algorithm development. Moreover, accurate quality models can be used in AI systems to predict audio quality -- a critical part of customer experience -- in situations where using human subjects is costly (e.g., in day-to-day product development) or impractical such as in audio transmission network monitoring and informing deep learning audio algorithms.

The complex nature of quality perception requires users and developers of these models to possess specific domain knowledge that extends beyond the general machine learning set of skills. This expertise includes experimental design in the domain of subjective audio quality assessment, data collection, data augmentation and filtering, in addition to model design and cross-validation.

The workshop aims to shed light on historical approaches to addressing these challenges by marrying aspects of machine learning and auditory perception knowledge. Furthermore, it will provide insights into state-of-the-art techniques and current related research topics in data-driven quality perception modelling.

The main goal of the workshop is to offer its attendees —- whether users or developers —- the necessary tools to assess the suitability of existing ML-based quality models for their use case.
Speakers
avatar for Pablo Delgado

Pablo Delgado

Fraunhofer IIS
Pablo Delgado is part of the scientific staff of the Advanced Audio Research Group at the Fraunhofer Institute for Integrated Circuits (IIS) in Erlangen, Germany. He specializes in psychoacoustics applied to audio and speech coding, as well as machine learning applications in audio... Read More →
avatar for Jan Skoglund

Jan Skoglund

Google
Jan Skoglund leads a team at Google in San Francisco, CA, developing speech and audio signal processing components for capture, real-time communication, storage, and rendering. These components have been deployed in Google software products such as Meet and hardware products such... Read More →
avatar for Phill Williams

Phill Williams

Audio Algorithms, Netflix
Phill is member of Netflix's Audio Algorithms team, working on all aspects of the audio delivery toolchain to get the best possible sound to everybody who watches Netflix, all over the world, all of the time.Prior to working at Netflix, Phill worked at Dolby Laboratories, as a contributor... Read More →
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Wednesday October 9, 2024 9:00am - 10:30am EDT
1E16

9:30am EDT

Acoustic modeling and designing of emergency sound systems in road tunnels
Wednesday October 9, 2024 9:30am - 9:50am EDT
Road tunnels are subject to strong regulations and high life-safety standards. Requirements importantly include the speech intelligibility of the emergency sound system. However, a road tunnel is an acoustically challenging environment due to extreme reverberation, high noise floors, and its tube-like geometry. Designing an adequate sound system is a challenge and requires extensive acoustic simulation. This article summarizes recent design work on several major road tunnel projects and gives practical guidelines for the successful completion of similar projects. The project includes several tunnels, each of several kilometers’ length with one to five lanes, transitions and sections, having a total length of 33 km. For each tunnel, first a working acoustic model had to be developed before the sound system itself could be designed and optimized. On-site measurements were conducted to establish data for background noise including jet fans and various traffic situations. Critical environmental parameters were measured and reverberation times were recorded using large balloon bursts. Sprayed concrete, road surface, as well as other finishes were modeled or estimated based on publicly available data for absorption and scattering characteristics. To establish the geometrical model, each tunnel was subdivided into manageable segments of roughly 1-2 km in length based on theoretical considerations. After calibrating the model, the sound system was designed as a large number of loudspeaker lines evenly distributed along the tunnel. Level and delay alignments as well as filter adjustments were applied to achieve the required average STI of 0.48. Validation by measurement showed good correlation with modeling results.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Authors
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
avatar for Tim Kuschel

Tim Kuschel

Acoustic Consultant, GUZ BOX design + audio
Experienced Acoustic Consultant with a demonstrated history of working in the architecture & planning industry. Skilled in architectural documentation, audio system design, acoustics and extensive experience using AFMG's acoustic modelling software EASE, Tim provides professional... Read More →
Wednesday October 9, 2024 9:30am - 9:50am EDT
1E03

9:50am EDT

Sound immission modeling of open-air sound systems
Wednesday October 9, 2024 9:50am - 10:10am EDT
Noise emission into the neighborhood is often a major concern when designing the configuration of an open-air sound system. In order for events to be approved, advance studies have to show that expected immission levels comply with given regulations and requirements of local authorities. For this purpose, certified engineering offices use dedicated software tools for modeling environmental noise propagation. However, predicting the radiation of modern sound systems is different from classical noise sources, such as trains or industrial plants. Sound systems and their directional patterns are modeled in electro-acoustic simulation software that can be fairly precise but that typically does not address environmental issues.
This paper proposes to use a simple data exchange format that can act as an open interface between sound system modeling tools and noise immission software. It is shown that most immission studies are conducted at points in the far field of the sound system. Far-field directivity data for the sound system is therefore a suitable solution if it is accompanied by a corresponding absolute level reference. The proposed approach has not only the advantage of being accurate for the given application but also involves low computational costs and is fully compliant with the existing framework of outdoor noise modeling standards. Concerns related to documentation and to the protection of proprietary signal processing settings are resolved as well. The proposed approach was validated by measurements at a number of outdoor concerts. Results are shown to be practically accurate within the given limits of uncertainty.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Authors
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Wednesday October 9, 2024 9:50am - 10:10am EDT
1E03

10:00am EDT

MoIP Security: From Quick and Dirty to Rock Solid
Wednesday October 9, 2024 10:00am - 10:30am EDT
Building on previous "101" presentations, this session focuses on advanced techniques for securing Media over IP (MoIP) networks. Attendees will learn to configure switches and media gateways, such as the Prodigy from DirectOut, for a secure setup. Emphasizing proactive problem prevention, the session covers basic to advanced security measures, including network segmentation, firewalls, and access restriction. Real-world case studies and future trends in MoIP security will be discussed. This presentation aims to transform your MoIP network from functional to fortified, ensuring both efficiency and security.
Speakers
avatar for Nicolas Sturmel

Nicolas Sturmel

Directout GmbH
Wednesday October 9, 2024 10:00am - 10:30am EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

10:00am EDT

Student Recording Competition Category 2
Wednesday October 9, 2024 10:00am - 10:45am EDT
Traditional Studio Recording finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
avatar for Andrew Schneider

Andrew Schneider

Recorderer / Mixerer / Producerer / Studio Ownerer, ACRE Audio
Andrew Schneider is a Brooklyn based Producer, Mixer and Engineer. He has a strong passion for working with independent artists to create unique recordings. American, Canadian and Swedish Grammy nominated, Andrew's production work includes recordings for labels Ipecac, Thrilljockey... Read More →
Wednesday October 9, 2024 10:00am - 10:45am EDT
1E07

10:00am EDT

Fania Records at 60
Wednesday October 9, 2024 10:00am - 11:00am EDT
Fania Records is often called the Motown of Latin music. With a who's who amongst salsa royalty, they hold a catalog of some of the greatest talent in Latin Music. From their humble beginnings selling cds out of cars, to Yankee Stadium and then Live in Zaire for the Rumble in the Jungle, Fania reached the top. Now at 60, Fania records is re-mastering much of their catalog. Join Grammy Winning Engineer JC "Mr. Sonic" Losada as he moderates a discussing with one of the re-mastering engineers Kevin Gray to hear more about the tech behind handling the Fania catalog.
Wednesday October 9, 2024 10:00am - 11:00am EDT
Stage

10:00am EDT

Technical Committee Meeting: Automotive Audio
Wednesday October 9, 2024 10:00am - 11:00am EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.

This Meeting will take place online and in-person. To Join online click this link: https://us02web.zoom.us/j/84593160017?pwd=IOfBmz7NYIaPf0ePwFNkSFWtOGbFWA.1
Wednesday October 9, 2024 10:00am - 11:00am EDT
2D02/03

10:00am EDT

Exhibit Hall
Wednesday October 9, 2024 10:00am - 6:00pm EDT
Step into the heart of innovation at the Audio Engineering Society’s Annual Conference Exhibit Hall. This dynamic space brings together leading companies and cutting-edge technologies from across the audio engineering industry. Attendees will have the opportunity to explore the latest advancements in audio equipment, software, and services, engage with industry experts, and discover new solutions to enhance their projects. Whether you’re looking to network with professionals, gain insights from live demonstrations, or simply stay ahead of the curve, the Exhibit Hall is the place to be. Don’t miss this chance to immerse yourself in the future of audio engineering!
Wednesday October 9, 2024 10:00am - 6:00pm EDT
Exhibit Hall

10:10am EDT

Virtual Acoustics Technology in the Recording Studio-A System Update
Wednesday October 9, 2024 10:10am - 10:30am EDT
This paper describes ongoing efforts toward optimizing the Virtual Acoustics Technology (VAT) system installed in the Immersive Media Lab at McGill University. Following the integration of the CAVIAR cancelling auralizer for feedback suppression, this current iteration of the active acoustics system is able to flexibly support the creation of virtual environments via the convolution of Spatial Room Impulse Responses (SRIRs) with real-time microphone signals. While the system has been successfully used for both recordings and live performances, we have nevertheless been looking to improve upon its “stability from feedback” and “natural sound quality,” two significant attributes of active acoustics systems [1]. We have implemented new software controls and microphone input methods to increase our ratio of gain before feedback, while additionally repositioning and adding loudspeakers to the system to generate a more even room coverage. Following these additions, we continue to evaluate the space through objective measurements and feedback from musicians and listeners.
[1] M. A. Poletti, “Active Acoustic Systems for the Control of Room Acoustics,” in Proceedings of the International Symposium on Room Acoustics, Melbourne, Australia, Aug. 2010.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Authors
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Wednesday October 9, 2024 10:10am - 10:30am EDT
1E03

10:15am EDT

The Soundsystem as Survival and Resistance: A Black, Afro-Asian, and Indigenous Audio History
Wednesday October 9, 2024 10:15am - 11:00am EDT
Short Abstract:
This presentation explores Black, Indigenous, Afro-Asian, and Afro-Indigenous cross-cultural collaboration leading to influential innovations in sonic creative technology. Centering on the soundsystem as a form of self-determination and as a contemporary art form rooted in survival and resistance, the presentation aims to illuminate meaningful connections between ancestral and state-of-the-art sound technologies through the lens of critical decolonization theory. Highlighting innovations ranging from parametric EQ, radio, and the sound system as both cultural and analog technologies, the presentation shows how these could only exist through mutual influence and the building of nations and kinship across cultures bound by mutual experiences of oppression. These artistic traditions in music, gathering, and engineering emerged and continue to act as dynamic tools of survivance within Black and globally Indigenous communities. By skillfully using these tools, practitioners of these electronic and sonic art forms destabilize racialized boundaries and empower the collective ownership of cultural narratives—a fearless pursuit reflecting a shared vision of Black and Indigenous self-determination, liberation, and futurity for kin past, present, and future.

Extension:
The presentation’s audiovisual materials will comprise two aspects: archival materials from music and resistance movements in the featured locations and contemporary artworks dealing with identity politics and protest in connection with music. Featured locations will include, but are not limited to: Indian Country, USA & Canada; Kingston, Jamaica; Notting Hill, UK; and Detroit, MI. Timeline of technological innovation will center audio advancements in amplification and transmission within the post-transistor era (1947 - present), with a historical focus primarily from 1947 to 1996. Featured BIPOC artists and innovators in electrical engineering in the sonic arts will include and are not limited to: Hedley Jones, Tom Wong, and Don Lewis. Featured entrepreneurial artists in audio engineering, radio, recording, music and storytelling will include, but are not limited to: Vy Higgenson, Kemistry & Storm, Jeff Mills, K-Hand, and Patricia Chin. Featured Black, Asian and Indigenous owned radio stations and music labels that function as both mutual aid and community spaces will include, but are not limited to: Bvlbancha Liberation Radio, Cool Runnings Music, KTNN, VP Records, and the Sound System Futures Programme.

The objective of this session is to provide a brief overview mapping sound histories to the BIPOC innovators of origin with the aim of sparking generative discussion on Black and Indigenous creative technologies as both a cultural and computational history valuable and critical to the healing of community and dismantling of hegemonic societal structures. The session is structured largely in response to a critical lack of representation regarding BIPOC engineering history and legacy, and approaches this topic as a celebration of cross-cultural collaboration in a historical paradigm that often centers cross-cultural interactions as conflict-oriented. At the presentation’s conclusion participants will be encouraged to reflect on their own cross-cultural legacies as sources of innovation, as well as reflect on the unactivated dreams they may hold from their ancestors and how they might activate these dreams through sound, collective action, and community gathering.
Speakers
Wednesday October 9, 2024 10:15am - 11:00am EDT
1E08

10:15am EDT

KEYQUALIZATION New EQ Technique and KEY-QAlpha Product Demonstration by Roey Shamir
Wednesday October 9, 2024 10:15am - 11:15am EDT
Building on the industry standard that saw the development of Parametric Equalization as the first step in providing the audio engineer with complete control of spectrum modification, this presentation concerns the application of a new update, built on Parametric equalization technique to professional audio control. Invented in 1998 by Roey Shamir, now is a product in its alpha stage of development. The technique and product will be fully demonstrated by the inventor and well known mix producer (Roey Shamir).
The co developer of the hardware Useful Arts Audio / Peter Swann will also discuss matters of the hardware while its being developed. Peter  built his first console in 1982, and has been working ever since to create signal processing devices that don’t just measure well on the bench, but translate sound to compliment the perceptual mechanisms of the human ear and brain. In 2017, he launched Useful Arts Audio, which manufactures specialty front-end gear designed to accentuate musical performances by taking advantage of the natural emphasis the brain places on dynamics and harmonic content. Mr. Swann creates original designs aimed at translating incoming musical signals into compelling listening experiences.

Only the proof of concept Software will be demonstrated the hardware will not be ready for AES NY Convention Fall 2024.

There will be a well known song played in immersive 7.1.4 with examples of the new technique in the elements of the immersive mix.
Speakers
avatar for Peter Swann

Peter Swann

Peter Swann built his first console in 1982, and has been working ever since to create signal processing devices that don’t just measure well on the bench, but translate sound to compliment the perceptual mechanisms of the human ear and brain.  In 2017, he launched Useful Arts... Read More →
avatar for Roey Shamir

Roey Shamir

Mixmaster, INFX PRODUCTIONS INC.
• Freelance Audio Engineer, Mixer and Producer • Audio Engineer Mixer, Producer, Remixer for extensive major label records and filmproductions (Mix to Picture) • Mixed and Recorded music / albums for numerous RIAA multi platinum and Grammynominated... Read More →
Wednesday October 9, 2024 10:15am - 11:15am EDT
1E06

10:20am EDT

ADM-OSC 1.0
Wednesday October 9, 2024 10:20am - 10:40am EDT
Spatial and immersive audio has become increasingly mainstream, presented in concert halls and more recently through music streaming services. There is a diverse ecosystem of hardware and software controllers and renderers in both live and studio settings that would benefit from a standardized communication protocol. In 2019 a group of industry stakeholders began designing ADM-OSC to fill this need. ADM-OSC is a standard for transmitting metadata for object-based audio by implementing a namespace in parallel with the Audio Definition Model (ADM), a metadata standard developed in the broadcast industry. Open Sound Control (OSC) is a well-established data transport protocol developed for flexible and accurate communication of real-time performance data. By leveraging these open standards, we have created a lightweight specification that can be easily implemented in audio software, plugins, game engines, consoles, and controllers.
ADM-OSC has reached a level of maturity over multiple implementations be ready for an official 1.0 release. This paper will discuss the design of ADM-OSC 1.0 and how it was developed to facilitate interoperability for a range of stakeholders and use cases. The core address space for position data is described, as well as extensions for live control data. We conclude with an overview of future ADM-OSC development, including next steps in bringing together ideas and discussion from multiple industry partners.
Moderators Speakers
avatar for Michael Zbyszynski

Michael Zbyszynski

Software Development Engineer, L-Acoustics
Michael Zbyszyński is musician, researcher, teacher and developer in the field of contemporary electroacoustic music. He is currently part of the Creative Technologies R&D group at L-Acoustics. As a musician, his work spans from brass bands to symphony orchestras, including composition... Read More →
Authors
avatar for Michael Zbyszynski

Michael Zbyszynski

Software Development Engineer, L-Acoustics
Michael Zbyszyński is musician, researcher, teacher and developer in the field of contemporary electroacoustic music. He is currently part of the Creative Technologies R&D group at L-Acoustics. As a musician, his work spans from brass bands to symphony orchestras, including composition... Read More →
avatar for Hugo Larin

Hugo Larin

Senior Mgr. Business Development | FLUX:: GPLM, Harman International
Hugo Larin is a key collaborator to the FLUX: SPAT Revolution project and has deep roots in audio mixing, design and operation, as well as in networked control and data distribution. He leads the FLUX:: business development at HARMAN. His recent involvements and interests include object-based spatial audio mixing workflows, interoperability... Read More →
avatar for Lucas Zwicker

Lucas Zwicker

Senior Director, Workflow and Integration, CTO Office, Lawo AG
Lucas joined Lawo in 2014, having previously worked as a freelancer in the live sound and entertainment industry for several years. He holds a degree in event technology and a Bachelor of Engineering in electrical engineering and information technology from the University of Applied... Read More →
avatar for Mathieu Delquignies

Mathieu Delquignies

Education & Application Support France, d&b audiotechnik
Mathieu holds a Bachelors's degree in applied physics from Paris 7 University and Master's degree in sound engineering from ENS Louis Lumière in 2003. He has years of diverse freelance mixing and system designer experiences internationally, as well as loudspeakers, amplifiers, dsp... Read More →
Wednesday October 9, 2024 10:20am - 10:40am EDT
1E04

10:30am EDT

A General Overview of Methods for Generating Room Impulse Responses
Wednesday October 9, 2024 10:30am - 10:50am EDT
The utilization of room impulse responses has proven valuable for both the acoustic assessment of indoor environments and music production. Various techniques have been devised over time to capture these responses. Although algorithmic solutions have been in existence since the 1960s for generating synthetic reverberation in real time, they continue to be computationally demanding and in general lack the accuracy in comparison to measured authentic Room Impulse Responses (RIR). In recent times, machine learning has found application in diverse fields, including acoustics, leading to the development of techniques for generating RIRs. This paper provides a general overview, of approaches and methods for generating RIRs, categorized into algorithmic and machine learning techniques, with a particular emphasis on the latter. Discussion covers the acoustical attributes of rooms relevant to perceptual testing and methodologies for comparing RIRs. An examination of disparities between captured and generated RIRs is included to better delineate the key acoustic properties characterizing a room. The paper is designed to offer a general overview for those interested in RIR generation for music production purposes, with future work considerations also explored.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
Authors
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Wednesday October 9, 2024 10:30am - 10:50am EDT
1E03

10:30am EDT

Revolutionizing Media Workflows: A modern approach to audio processing in a dynamic environment
Wednesday October 9, 2024 10:30am - 11:00am EDT
This presentation explores advanced cloud-technology-based audio processing, highlighting platform-agnostic solutions. It emphasizes the integration of SMPTE ST2110, NDI, and SRT for seamless, high-quality media transport. Key topics include the benefits of scalability, real-time processing capabilities, and integration across diverse systems, offering a comprehensive overview of modern, efficient and dynamic workflows in media production and distribution.
Speakers
avatar for Lucas Zwicker

Lucas Zwicker

Senior Director, Workflow and Integration, CTO Office, Lawo AG
Lucas joined Lawo in 2014, having previously worked as a freelancer in the live sound and entertainment industry for several years. He holds a degree in event technology and a Bachelor of Engineering in electrical engineering and information technology from the University of Applied... Read More →
Wednesday October 9, 2024 10:30am - 11:00am EDT
AES and AIMS Alliance Program Room
  Sponsor Session

10:30am EDT

Active Dynamics
Wednesday October 9, 2024 10:30am - 11:30am EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Wednesday October 9, 2024 10:30am - 11:30am EDT
1E05

10:40am EDT

Analysis of Ultra Wideband Wireless Audio Transceivers for High-Resolution Audio Transmission
Wednesday October 9, 2024 10:40am - 11:00am EDT
Modern Ultra Wide-Band Wireless (UWB) transceiver radio systems enhance digital audio wireless transmission by eliminating the need for audio data compression required by narrowband technologies such as Bluetooth. UWB systems, characterized by their high bandwidth, bypass compression computation delays, enabling audio data transmission from transmitter to receiver in under 10 milliseconds. This paper presents an analytical study of audio signals transmitted using a contemporary UWB transceiver system. The analysis confirms that UWB technology can deliver high-resolution (96kHz/24-bit) audio that is free from artifacts and comparable in quality to a wired digital link. This study underscores the potential of UWB systems to revolutionize wireless audio transmission by maintaining integrity and reducing latency, aligning with the rigorous demands of high-fidelity audio applications.
Moderators Speakers
avatar for Jeff Anderson

Jeff Anderson

SPARK Microsystems
Authors
avatar for Jeff Anderson

Jeff Anderson

SPARK Microsystems
Wednesday October 9, 2024 10:40am - 11:00am EDT
1E04

10:45am EDT

De-Mixing: Should we just because we can?
Wednesday October 9, 2024 10:45am - 11:30am EDT
Multi-Grammy Award winners, Michael Graves (Osiris Studio) and Cheryl Pawelski (Omnivore Recordings) will lead a discussion exploring the intersection of historical/catalog recordings and technology.

We’re at a point in technological developments where tech tools can be used to fix flawed media, restore poor recordings or de-mix to remix sources.

We will share questions we ask ourselves when we consider employing new technology and discuss the responsibility we feel toward the art, the artist and the artists’ intent.

To help aid the discussion, audio samples will be played by:

Chris Bell (I Am The Cosmos)
Hank Williams (The Garden Spot Programs, 1950)
Art Pepper (The Complete Maiden Voyage Recordings)
Janis Joplin & Jorma Kaukonen (The Legendary Typewriter Tape, 1964)
Various Artists (Written In Their Soul: The Stax Songwriter Demos)
Speakers
avatar for Michael Graves

Michael Graves

Mastering Engineer | Audio Restoration Specialist, Osiris Studio
Michael Graves is a multiple GRAMMY-Award winning mastering engineer. During his career, Graves has worked with clients worldwide and has gained a reputation as one of the leading audio restoration and re-mastering specialists working today, earning 5 GRAMMY Awards and 14 nominations... Read More →
avatar for Cheryl Pawelski

Cheryl Pawelski

Co-Founder/Co-Owner, Omnivore Recordings
Cheryl Pawelski is a four-time Grammy Award-winning producer who has for years been entrusted with preserving, curating, and championing some of music’s greatest legacies. Before co-founding Omnivore Entertainment Group, she held positions at Rhino, Concord, and EMI-Capitol Records... Read More →
Wednesday October 9, 2024 10:45am - 11:30am EDT
1E09

10:45am EDT

BBC Pop Hub - Moving Europe's largest radio station
Wednesday October 9, 2024 10:45am - 11:30am EDT
BBC Radio 2 is Europe’s largest radio station. In this session Jamie Laundon, Solution Lead at BBC Technology & Media Operations, will guide us through the move into the new, state-of-the-art "Pop Hub" studios on 8th floor of BBC Broadcasting House in London.

The project was delivered to extremely tight timescales and with no impact on our listeners. These studios have been designed as dynamic, collaborative spaces, with the aim to inspire our teams to produce their best work, and to reflect an energy and buzz that will be transmitted to our audiences.

The studios have been designed to be simple to use yet also feature complex workflows that permit studios to used in pairs and give additional assistance to presenters and musicians as required.

In this session you’ll learn about managing design decisions when working on an existing site, and the use of realtime media networks, AES67 and Dante to join infrastructure together. As well as audio equipment selection, we will also cover furniture design and some interesting challenges around accessibility, sight lines and camera angles. We will also cover the tools used to visualise radio, and create well-lit branded spaces that suit any of the BBC’s pop music networks. We will also talk about the collaboration and skill sharing between our audio, technology and operations teams to help deliver against very tight timelines.
Speakers
Wednesday October 9, 2024 10:45am - 11:30am EDT
1E16

10:45am EDT

How Audio Engineering can assist in Neurodivergent Hearing Loss Prevention
Wednesday October 9, 2024 10:45am - 11:30am EDT
Within live and studio sound engineering, the use of compression, normalization and parametric EQ is employed to perfect the sound of musicians we work with. These effects can impact the tone and timbre of an audio sample or even eliminate unwanted noise. Even the resistance of the electrical equipment used to aid us in our work as audio engineers has an impact on the quality of our end product. Many audio engineers and audio companies have used our understanding of the technology within our industry and applied it to the use of hearing mechanisms such as in-ear monitors or ear plugs.

Around 3-9% of adults struggle with some form of hearing disability or hearing loss. While the science and studies of audiology have helped us find ways to assist those disabled in this way, hearing disabilities tend to be comorbid with neurological or psychological disorders that are harder to manage when combined. PTSD, dementia, autism, as well as several other disorders severely impact the way our brains interact with auditory and kinesthetic stimuli that make current hearing aid technology more of a discomfort than a help to patients with these comorbidities.

After multiple observations of audio engineering techniques and the equipment we use, peer reviewed studying of clientele with these comorbidities, and applications of techniques to hearing sciences, we wish to display some of those applications and report our findings so as to make the world more accessible to those with hearing disabilities, neurological/psychological disorders, and those with both.
Wednesday October 9, 2024 10:45am - 11:30am EDT
1E15

10:45am EDT

Student Recording Competition Category 3
Wednesday October 9, 2024 10:45am - 11:45am EDT
Modern Studio Recording finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
MW

Marek Walaszek

General, Addicted To Music
Wednesday October 9, 2024 10:45am - 11:45am EDT
1E07

11:00am EDT

Reimagining Delay Effects: Integrating Generative AI for Creative Control
Wednesday October 9, 2024 11:00am - 11:20am EDT
This paper presents a novel generative delay effect that utilizes generative AI to create unique variations of a melody with each new echo. Unlike traditional delay effects, where repetitions are identical to the original input, this effect generates variations in pitch and rhythm, enhancing creative possibilities for artists. The significance of this innovation lies in addressing artists' concerns about generative AI potentially replacing their roles. By integrating generative AI into the creative process, artists retain control and collaborate with the technology, rather than being supplanted by it. The paper outlines the processing methodology, which involves training a Long Short-Term Memory (LSTM) neural network on a dataset of publicly available music. The network generates output melodies based on input characteristics, employing a specialized notation language for music. Additionally, the implementation of this machine learning model within a delay plugin's architecture is discussed, focusing on parameters such as buffer length and tail length. The integration of the model into the broader plugin framework highlights the practical aspects of utilizing generative AI in audio effects. The paper also explores the feasibility of deploying this technology on microcontrollers for use in instruments and effects pedals. By leveraging low-power AI libraries, this advanced functionality can be achieved with minimal storage requirements, demonstrating the efficiency and versatility of the approach. Finally, a demonstration of an early version of the generative delay effect will be presented.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers Authors
Wednesday October 9, 2024 11:00am - 11:20am EDT
1E03

11:00am EDT

Migrating to compute for live media production processing – the impact on audio
Wednesday October 9, 2024 11:00am - 11:30am EDT
The industry is on an sure and steady move to using compute for the complete chain of live media production processing. This evolution from dedicated hardware products provides significant flexibility yet comes with challenges relating to the precision of behavior needed to process broadcast audio and video. This presentation will outline the techniques being used for the transfer and processing of each media flow and examine the optimizations needed to ensure evolving systems are both feature-capable and fully usable.
Speakers
Wednesday October 9, 2024 11:00am - 11:30am EDT
AES and AIMS Alliance Program Room
  Sponsor Session

11:00am EDT

Ulrike & Jim: Enveloping Masterclass
Wednesday October 9, 2024 11:00am - 12:00pm EDT
Ulrike Schwarz and Jim Anderson discuss recordings from their extraordinary catalogues, including high resolution listening examples in formats 5.1, 9.1 and 7.1.4.

This masterclass series, featuring remarkable recording artists, is a chance to hear surround and 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Ulrike Schwarz

Ulrike Schwarz

Engineer/Producer, Co-Founder, Anderson Audio New York
Engineer/Producer, Co-Founder
avatar for Jim Anderson

Jim Anderson

Producer/Engineer, Anderson Audio New York
Producer/Engineer
Wednesday October 9, 2024 11:00am - 12:00pm EDT
3D06

11:00am EDT

Technical Committee Meeting: Acoustics and Sound Reinforcement
Wednesday October 9, 2024 11:00am - 12:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.
Wednesday October 9, 2024 11:00am - 12:00pm EDT
2D02/03

11:15am EDT

Immersive Voice and Audio Services (IVAS) Codec – A deeper look into the new 3GPP Standard for Immersive Communication
Wednesday October 9, 2024 11:15am - 12:45pm EDT
3GPP, the worldwide partnership project among seven major regional telecommunications standard development organizations responsible for mobile communication standards, has recently completed the standardization of a new codec intended for Immersive Voice and Audio Services (IVAS) in 5G systems. The new IVAS codec is a feature of 3GPP Release 18 (5G-Advanced) and enables completely new service scenarios by providing capabilities for interactive stereo and immersive audio communications, content sharing and distribution. The addressed service scenarios include conversational voice with stereo and immersive telephony/conferencing, XR (VR/AR/MR) communications, i.e., XR telephony and conferencing, and live and non-live streaming of user-generated immersive and XR content.
The IVAS Codec is an extension of the 3GPP EVS codec. Beyond support for mono voice and audio content, IVAS also brings support for coding and rendering of stereo, multichannel-based audio, scene-based audio (Ambisonics), object-based audio and metadata-assisted spatial audio – a new parametric audio format designed for direct spatial audio pick-up from smartphones. The performance of the codec was evaluated in the course of the 3GPP standardization in 23 experiments carried out by two independent laboratories each.
The workshop will introduce the IVAS Codec starting with a brief overview of standardization process and state of the art. The codec architecture, supported audio formats, and main technology blocks behind the supported coding and rendering options will be presented and discussed in detail together with test results and capabilities of the new codec in terms of compression efficiency and audio quality. In addition, the panelists will discuss new service scenarios for immersive audio communication and other mobile use-cases. If possible, immersive audio demonstrations will be carried out during the workshop. The panelists will collect and address questions and other input from the audience.
Speakers
avatar for Adriana Vasilache

Adriana Vasilache

Nokia Technologies
AF

Andrea Felice Genovese

Andrea Genovese is a Senior Research Engineer at Qualcomm
Wednesday October 9, 2024 11:15am - 12:45pm EDT
1E08

11:20am EDT

Rediscovering Xenakis: Decoding the Free Stochastic Music Program's Output File
Wednesday October 9, 2024 11:20am - 11:40am EDT
Composer Iannis Xenakis (1922-2001), in his book Formalized Music (1963), includes a piece of Fortran IV computer code which produces a composition of stochastically generated music. This early example of indeterminate music made by a computer was in fulfillment of his goal of creating a musical composition of minimum constraints by letting the computer stochastically generate parameters. Stochasticity is used on all levels of the composition, macro (note density and length of movements) and micro (length of notes and pitches). This paper carefully analyzes the output composition and the variations the program might produce. Efforts are then made to convert these compositions into MIDI format in order to promote their preservation and increase their accessibility. The preservation of the Free Stochastic Music Program is beneficial to understand one of the first examples of indeterminate computer music, with similar techniques being found in modern day music creation software.
Moderators Speakers Authors
Wednesday October 9, 2024 11:20am - 11:40am EDT
1E04

11:20am EDT

Acoustic Characteristics of Parasaurolophus Crest: Experimental Results from a simplified anatomical model
Wednesday October 9, 2024 11:20am - 11:40am EDT
This study presents a revised acoustic model of the Parasaurolophus crest, incorporating both the main airway and lateral diverticulum, based on previous anatomical models and recent findings. A physical device, as a simplified model of the crest, was constructed using a coupled piping system, and frequency sweeps were conducted to investigate its resonance behavior. Data were collected using a minimally invasive microphone, with a control group consisting of a simple open pipe for comparison. The results show that the frequency response of the experimental model aligns with that of the control pipe at many frequencies, but notable shifts and peak-splitting behavior were observed, suggesting a more active role of the lateral diverticulum in shaping the acoustic response than previously thought. These findings challenge earlier closed-pipe approaches, indicating that complex interactions between the main airway and lateral diverticulum generate additional resonant frequencies absent in the control pipe. The study provides empirical data that offer new insights into the resonance characteristics of the Parasaurolophus crest and contribute to understanding its auditory range, particularly for low-frequency sounds.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers Authors
Wednesday October 9, 2024 11:20am - 11:40am EDT
1E03

11:30am EDT

Good Enough Is Not Enough
Wednesday October 9, 2024 11:30am - 12:00pm EDT
For media or content delivery, “good enough” is not enough anymore. Just because signal distribution is over the internet doesn’t mean audiences will tolerate jitter or latency -- and they shouldn’t have to. New IP protocols, including SRT, NDI and the newest IPMX, are melding broadcast workflows with “traditional” AV technology. Live streaming has matured into a reliable and secure AV transport method, and it’s now possible to cost-effectively produce 1080P or 4K HDR content from a conference room camera, and then mix and integrate that with a broadcast-quality stream. Also to be covered: why IPMX will emerge as the new AVoIP standard.
Speakers
KH

Kevin Henneman

KMH Integration
Wednesday October 9, 2024 11:30am - 12:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

11:30am EDT

New Trends in Fully Virtual Design and Prototyping of Speaker Drivers
Wednesday October 9, 2024 11:30am - 12:30pm EDT
The audio industry is increasingly focused on fully virtual prototyping of devices, designs, and measurement setups. New modeling trends include performing detailed nonlinear distortion analysis, incorporating frequency-dependent properties in time-domain studies, creating lumped components for real-time system analysis, and virtually developing and testing innovative materials. In this talk, we will explore two examples to discuss the emerging modeling techniques and strategies that enable these advanced analyses and address challenges in speaker driver design.
Speakers
JH

Jinlan Huang

Applications Engineer, COMSOL, Inc.
Wednesday October 9, 2024 11:30am - 12:30pm EDT
Stage
  Sponsor Session

11:40am EDT

Unveiling the Impact of K-Pop Song Lyrics on Adolescent Self-Harm Rates: A Pilot Study of Gender Disparities
Wednesday October 9, 2024 11:40am - 12:00pm EDT
This study investigates the phenomenon of increased emergency room visits by teenagers following the broadcast of ‘High School Rapper 2’, which features self-harm in its lyrics. The rate of emergency room visits for self-harm among adolescents aged 10 to 24 notably increased during and shortly after the program. Specifically, in the 10-14 age group, visits per 100,000 population tripled from 0.9 to 3.1. For those aged 15-19, the rate rose from 5.7 to 10.8, and in the 20-24 age group, it increased from 7.3 to 11.0. This study aims to clarify the relationship between lyrics and self-harm rates among adolescents. We analyzed the lyrics of the top 20 songs performed on Melon, a popular music streaming platform for teenagers, from April to October 2018. Using Python's KoNLPy and Kkma libraries for tokenization, part-of-speech tagging, and stop-word filtering, we identified the top 50 frequently appearing words and narrowed them down to the top 20 contextually significant words. A correlation analysis (Pearson R) with "Emergency Department Visits due to Self-Harm" data revealed that the words 'sway', 'think', 'freedom' and 'today' had a statistically significant correlation (p < 0.05). Additionally, we found that males’ self-harm tendency was less influenced by the broadcast compared to females. This research suggests a computational approach to understanding the influence of music lyrics on adolescent self-harm behavior. This pilot study demonstrated correlations between specific words in K-pop lyrics and increased adolescent self-harm rates, with notable gender differences.
Moderators Speakers Authors
avatar for Sungjoon Kim

Sungjoon Kim

Research Intern, Korea Advanced Institute of Science and Technology
Wednesday October 9, 2024 11:40am - 12:00pm EDT
1E04

11:40am EDT

Interpreting user-generated audio from war zones
Wednesday October 9, 2024 11:40am - 12:00pm EDT
Increasingly, civilian inhabitants and combatants in conflict areas use their mobile phones to record video and audio of armed attacks. These user-generated recordings (UGRs) often provide the only source of immediate information about armed conflicts because access by professional journalists is highly restricted. Audio forensic analysis of these UGRs can help document the circumstances and aftermath of war zone incidents, but consumer off-the-shelf recording devices are not designed for battlefield circumstances and sound levels, nor do the battlefield circumstances provide clear, noise-free audio. Moreover, as with any user-generated material that generally does not have a documented chain-of-custody, there are forensic concerns about authenticity, misinformation, and propaganda that must be considered. In this paper we present several case studies of UGRs from armed conflict areas and describe several methods to assess the quality and integrity of the recorded audio. We also include several recommendations for amateurs who make UGRs so that the recorded material is more easily authenticated and corroborated. Audio and video examples are presented.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Authors
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Wednesday October 9, 2024 11:40am - 12:00pm EDT
1E03

11:45am EDT

Be a Leader at AES
Wednesday October 9, 2024 11:45am - 12:45pm EDT
Are you interested in a leadership position within AES? Whether it's a role in your local section as a chair or secretary, a committee, or our Board of Governors, join our panel of experienced AES leaders to find out what is involved and whether a role in leadership is right for you.







Speakers
avatar for David Bowles

David Bowles

Owner, Swineshead Productions, LLC
David v.R Bowles formed Swineshead Productions, LLC as a classical recording production company in 1995. His recordings have been GRAMMY- and JUNO-nominated and critically acclaimed worldwide. His releases in 3D Dolby Atmos can be found on Avie, OutHere Music (Delos) and Navona labels.Mr... Read More →
avatar for Leslie Gaston-Bird

Leslie Gaston-Bird

Owner, Mix Messiah Productions
Leslie Gaston-Bird (AMPS, MPSE) is author of the book "Women in Audio", part of the AES Presents series and published by Focal Press (Routledge). She is a voting member of the Recording Academy (The Grammys®) and its P&E (Producers and Engineers) Wing. Currently, she is a freelance... Read More →
avatar for Marcela Zorro

Marcela Zorro

Universidad de los Andes
Marcela is an audio engineer specializing in the recording and production of classical music. She currently teaches in the audio production emphasis of the Music Department at Universidad de los Andes in Bogotá, Colombia. Marcela has been an active member of the AES since 2004.  She... Read More →
avatar for Gary Gottlieb

Gary Gottlieb

AES President-Elect, Mendocino College
President-Elect, Co-Chair of the Events Coordination Committee, Chair of the Conference Policy Committee, and former Vice President of the Eastern Region, US and Canada; AES Fellow, Engineer, Author, Educator and Guest Speaker Gary Gottlieb refers to himself as a music generalist... Read More →
avatar for Angela Piva

Angela Piva

Angela Piva, Audio Pro/Audio Professor, highly skilled in all aspects of music & audio production, recording, mixing and mastering with over 35 years of professional audio engineering experience and accolades. Known as an innovator in sound technology, and for contributing to the... Read More →
Wednesday October 9, 2024 11:45am - 12:45pm EDT
1E15

11:45am EDT

Navigating the Legal Landscape of AI in Music and Tech
Wednesday October 9, 2024 11:45am - 12:45pm EDT
Artificial intelligence (AI) and machine learning (ML) is revolutionizing the music industry, from creation to distribution to consumption. These technologies, and the rapid growth, raise complex legal issues that parties should be aware of to ethically foster innovation while protecting legal rights. This panel will explore the legal challenges and implications of AI in music and music tech, providing insights from industry experts, legal professionals, and technologists, including discussion of topics on intellectual property rights, data privacy, the regulatory landscape, and how to navigate the complex field. Join us for an insightful panel that delves into the intersection of AI, music, and law, offering practical guidance and perspectives on protecting rights, legally and ethically sourcing datasets, and navigating this rapidly evolving landscape.

Topics:
1. Intellectual Property Rights: Understanding copyright issues related to AI-generated music. Was the ML model dataset trained on copyrighted works? Did the company clear the rights to train the model on the material used? Who owns the rights to a song composed by an AI? To what extent, if any? What are the implications for sampling and remixing AI-generated content? How AI impacts the traditional models of licensing and royalties. What changes are necessary to ensure fair compensation for creators?
2. Data Privacy and Security: Addressing concerns about data used in training AI models for music tech. What are the best practices for protecting artists' and users' data? Is data collected and stored properly? Are proper company policies in place?
3. Regulatory Landscape: Reviewing current laws and proposed regulations affecting AI in the music industry. How are different jurisdictions approaching these issues, and what can we expect in the future? Discussing the EU AI Act and other regulations on the horizon.
4. Ethical and Legal Accountability: Exploring the responsibilities of AI developers and users in the music industry. How can we ensure ethical use of AI technologies in music creation and distribution?
Speakers
avatar for Heather Rafter

Heather Rafter

Co-founder & Principal, RafterMarsh
Heather Dembert Rafter has been providing legal and business development services to the audio, music technology, and digital media industries for over twenty-five years. As principal counsel at RafterMarsh US, she leads the RM team in providing sophisticated corporate and IP advice... Read More →
DG

Denis Geokdag

CEO, Zynaptiq GmbH
avatar for Philipp Lengeling

Philipp Lengeling

Senior Counsel, RafterMarsh
Philipp G. Lengeling, Mag. iur., LL.M. (New York), Esq. is an attorney based in New York (U.S.A.) and Hamburg (Germany), who is heading the New York and Hamburg based offices for RafterMarsh, a transatlantic boutique law firm (California, New York, U.K., Germany).
avatar for Zack Klein

Zack Klein

Associate Counsel, RafterMarsh
Zack is an attorney licensed in New York and New Jersey, and is an associate with RafterMarsh US with focuses in the fields of tech, music law, intellectual property, licensing, as well as work in Web3. Zack has degrees in music technology and electrical engineering and his background... Read More →
Wednesday October 9, 2024 11:45am - 12:45pm EDT
1E09

11:45am EDT

Fix The Mix: Building a Thriving Career Behind the Board
Wednesday October 9, 2024 11:45am - 12:45pm EDT
Presented by We Are Moving The Needle, this panel brings together industry leaders from A&R, management, and artist relations to discuss building an impactful career behind the board. These experienced leaders will share the inside scoop on strategies to build your profile as a producer and engineer from navigating rates to defining your brand to attracting a manager. Whether you're an aspiring engineer or a seasoned pro, this conversation offers valuable insights, practical tips, and a chance to connect with a vibrant community of music industry leaders.
Speakers
LW

Lysee Webb

Founder, Van Pelt Management
avatar for Lee Dannay

Lee Dannay

VP of A&R, Thirty Tigers
Lee Dannay's 20 plus year career includes VP/A&R positions at Epic and Columbia Records, VP A&R in music publishing at Warner/Chappell Music, and Senior Music Producer at America’s Got Talent. She has signed and developed artists ranging from John Mayer, Shawn Mullins, Five For... Read More →
avatar for Samantha Sichel

Samantha Sichel

Head of Social Product & Digital Innovation, Live Nation
Live Nation Entertainment; Senior Manager of Marketing Solutions/Digital Business Development Samantha Sichel leads Digital Marketing Solutions for Live Nation Entertainment. This includes working with the Live Nation Sales and Account Management teams to conceptualize and budget... Read More →
Wednesday October 9, 2024 11:45am - 12:45pm EDT
1E16

12:00pm EDT

Experimental analysis of a car loudspeaker model based on imposed vibration velocity: effect of membrane discretization
Wednesday October 9, 2024 12:00pm - 12:20pm EDT
Nowadays, the research about the improvement of the interior sound quality of road vehicles is a relevant task. The cabin is an acoustically challenging environment due to the complex geometry, the different acoustic properties of the materials of cabin components and the presence of audio systems based on multiple loudspeaker units. This paper aims at presenting a simplified modelling approach designed to introduce the boundary condition imposed by a loudspeaker to the cabin system in the context of virtual acoustic analysis. The proposed model is discussed and compared with experimental measurements obtained from a test-case loudspeaker.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers Authors
Wednesday October 9, 2024 12:00pm - 12:20pm EDT
1E03

12:00pm EDT

How do AES67 and SMPTE ST 2110 relate to IPMX
Wednesday October 9, 2024 12:00pm - 12:30pm EDT
Originally, the definition of IPMX started from the proven grounds of AES67 and SMPTE ST 2110. While AES67 and ST 2110 were build around the interoperability requirements of the broadcast realm, it became immediately apparent that for application to the world of ProAV enhanced functionality would be required, while other speciciations indispensible to the broadcast world would be too complex or rigid for ProAV application - namely the very strict timing and synchronization requirements only made possible by the use of PTP. During development of the IPMX specifications within the VSF group it became clear that certain constraints and enhancements have to be made with respect to AES67 and ST 2110 to enable wider adoption in the ProAV domain. IPMX is now taking shape with several companies having announced product availability. While the larger focus of IPMX interoperability is on the video side of things, audio is certainly a very important aspect. And with over 5000 AES67- or ST 2110-compatible audio devices in the market, it would certainly be a massive adoption accelerator if these products could interoperate in an IPMX environment. In this presentation Andreas will discuss how the addditional constraints and enhancements of the IPMX specifications affect interoperability with AES67 and ST 2110 and how these devices can be used in IPMX environments.
Speakers
Wednesday October 9, 2024 12:00pm - 12:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

12:00pm EDT

Student Recording Competition Category 4
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
Sound for Visual Media finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
avatar for Mary Mazurek

Mary Mazurek

Audio Educator/ Recording Engineer, University of Lethbridge
Audio Educator/ Recording Engineer
avatar for Jamie Baker

Jamie Baker

Professor of Sound Design, Hoboken Post LLC
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
1E06

12:00pm EDT

Implementing WHO Hearing Health Guidelines in Small Venues & Houses of Worship
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
This workshop falls under the category of an AES Tutorial. This workshop outlines the tested method implemented in a small house of worship, and aims to demystify WHO standards as well as communicate an effective solution to implement hearing health monitoring.

Implementing affordable and efficient hearing health monitoring in small music venues is crucial to align with the World Health Organization (WHO) guidelines aimed at preventing hearing damage due to prolonged exposure to high sound levels, which have become much more standardized in recent years. This tutorial will address the unique challenges faced by small music venues, especially houses of worship, which often operate under tighter budgets and where user-friendly permanent solutions are key to long term success. The WHO guidelines emphasize regular monitoring of sound levels, provision of adequate hearing protection, and educating both patrons and staff about the risks of hearing loss and the measures that can be taken to mitigate these risks. This workshop will provide a detailed, practical, and cost-effective roadmap for small venues to implement these recommendations without compromising on the quality of the musical experience. We will walk through interpreting the WHO and AES hearing health standards for live venues, and making sense of large technical documents to communicate to volunteers and newer employees, as well as outline a real world tested solution to implementing hearing health monitoring. Another area of focus will be a brief explanation of how human hearing works, and how it can be damaged. The tutorial will demonstrate how to collect and interpret data, and practical everyman solutions to adapt to hearing health challenges as they arise. Another critical aspect of the tutorial will be on hearing protection strategies. Small music venues can provide disposable earplugs to patrons and staff at a minimal cost or invest in reusable earplugs that offer better sound quality while protecting hearing. The tutorial will discuss the different types of earplugs available, their cost, and how to encourage their use among patrons without detracting from the concert experience. Education is a cornerstone of the WHO guidelines, and this tutorial will explore effective ways to raise awareness about hearing health. This includes training staff to recognize the signs of excessive sound exposure and understand the importance of hearing protection, as well as developing clear and engaging signage and informational materials for patrons. By the end of the workshop, attendees will have a comprehensive understanding of how to implement a hearing health monitoring system in their venues that is both practical and affordable, aligned with the WHO guidelines, and tailored to the unique needs of small music venues. They will leave equipped with actionable strategies, practical tools, and a network of peers committed to promoting hearing health in the music community. This tutorial not only aims to protect the hearing of patrons and staff but also to enhance the overall quality of the live music experience by fostering a safer and more aware listening environment.
Speakers
TS

Tommy Spurgeon

Physics Student & Undergraduate Researcher, University of South Carolina
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
1E07

12:00pm EDT

Technical Committee Meeting: Coding of Audio Signals
Wednesday October 9, 2024 12:00pm - 1:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.

This Meeting will take place online and in-person. To Join online click this link: https://us02web.zoom.us/j/82216397103?pwd=L0VhaUlBK3JjcTJwai9jdjV6VUh2Zz09
Wednesday October 9, 2024 12:00pm - 1:00pm EDT
2D02/03

12:20pm EDT

A novel derivative-based approach for the automatic detection of time-reversed audio in the MPAI/IEEE-CAE ARP international standard
Wednesday October 9, 2024 12:20pm - 12:50pm EDT
The Moving Picture, Audio and Data Coding by Artificial Intelligence (MPAI) Context-based Audio Enhancement (CAE) Audio Recording Preservation (ARP) standard provides the technical specifications for a comprehensive framework for digitizing and preserving analog audio, specifically focusing on documents recorded on open-reel tapes. This paper introduces a novel, envelope derivative-based method incorporated within the ARP standard to detect reverse audio sections during the digitization process. The primary objective of this method is to automatically identify segments of audio recorded in reverse. Leveraging advanced derivative-based signal processing algorithms, the system enhances its capability to detect and reverse such sections, thereby reducing errors during analog-to-digital (A/D) conversion. This feature not only aids in identifying and correcting digitization errors but also improves the efficiency of large-scale audio document digitization projects. The system's performance has been evaluated using a diverse dataset encompassing various musical genres and digitized tapes, demonstrating its effectiveness across different types of audio content.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Authors
Wednesday October 9, 2024 12:20pm - 12:50pm EDT
1E03

12:45pm EDT

Grammy Producers & Engineers' Wing: Songs That Changed My Life Vol. II
Wednesday October 9, 2024 12:45pm - 1:45pm EDT
Some songs are hits, some we just love, and some have changed our lives. Our panelists break down the DNA of their favorite tracks and explain what moved them, what grabbed them, and why these songs left a life-long impression. Back by popular demand, this presentation is guaranteed to make you feel good about being in the recording business.
Speakers
avatar for Maureen Droney

Maureen Droney

Vice President, The Recording Academy Producers & Engineering Wing
Maureen Droney is Vice President of the Recording Academy (GRAMMY Awards) Producers & Engineers Wing. A former recording engineer herself, Maureen has worked on GRAMMY-winning recordings for artists including Aretha Franklin, Whitney Houston and Santana—as well as on numerous other... Read More →
Wednesday October 9, 2024 12:45pm - 1:45pm EDT
Stage

1:00pm EDT

AES67 and ST 2110-30 Connection Management in AES70
Wednesday October 9, 2024 1:00pm - 1:30pm EDT
This presentation will describe the features of AES70-21, the recently published AES70 standard for management and monitoring of AES67 and ST 2110-30 audio stream connections.
Speakers
avatar for Jeff Berryman

Jeff Berryman

Senior Scientist, Bosch Communications
Jeff Berryman is Senior Scientist at Bosch Communications Systems.  Over the last 50 years, he has had a dual career in audio and information technology.  His audio experience is in professional loudspeaker and sound reinforcement systems design, focusing on concert touring and... Read More →
Wednesday October 9, 2024 1:00pm - 1:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

1:00pm EDT

Book Signing: Women in Audio by Leslie Gaston-Bird
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
Join us at the AES membership booth for an exclusive book signing event with AES President, Dr Leslie Gaston-Bird, the author of Women in Audio. This book features almost 100 profiles and stories of audio engineers who are women and have achieved success throughout the history of the trade. Beginning with a historical view, the book covers the achievements of women in various audio professions and then focuses on organizations that support and train women and girls in the industry. What follows are eight chapters divided by discipline, highlighting accomplished women in various audio fields: radio; sound for film and television; music recording and electronic music; hardware and software design; acoustics; live sound and sound for theater; education; audio for games, virtual reality, augmented reality, and mixed reality, as well as immersive sound.

Women in Audio - 1st Edition - Leslie Gaston-Bird - Routledge Book
Speakers
avatar for Leslie Gaston-Bird

Leslie Gaston-Bird

Owner, Mix Messiah Productions
Leslie Gaston-Bird (AMPS, MPSE) is author of the book "Women in Audio", part of the AES Presents series and published by Focal Press (Routledge). She is a voting member of the Recording Academy (The Grammys®) and its P&E (Producers and Engineers) Wing. Currently, she is a freelance... Read More →
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
AES Membership Booth

1:00pm EDT

Best Practices for Creating, Mixing, and Producing in Immersive Audio for Music, Film, TV, Podcasting and More
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
Join Focusrite Group’s Lee Whitmore in an interactive panel conversation and listening session with 19-time GRAMMY and Latin GRAMMY winner Rafa Sardina, Dolby’s Tom McAndrew, and ADAM Audio’s Mark Cummins. Listen to Dolby Atmos Music mixes by Rafa Sardina, and chat with industry experts about best practices for creating, mixing, and playback of immersive audio for music, film, TV, gaming, audiobooks and podcasting, as well as considerations for your studio and workflows.

Also held on 10/10 at 12:45pm
Speakers
MC

Mark Cummins

Head of Product Marketing, ADAM Audio
avatar for Tom McAndrew

Tom McAndrew

Senior Technical Manager – Content Relations, Dolby Laboratories
avatar for Rafa Sardina

Rafa Sardina

Rafa Sardina | Afterhours Studios / Abbey Road Institute Los Angeles
Los Angeles based 19-Time Grammy® & Latin Grammy® Winner Producer/Mixer/Engineer Rafa Sardina has built an eclectic and impressive client list over the years: Stevie Wonder, Camila Cabello, D’Angelo, Dr. Dre, Rosalía, John Legend, Elvis Costello & The Roots, Lady Gaga, Michael... Read More →
avatar for Dr. Lee Whitmore

Dr. Lee Whitmore

VP, Education and B2B, Americas, Focusrite Group
Dr. Lee Whitmore is a thought leader in music, audio, creative digital media, and education. He’s the Vice President for business-to-business (B2B), including education, pro, post, and government, at Focusrite Group, which includes FocusriteNovationSequential and Oberh... Read More →
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
1E06

1:00pm EDT

Technical Committee Meeting: Spatial Audio
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.

This Meeting will take place online and in-person. To Join online click this link: https://us02web.zoom.us/j/81754581433?pwd=hET9ewp9peRTCVsLGGAmKSuirkNh6J.1
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
2D02/03

1:00pm EDT

dLive Certification
Wednesday October 9, 2024 1:00pm - 4:00pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.
 
The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Wednesday October 9, 2024 1:00pm - 4:00pm EDT
1E05

1:30pm EDT

Catena Control Plane Standardization Update
Wednesday October 9, 2024 1:30pm - 2:00pm EDT
We will provide an update on progress within SMPTE on the Catena initiative. The goal of this initiative is to create a standardized, open source control plane for cloud, on-prem and hybrid environments.
Speakers
CL

Chris Lennon

Ross Video
Wednesday October 9, 2024 1:30pm - 2:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

2:00pm EDT

Auditory Envelopment and Affective Touch Hypothesis
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
Anticipation and pleasure in response to music listening can lead to dopamine release in the human striatal system, a neural midbrain reward and motivation cluster. The sensation of auditory envelopment, however, may also in itself have a stimulating effect. Theoretical reason and circumstantial evidence are given why this could be the case, thereby possibly constituting an auditory complement to a newly discovered and studied percept, affective touch, originating from C-tactile fibres in the skin, stimulated certain ways. In a pilot test, abstract sounds were used to determine the audibility of low frequency inter-aural fluctuation. Naïve subjects aged between 6 and 96 years were all sensitive to the conditions tested and asked to characterize the stimuli in their own words. Based on these results, controlling low frequency inter-aural fluctuation in listeners should be a priority when recording, mixing, distributing and reproducing audio.
Moderators Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
Authors
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
1E03

2:00pm EDT

Convergence - The vision of one AV network
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
The landscape of Media Networks has become fragmented with standards and solutions which are partly incompatible. This often forces to implement multiple parallel networks into ProAV systems which causes huge cost and efforts and leaves the customer with very inflexible systems. In contrast, the promise and vision of media networking has been the opposite: To bring data and media traffic into one network structure with opportunities for flexible connections and interoperability. The session shows the dimension of the problem and possible paths for solutions. The session will explain the idea of ‘convergence’ with examples from actual productions.
Speakers
avatar for Henning Kaltheuner

Henning Kaltheuner

Owner, Independent consultant
Independent consultant for the ProAV industry. After > 25y of practical Live Sound Engineering, Product Management and Product Design, my main experience and focus of my consultancy is Strategic Consulting and Market Research. I am expert in customer motivations and user needs, brand... Read More →
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

2:00pm EDT

Acoustical Simulation and Calculation Techniques for Small Spaces
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
This presentation provides an insightful overview of statistical acoustic calculation techniques and geometrical acoustic modeling. It addresses the unique challenges associated with small to medium-sized spaces and delves into the intricacies of contemporary ray tracing software tools. Attendees will also learn about the latest advancements of using these tools in immersive audio environments. Additionally, the presentation will showcase comparative low-frequency analysis calculations obtained using a variety of tools, including boundary element methods (BEM), finite element
methods (FEM) and others. The session includes a series of real-world case studies to illustrate these concepts in practical applications.
Speakers
avatar for Peter D'Antonio

Peter D'Antonio

Director of Research, REDIacoustics
Dr. Peter D’Antonio is a pioneering sound diffusion expert. He received his B.S. from St. John’s University in 1963 and his Ph.D. from the Polytechnic Institute of Brooklyn, in 1967. During his scientific career as a diffraction physicist, Dr. D’Antonio was a member of the Laboratory... Read More →
avatar for Dirk Noy

Dirk Noy

Managing Director Europe, WSDG
BioDirk Noy has a Master of Science (MSc) Diploma in Experimental Solid State Physics from the University of Basel, Switzerland and is a graduate from Full Sail University, Orlando, USA, where he was one of John Storyk’s students.Since joining the consulting firm Walters-Storyk... Read More →
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
1E08

2:00pm EDT

You're a Gear Geek - Just Like Us! Staging Gear in Trade Shows, Print Media and Online
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
Drawing on research to published in our forthcoming book, Gear: Cultures of Audio and Music Technologies (The MIT Press), this workshop focuses on gear in trade show, print media, and online fora. Samantha Bennett and Eliot Bates first introduce 'gear', hardware professional audio recording technologies, and 'gear cultures', those being 'milieux where sociability is centred around audio technology objects’ (Bates and Bennett, 2022). With special focus on Part III of our book, this workshop examines the myriad ways we see gear 'staged' - in events including AES and NAMM, in diverse print media, including the Sweetwater catalogue and Sound on Sound, and online in fora including Gearspace.

So much of the fetishistic and technostalgic qualities associated with gear today exceed the materials and design; and it is precisely these fetish ideologies that become central to gear sociability. So how does gear become social, and where do gear cultures gather? Since the mid 1990s, we have seen technological objects transformed into gear when staged within these trade show / media / online milieux. In these spaces, gear is gassed to the point where it attains fetish status. When staged, gear is sometimes framed in sex and war metaphor (Lakoff and Johnson, 1980), and heavily draws on heritage, canon, and iconicity to amplify and maintain gear discourses. In all these trade show / print media / online fora, we see the erasure of women’s labour in order to maintain hegemonic masculinities that gear cultures rely upon. In this workshop, Samantha and Eliot use a range of examples - from online gear lovers gassing over gear components, to Foxy Stardust selling us a DAW - to show how gear is called upon to do extensive work in structuring social relations, and how these gear-centric social formations produce gear cultures.
Speakers
avatar for Samantha Bennett

Samantha Bennett

The Australian National University
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
1E16

2:00pm EDT

Soundgirls Presents Best Skills for Live Sound
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
Join SoundGirls as we explore the different skill sets required for various production environments, from clubs to arena and stadium tours, as well as corporate events. We'll discuss what it takes to become the ideal candidate in each setting and how to stay relevant as a freelance live sound engineer.

In addition, we'll cover the importance of education in audio engineering, focusing on career development, networking, certifications, and the value of continuing education.

The panel will include members from SoundGirls.


----

Barbara Adams is an audio engineer and educator with 30 years of experience in the music industry. She specializes in live sound and production management. Her strong and varied experience also includes recording and artist management. She has worked in Philadelphia’s top music venues, toured and worked festivals as FOH and monitor engineer, and educated and mentored the next generation since 2010.

Barbara is an Associate Teaching Professor at Rowan University and a vital part of their music industry program. In addition to teaching classes in recording technology and live sound, she helps students find opportunities to gain valuable experience both in and out of the university community.

Barbara advocates for creating more diversity in this industry while she inspires the next generation of techs to follow their dreams in the music industry. She is a member of the Audio Engineering Society, Recording Academy, and the Philadelphia chapter president of Soundgirls.


Carolyn Slothour is a freelance live sound engineer based in New York with 10 years of experience in the audio industry with a specialization in FOH, monitor, and systems engineering. She has recently worked with artists such as Matchbox Twenty, Rob Thomas and Michael Bublé, and has mixed at festivals including Download, When We Were Young, Corona Capital, Slam Dunk and Lollapalooza. In addition to engineering, she develops curriculum and teaches live sound engineering courses for a non-profit organization whose mission is to improve accessibility and diversity within the audio industry. When not working, Carolyn enjoys making electronic music and spinning fire.


Kitzy is an audio engineer with experience in live sound, recording, post-production, and podcasting. As a touring FOH engineer, she has worked with artists like Maude Latour, Devon Again, Leanna Firestone, and Bartees Strange. In addition to her work as a touring engineer, Kitzy owns a small audio production & rental company in Philadelphia.

Kitzy was the co-host of the independent music podcast Left of the Dial, as well as the engineer behind the board for Left of the Dial's live in-studio sessions. She is also one of the co-hosts for the upcoming relaunch of the SoundGirls podcast.

Kitzy strives to be an advocate for people who are gender non-conforming, neurodivergent, and/or disabled in the live sound and production industry. When she isn't working, Kitzy enjoys fixing and riding motorcycles, playing with her housemate's dog, and working on her own music.
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
1E09

2:00pm EDT

Is Audio AI Changing Audio Education?
Wednesday October 9, 2024 2:00pm - 3:15pm EDT
Quickly developing, it's often difficult to keep up with the latest in AI, and how AI based products are permeating into the industry. What AI based products are in use? Are they changing what we need to teach? What do we think will be coming next? Join us as a panel of educators and professionals (connected to AI in the industry) discuss these things, and answer your questions about and the state of AI in the industry. This panel is presented by the Technical Committee on Machine Learning and Artificla Intelligence and the Education Committee.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Brecht De Man

Brecht De Man

Head of Research, PXL University of Applied Sciences and Arts
Brecht is an audio engineer with a broad background comprising research, software development, management and creative practice. He holds a PhD from the Centre for Digital Music at Queen Mary University of London on the topic of intelligent software tools for music production, and... Read More →
MD

Michele Darling

Chair, Electronic Production and Design, Berklee College of Music
Wednesday October 9, 2024 2:00pm - 3:15pm EDT
1E07

2:00pm EDT

Challenges and Opportunities of Binaural Audio
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
This workshop session will bring together leading sound engineers to explore the challenges and opportunities in binaural audio. The session will begin with an overview of the various binaural rendering methods currently used in the industry. The panelists will share their experiences with immersive audio, focusing on the intricacies of creating high-quality binaural mixes. They will discuss the philosophies behind binaural delivery, such as realistic mix room experiences versus creative purposes, as well as monitoring and quality control methods, and the perceived quality issues in current binaural rendering techniques around tonal fidelity and externalisation. The discussion will cover various applications of binaural audio, including music production, film, and broadcasting/streaming.
Speakers
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
avatar for Vicki Melchior

Vicki Melchior

Consultant, Audio DSP and Software
avatar for Bob Katz

Bob Katz

Digital Domain, Inc.
Bob Katz, workshop co-chair, has been an audio engineer since 1971 and AES member since 1975. He’s played the Bb Clarinet since age 10. Currently President and mastering engineer at Digital Domain, he is an AES Fellow, engineered 3 Grammy-winning albums and numerous nominees. He... Read More →
avatar for Morten Lindberg

Morten Lindberg

Producer and Engineer, 2L (Lindberg Lyd, Norway)
Recording Producer and Balance Engineer with 43 GRAMMY-nominations, 35 of these in craft categories Best Engineered Album, Best Surround Sound Album, Best Immersive Audio Album and Producer of the Year. Founder and CEO of the record label 2L. Grammy Award-winner 2020.
avatar for Justin Gray

Justin Gray

Chief Engineer, Synthesis Sound
Justin Gray is an award-winning producer, mixing engineer and mastering engineer based in Toronto. Justin is a leader in the field of immersive audio mixing & mastering, where he works with artists, producers and labels worldwide to bring their music to life in the Dolby Atmos and... Read More →
avatar for Andrew Scheps

Andrew Scheps

Owner, Tonequake Records
Andrew Scheps has worked with some of the biggest bands in the world: Green Day, Red Hot Chili Peppers, Weezer, Audioslave, Black Sabbath, Metallica, Linkin Park, Hozier, Kaleo and U2. He’s worked with legends such as Johnny Cash, Neil Diamond and Iggy Pop, as well as indie artists... Read More →
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
1E15

2:00pm EDT

The Technical Legacy of Dr. Robert Moog
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
This year marks the 60th anniversary of Dr. Robert Moog's AES paper, "Voltage-Controlled Electronic Music Modules" (paper number 346), as well as the debut of the Moog modular synthesizer. This along with a subsequent paper, "A Voltage-Controlled Low-Pass High-Pass Filter for Audio Signal Processing" (paper number 413) presented in 1965, form the basis for generations of electronic instruments that followed. Throughout his career, Bob Moog worked closely with artists to solve technical challenges that led to musically satisfying instruments. In this session we'll reflect on the development of these designs as well as the impact they have had on the industry from some of those who worked with Moog as well as designers who have built on this technical foundation.

This panel will be followed by a book signing at the AES Booth.

A Brief Testimonial on the Occasion of the 60thAnniversary of Bob Moog’s Historic AES Presentation on his Prototype Voltage-Controlled Modular System
By Albert Glinsky, author of Switched On: Bob Moog and the Synthesizer Revolution, Oxford University Press, 2022

Sixty years ago this week, Robert Arthur Moog, a shy, unassuming young man of 30, not generally known to anyone outside the world of theremin kits, sat at his little AES display at the Barbizon-Plaza Hotel in Manhattan, his table draped with his wife’s Indian print bedspread. Sitting atop it was his self-described “Abominatron”—four handmade modules with sheet metal panels hastily spray-painted, with jacks and knobs identified only by paper labels stuck on with rubber cement. He was surrounded on all sides by the titans of the audio electronics industry: Ampex, Scully, 3M. It was intimidating to a fault; he sat, hoping a few passersby would care to don a pair of earphones to sample the aural cherry pies or pumpkins of his State Fair-like presentation—a pleasant curiosity at best, he figured. Then something unexpected happened. As he famously recalled, a prominent visitor approached and uttered the words, “I’d like a couple of these and a couple of these.” He was in business, and he hadn’t planned on anything like that. History was made right then and there.


During this same 16th Annual AES convention week, on Wednesday, October 14th, at 9:30 in the morning, a session titled, MUSIC AND ELECTRONICS was chaired by the venerable Harald Bode. The second presenter of the morning, a certain Robert A. Moog of the R. A. Moog Company of Trumansburg NY, strode to the podium and read a 9-page typed manuscript titled, “Voltage-Controlled Electronic Music Modules.” Footnotes at the bottom of his last page credited a few predecessors on whose giants’ shoulders he felt he was standing—figures like Myron Schaeffer, Hugh LeCaine, and Bode himself. And the crowd was formidable: Le Caine was there, Bode, of course, and Harry F. Olson and Herbert Belar, inventors of the already classic RCA Mark I and Mark II synthesizers.


After the convention, Bob began selling custom variations on his prototype modules, and the rest, of course became history. But what some may not realize—something that really struck me as Bob’s biographer—is the influence, not just of the instruments themselves, but of Bob’s generosity in sharing the “recipe,” as I like to call it, for this groundbreaking system that would revolutionize music and sound art. Less than a year after his AES presentation, Bob published a detailed 7-page article in the July 1965 issue of the Journal of the Audio Engineering Society, replete with generous block diagrams, schematics, and photos. In the proprietary world of our current technology, this selfless, open-source act is astounding. Not that others hadn’t done this sort of thing before, and Bob had certainly come of age in the world of hobbyist magazines, DIY projects, and the brotherhood of tinkerers. But with the commercial potential of Bob’s system, it fairly gave the whole game away. Bob had inadvertently presented his own giant’s shoulders to the world, upon which so many others would stand—some eventually acknowledging the pedestal on which they held their own inventions, some not. But then, Bob’s passion was sheer invention, and often, as we know, to the detriment of profit and the big picture. But that was Bob: generosity of spirit, and the delight of scattering his seeds widely, even if some of it occasionally took root in competitors’ pastures. It’s not a stretch to say, and I say it with full confidence and conviction based on the facts: Buchla, ARP, EMS, and numerous others that followed—none of them would have been possible without Bob’s published recipe. The timeline bears it out, and some have openly admitted it.


The seminal genius of Bob’s work (and there are so many aspects to it), might be summed up by two statements he made in his AES talk that October morning in 1964. After citing several early historical attempts at producing music electronically, he explained, “The failure of any of these instruments to attain widespread success can be attributed at least partially, to the limitation on the degree of control which the performer is able to exercise over the instrument’s sounds.” He concluded his paper by saying, “In particular, the setting up of prototype experimental musical instruments, and the remote-control processing of conventional audio signals are ideal applications for the voltage-controlled modules.” Ideal indeed!! Amen Bob. And what a legacy you’ve left us. Thank you, and congratulations on such a rich 60-year legacy!!
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
MM

Michelle Moog-Koussa

Executive Director, Bob Moog Foundation
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
Stage

2:00pm EDT

A study on the relative accuracy and robustness of the convolutional recurrent neural network based approach to binaural sound source localisation.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Binaural sound source localization is the task of finding the location of a sound source using binaural audio as affected by the head-related transfer functions (HRTFs) of a binaural array. The most common approach to this is to train a convolutional neural network directly on the magnitude and phase of the binaural audio. Recurrent layers can then also be introduced to allow for consideration of the temporal context of the binaural data, as to create a convolutional recurrent neural network (CRNN).
This work compares the relative performance of this approach for speech localization on the horizontal plane using four different CRNN models based on different types of recurrent layers; Conv-GRU, Conv-BiGRU, Conv-LSTM, and Conv-BiLSTM, as well as a baseline system of a more conventional CNN with no recurrent layers. These systems were trained and tested on datasets of binaural audio created by convolution of speech samples with BRIRs of 120 rooms, for 50 azimuthal directions. Additive noise created from additional sound sources on the horizontal plane were also added to the signal.
Results show a clear preference for use of CRNN over CNN, with overall localization error and front-back confusion being reduced, with it additionally being seen that such systems are less effected by increasing reverb time and reduced signal to noise ratio. Comparing the recurrent layers also reveals that LSTM based layers see the best overall localisation performance, while layers with bidirectionality are more robust, and so overall finding a preference for Conv-BiLSTM for the task.
Speakers
avatar for Jago T. Reed-Jones

Jago T. Reed-Jones

Research & Development Engineer, Audioscenic
I am a Research & Development Engineer at Audioscenic, where we are bringing spatial audio to people's homes using binaural audio over loudspeakers. In addition, I am finishing a PhD at Liverpool John Moores University looking at use of neural networks to achieve binaural sound source... Read More →
Authors
avatar for Jago T. Reed-Jones

Jago T. Reed-Jones

Research & Development Engineer, Audioscenic
I am a Research & Development Engineer at Audioscenic, where we are bringing spatial audio to people's homes using binaural audio over loudspeakers. In addition, I am finishing a PhD at Liverpool John Moores University looking at use of neural networks to achieve binaural sound source... Read More →
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Authoring Inter-Compatible Flexible Audio for Mass Personalization
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
The popularity of internet-based services for delivering media, i.e., audio and video on demand, creates an opportunity to offer personalized media experiences to audiences. A significant proportion of users are experiencing reduced enjoyment, or facing inaccessibility, due to combinations of different impairments, languages, devices and connectivity, and preferences. Audio currently leads video in the maturity of object-based production and distribution systems; thus, we sought to leverage existing object-based audio tools and frameworks to explore creation and delivery of personalized versions. In this practice-based investigation of personalization affordances, an immersive children’s fantasy radio drama was re-authored within five "dimensions of personalization", motivated by an analysis of under-served audiences, to enable us to develop an understanding of what and how to author personalized media. The dimensions were Duration, Language, Style, Device and Clarity. Our authoring approaches were designed to result in alternative versions that are inherently inter-compatible. The ability to combine them exploits the properties of object-based audio and creates the potential for mass personalization, even with a modest number of options within each of the dimensions. The result of this investigation is a structured set of adaptations, based around a common narrative. Future work will develop automation and smart network integration to trial the delivery of such personalized experiences at scale, and thereby validate the benefits that such forms of media personalization can bring.
Speakers
avatar for Craig Cieciura

Craig Cieciura

Research Fellow, University of Surrey
Craig graduated from the Music and Sound Recording (Tonmeister) course at The University of Surrey in 2016. He then completed his PhD at the same institution in 2022. His PhD topic concerned reproduction of object-based audio in the domestic environment using combinations of installed... Read More →
Authors
avatar for Craig Cieciura

Craig Cieciura

Research Fellow, University of Surrey
Craig graduated from the Music and Sound Recording (Tonmeister) course at The University of Surrey in 2016. He then completed his PhD at the same institution in 2022. His PhD topic concerned reproduction of object-based audio in the domestic environment using combinations of installed... Read More →
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Automatic Corrective EQ for Measurement Microphone Emulation
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
This study investigates automatic corrective equalisation (EQ) to adjust the frequency response of microphones with non-flat responses to match that of a flat frequency response measurement microphone. Non-flat responses in microphones can cause significant colouration, necessitating correction for accurate sound capture, particularly in spectral analysis scenarios. To address this, 10 non-flat microphones were profiled in an anechoic chamber, and a 1/3 octave digital graphic equaliser (GEQ) was employed to align their spectra with that of an industry-standard reference microphone. The system's performance was evaluated using acoustic guitar recordings, measuring spectral similarity between Pre- and Post-Corrected recordings and the reference microphone with objective metrics. Results demonstrated improvements in spectral similarity across the metrics, confirming the method's effectiveness in correcting frequency response irregularities. However, limitations such as the inability to account for proximity effects suggest the need for further refinement and validation in diverse acoustic environments.
Speakers
avatar for Matthew Cheshire

Matthew Cheshire

Lecturer in Audio Engineering / Researcher, Birmingham City University
Authors
avatar for Matthew Cheshire

Matthew Cheshire

Lecturer in Audio Engineering / Researcher, Birmingham City University
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Corelink Audio: The Development of A JUCE-based Networked Music Performance Solution
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Existing Networked Music Performance (NMP) solutions offer high-fidelity audio with minimal latency but often lack the versatility needed for multi-plugin configurations and multi-modal integrations, while ensuring ease of use and future development. This paper presents a novel NMP solution called Corelink Audio which aims to address these limitations. Corelink Audio is built on the Corelink network framework, a data agnostic communication protocol, and JUCE, an audio plugin development framework. Corelink audio can be used flexibly as an Audio Unit (AU) or Virtual Studio Technology (VST) plugin inside different host environments and can be easily integrated with other audio or non-audio data streams. Users can also extend the Corelink Audio codebase to tailor it to their own specific needs using the available JUCE and Corelink documentation. This paper details the architecture and technical specifics of the software. The performance of the system, including latency measurements and audio artifacts under varying conditions, is also evaluated and discussed.
Speakers
avatar for Zack Nguyen

Zack Nguyen

Software Developer, NYU IT
Authors
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Do We Need a Source Separation for Automated Subtitling on Media Contents?
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
In this paper, we investigate the efficacy of a speech and music source separation technique on an automated subtitling system under different signal-to-noise ratios (SNRs). To this end, we compare the generated subtitle errors by measuring the word error rate (WER) with and without source separation applied to speech in music. Experiments are first performed on a dataset by mixing speech from the LibriSpeech dataset with music from the MUSDB18-HQ dataset. Accordingly, it is revealed that when the SNR is below 5 dB, using separated speech yields the lowest subtitle error. By contrast, when the SNR exceeds 5 dB, using the mixture audio shows the lowest subtitle error. On the basis of these findings, we propose an automated subtitling system that dynamically chooses between using mixture audio or separated speech to generate subtitles. The system utilizes the estimated SNR as a threshold to decide whether to apply source separation, achieving the lowest average WER under various SNR conditions.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Miniaturized Full-Range MEMS Speaker for In-Ear Headphones
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
This paper reports a highly miniaturized loudspeaker for in-ear applications. It is manufactured using MEMS (micro-electromechanical systems) technology, a technology that is well known e.g. from MEMS microphones and accelerometers. The speaker is based on a mechanically open design, featuring a perfectly decoupled piezoelectric bending actuator. To provide good acoustic sealing, the actuator is surrounded by an acoustic shield. Measurements performed on loudspeaker prototypes attached to an artificial ear simulator showed that despite the small size of only 2.4 x 2.4 mm², sound pressure levels of 105 dB can be achieved within the full audio range. Moreover, low total harmonic distortion (THD) of less than 0.4 % at 90 dB and 1 kHz have been observed.
Speakers
avatar for Fabian Stoppel

Fabian Stoppel

Head Acoustic Systems and Microactuators, Fraunhofer Institute for Silicon Technology
Fabian Stoppel is an expert in the field of micro-electro-mechanical systems (MEMS) and holds a PhD degree in electrical engineering. Since 2010 he is with the Fraunhofer Institute for Silicon Technology (ISIT), Germany, where he is heading the group for acoustic micro systems. Dr... Read More →
Authors
avatar for Fabian Stoppel

Fabian Stoppel

Head Acoustic Systems and Microactuators, Fraunhofer Institute for Silicon Technology
Fabian Stoppel is an expert in the field of micro-electro-mechanical systems (MEMS) and holds a PhD degree in electrical engineering. Since 2010 he is with the Fraunhofer Institute for Silicon Technology (ISIT), Germany, where he is heading the group for acoustic micro systems. Dr... Read More →
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Music Auto-tagging in the long tail: A few-shot approach
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
In the realm of digital music, using tags to efficiently organize and retrieve music from extensive databases is crucial for music catalog owners. Human tagging by experts is labor-intensive but mostly accurate, whereas automatic tagging through supervised learning has approached satisfying accuracy but is restricted to a predefined set of training tags. Few-shot learning offers a viable solution to expand beyond this small set of predefined tags by enabling models to learn from only a few human-provided examples to understand tag meanings and subsequently apply these tags autonomously. We propose to integrate few-shot learning methodology into multi-label music auto-tagging by using features from pre-trained models as inputs to a lightweight linear classifier, also known as a linear probe. We investigate different popular pre-trained features, as well as different few-shot parametrizations with varying numbers of classes and samples per class. Our experiments demonstrate that a simple model with pre-trained features can achieve performance close to state-of-the-art models while using significantly less training data, such as 20 samples per tag. Additionally, our linear probe performs competitively with leading models when trained on the entire training dataset. The results show that this transfer learning-based few-shot approach could effectively address the issue of automatically assigning long-tail tags with only limited labeled data.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

On-Device Automatic Speech Remastering Solution in Real Time
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
With the development of AI technology, there are many attempts to provide new experiences to users by applying AI technology to various multimedia devices. Most of these technologies are provided through server-based AI models due to the large model size. In particular, most of the audio AI technologies are applied through apps and it is server-based in offline AI models. However, there is no doubt that AI technology which can be implemented in real time is important and attractive for streaming service devices such as TVs. This paper introduces an on-device automatic speech remastering solution. The automatic speech remastering solution indicates extracting speech in real-time from the on-device and automatically adjusts the speech level considering the current background sound and volume level of the device. In addition, the automatic speech normalization technique that reduces the variance in speech level for each content is applied. The proposed solution provides users with a high understanding and immersion in the contents by automatically improving the delivery of speech and normalizing speech levels without manually controlling the volume level. There are three key points in this paper. The first is a deep learning speech extraction model that can be implemented in real-time on TV devices, the second is an optimized implementation method using the DSP and NPU, and last is audio signal processing for the speech remastering to improve speech intelligibility.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Real-time Recognition of Speech Emotion for Human-robot Interaction
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
In this paper, we propose a novel method for real-time speech emotion recognition (SER) tailored for human-robot interaction. Traditional SER techniques, which analyze entire utterances, often struggle in real-time scenarios due to their high latency. To overcome this challenge, the proposed method breaks down speech into short, overlapping segments and uses a soft voting mechanism to aggregate emotion probabilities in real time. The proposed real-time method is applied to an SER model comprising the pre-trained wav2vec 2.0 and a convolutional network for feature extraction and emotion classification, respectively. The performance of the proposed method was evaluated on the KEMDy19 dataset, a Korean emotion dataset focusing on four key emotions: anger, happiness, neutrality, and sadness. Consequently, applying the real-time method, which processed each segment with a duration of 0.5 or 3.0 seconds, resulted in relative reduction of unweighted accuracy by 10.61% or 5.08%, respectively, compared to the method that processed entire utterances. However, the real-time factor (RTF) was significantly improved.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:15pm EDT

Mixing for Dolby Atmos Music
Wednesday October 9, 2024 2:15pm - 3:15pm EDT
Joseph J. Chudyk is a multi-platinum mixer and educator with significant experience mixing albums in immersive audio. In addition to teaching at Nazareth College in Rochester, NY, some of his recent notable Atmos mixes include work with Elle King and Miranda Lambert, and Mitchell Tenpenny. In this session Joe will play some of his Dolby Atmos Music mixes, talking about his experience, sharing tips and best practices, and more, with ample opportunities for attendees to ask questions.
Speakers
JC

Joe Chudyk

Joseph J. Chudyk is a multi-platinum mixer having worked with Kane Brown, Mitchell Tenpenny, Teddy Swims, Ghost, IDK, Newsboys, MxPx, Nathaniel Rateliff, Joe Pug, The National Reserve with members from the likes of Gregory Alan Isakov, Jeff Beck, Polar Bear Club, Air Dubai, OneRepublic... Read More →
Wednesday October 9, 2024 2:15pm - 3:15pm EDT
1E06

2:30pm EDT

A framework for high spatial-density auditory data displays using Matlab and Reaper
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
This research aimed to develop a software framework to study and optimize mapping strategies for complex data presented with auditory displays with high spatial resolution, such as wave-field synthesis and higher-order ambisonics systems. Our wave field synthesis system, the Collaborative-Research Augmented Immersive Virtual Environment Laboratory (CRAIVE-Lab), has a 128 full-range loudspeaker system along the circumference of the lab. We decided to use available software music synthesizers because they are built for excellent sound quality, and much knowledge exists on how to program analog synthesizers and other common methods for a desired sound output. At the scale of 128 channels, feeding 128 synthesizers with complex data was not practical for us for initial explorations because of computational resources and the complexity of data flow infrastructure. The proposed framework was programmed in Matlab, using Weather data from the NOAA database for initial exploration. Data is processed from 128 weather stations from East to West in the US spatially aligned with latitude. A MIDI script, in sequential order for all 128 channels, is compiled from converted weather parameters like temperature, precipitation amount, humidity, and wind speed. The MIDI file is then imported into Reaper to render a single sound file using software synthesizers that are operated with the MIDI file control data instructions. The rendered file is automatically cut into the 128 channels in Matlab and reimported into Reaper for audio playback.
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
1E03

2:30pm EDT

Converged AV Networking
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
The 'Converged AV Networking' presentation goes deeper into how modern AV protocols can harmoniously operate on a unified network infrastructure. It examines the complexities and challenges associated with clock synchronization, queuing mechanisms, and specific protocol requirements. Key topics covered include AES67 for audio over IP, ST2110 for professional media over managed IP networks, AVB/Milan for audio video bridging, PtPv2 for precise time protocol, gPtP for general precision time protocol, and QoS strategies like credit-based and strict priority queuing.
Speakers
avatar for Bart Swinnen

Bart Swinnen

CEO, Luminex Network Intelligence
Bart Swinnen, founder and CEO of Luminex Network Intelligence.Bart is for more than 2 decades closely involved in making AV Networking easy in the professional lighting, audio and video industry.With Luminex he provides a range of network switch products optimized for converged AV... Read More →
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
3D04 (AES and AIMS Alliance Program Room)

3:00pm EDT

Immersive Voice and Audio Services (IVAS) Codec – The New 3GPP Standard for Immersive Communication
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
The recently standardized 3GPP codec for Immersive Voice and Audio Services (IVAS) is the first fully immersive communication codec designed for 5G mobile systems. The IVAS codec is an extension of the mono 3GPP EVS codec and offers additional support for coding and rendering of stereo, multi-channel, scene-based audio (Ambisonics), objects and metadata-assisted spatial audio. The IVAS codec enables completely new service scenarios with interactive stereo and immersive audio in communication, content sharing and distribution. This paper provides an overview of the underlying architecture and new audio coding and rendering technologies. Listening test results show the performance of the new codec in terms of compression efficiency and audio quality.
Moderators Speakers Authors
avatar for Adriana Vasilache

Adriana Vasilache

Nokia Technologies
avatar for Andrea Genovese

Andrea Genovese

Research, Qualcomm
Andrea Genovese is a Senior Researcher Engineer at Qualcomm Technologies Inc. working in Multimedia R&D. Andrea specializes in spatial audio and psychoacoustics, acoustic simulations, networked immersive distributed audio, and signal processing for environmental awareness. In 2023... Read More →
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
1E03

3:00pm EDT

Enabling Next Generation Audio - SMPTE ST 2110-41 with Serialized ADM Audio Metadata
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
One of achievements of the SMPTE ST 2110 suite of standards is to bring flexibility and agility to modern IP media systems. However, until now the format of the audio has been mostly static. With the recent publication of SMPTE ST 2110-41 and ST 2127-2, metadata can now be carried alongside the audio allowing adaptation to multiple audio formats. This enables more personalized and accessible services and brings operational benefits. The presentation will explain the opportunities of this transition and present a Paris Olympics case-study that used the first commercial products.
Speakers
avatar for James Cowdery

James Cowdery

Dolby Laboratories
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

3:00pm EDT

Technical Committee Meeting: Fiber Optics for Audio
Wednesday October 9, 2024 3:00pm - 4:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.
Wednesday October 9, 2024 3:00pm - 4:00pm EDT
2D02/03

3:15pm EDT

Remote Live Events for Broadcast-music/spoken word
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
With upwards of fifty remotes each year, NY Public Radio broadcasts a great variety of events from the very simple to large scale events that can involve a many different challenges, including connectivity issues and internet issues with getting audio back to the studios for live broadcasts, dealing with microphone techniques for the greatest orchestras in the world at Carnegie Hall to dealing with the elements with live broadcasts in Central Park or other outdoor locations around NYC, dealing with cooperating with p.a.'s and unions at various sites, to broadcasts at a politician's residence. Each of these kinds of events might entail unique technical and personnel challenges, but there's also a lot of common sense approaches to working with each of these challenges, as well as applying new technologies as they come along to simplify things, and hopefully make things easier over the years to deal with these challenges.

We will discuss in detail two recent events, "This Land" a live broadcast of Classical, jazz, and Americana from Brooklyn Bridge Park and World Orchestra Week (WOW!) a Youth orchestra festival from Carnegie Hall, contrasting indoor vs outdoor and amplified vs unamplified.
Speakers
avatar for Edward Haber

Edward Haber

For 43 years I was an engineer at WNYC, then New York Public Radio (encompassing WNYC, WQXR, and NJPR); and for the last 36 of those years in charge of WNYC and latterly WQXR's remote recording, including recording the best orchestras in the world as part of WQXR's Carnegie Live radio... Read More →
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
1E08

3:15pm EDT

Gender Equity Townhall
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
The Annenberg Inclusion Initiative released a study demonstrating the dismal gender balance that continues to plague our industry. While there has been much positive change in the past few decades, there is still a lot of work to be done. Join representatives of the leading gender equity organizations including Women's Audio Mission, We Are Moving the Needle, She Is the Music, Soundgirls, and We Make Noise in a discussion on our common goals to change the gender imbalance in the industry moderated by Singer, Songwriter and Producer Eve Horne, founder of We Are the Unheard.
Speakers
avatar for Jeanne Montalvo

Jeanne Montalvo

Engineer/Producer, Self
Jeanne Montalvo is a Grammy-nominated audio engineer and award-winning radio producer. She was selected amongst thousands of applicants as the 2018 EQL resident at Spotify Studios and Electric Lady Studios in New York City, assisting in the recording process for artists like John... Read More →
avatar for Terri Winston

Terri Winston

Executive Director, Women's Audio Mission
Women's Audio Mission (WAM) is dedicated to closing the chronic gender gap in the music, recording and technology industries. Less than 5% of the people creating and shaping the sounds that make up the soundtrack of our lives are women and gender-expansive people. WAM is “changing... Read More →
EB

Erin Barra

MAMMA BARRA LLC; Artist/Manager If Alicia Keys and Nelly Furtado collided with Daft Punk and Esthero in the vast circuitry of an Akai APC 40, you'd probably end up with an experience as rich and soulful as Erin Barra. Joining the likes of MNDR and Tokimonsta, Erin is part of a... Read More →
avatar for Ann Mincieli

Ann Mincieli

Producer, Jungle City Studios
Ann Mincieli, longtime engineer and studio coordinator for Alicia Keys, has channelled her talents into making Jungle a technical prowess featuring the very best in vintage and modern technology. With over twenty years of experience in the music business, Ann Mincieli has travelled... Read More →
avatar for Eve Horne

Eve Horne

Peak Music
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
1E09

3:15pm EDT

Kseniya Kawko: Enveloping Masterclass
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
The masterclass consists of two parts, the first dedicated to jazz. Kseniya details 3D studio productions and live recordings from a German club “Unterfahrt”, featuring jazz legends among large ensembles and stunning new bands. The second part presents new msm-studios 3D tracks, including from the latest Boris Blank and Peter Gabriel albums.

This masterclass series, featuring remarkable recording artists, is a chance to hear 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Kseniya Kawko

Kseniya Kawko

Tonmeister, msm-studios & Jazz Club Unterfahrt
Kseniya Kawko is a producer and recording engineer specialized in classical music and jazz. She holds Master of Music degrees from two world-renowned audio programs: Sound Recording, McGill University (Montréal, Canada) and Musikregie / Tonmeister, Hochschule für Musik Detmold (Germany... Read More →
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
3D06

3:15pm EDT

Implementing WHO safe listening standards: Insights from the Serendipity Arts Festival
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
This tutorial presents a detailed examination of the first known large-scale implementation of the WHO Global Standard for Safe Listening Venues and Events at the 2023 Serendipity Arts Festival, India's largest multidisciplinary arts event. The case study highlights the methods used to monitor and manage sound levels, the design and deployment of sound systems, and the provision of personal hearing protection, training and information. The session will delve into the practical challenges encountered, the strategies employed to adhere to the WHO Global Standard, and the outcomes of these efforts. Attendees will gain an understanding of the complexities involved in applying safe listening principles in a live event context and the implications for future large-scale events.
Speakers
avatar for Adam Hill

Adam Hill

Associate Professor of Electroacoustics, University of Derby
Adam Hill is an Associate Professor of Electroacoustics at the University of Derby where he leads the Electro-Acoustics Research Lab (EARLab) and runs the MSc Audio Engineering program. He received a Ph.D. from the University of Essex, an M.Sc. in Acoustics and Music Technology from... Read More →
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
1E16

3:30pm EDT

Book Signing: The Technical Legacy of Dr. Robert Moog
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
A book signing of Switched On: Bob Moog and the Synthesizer Revolution by Albert Glinsky will be held following "The Technical Legacy of Dr. Robert Moog" panel on the main stage. Present for the signing will be Larry Fast, legendary synthesist, Steve Dunnington, Vice President of Engineering, Moog Music, Erik Norlander, Director of Keyboard Development for the inMusic Family of Brands, and Michelle Moog-Koussa, Executive Director of the Bob Moog Foundation. Copies of the biography will be for sale at the signing.
Speakers
MM

Michelle Moog-Koussa

Executive Director, Bob Moog Foundation
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
AES Membership Booth

3:30pm EDT

Enhancing Spatial Post-Filters through Non-Linear Combinations
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
This paper introduces a method to enhance the spatial selectivity of spatial post-filters estimated with first-order directional signals. The approach involves applying non-linear transformations on two different spatial post-filters and combining them with weights found by convex optimization of the resulting directivity patterns. The estimation of the post-filters is carried out similarly to the Cross Pattern Coherence (CroPaC) algorithm. The performance of the proposed method is evaluated in a two- and three-speaker scenario with different reverberation times and angular distances of the interfering speaker. The signal-to-interference, signal-to-distortion, and signal-to-artifact ratios are used for evaluation. The results show that the proposed method can improve the spatial selectivity of the post-filter estimated with first-order beampatterns. Using first-order patterns only, it even achieves better spatial separation than the original CroPaC post-filter estimated using first- and second-order signals.
Moderators Speakers
avatar for Stefan Wirler

Stefan Wirler

Aalto University
I'm a PhD student at the Aalto Acoustics Labs, in the Group of Ville Pulkki. I started my PhD studies in 2020. I received my Masters degree in Electrical Engineering from the Friedrich-Alexander University Erlangen-Nuremberg (FAU). My Master thesis was condicted at the AudioLabs in... Read More →
Authors
avatar for Stefan Wirler

Stefan Wirler

Aalto University
I'm a PhD student at the Aalto Acoustics Labs, in the Group of Ville Pulkki. I started my PhD studies in 2020. I received my Masters degree in Electrical Engineering from the Friedrich-Alexander University Erlangen-Nuremberg (FAU). My Master thesis was condicted at the AudioLabs in... Read More →
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
1E03

3:30pm EDT

Audio for Remote Live Production
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
More and more of live production is being done in a location-agnostic manner, with the processing, storage and control remote from the acquisition location. Whilst the video transport requirements often gather the headlines, the audio requirements can be far more complex. This presentation will unpack the technologies used to connect and transport audio, examine the dependencies with video and the different systems requirements in terms of timing references and how to have as much resilience as possible.
Speakers
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

3:30pm EDT

Newly Discovered Jazz Recordings at New York Public Library
Wednesday October 9, 2024 3:30pm - 4:15pm EDT
Last year saw the release of the Grammy-nominated album “Evenings at the Village Gate – John Coltrane with Eric Dolphy” on Verve. It was the result of many years of intensive detective work by Jessica Wood, Assistant Curator at New York Public Library’s Music & Recorded Sound Division. A long-missing cache of tapes from Richard Striker’s Institute of Sound, recorded by legendary New York engineer Richard Alderson, had been discovered and located by Wood.

Wood, Alderson, and NYPL Audio Preservation Engineer Jeff Willens then went through hundreds of tapes. Among them were unreleased studio and/or live recordings by Bob Dylan, John Coltrane, Nina Simone, Art Blakey, Thelonious Monk, and many other jazz and classical artists, as well as performers as diverse as Tom Paxton, Moondog, and poet Hugh Romney (aka Wavy Gravy).

In this talk, we will explain how this collection of vital, previously unknown recorded documents of the 1960s New York music scene was “lost,” located, and reclaimed by NYPL. We will also tell the story about how two of these unheard reels were brought to the public, filling some gaps in the Coltrane canon. And we will also share why a maverick NYC engineer at the top of his game, left the city and his invaluable tapes behind.

We will be playing samples from the Striker Collection during the talk and inviting attendees to help us “crowdsource” some of the unidentified recordings therein.
Speakers
avatar for Jeff Willens

Jeff Willens

Media Preservation Engineer, New York Public Library
avatar for Jessica Wood

Jessica Wood

Assistant Curator, Music Division and Rodgers & Hammerstein Archives of Recorded Sound, New York Public Library
Jessica is the Assistant Curator for the Music Division as well as the Rodgers & Hammerstein Archives of Recorded Sound at the New York Public Library for the Performing Arts. She previously worked as a project archivist at the Library of Congress' Music Division and as a sound recordings... Read More →
Wednesday October 9, 2024 3:30pm - 4:15pm EDT
1E07

3:30pm EDT

Student Mix Critiques
Wednesday October 9, 2024 3:30pm - 4:30pm EDT
These non-competitive sessions are an opportunity for AES student members at all stages in their studies to receive feedback on their mixes from a panel of industry professionals – live and in-person. Join us to hear mixes by other students, and get tips, tricks, and advice to push your skills to the next level! Advance submission procedures for each event will be posted at www.aesstudents.org. Very limited sign up will be available on-site. Maybe one of your mixes can be featured!
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
Wednesday October 9, 2024 3:30pm - 4:30pm EDT
1E06

4:00pm EDT

The Impact of Height Microphone Layer Position on Perceived Realism of Organ Recording Reproduction
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
For on-site immersive recordings, height microphones are often placed carefully to avoid a distorted or unrealistic image, with many established immersive microphone arrays placing the height microphones 1.5 m or less above the horizontal layer. However, with an instrument so acoustically symbiotic with its space as the pipe organ, the impact of non-coincident height microphone placement has not previously been explored in-depth. Despite this, the pipe organ's radiation characteristics may benefit from non-coincident height microphone placement, providing subjectively improved tone color without sacrificing perceived realism. Subjective listening tests were conducted comparing a pipe organ recording with coincident and non-coincident height microphone positions. The findings of this case study conclude that non-coincident height microphone placement does not significantly impact perceived realism of the immersive organ recording.
Moderators Speakers
avatar for Jessica Luo

Jessica Luo

Graduate Student, New York University
Authors
avatar for Jessica Luo

Jessica Luo

Graduate Student, New York University
avatar for Garrett Treanor

Garrett Treanor

New York University
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
1E03

4:00pm EDT

Realtime distance collaboration in the performing arts
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
Is it possible for musicians to play together over the Internet in real time? The answer is yes. It is possible up to a distance of 3000 km. When using MVTP (Modular Video Transmisson Platform) technology based on the use of optimized JPEG XS codec running on a powerful FPGA. The presentation describes the workflow and user case of using Ultra low latency CESNET MVTP technology in a real environment. During a shared musical performance between two or three locations or during remote teaching or workshops at a music university.
Speakers
PK

Petr Krkavec

DISK Systems
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

4:00pm EDT

Designing an Immersive Audio Production Studio - an in depth Study of Rue Boyer, Paris
Wednesday October 9, 2024 4:00pm - 5:30pm EDT
With consumers able to experience immersive audio more easily and cost-effectively than ever, and both Spotify as well as Apple Music platforms offering playout facilities for immersive content, demand is seemingly continuing to accelerate. The presentation offers a discussion of acoustical and architectural requirements and challenges as well as the relevant standards to adhere to. As a case study, the presentation will review Paris' newest studio, including the design strategies as well as advanced predictive acoustic analysis involved in creating the studio's design. It will also the
studio's unique elevator assisted “disappearing console” strategy for its Studio A Control Room changing into a multi-purpose live room and classroom.
Speakers
avatar for Sergio Molho

Sergio Molho

Partner, Director of Business Development, WSDG Walters-Storyk Design Group
Partner, Director of Business Development
avatar for John Storyk

John Storyk

Founding Partner, WSDG
Since graduating from Princeton as an architect in 1968, and co-founding WSDG (Walters-Storyk Design Group) with his wife, Beth Walters in 1986, John Storyk has been involved with the design of over 3500 recording studios, corporate conference rooms, performance venues, broadcast... Read More →
avatar for Dirk Noy

Dirk Noy

Managing Director Europe, WSDG
BioDirk Noy has a Master of Science (MSc) Diploma in Experimental Solid State Physics from the University of Basel, Switzerland and is a graduate from Full Sail University, Orlando, USA, where he was one of John Storyk’s students.Since joining the consulting firm Walters-Storyk... Read More →
Wednesday October 9, 2024 4:00pm - 5:30pm EDT
1E15

4:15pm EDT

Immediate Family Film at AES NY 2024
Wednesday October 9, 2024 4:15pm - 5:15pm EDT
The Immediate Family (the modern iteration of a legendary studio ensemble known as “The Section”) is a unique group of iconic musicians composed of four of the most recorded, respected, and sought-after players in modern music: Danny Kortchmar (guitar and vocals), Waddy Wachtel (guitar and vocals), Leland Sklar (bass), Russ Kunkel (drums) and the addition of prominent touring, session guitarist and songwriter Steve Postell (guitar and vocals). Frequent collaborators both in the studio and on stage, their work can be heard on albums from Jackson Browne, James Taylor, Linda Ronstadt, Stevie Nicks, Keith Richards, Warren Zevon, Graham Nash, Neil Young, David Crosby, and many, many more. Kortchmar, Sklar, and Kunkel have worked together since the early ‘70s and made up three-quarters of The Section, best known for both their studio and live work in support of some of the top-selling singer/songwriters and solo singers in the history of music, as well as their acclaimed instrumental albums. (By the mid-70s, they were later joined by Wachtel). The film reveals the machinery behind the booming era of the singer-songwriter, when the talents of these four musicians were in furious demand. Collectively, these musicians helped define the sound of a generation. The band is managed by Industry veterans, Fred Croshal of Croshal Entertainment Group and David Helfant of Arpeggio Entertainment, who also served as Executive Producers of the band’s Documentary. Join band members Danny Kortchmar and Waddy Wachtel, and record producers Russ Titelman and Niko Bolas, and acclaimed filmmaker Denny Tedesco to discuss this award- winning documentary.
Speakers
avatar for Danny Kortchmar

Danny Kortchmar

Danny Kortchmar is a guitarist, songwriter, and producer, renowned for his work with some of the biggest names in rock and pop music. Kortchmar became a prominent figure in the 1970s as part of the West Coast music scene. He has collaborated extensively with artists like James Taylor... Read More →
avatar for Denny Tedesco

Denny Tedesco

Acclaimed filmmaker Denny Tedesco (director of The Wrecking Crew and now the full-length documentary Immediate Family), grew up in Los Angeles. In 1996, Denny set out to tell the story of his terminally ill father, legendary session guitarist Tommy Tedesco.  “The Wrecking Crew... Read More →
avatar for Niko Bolas

Niko Bolas

Niko Bolas is an award-winning music producer and recording engineer with credits that include work with such artists as LeAnn Rimes, The Mavericks, Prince, Neil Young, Tyler Bank, Boz Scaggs, Sheryl Crow, Toto, Los Lobos, Jewel, OneRepublic, Johnny Cash, and Keith Urban.He is also... Read More →
avatar for Russ Titelman

Russ Titelman

Russ Titelman is an acclaimed American record producer, known for his versatile work across rock, pop, and jazz genres. Titelman began his career working with Phil Spector as a guitarist and singer in the early 60’s. Then, after being signed to Don Kirchner's publishing company... Read More →

Wednesday October 9, 2024 4:15pm - 5:15pm EDT
Stage

4:30pm EDT

Spatial Matrix Synthesis
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
Spatial Matrix synthesis is presented in this paper. This modulation synthesis technique creates acoustic velocity fields from acoustic pressure signals by using spatial transformation matrices, thus generating complete sound fields for spatial audio. The analysis presented here focuses on orthogonal rotation matrices in both two and three dimensions and compares the results in each scenario with other sound modulation synthesis methods, including amplitude and frequency modulation. As an alternative method for spatial sound synthesis that exclusively modifies the acoustic velocity vector through effects comparable to those created by both amplitude and frequency modulations, Spatial Matrix synthesis is argued to generate inherently spatial sounds, giving this method the potential to become a new musical instrument for spatial music.
Moderators Speakers
avatar for Timothy Schmele

Timothy Schmele

Researcher, Eurecat
Researcher in audio, audio technology, immersive audio, sonification and composition practices. Composer of electroacoustic music.
Authors
avatar for Timothy Schmele

Timothy Schmele

Researcher, Eurecat
Researcher in audio, audio technology, immersive audio, sonification and composition practices. Composer of electroacoustic music.
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
1E03

4:30pm EDT

IPMX Site Open “By Invitation Only” to Developers, Integrators, and End Users in New York City
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
“The Gateway” is an initiative by the Swedish-American Chamber of Commerce in the heart of Manhattan, New York on the 29th floor of the 900 Third Avenue tower. The facility offers event spaces and conference rooms for members, but it is also the locus of a constantly evolving panoply of latest technologies from the Pro AV and IT world assembled and managed by HarperVision Associates.In addition to cutting edge interoperability demos and first looks at brand new products, the Gateway NYC and HarperVision also host key AV/IT events such as Collaboration Week by the IMCCA (Interactive Multimedia & Collaborative Communications Alliance).Greg Harper, President of HarperVision Associates and CIO of the Gateway, has also recently led an initiative involving IPMX, a suite of open specifications and standards that build upon SMPTE ST 2110, NMOS, and AES67. This includes a live interoperability demonstration consisting of several brands including: converters, presentation switchers, routing software, and more.This growing interop site is a gathering place to witness working products, conduct events and presentations, and to showcase IPMX in an enterprise AV/IT setting. Businesses, government users, equipment vendors and developers, integrators, and educators alike are encouraged to communicate with Greg about site visits, events, and other initiatives.
Speakers
avatar for Greg Harper

Greg Harper

President, HarperVision
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

4:30pm EDT

Ask Me Anything! (About Starting Your Career)
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
Aimed at students, novices, and anyone thinking of transitioning into working in the audio industry as either a professional or hobbyist/amateur. What best prepared our panel of professionals for entry into their career? We will discuss skills, knowledge, work habits, obstacles, working conditions, goals and realities, and whatever else YOU want us to discuss. Join us for a relaxed and informal discussion featuring professionals from different sectors of the industry (including music production, live sound, broadcasting, and research/development, and maybe more!) – and bring your questions – a good portion of the discussion will be guided by audience interests.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E16

4:30pm EDT

7 Non Traditional Production Methods w/ QuestionATL
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
Some engineers and producers take advantage of a few of the customizations and optimizations of their machines (laptops, desktops, mobile recorders, tablets, etc.) However, unlocking the deeper settings of your production machine can lead to greater efficiency and productivity with cutting-edge applications and technologies. QuestionATL will take attendees through a variety of production optimizations–including mobile devices when working remotely with artists who need to provide high-resolution content without having access to a studio or professional-grade gear.
Speakers
avatar for Question Atl

Question Atl

QuestionATL Music
QuestionATL is a blind Rap Artist & Producer. He is self-taught on several instruments, began freestyling at 5 and making beats at 12. He has won over 20 beat battles and produced for several artists. QuestionATL released his debut project, The Dream, on all streaming platforms as... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E08

4:30pm EDT

Crafting Modern Hip Hop/Rap: A Hands-On Guide
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
This 45-minute workshop is designed for music production students eager to dive into the world of modern hip hop and rap. With hard-hitting beats, intricate rhythms, and innovative production techniques, hip hop/rap continues to dominate the music scene. Our session will cover essential techniques and creative tips to help you produce tracks that capture the essence of contemporary hip hop and rap.

Workshop Overview

Title: Crafting Modern Hip Hop/Rap: A Hands-On Guide

Duration: 45 minutes

Target Audience: Music production students and aspiring producers

Prerequisites: Basic knowledge of DAWs (Digital Audio Workstations) and audio production

Objectives

By the end of this workshop, participants will:

Understand the key elements of modern hip hop and rap music.
Learn how to create hard-hitting beats and basslines.
Explore techniques for crafting catchy hooks and verses.
Gain insights into mixing techniques to achieve a polished hip hop/rap sound.

Workshop Content

Introduction to Modern Hip Hop/Rap (5 minutes)
Brief history and evolution of the genre
Key characteristics: Beats, basslines, and lyrical flow
Influential artists and iconic tracks

Creating Hard-Hitting Beats (10 minutes)
Overview of drum machines and sample packs
Layering and sequencing drum sounds
Practical demo: Designing a punchy beat

Crafting Catchy Hooks and Verses (10 minutes)
Importance of hooks in hip hop and rap
Techniques for writing memorable verses and choruses
Practical demo: Building a hook and verse from scratch

Basslines and 808s (10 minutes)
Essential bass patterns and 808 techniques
Using sub-bass and distortion for impact
Practical demo: Programming a powerful bassline

Mixing for Hip Hop/Rap (5 minutes)
Key principles of mixing for clarity and impact
Using effects to enhance the vibe: Reverb, delay, and saturation
Practical demo: Quick mixing tips for a hip hop/rap track

Q&A and Wrap-Up (5 minutes)
Recap of key points
Q&A session for participants
Additional resources and next steps
Speakers
avatar for Tyron Williams

Tyron Williams

Music Educator, Certified Banga Productions LLC
Greeting everyone. My name is Tyron Williams, better known as TWiZ, an accomplished GRAMMY nominated Music Educator, record producer, audio engineer, and content creator hailing from Savannah, Georgia. Currently, I live in the NYC area extending my expertise to aspiring young artists... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E07

4:30pm EDT

George Massenburg: Enveloping Masterclass
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
George plays high resolution stereo, 5.1 and 3D recordings from his fabulous back catalogue, commenting on production techniques, dynamics processing and the most appropriate listening format considering each track.

This masterclass series, featuring remarkable recording artists, is a chance to hear 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for George Massenburg

George Massenburg

Associate Professor of Sound Recording, Massenburg Design Works
George Y. Massenburg is a Grammy award-winning recording engineer and inventor. Working principally in Baltimore, Los Angeles, Nashville, and Macon, Georgia, Massenburg is widely known for submitting a paper to the Audio Engineering Society in 1972 regarding the parametric equali... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
3D06

4:30pm EDT

Vocal Production with Simone Torres
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
Join Simone as she dives into the art of vocal production, sharing her unique process and insights. Discover the psychology behind the craft, learn how she creates safe spaces for her artists, and get your questions answered in a live Q&A session!

_____

Simone Torres is a Diamond, Grammy Nominated, and Latin Grammy honored engineer, vocal producer, and artist from New York. She has worked with artists such as Becky G, Chlöe Bailey, Sofia Carson, Normani, Anitta, Cardi B, Camila Cabello, Monsta X, Jessie J, Billy Porter, The Backstreet Boys, Dua Lipa, and more.

Some notable accomplishments include being honored by the Latin Recording Academy as one of their Leading Ladies of Entertainment (2023), Vocal Producing Sofia Carson's "Slow Dance" with BTS' Jimin, Becky G’s “Coming Your Way,” and “La Nena,” Normani’s “Motivation” and Engineering Cardi B’s 11x Platinum “I Like It” and 4x Platinum “Be Careful.”

IG: @iamsimonetorres
Speakers
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E09

4:45pm EDT

Composing and Producing Music for Immersive Audio
Wednesday October 9, 2024 4:45pm - 5:30pm EDT
Justin Gray is a leader in Immersive Audio music production and has presented at numerous AES events on the topic. As an early adopter of object-based technology for music production, Justin has mixed and mastered over 2000 songs in immersive formats (Dolby Atmos and Sony 360).

In 2020, Justin began composing new music for a project specifically conceived for immersive production. This project was recorded over 40 days in a world-class studio in Toronto, Canada, and included over 40 musicians from across the globe. The music draws influence from a diverse range of musical traditions and was composed, orchestrated, recorded, edited, mixed, and mastered with immersive intent at each stage.

Justin has drawn upon his broad musical experience as a composer, arranger, performer, recording engineer, producer, and engineer to bring this musical project to life. He is inspired to share the music, philosophy, and immersive production process with the AES community.

As Immersive Audio continues to expand its scope in the global music community, Justin believes that the next step is for artists to integrate immersive design into their creative practice. This can begin at the initial composition stage, and carry through each stage of a musical production.

In this workshop, Justin will share examples from the project in Dolby Atmos. He will also explore composing and orchestrating music with immersive intent, a rarely discussed topic in this field. The presentation will also include an in-depth look into immersive production techniques, including contemporary immersive recording techniques and immersive mixing philosophy/aesthetics.

Each performance in the project was filmed at a cinematic level, further supporting the learning experience.

Justin also rigorously documented the technical aspects of the project and will have additional visual content to support the presentation.

This AES event will be one of the first public previews of the music. The full album will be released in various cities worldwide in Dolby Atmos movie theatres in early 2025.



Speakers
avatar for Justin Gray

Justin Gray

Chief Engineer, Synthesis Sound
Justin Gray is an award-winning producer, mixing engineer and mastering engineer based in Toronto. Justin is a leader in the field of immersive audio mixing & mastering, where he works with artists, producers and labels worldwide to bring their music to life in the Dolby Atmos and... Read More →
Wednesday October 9, 2024 4:45pm - 5:30pm EDT
1E06

6:00pm EDT

Richard Heyser Memorial Lecture: Doohickii Digitali Redux
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
Anthony Agnello and Richard Factor have over 100 combined years as members of the AES. They have been fortunate to witness and, in some cases, perpetrate the evolution of audio engineering from electro-mechanical to digital. In May of 1979, they published an article in db Magazine, “Doohickii Digitali”, that concluded:

"… we have summarized some of the possibilities of digital signal processing (and completely ignored digital recording, a probably more important subject).  While we don’t have the “all digital” studio yet, it can be reasonably foreseen that before too long, all the components will be available to make one, so that the only analog elements remaining will be the microphone and speaker.  Since these are the primary limitations of the recording process at present, we anxiously await the development of a standard digital I/O bus for the human element."


In this lecture, they will pick up not only where they left off, but also before they began.
Speakers
avatar for Anthony Agnello

Anthony Agnello

Managing Director, Eventide
Tony Agnello was born in Brooklyn, graduated from Brooklyn Technical High School in 1966, received the BSEE from City College of NY in 1971, the MSEE from the City University of NY in 1974 followed by post graduate studies in Digital Signal Processing at Brooklyn’s Polytechnical... Read More →
avatar for Richard Factor

Richard Factor

Chairman / Prime Fossil, Eventide
Richard Factor was born in 1945 and missed being a baby boomer by only weeks.  He lived and was schooled in Manhattan and the Bronx, NYC, until he moved to New Jersey at age 40 and then to Sedona, Arizona in his 60s.  He obtained a degree in Economics but preferred to pursue broadcasting... Read More →
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
1E08

6:00pm EDT

Tech Tour: Lincoln Center - David Rubenstein Atrium & David Geffen Hall
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
All Tech Tours are full. Know Before You Go emails were sent to accepted Tech Tour registrants with information about the tour.  

Duration: 1.5 hours
Capacity: 20 (Once capacity is reached, registrants will be added to a wait list

Tour Description: Join Programming Senior Manager, Godfrey Palaia, and Technical Director, Rudy R. Rodriguez, for an exclusive tour of two iconic spaces at Lincoln Center. Godfrey will guide participants through the vibrant David Rubinstein Atrium, showcasing its unique programming and design. Rudy will then lead the group to David Geffen Hall, offering insights into the cutting-edge technology and production infrastructure that power its world-class performances.
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
Offsite

7:00pm EDT

DEI Mixer at Tailor Public House
Wednesday October 9, 2024 7:00pm - 9:00pm EDT
Experience the vibrant convergence of diversity, equity, and inclusion within the audio engineering community at the AES Show 2024 DEI Mixer.

Join us for an engaging evening of networking, insightful conversations, and a celebration of our shared commitment to creating a more inclusive soundscape. Connect with fellow professionals, share your experiences, and explore the diverse perspectives that shape our industry. Together, we are harmonizing the future of audio engineering.

We are excited to have We Are Moving the Needle as a co-organizer for this event, reinforcing our focus on gender equity in the music industry. 

Additionally, we are honored to have long-time DEI advocate Rebekah Wilson, CEO of Source Elements, and AES President Leslie Gaston-Bird delivering a powerful opening to set the tone for the evening.


Light refreshments and beverages will be provided. This event will be held offsite at
Tailor Public House. Sponsored by: Source Elements 
Wednesday October 9, 2024 7:00pm - 9:00pm EDT
Offsite
 
Thursday, October 10
 

8:00am EDT

Attendee Registration
Thursday October 10, 2024 8:00am - 5:30pm EDT
Thursday October 10, 2024 8:00am - 5:30pm EDT
Crystal Palace South

9:00am EDT

Overcoming Obsolescence: Repair and maintenance of legacy playback machines
Thursday October 10, 2024 9:00am - 9:45am EDT
Since the 1980s, several digital recording systems were produced that used open reel tapes (DASH standard and ProDigi), cassette tapes, DAT, Video8, VHS (ADAT), U-Matic, etc and the more recent MOD (magnetic optical disk).

All these formats became obsolete with the introductions of Hard Disk and software-based recording systems.

Today, many of these formats can still be found in our archives and need to be preserved, but the real issue is having the machines to read them. This is why the second-hand market is relied upon.

But what happens when these machines are broken? Are there still resources available to repair them? This is the challenge that the world of archiving, and beyond, must face.

Gabriele Franzoso and Bob Shuster will discuss creative solutions for repairing and maintaining these legacy playback machines
Speakers
avatar for Gabriele Franzoso

Gabriele Franzoso

Audio preservation engineer and repairer and maintenance audio machinery, Swiss National Sound Archive
I started as repair technician for some of the most important audio brands. Later, I worked as assistant audio engineer and maintenance technician in a renowned recording studio. During that time, I also worked occasionally as live sound engineer. For a short time, I worked as electronic... Read More →
Thursday October 10, 2024 9:00am - 9:45am EDT
1E07

9:00am EDT

Immersive is out, interactive and flexible is in!
Thursday October 10, 2024 9:00am - 9:45am EDT
The workshop will offer an insight into what develops beside Dolby Atmos and immersive live applications. A brand new LED Dome project in LA will be introduced and how audio will be managed not only on a 40-channel custom configuration but also in other areas of the venue like different bars, launches, and broadcasts. The challenge is dealing with various spaces and formats, interactively, and how to create content for such a location.
Speakers
avatar for Tom Ammermann

Tom Ammermann

New Audio Technology
Grammy-nominated music producer, Tom Ammermann, began his journey as a musician and music producer in the 1980s.At the turn of the 21st Century, Tom produced unique surround audio productions for music and film projects as well as pioneering the very first surround mixes for headphones... Read More →
Thursday October 10, 2024 9:00am - 9:45am EDT
1E06

9:00am EDT

Working Class Audio Live at AES
Thursday October 10, 2024 9:00am - 10:00am EDT
Working Class Audio is coming to AES! This exciting and beloved show will be recording a live podcast at the show. Join host Matt Boudreau with his special guests as they discuss navigating the world of recording with a working class perspective.
Speakers
MB

Matt Boudreau

An audio engineer at Matt Boudreau MIXING|MASTERING + Producer/Host/Author Working Class Audio Podcast.
Thursday October 10, 2024 9:00am - 10:00am EDT
1E09

9:00am EDT

Sustainability in Music Technology Manufacturing
Thursday October 10, 2024 9:00am - 10:00am EDT
While every part of the music industry is examining the issue of sustainability, this panel will explore specifically the ways that manufacturers are approaching this. Many are considering every aspect of the design and manufacturing process from the choice of materials, manufacturing techniques and packaging to eco-friendly distribution and shipping methods in an effort to ensure a sustainable future.
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
avatar for Jesse Terry

Jesse Terry

Head of Hardware, Ableton Inc
Jesse Terry is the Head of Hardware at Ableton, and leads the team designing and manufacturing Ableton Push. He joined Ableton in 2005 working in artist and partner relations, which lead him to help design the first dedicated Ableton controller, the APC40 (made in collaboration with... Read More →
avatar for Will Eggleston

Will Eggleston

Marketing Director, Genelec
Will Eggleston likes to refer to himself as “Chief Cook and Bottle Washer” at Genelec Inc in Natick, Massachusetts.  Having spent nearly 50 years in the audio business as a recording engineer for broadcast and music, product development specialist and product manager, Will has... Read More →
Thursday October 10, 2024 9:00am - 10:00am EDT
1E15

9:00am EDT

Towards AI-augmented Live Music Performances
Thursday October 10, 2024 9:00am - 10:00am EDT
There is a renewed need to study the development of AI systems that the artistic community can embrace as empowering tools rather than replacements for human creativity. Previous research suggests that novel AI technologies can greatly help artists expand the range of their musical expressivity and display new kinds of virtuosity during live music performances[1,2]. However, the advent of powerful AI systems, in particular generative models, has also attracted criticism and sparked concern among artists who fear for their artistic integrity and financial viability[3,4]. This is further exacerbated by a widening gap in technological innovation between private companies and research-focused academic institutions.

In this context, we need to pay specific attention to topics such as artistic consent, data collection[5], as well as audience approval. Furthermore, we deeply believe in the importance of integrating lighting and visual arts to effectively convey the role and impact of AI-generated content in human-AI performances to broader audiences.

This workshop will bring together researchers and professionals from music, lighting, visual arts, and artificial intelligence to explore the latest advancements in AI technologies and their transformative potential for live music performances. In particular, discussions will touch on the controllability requirements of AI-augmented instruments, the associated visualization methods, and the sociocultural impact of these technologies[6]. We will focus on the limitations of such technologies, as well as ethical considerations. As an example, we will also discuss the outcomes of the innovative Human-AI co-created concert that we produced with Jordan Rudess on September, 21st, 2024.

References:
[1] Blanchard, Lancelot, Naseck, Perry, Egozy, Eran, and Paradiso, Joe. “Developing Symbiotic Virtuosity: AI-augmented Musical Instruments and Their Use in Live Music Performances.” MIT Press, 2024.
[2] Martelloni, Andrea, McPherson, Andrew P, and Barthet, Mathieu. “Real-time Percussive Technique Recognition and Embedding Learning for the Acoustic Guitar.” arXiv, 2023.
[3] Morreale, Fabio. “Where does the buck stop? Ethical and political issues with AI in music creation.” Transactions of the International Society for Music Information Retrieval, 2021.
[4] Rys, Dan. “Billie Eilish, Pearl Jam, Nicki Minaj Among 200 Artists Calling for Responsible AI Music Practices.” Billboard, April 2, 2024.
[5] Morreale, Fabio, Sharma, Megha, and Wei, I-Chieh. “Data Collection in Music Generation Training Sets: A Critical Analysis.” International Society for Music Information Retrieval, 2023.
[6] Born, Georgina, Morris, Jeremy, Diaz, Fernando, and Anderson, Ashton. “Artificial intelligence, music recommendation, and the curation of culture". White paper: University of Toronto, Schwartz Reisman Institute for Technology and Society, CIFAR AI & Society program, 2021.
Speakers
avatar for Lancelot Blanchard

Lancelot Blanchard

Research Assistant, MIT Media Lab
Musician, Engineer, and AI Researcher. Working at the intersection of Generative AI and musical instruments for live music performances.
avatar for Perry Naseck

Perry Naseck

Research Assistant, MIT Media Lab
Artist and engineer working in interactive, kinetic, light- and time-based media. Specialization in interaction, orchestration, and animation of systems of sensors and actuators.
avatar for Jordan Rudess

Jordan Rudess

Keyboardist, Dream Theater
Voted “Best Keyboardist of All Time” by Music Radar Magazine, Jordan Rudess is best known as the keyboardist/multi-instrumentalist extraordinaire for platinum-selling Grammy Award–winning rock band, Dream Theater. A classical prodigy who began his studies at the Juilliard School... Read More →
PS

Pedro Sarmento

PhD candidate, Queen Mary University of London
avatar for Eran Egozy

Eran Egozy

Professor of the Practice, Music Technology, Massachusetts Institute of Technology
Thursday October 10, 2024 9:00am - 10:00am EDT
1E16

9:00am EDT

DEI work in Audio Engineering Higher Education
Thursday October 10, 2024 9:00am - 10:30am EDT
This panel introduces DEI pedagogical examples in teaching audio engineering and music technology from the University of Colorado, the University of Lethbridge, Georgia Tech, and the University of Michigan, which incorporate topics that address diversity, equity, inclusion (DEI), and accessibility in the broader field of audio engineering and music technology. These topics include 1) how Open Educational Resources (OER) pedagogy can help bridge the Music Technology Educational Gap and increase accessibility; 2) how to create a safe learning space and build community in the classroom where ideas can be shared and valued; 3) How the Project Studio Music Technology course contributes female-identifying, trans, and non-binary musicians and engineers to electronic music and how this process-based learning method impacts students from diverse backgrounds. 4) how the new course called "Diversity in Music Technology" brings students in new DEI experiences and how the course was structured to include community-building. Through this panel discussion, we strive to improve accessibility, welcoming diverse perspectives, and radiating inclusiveness to all races, genders, and gender identities. Meanwhile, we aim to provide some insights for educators to improve students' mental health and well-being during the post-COVID era.
Speakers
avatar for Jiayue Cecilia Wu

Jiayue Cecilia Wu

Assistant Professor, Graduate Program Director (MSRA), University of Colorado Denver
Originally from Beijing, Dr. Jiayue Cecilia Wu (AKA: 武小慈) is a scholar, composer, audio engineer, and multimedia technologist. Her work focuses on how technology can augment the healing power of music. She earned her Bachelor of Science degree in Design and Engineering in 2000... Read More →
avatar for Mary Mazurek

Mary Mazurek

Audio Educator/ Recording Engineer, University of Lethbridge
Audio Educator/ Recording Engineer
avatar for Alexandria Smith

Alexandria Smith

Assistant Professor, Georgia Institute of Technology
Praised by The New York Times for her “appealingly melancholic sound” and “entertaining array of distortion effects,” Alexandria Smith is a multimedia artist, audio engineer, scholar, trumpeter, and educator who enjoys working at the intersection of all these disciplines... Read More →
avatar for Zeynep Özcan

Zeynep Özcan

Assistant Professor, University of Michigan
Thursday October 10, 2024 9:00am - 10:30am EDT
1E08

9:00am EDT

dLive Certification
Thursday October 10, 2024 9:00am - 12:00pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.
 
The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Thursday October 10, 2024 9:00am - 12:00pm EDT
1E05

10:00am EDT

Delay detection in hearing with moving audio objects at various azimuths and bandwidths
Thursday October 10, 2024 10:00am - 10:20am EDT
In order to design efficient binaural rendering systems for 3D audio, it’s important to understand how delays in updating the relative directions of sound sources from a listener to compensate for listener’s head movements affect the sense of realism. However, sufficient study has not yet been done with this problem. We therefore investigated the delay detection capability (threshold) of hearing during localization of audio objects. We used moving sound sources emitted from loudspeakers emulating smooth update of head related transfer functions (HRTFs) and delayed update of HRTFs. We investigated the delay detection threshold with different bandwidth, direction, and speed of sound source signals. As a result, the delay detection thresholds in this experiment were found to be approximately 100ms to 500ms, and it was observed that the detection thresholds vary depending on the bandwidth and direction of the sound source. On the other hand, no significant variation in detection thresholds was observed based on the speed of sound source movement.
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Speakers
avatar for Masayuki Nishiguchi

Masayuki Nishiguchi

Professor, Akita Prefectural University
Masayuki Nishiguchi received his B.E., M.S., and Ph.D. degrees from Tokyo Institute of Technology, University of California Santa Barbara, and Tokyo Institute of Technology, in 1981, 1989, and 2006 respectively.  He was with Sony corporation from 1981 to 2015, where he was involved... Read More →
Authors
avatar for Masayuki Nishiguchi

Masayuki Nishiguchi

Professor, Akita Prefectural University
Masayuki Nishiguchi received his B.E., M.S., and Ph.D. degrees from Tokyo Institute of Technology, University of California Santa Barbara, and Tokyo Institute of Technology, in 1981, 1989, and 2006 respectively.  He was with Sony corporation from 1981 to 2015, where he was involved... Read More →
Thursday October 10, 2024 10:00am - 10:20am EDT
1E03

10:00am EDT

Updates from the AES’ Technical Committee for Broadcast and Online Delivery
Thursday October 10, 2024 10:00am - 11:00am EDT
We will discuss Loudness Management, Dialog Intelligibility, and the new Loudness Education Website. This committee meets weekly and has produced a number of significant Technical Papers including TD-1008. Also recently, they wrote the standards AES71 and AES77.
Speakers
avatar for David Bialik

David Bialik

TOA Electronics
Audio has been an important part of David's professional work, beginning with U.S. major market radio engineering in 1983, becoming chief engineer at United Broadcastings’ WKDM in 1991. He worked at NPR affiliate WAMU-FM and the NAB Science and Technology Department, publishing... Read More →
RB

Robert Bleidt

Streamcrest Associates
avatar for Jim Starzynski

Jim Starzynski

Director and Principal Audio Engineer, NBCUniversal
Jim Starzynski is Director and Principal Audio Engineer for NBCUniversal, overseeing audio technologies and practices for NBCUniversal television properties. Jim is responsible for establishing NBC’s audio strategy for digital television and works on advanced technologies.Jim is... Read More →
avatar for John Kean

John Kean

Senior Engineer, NPR (retired)
John applies 40 years of experience in television and radio technology.  As Senior Technologist at National Public Radio until 2015, he directed network projects and technical studies at NPR Labs.  He started in radio as an FM chief engineer in San Diego, leaving to join NPR in... Read More →
CG

Cornelius Gould

Angry Audio
Thursday October 10, 2024 10:00am - 11:00am EDT
1E07

10:00am EDT

The Science and Practice of Recording Rockets
Thursday October 10, 2024 10:00am - 11:00am EDT
This dynamic presentation will showcase methodology behind the capture and studio presentation of the most significant rocket launches in the world into immersive audio. Audio educator, producer, & AES member Gabe Herman and former NASA extraterrestrial audio engineer Jason Achilles will share their recent experience at the launch pad of Starship IFT-4, highlighting the unique challenges involved in planning, execution, and post-production. Students are encouraged to learn how their passion for audio and acoustic science can lead to unexpected & thrilling opportunities. Attendees will witness a roof-shaking immersive audio demonstration and join in remote Q&A with fellow team members Andrew Keating (Cosmic Perspectives) and Benny Burtt (Skywalker Sound).
Speakers
avatar for Gabe Herman

Gabe Herman

Assistant Professor, University of Hartford
Gabe Herman is a music producer, audio engineer, sound designer and educator.Gabe has engineered and produced numerous music recordings in a variety of genres spanning from classical ensembles to modern rock, contemporary world beat music, jazz, blues, R&B and more on just about every... Read More →
Thursday October 10, 2024 10:00am - 11:00am EDT
1E06

10:00am EDT

Full-Frequency Room Impulse Response Simulation: Combining Ray Tracing and Finite Element Methods
Thursday October 10, 2024 10:00am - 11:00am EDT
Room acoustics simulation is used in the audio industry for designing rooms, evaluating acoustic treatments, determining optimal placement of speakers and microphones, and auralization. Due to the wavelength and size scales, it is difficult to accurately model the room’s response over the entire frequency range. This presentation will focus on using a hybrid approach, where the results from multiple numerical methods are implemented into a single model, thereby improving accuracy and maintaining feasibility. This presentation will include detailed discussion about finite element modeling, ray tracing, appropriate source conditions, and boundary conditions. Results from the model are compared to measurements of Acentech’s Founders Room Listening Lab in Cambridge, MA.
Speakers
Thursday October 10, 2024 10:00am - 11:00am EDT
Stage
  Sponsor Session

10:00am EDT

Acoustics in Live Sound
Thursday October 10, 2024 10:00am - 12:00pm EDT
Acoustics in Live Sound looks at roadhouses and touring engineers in order to determine how acoustics is applied and whether this is a conscious or subconscious choice. This is done through various sources with a strong emphasis of interviews with professionals in the field. A look at an engineers’ workflow, tools, a cost-benefit analysis of the tools, looking at different types of spaces, ways to improve the spaces, and delays and equalization are combined to give a comprehensive look as to what an engineer is doing and what can make their job easier and mix better. For the beginning engineer, this is a comprehensive guide as to what tools and skills are needed as it relates to acoustics. For the more seasoned professional, it is a different way to think about how problems are being approached within the field and how solutions could be more heavily based on the statistics gathered from acoustics.
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Design and Training of an Intelligent Switchless Guitar Distortion Pedal
Thursday October 10, 2024 10:00am - 12:00pm EDT
Guitar effects pedals are designed to provide an alteration to a guitar signal through electronic means and are often controlled by a footswitch that routes the signal either through the effect or directly to the output through a 'clean' channel. Because players often switch the effects on and off during different portions of a song and different playing styles, our goal in this paper is to create a trainable guitar effect pedal that classifies the incoming guitar signal into two or more playing style classes and route the signal to the bypass channel or effect channel depending on the class. A training data set is collected that consists of recorded single notes and power chords. The neural network algorithm is able to distinguish between these two playing styles with 95\% accuracy in the test set. An electronic system is designed with a Raspberry Pi Pico, preamplifiers, multiplexers, and a distortion effect that runs a neural network trained using Edge Impulse software that runs the classification and signal routing in real-time.
Speakers
avatar for David Anderson

David Anderson

Assistant Professor, University of Minnesota Duluth
Authors
avatar for David Anderson

David Anderson

Assistant Professor, University of Minnesota Duluth
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Detection and Perception of Naturalness in Drum Sounds Processed with Dynamic Range Control
Thursday October 10, 2024 10:00am - 12:00pm EDT
This study examines whether trained listeners could identify and judge the “naturalness” of drum samples processed with Dynamic Range Compression (DRC) when paired with unprocessed samples. A two-part experiment was conducted utilizing a paired comparison of a 10-second reference drum loop with no processing and a set of drum loops with varying degrees of DRC. In Part 1, subjects were instructed to identify which sample of the two they believed to have DRC applied. They were then asked to identify which sample they believed sounded “more natural” in Part 2. Out of 18 comparisons for Part 1, only three demonstrated reliable identification of the target variable. Results from Part 2 showed that the subjects perceived DRC processed samples as natural and unprocessed. However, while inconclusive, results here may suggest that listeners perceived the processed drum sound to be equally as natural as the original.
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Headphones vs. Loudspeakers: Finding the More Effective Monitoring System
Thursday October 10, 2024 10:00am - 12:00pm EDT
There has been some debate about whether headphones or loudspeakers are the more effective monitoring system. However, there has been little discussion of the effectiveness of the two mediums in the context of mixing and music production tasks. The purpose of this study was to examine how the monitoring systems influenced users’ performance in production tasks using both mediums. An experiment was designed, in which the subject was asked to match the boost level of a given frequency band in a reference recording until the boost sounded identical to the reference. A group of six audio engineering graduate students completed the experiment using headphones and loudspeakers. Subjects’ adjustment levels across eight frequency bands were collected. Results suggest that both monitoring systems were equally effective in the context of adjusting equalizer settings. Future research is called for which will include more subjects with diverse levels of expertise in various production-oriented tasks to better understand the potential effect of each playback system.
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Intentional Audio Engineering Flaws: The Process of Recording Practice Exercises for Mixing Students
Thursday October 10, 2024 10:00am - 12:00pm EDT
There are plenty of resources, both scholarly and otherwise, concerning how to mix audio. Students can find information on the process of mixing and descriptions of the tools used by mixing engineers like level, panning, equalization, compression, reverb, and delay through an online search or more in-depth textbooks by authors like Izhaki [1], Senior [2], Owsinski [3], Case [4, 5], Stavrou [6], Moylan [7,] and Gibson [8]. However, any professional mixing engineer knows that simply reading and understanding such materials is not enough to develop the skills necessary to become an excellent mixer. Much like developing proficiency and expertise on an instrument, understanding the theory and technical knowledge of mixing is not enough to become a skilled mixer. In order to develop exceptional mixing skills, practice is essential. This begs the question; how should an aspiring mixer practice the art and craft of mixing?
The discussion of this topic is absent from a large portion of the literature on mixing except for Merchant’s 2011 and 2013 Audio Engineering Society convention papers in which Colvin’s concept of deliberate practice is applied to teaching students how to mix [9, 10, 11]. In order for students to carry out deliberate practice in the mixing classroom, quality multitrack recordings are necessary that help students focus on using specific mixing tools and techniques. This paper looks at the process of recording a series of mixing exercises with inherent audio engineering challenges for students to overcome.
Speakers Authors
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Investigation of the impact of pink and white noise in an auditory threshold of detection experiment
Thursday October 10, 2024 10:00am - 12:00pm EDT
Many listening experiments employ so-called "pink" and "white" noise, including those to find the threshold of detection (TOD). However, little research exists that compares the impacts of these two noises on a task. This study is an investigation of the TODs at whole and third octave bands using pink and white noise. A modified up-down method was utilized to determine the TOD in white and pink noise at eight frequency levels using a dB boosted bell filter convolved with the original signal. Six graduate students in audio engineering participated. Subjects were presented with an unfiltered reference signal followed by a filtered or unfiltered signal and then were instructed to respond “same” or “different”. Correct answers decreased the filter boost while incorrect answers resulted in reversals that would increase the filter boost until the subject answered correctly. A trial would conclude when a total of ten reversals occurred. The filter boost levels of the last five reversals were collected to obtain an average threshold value. Results show no significant differences between white and pink noise TOD levels at all frequency intervals. Additionally, the JND between the original and the boosted signal was observed to be ±3 dB, consistent with existing literature. Therefore, it can be suggested that white or pink noise can be equivalently employed in listening tests such as TOD measurements without impact on perceptual performance.
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

New Music Instrument Design & Creative Coding: Assessment of a Sensor-Based Audio and Visual System
Thursday October 10, 2024 10:00am - 12:00pm EDT
Technological developments within the modern age provide near limitless possibilities when it comes to the design of new musical instruments and systems. The aim of the current study is to investigate the design, development, and testing of a unique audio-visual musical instrument. Five participants tested an initial prototype and completed semi-structured interviews providing user feedback on the product. A qualitative analysis of the interviews indicate findings related to two primary areas: enjoyability and functionality. Aspects of the prototype commonly expressed by participants as being enjoyable related to the product’s novelty, simplicity, use of distance sensors, and available interactivity with hardware components. Participant feedback commonly expressed related to improving the prototype’s functionality included adding delay and volume modules, making the product surround sound compatible, and increasing lower register control for certain sensors. These results informed design decisions in a secondary prototype building stage. The current study’s results suggest that designers and developers of new musical instruments should value novel rather than traditional instrument components, hardware rather than software-based interfaces, and simple rather than complex design controls. Ultimately, this paper provides guiding principles, broad themes, and important considerations in the development of enjoyable, new musical instruments and systems.
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Subjective Evaluation of Emotions in Music Generated by Artificial Intelligence
Thursday October 10, 2024 10:00am - 12:00pm EDT
Artificial Intelligence (AI) models offer consumers a resource that will generate music in a variety of genres and with a range of emotions with only a text prompt. However, emotion is a complex human phenomenon which becomes even more complex when attempted to convey through music. There is limited research assessing AI’s capability to generate music with emotion. Utilizing specified target emotions this study examined the validity of those emotions as expressed in AI-generated musical samples. Seven audio engineering graduate students listened to 144 AI-generated musical examples with sixteen emotions in three genres and reported their impression of the most appropriate emotion for each stimulus. Using Cohen’s kappa minimal agreement was found between subjects and AI. Results suggest that generating music with a specific emotion is still challenging for AI. Additionally, the AI model here appeared to operate with a predetermined group of musical samples linked to similar emotions. Discussion includes how this rapidly changing technology might be better studied in the future.
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Exploring Immersive Recording and Reproduction Methods for Pipe Organs
Thursday October 10, 2024 10:00am - 12:00pm EDT
The pipe organ is often underrepresented in classical music recordings, particularly in immersive audio formats. This study explores an innovative approach to recording and reproducing the organ for immersive formats using a modified Bowles Array. Key considerations for immersive recording and reproduction are examined, including the balance between aestheticism and realism, the role of the LFE channel, and the acoustic characteristics of the recording and reproduction environments, as well as the instrument itself. The findings aim to enhance the immersive audio experience of pipe organ music and provide valuable insights for developing standards in immersive recording and reproduction methods for pipe organ performances.
Speakers
avatar for Garrett Treanor

Garrett Treanor

New York University
Authors
avatar for Garrett Treanor

Garrett Treanor

New York University
avatar for Jeremy Tsuaye

Jeremy Tsuaye

New York University
avatar for Jessica Luo

Jessica Luo

Graduate Student, New York University
avatar for Yi Wu

Yi Wu

PhD Student, New York University
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Pilot Study on Creating Fundamental Methodologies for Capturing Spatial Room Information for XR Listening Environments
Thursday October 10, 2024 10:00am - 12:00pm EDT
Holophonic recording techniques were created and observed to see which techniques would best transcribe the three-dimensional characteristics of the room, allowing a listener 6-degrees-of-freedom within a virtual environment. Two recordings were conducted, one consisting of a solo harp and the second consisting of an African gyil duet. Three microphone systems were prepared to capture the instruments: XYZ array, Hamasaki-Ambeo array, and XYZ with an additional XY side cluster. The recordings were spatialized in Unity using the Steam Audio Plug-in. A preliminary subjective evaluation was conducted with expert subjects ranking each of the systems based on certain attributes. Results showed some favor toward the XYZ array due to the better capture of reverberation, however, this does not exclude the Hamasaki-Ambeo from being a system that would be suitable for holophonic ambience reproduction in virtual environments. These results will be used as a foundation to a long-term research of building a methodology for holophonic recording.
Speakers
avatar for Parichat Songmuang

Parichat Songmuang

Studio Manager/PhD Student, New York University
Parichat Songmuang graduated from New York University with her Master of Music degree in Music Technology at New York University and Advanced Certificate in Tonmeister Studies. As an undergraduate, she studied for her Bachelor of Science in Electronics Media and Film with a concentration... Read More →
Authors
avatar for Parichat Songmuang

Parichat Songmuang

Studio Manager/PhD Student, New York University
Parichat Songmuang graduated from New York University with her Master of Music degree in Music Technology at New York University and Advanced Certificate in Tonmeister Studies. As an undergraduate, she studied for her Bachelor of Science in Electronics Media and Film with a concentration... Read More →
Thursday October 10, 2024 10:00am - 12:00pm EDT
Poster

10:00am EDT

Exhibit Hall
Thursday October 10, 2024 10:00am - 5:00pm EDT
Step into the heart of innovation at the Audio Engineering Society’s Annual Conference Exhibit Hall. This dynamic space brings together leading companies and cutting-edge technologies from across the audio engineering industry. Attendees will have the opportunity to explore the latest advancements in audio equipment, software, and services, engage with industry experts, and discover new solutions to enhance their projects. Whether you’re looking to network with professionals, gain insights from live demonstrations, or simply stay ahead of the curve, the Exhibit Hall is the place to be. Don’t miss this chance to immerse yourself in the future of audio engineering!
Thursday October 10, 2024 10:00am - 5:00pm EDT
Exhibit Hall

10:15am EDT

Da Fingaz Presents: The Art of Sync: Navigating the Balance Between Creativity and Commerce
Thursday October 10, 2024 10:15am - 11:15am EDT
Let's explore the dynamic world of sync licensing, bringing together key players from across the industry, including advertising, film, and television. Our discussion will dive into the creative and business sides of sync, sharing insights on how to secure placements, communicate effectively with supervisors, and maximize opportunities for music in visual media. Whether you're an aspiring composer or an industry veteran, this session will provide actionable advice and unique perspectives from top professionals working in sync today.
Speakers
JC

Joshua Christie

Music Ronin Music Supervision
RS

Rob Schwartz

Co-Founder, Chief Music Officer, ACX1 Studios (at Caesar's Pier)
JR

Joshua Rabinowitz

President, Brooklyn Music Experience
avatar for Marcus Manderson

Marcus Manderson

Music Composer, Producer, Da Fingaz
I'm a versatile composer and producer, creating music in the epic cinematic, hip hop, pop, R&B, and more genres.Highlights include composing for Disney theme parks, composing music for the film "Hemisphere" (streaming on Amazon Prime Video, Google Play, Tubi), and securing sync placements... Read More →
Thursday October 10, 2024 10:15am - 11:15am EDT
1E15

10:15am EDT

Bob Katz & Hyunkook Lee: 3D for Two
Thursday October 10, 2024 10:15am - 11:15am EDT
In this 3D for Two, Bob Katz and Hyunkook Lee take turns when playing high resolution music examples and presenting two production techniques: Bob created recent 3D mixes based on headphone monitoring, in contrast to Hyunkook's microphone recordings.

We listen and discuss how each method comes across in this room, touching also upon topics such as integrated microphone pickup, level-panning, use of LFE; and the current distribution formats available for 3D music.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
avatar for Bob Katz

Bob Katz

Digital Domain, Inc.
Bob Katz, workshop co-chair, has been an audio engineer since 1971 and AES member since 1975. He’s played the Bb Clarinet since age 10. Currently President and mastering engineer at Digital Domain, he is an AES Fellow, engineered 3 Grammy-winning albums and numerous nominees. He... Read More →
Thursday October 10, 2024 10:15am - 11:15am EDT
3D06

10:15am EDT

Putting a 1970s recording studio on stage, Sound Design for Stereophonic.
Thursday October 10, 2024 10:15am - 11:15am EDT
Come join the team from the Tony award winning stage play Stereophonic at AES NY 2024, to learn what it takes to put a 1970s recording studio on stage!
Speakers
avatar for John McKenna

John McKenna

sndwrks LLC
Professionally, John (he/him) practices sound design for Broadway plays and musicals in addition to software engineering. When not busy with work, John designs and builds innovative furniture, creates purpose-designed products using 3D printing, and enjoys playing with his cat, Nori.He... Read More →
RR

Ryan Rumery

Ryan Rumery
Thursday October 10, 2024 10:15am - 11:15am EDT
1E16

10:15am EDT

Dolby Atmos in Higher Education
Thursday October 10, 2024 10:15am - 11:30am EDT
Dolby Atmos has been the delivery format of choice for visual media since 2012. But with the adoption of Dolby Atmos for Music by streaming platforms such as Apple Music, Tidal, and Amazon, along with easily accessible tools for creation, the demand for content has exploded. However, the transition to multichannel and object-based production comes with increased complexity, and familiar stereo recording and mixing techniques do not necessarily work in the immersive space requiring the development of new recording and mixing methodologies, software, and hardware. It is clear that immersive audio has become an essential part of an audio engineer’s tool set. Both experienced and novice engineers must be educated in the technology and techniques to create and deliver immersive audio content.

This panel of educators will discuss their approach to teaching Dolby Atmos, the challenge of creating a curriculum in this rapidly-changing field, the role of research, and how the study of immersive audio formats fit into the study of audio.
Speakers
avatar for Konrad Strauss

Konrad Strauss

Professor, Indiana University Jacobs School of Music
Konrad Strauss is a Professor of Music in the Department of Audio Engineering and Sound Production at Indiana University’s Jacobs School of Music. He served as department chair and director of Recording Arts from 2001 to 2022. Prior to joining the faculty of IU, he worked as an... Read More →
avatar for Scott Metcalfe

Scott Metcalfe

Professor, Dept Chair, Peabody Conservatory, Johns Hopkins University
Scott Metcalfe is an audio engineer and music technologist based in Baltimore, MD where he is Director of Recording Arts and Sciences at the Peabody Institute, Program Coordinator for the Sound Concentration of Hopkins’ Advanced Academic Programs MA degree in Film and Media, and... Read More →
avatar for Justin Gray

Justin Gray

Chief Engineer, Synthesis Sound
Justin Gray is an award-winning producer, mixing engineer and mastering engineer based in Toronto. Justin is a leader in the field of immersive audio mixing & mastering, where he works with artists, producers and labels worldwide to bring their music to life in the Dolby Atmos and... Read More →
Thursday October 10, 2024 10:15am - 11:30am EDT
1E09

10:20am EDT

Expanding and Analyzing ODAQ - The Open Dataset of Audio Quality
Thursday October 10, 2024 10:20am - 10:40am EDT
Datasets of processed audio signals along with subjective quality scores are instrumental for research into perception-based audio processing algorithms and objective audio quality metrics. However, openly available datasets are scarce due to listening test effort and copyright concerns limiting the distribution of audio material in existing datasets. To address this problem, Open Dataset of Audio Quality (ODAQ) was introduced, containing audio material along with extensive subjective test results with permissive licenses. The dataset comprises processed audio material with six different classes of signal impairments in multiple levels of processing strength covering a wide range of quality levels. The subjective quality evaluation has recently been extended and now comprises results from three international laboratories providing a total 42 listeners and 10080 subjective scores overall. Furthermore, ODAQ was recently expanded by a performance evaluation of common objective metrics for perceptual quality evaluation in their ability to predict subjective scores. The wide variety of audio material and test subjects in the test provides insight into influences and biases in subjective evaluation, which we investigated by statistical analysis, finding listener-based, training-based and lab-based influences. We also demonstrate the methodology for contributing to ODAQ, and make a request for additional contributors. In conclusion, the diversity of the processing methods and quality levels, along with a large pool of international listeners and permissive licenses make ODAQ particularly suited for further research into subjective and objective audio quality.
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Authors
avatar for Christoph Thompson

Christoph Thompson

Director of Music Media Production, Ball State University
Christoph Thompson is vice-chair of the AES audio education committee. He is the chair of the AES Student Design Competition and the Matlab Plugin Design Competition. He is the director of the music media production program at Ball State University. His research topics include audio... Read More →
avatar for Pablo Delgado

Pablo Delgado

Fraunhofer IIS
Pablo Delgado is part of the scientific staff of the Advanced Audio Research Group at the Fraunhofer Institute for Integrated Circuits (IIS) in Erlangen, Germany. He specializes in psychoacoustics applied to audio and speech coding, as well as machine learning applications in audio... Read More →
Thursday October 10, 2024 10:20am - 10:40am EDT
1E03

10:40am EDT

Perceptual Evaluation of Hybrid Immersive Audio Systems in Orchestral Settings
Thursday October 10, 2024 10:40am - 11:00am EDT
This study investigates the perceptual strengths and weaknesses of various immersive audio capture techniques within an orchestral setting, employing channel-based, object-based, and scene-based methodologies concurrently. Conducted at McGill University’s Pollack Hall in Montreal, Canada, the research featured orchestral works by Boulanger, Prokofiev, and Schubert, performed by the McGill Symphony Orchestra in April 2024.
The innovative aspect of this study lies in the simultaneous use of multiple recording techniques, employing traditional microphone setups such as a Decca tree with outriggers, alongside an experimental pyramidal immersive capture system and a 6th order Ambisonic em64 “Eigenmike.” These diverse methodologies were selected to capture the performance with high fidelity and spatial accuracy, detailing both the performance's nuances and the sonic characteristics imparted by the room. The capture of this interplay is the focus of this study.
The project aimed to document the hall's sound quality in its last orchestral performance before closing for 2 years for renovations, providing the methodology and documentation needed for future comparative recordings of the acoustics before and after. The pyramidal system, designed with exaggerated spacing, improves decorrelation at low frequencies, allowing for the impression of a large room within a smaller listening space. Meanwhile, Ambisonic recordings provided insights into single-point versus spaced multi-viewpoint capture.
Preliminary results from informal subjective listening sessions suggest that combining different systems offers potential advantages over any single method alone, supporting exploration of hybrid solutions as a promising area of study for audio recording, enhancing the realism and spatial immersion of orchestral music recordings.
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Speakers Authors
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Thursday October 10, 2024 10:40am - 11:00am EDT
1E03

10:45am EDT

Take the Blue Pill-Interactive and Adaptive Music In the 21st century
Thursday October 10, 2024 10:45am - 11:30am EDT
Games and music go together like macaroni and cheese. Music is a large part of what makes games fun. So, what are some of the hot button issues that composers folks need to know about game music design and production? Good questions!

This is a must-attend for anyone in the field interested in trying to navigate the waters and steer a path toward more immersive and creative game music. Attendees will gain a more complete understanding of the issues that composers and game designers face when working together in teams. All who attend will gain a better understanding of how great music makes games more fun and increases playability.

This is for anyone involved with game audio -- calling all music composers, game designers, sound designers, animators, audio producers, programmers and technical implementation specialists! It's a must for industry professionals, students, and teachers alike, all who are trying to navigate the waters and steer a path toward deeper creativity and understanding
Speakers
avatar for Steven Horowitz

Steven Horowitz

Executive director of Technology and Applied Composition (TAC) program at SFCM, SFCM
Critically acclaimed as "One of the foremost figures in the field" Composer Steven Horowitz is the executive director of Technology and Applied Composition (TAC) program at SFCM, a position he took over in January, 2024.Horowitz comes to SFCM after 23 years as Audio Director at Nickelodeon... Read More →
Thursday October 10, 2024 10:45am - 11:30am EDT
1E08

11:00am EDT

Perception of the missing fundamental in vibrational complex tones
Thursday October 10, 2024 11:00am - 11:20am EDT
We present a study on the perception of the missing fundamental in vibrational complex tones. When asked to match an audible frequency to the frequency of a vibrational tone with a missing fundamental frequency, participants in two experiments associated the audible frequency with lower frequencies than those present in the vibration, often corresponding to the missing fundamental of the vibrational tone. This association was found regardless of whether the vibration was presented on the back (first experiment) or the feet (second experiment). One possible application of this finding could be the reinforcement of low frequencies via vibrational motors even when such motors have high resonance frequencies
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Speakers Authors
Thursday October 10, 2024 11:00am - 11:20am EDT
1E03

11:00am EDT

Objective Audio Quality Evaluation of UWB Wireless Transmission
Thursday October 10, 2024 11:00am - 11:45am EDT
Presentation of audio transmission data comparison of UWB and straight S/PDIF wire link at data rates up to 24 bits and 96Ks/s showing that UWB is nearly indistinguishable from a wire. If time permits, a comparison to alternative wireless transmission technologies may be presented.
Speakers
avatar for Jeff Anderson

Jeff Anderson

SPARK Microsystems
Thursday October 10, 2024 11:00am - 11:45am EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

11:15am EDT

Student Mix Critiques
Thursday October 10, 2024 11:15am - 12:15pm EDT
These non-competitive sessions are an opportunity for AES student members at all stages in their studies to receive feedback on their mixes from a panel of industry professionals – live and in-person. Join us to hear mixes by other students, and get tips, tricks, and advice to push your skills to the next level! Advance submission procedures for each event will be posted at www.aesstudents.org. Very limited sign up will be available on-site. Maybe one of your mixes can be featured!
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
Thursday October 10, 2024 11:15am - 12:15pm EDT
1E06

11:15am EDT

Produce Like A Pro at AES NY 2024
Thursday October 10, 2024 11:15am - 12:15pm EDT
Produce Like A Pro by Warren Huart is an exciting community and site dedicated to empowering engineers and producers with the tools to make better recordings and mixes. Now, Huart brings that knowledge and his group of special guests to the AES stage to go even more in depth live. Don't miss this exciting panel as you're sure to learn something new as well as be in the company of some of the best engineers in the business.
Speakers
WH

Warren Huart

Produce Like A Pro
Thursday October 10, 2024 11:15am - 12:15pm EDT
Stage

11:15am EDT

Case studies: Large scale sound reinforcement: Festivals featuring Roskilde, Outside Lands, Download, Power Trip and more
Thursday October 10, 2024 11:15am - 12:45pm EDT
Large scale systems face unique challenges of scale, shape, acoustic overflow restrictions, flown weight restrictions and yet still must produce high quality experiences for the audience. Join Bob McCarthy for a discussion of the large-scale outdoor systems used in festivals from the NY Philharmonic in Central Park through major rock festivals.
Speakers
avatar for Bob McCarthy

Bob McCarthy

Meyer Sound
Bob McCarthy’s involvement in the field of sound system optimization began in 1984 with the development of Source Independent Measurement™ (SIM™) with John Meyer at Meyer Sound Laboratories. Since that time he has been an active force in developing the tools and techniques of... Read More →
Thursday October 10, 2024 11:15am - 12:45pm EDT
1E07

11:20am EDT

Perceptual loudness compensation for evaluation of personalized earbud equalization
Thursday October 10, 2024 11:20am - 11:40am EDT
The ear canal geometries of individuals are very different, resulting in significant variations in SPL at the drum reference point (DRP). Knowledge of the personalized transfer function from a near-field microphone (NFM) in the earbud speaker tip to DRP allows personalized equalization (P.EQ). A method to perceptually compensate for loudness perception within the evaluation of different personalized equalization filters for earbuds has been developed. The method includes: measurements at NFM to estimate the personalized transfer function from the NFM point to DRP, calibration of the NFM microphone, the acquisition of the transfer function from the earbuds' speaker terminals to NFM, and the estimation of perceptual loudness in phones at DRP, when applying the different equalization filters. The loudness estimation was computed using the Moore-Glasberg method as implemented in ISO 532-2. The difference to DRP estimation was recursively adjusted using pink noise until the delta was within 0.1 dB. The corresponding gains were applied to the different conditions to be evaluated. A listening test was performed to evaluate three conditions using the described method for loudness compensation.
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Speakers
avatar for Adrian Celestinos

Adrian Celestinos

Samsung Research America
Authors
Thursday October 10, 2024 11:20am - 11:40am EDT
1E03

11:30am EDT

Building Your Career in Audio
Thursday October 10, 2024 11:30am - 12:30pm EDT
Women's Audio Mission presents a panel discussion featuring award-winning audio engineers reflecting on their career journeys and how they built successful careers. The discussion will include everything from essential skills to building your network. The panelists are award-winning recording engineers that have worked with artist such as Beyonce, Patti Smith, Lana Del Ray, Solange, Sheryl Crow and Kacey Musgraves in major Los Angeles, New York and Nashville studios and beyond.

Panelists include:

Gloria Kaba - Gloria Kaba is a Ghanaian-American sound engineer, producer, mixer and writer based in New York. She began her career as an intern at Battery Studios and later as an assistant engineer at MSR Studios where she worked with artists including Madonna, Beyonce, Kanye West, Salaam Remi, Aretha Franklin and Frank Ocean. She’s worked on A Tribe Called Quest’s final album We Got It From Here...Thank You For Your Service and Solange’s breakthrough release, A Seat At The Table, working closely with legendary hip-hop artist and producer Q-tip. Kaba was inducted into YouTube’s Black Voices Class of 2022; a fund championing 21 outstanding Black producers and songwriters.

Leslie Richter - Leslie Richter is a Nashville based producer and recording engineer and Chair of the Audio Engineering Program at Belmont University. After graduating from Middle Tennessee State University, she began her career as a staff engineer at the acclaimed Oceanway Studios in Nashville and worked with a wide array of artists including Sheryl Crow, Kacey Musgraves, Spoon, Sara Bareilles, Wynonna, John Hiatt and Ben Folds.

Vira Byramji - Vira Byramji is an audio engineer specializing in studio recording and mixing. Vira’s entrance into the industry was through the legendary Electric Lady Studios in New York, where she was the assistant manager and a staff engineer. There she worked with major labels and high profile clients such as Patti Smith, U2, Lana Del Rey, Sara Bareilles and HBO to name a few. She is also works in A&R at Epidemic Sound.

Moderated by Terri Winston, Executive Director of Women's Audio Mission. Women's Audio Mission (WAM) is dedicated to closing the chronic gender gap in the music, recording and technology industries. Less than 5% of the people creating and shaping the sounds that make up the soundtrack of our lives are women and gender-expansive people. WAM is “changing the face of sound” by providing training, mentoring and career advancement that inspires the next generation of women and gender-expansive music producers, recording engineers and audio technologists and radically changes the narrative in these industries.
Speakers
avatar for Terri Winston

Terri Winston

Executive Director, Women's Audio Mission
Women's Audio Mission (WAM) is dedicated to closing the chronic gender gap in the music, recording and technology industries. Less than 5% of the people creating and shaping the sounds that make up the soundtrack of our lives are women and gender-expansive people. WAM is “changing... Read More →
avatar for Gloria Kaba

Gloria Kaba

Studio Manager, Power Station at BerkleeNYC
Gloria Kaba is a Ghanaian-American sound engineer, producer, mixer, and writer with over a decade of experience in the studio, often operating under the moniker Redsoul. She’s worked on A Tribe Called Quest’s final album We Got It From Here...Thank You For Your Service and Solange’s... Read More →
Thursday October 10, 2024 11:30am - 12:30pm EDT
1E15

11:30am EDT

Expressive Control in Electronic Instruments
Thursday October 10, 2024 11:30am - 1:00pm EDT
One of the considerations in electronic instrument design has always been expressive control, how to give the player a deeper more intuitive connection with an instrument's sound engine. MIDI Polyphonic Expression (MPE) was adopted by the MIDI Manufacturers Association in 2018 as an enhancement to the original specification. Since then manufacturer support has grown and several new instruments incorporate this type of gestural control. This panel will examine the creative possibilities of MPE from a sound design persecutive, exploring strategies for three dimensional control of sound parameters, their practical ranges and opportunities for reinventing existing instrument categories with continuous pitch control as well as more complex timbral effects.
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
avatar for Jesse Terry

Jesse Terry

Head of Hardware, Ableton Inc
Jesse Terry is the Head of Hardware at Ableton, and leads the team designing and manufacturing Ableton Push. He joined Ableton in 2005 working in artist and partner relations, which lead him to help design the first dedicated Ableton controller, the APC40 (made in collaboration with... Read More →
avatar for Richard Graham

Richard Graham

Principal, Delta Sound Labs
Richard Graham, Ph.D., is a musician, technologist, educator, and entrepreneur. His academic and practical pursuits encompass computer-assisted music composition and instrumental performance. In 2017, he co-founded Delta Sound Labs, an audio technology venture that has developed and... Read More →
avatar for Pat Scandalis

Pat Scandalis

CTO/CEO, moForte Inc
Pat Scandalis is the CTO/CEO moForte Inc, the creator of GeoShred.  He is also the  Chairman MPE Committee in the MIDI Association and a Visiting Scholar, Stanford/CCRMA.  He holds a BSc in Physics from Cal Poly San Luis Obispo
Thursday October 10, 2024 11:30am - 1:00pm EDT
1E16

11:40am EDT

The audibility of true peak distortion (0 dBFS+)
Thursday October 10, 2024 11:40am - 12:00pm EDT
In a recent study, the authors interviewed five professional mastering engineers on the topic of contemporary loudness practices in music. Among the findings, all five mastering engineers targeted their peak levels very close to 0 dBFS and seemed somewhat unconcerned regarding true peak distortion emerging in the transcoding process, not adapting to the current recommendations of not exceeding -1 dB true peak. Furthermore, true peak measurements over the last four decades show that quite a few releases even measure true peaks above 0 dBFS in full quality. The aim of this study is to investigate the audibility of such overshoots by conducting a tailored listening test. The results indicate that even experienced and trained listeners may not be very sensitive to true peak distortion.
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Speakers
avatar for Pål Erik Jensen

Pål Erik Jensen

University College Teacher, Høyskolen Kristiania
TeachingAudio productionMusic studio productionPro ToolsGuitarBass
Authors
avatar for Pål Erik Jensen

Pål Erik Jensen

University College Teacher, Høyskolen Kristiania
TeachingAudio productionMusic studio productionPro ToolsGuitarBass
avatar for Tore Teigland

Tore Teigland

Professor, Kristiania University College
Thursday October 10, 2024 11:40am - 12:00pm EDT
1E03

11:45am EDT

Protect Your Ass(ets): Disaster Planning and Material Recovery for the Audio Engineer
Thursday October 10, 2024 11:45am - 12:30pm EDT
Climate change is real and it’s here to stay. Floods are more severe because of sea level rise and intensifying storms. Fires are more frequent as droughts get worse. Basically, the world is collapsing. Good news! There are some simple steps you can take to protect your audio assets. This panel of audio preservation experts will cover easy archival strategies for mitigating the effects of natural disasters. We will discuss the best practices for safe storage, whether it’s your historic audio collection or newly-created digital files. The panel will also discuss some techniques for recovering your audio assets if they get damaged in a disaster. We hope you never have to use any of this information, but just in case, it’s best to be prepared!

This presentation will also highlight a case study from the “HBCU Radio Preservation Program.”
Speakers
KF

Karl Fleck

Northeast Document Conservation Center
Karl Fleck is an audio preservation engineer at the Northeast Document Conservation Center (NEDCC) in Andover, Massachusetts. He specializes in the preservation/digitization of audio from magnetic and grooved formats. He presented “Speed and Configuration Changes: a Solution to... Read More →
avatar for Bryce Roe

Bryce Roe

Director of Audio Preservation Services, NEDCC
As the Director of Audio Preservation at NEDCC, Bryce supervises a staff of audio preservation engineers and specialists and manages NEDCC's audio preservation program, which uses both traditional playback and optical-scanning technologies (a.k.a., IRENE) to digitize at-risk audio... Read More →
avatar for Kelly Pribble

Kelly Pribble

Director of Media Recovery Technology, Iron Mountain Entertainment Services (IMES)
Kelly Pribble, Director of Media Recovery Technology at Iron Mountain Entertainment Services (IMES), is a veteran studio engineer, studio builder, archivist and inventor. In March 2022, Kelly was issued a Patent for Media Recovery Technology. Before joining Iron Mountain, Kelly... Read More →
Thursday October 10, 2024 11:45am - 12:30pm EDT
1E08

11:45am EDT

Re- Mixing Billy Joel's Classic Album in Atmos
Thursday October 10, 2024 11:45am - 12:45pm EDT
With the popularity of Dolby atmos growing, the need to re-mix classic albums has become popular. In this case study on Billy Joel's latest Atmos releases, our panel of engineers will go over their step by step process, from tape transfers, to finding the correct takes, restoring the stereo mix and opening it up to Dolby Atmos. We will touch on how much artistic liberty we have as well how to future proof a mix in an ever changing world.
Speakers
avatar for Oscar Zambrano

Oscar Zambrano

Chief Engineer / Partner, Zampol Productions
Oscar Zambrano was born and raised in Mexico City. After graduating from Berklee College of Music in 2003, he moved to New York where he started working at Sound on Sound. In 2004 he and Jorge Castellanos founded Zampol Productions, a recording studio that focuses on mastering, mixing... Read More →
avatar for Bradshaw Leigh

Bradshaw Leigh

Leigh Audio Design
Recording engineer/mixer started his career at New York's famed A&R Recording Studios. Mentored by Phil Ramone, Brad engineered acts including Julian Lennon, Billy Joel, and Paul Simon. As a freelance engineer, Brad has recorded and mixed for Jimmy Iovine, Russ Titleman, Tracy Chapman... Read More →
avatar for Adam Grover

Adam Grover

Mastering Engineer, Sterling Sound
Adam Grover is a Grammy-nominated mastering engineer at Sterling Sound. His eclectic discography includes projects with Zach Bryan, Stephen Sanchez, Rainbow Kitten Surprise, Surfaces, Joy Oladokun, Shaboozey, Jeris Johnson, and Maggie Lindemann. Adam leads Sterling Sound’s spatial... Read More →
Thursday October 10, 2024 11:45am - 12:45pm EDT
1E09

12:00pm EDT

Evaluation of sound colour in headphones used for monitoring
Thursday October 10, 2024 12:00pm - 12:20pm EDT
Extensive studies have been made into achieving generally enjoyable sound colour in headphone listening, but few publications have been written focusing on the demanding requirements of a single audio professional, and what they actually hear. The present paper describes a structured and practical method, based on in-room monitoring, for getting to know yourself as a headphone listener, and the particular model and pair you are using. Headphones provide fundamentally different listening results compared to in-room monitoring adhering to professional standards; considering imaging, auditory envelopment, localization, haptic cues etc. Moreover, in headphone listening there may be no direct connection between the frequency response measured with a generic manikin and what a given user hears. Finding out just how a pair of headphones deviates from neutral sound colour must therefore be achieved personally. An evaluation scheme based on an ultra-nearfield reference system is described, augmented by a defined test setup and procedure.
Moderators
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
Authors
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
Thursday October 10, 2024 12:00pm - 12:20pm EDT
1E03

12:15pm EDT

Student Closing/Awards Ceremony
Thursday October 10, 2024 12:15pm - 12:30pm EDT
Join us to close out the convention, and to find out who the competition winners are.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Angela Piva

Angela Piva

Angela Piva, Audio Pro/Audio Professor, highly skilled in all aspects of music & audio production, recording, mixing and mastering with over 35 years of professional audio engineering experience and accolades. Known as an innovator in sound technology, and for contributing to the... Read More →
Thursday October 10, 2024 12:15pm - 12:30pm EDT
1E06

12:30pm EDT

Education and Career Fair
Thursday October 10, 2024 12:30pm - 2:30pm EDT
The only education and career fair focused entirely on degree and certificate programs in audio around the world. Come meet professors, students, and college admissions representatives and discover how to advance your career as an audio professional!
Thursday October 10, 2024 12:30pm - 2:30pm EDT
Exhibit Hall

12:45pm EDT

Best Practices for Creating, Mixing, and Producing in Immersive Audio for Music, Film, TV, Podcasting and More
Thursday October 10, 2024 12:45pm - 1:45pm EDT
Join Focusrite Group’s Lee Whitmore in an interactive panel conversation and listening session with 19-time GRAMMY and Latin GRAMMY winner Rafa Sardina, Dolby’s Tom McAndrew, and ADAM Audio’s Mark Cummins. Listen to Dolby Atmos Music mixes by Rafa Sardina, and chat with industry experts about best practices for creating, mixing, and playback of immersive audio for music, film, TV, gaming, audiobooks and podcasting, as well as considerations for your studio and workflows.

Also held on 10/9 at 1pm
Speakers
avatar for Dr. Lee Whitmore

Dr. Lee Whitmore

VP, Education and B2B, Americas, Focusrite Group
Dr. Lee Whitmore is a thought leader in music, audio, creative digital media, and education. He’s the Vice President for business-to-business (B2B), including education, pro, post, and government, at Focusrite Group, which includes FocusriteNovationSequential and Oberh... Read More →
avatar for Tom McAndrew

Tom McAndrew

Senior Technical Manager – Content Relations, Dolby Laboratories
avatar for Rafa Sardina

Rafa Sardina

Rafa Sardina | Afterhours Studios / Abbey Road Institute Los Angeles
Los Angeles based 19-Time Grammy® & Latin Grammy® Winner Producer/Mixer/Engineer Rafa Sardina has built an eclectic and impressive client list over the years: Stevie Wonder, Camila Cabello, D’Angelo, Dr. Dre, Rosalía, John Legend, Elvis Costello & The Roots, Lady Gaga, Michael... Read More →
MC

Mark Cummins

Head of Product Marketing, ADAM Audio
Thursday October 10, 2024 12:45pm - 1:45pm EDT
1E06

1:00pm EDT

Best Practices Make Life Better For Everyone!
Thursday October 10, 2024 1:00pm - 1:45pm EDT
The Recording Academy Producers & Engineers Wing is presenting the latest (major!) revision of its widely used “Recommendations For Delivery of Recorded Music Projects.” Tips, Tricks and Sage Advice for managing your files, versions, naming conventions, etc.

A dedicated group of engineers devoted their time over four years to bring, this highly useful, practical document and presentation to fruition.
Moderators
avatar for Maureen Droney

Maureen Droney

Vice President, The Recording Academy Producers & Engineering Wing
Maureen Droney is Vice President of the Recording Academy (GRAMMY Awards) Producers & Engineers Wing. A former recording engineer herself, Maureen has worked on GRAMMY-winning recordings for artists including Aretha Franklin, Whitney Houston and Santana—as well as on numerous other... Read More →
Speakers
BK

Bob Koszela

Director, Studio Operations North America, Iron Mountain Entertainment Services
Bob Koszela is the Director, Studio Operations North America for Iron Mountain Entertainment Services.  A musician and songwriter, Bob brings over 30-years’ experience in archive media migration, major record label operations, A&R, marketing/promotion, new media, production, licensing/exploitation... Read More →
avatar for George Massenburg

George Massenburg

Associate Professor of Sound Recording, Massenburg Design Works
George Y. Massenburg is a Grammy award-winning recording engineer and inventor. Working principally in Baltimore, Los Angeles, Nashville, and Macon, Georgia, Massenburg is widely known for submitting a paper to the Audio Engineering Society in 1972 regarding the parametric equali... Read More →
avatar for Paul Womack

Paul Womack

Record Producer/Recording Engineer
A producer, engineer and sonic artist, Paul "Willie Green" Womack has built a discography boasting names such as Armand Hammer, Wiz Khalifa, The Alchemist, The Roots, Billy Woods, ELUCID and many more, and established himself as one of the top names in independent Hip-Hop & R&B. Currently... Read More →
Thursday October 10, 2024 1:00pm - 1:45pm EDT
3D04 (AES and AIMS Alliance Program Room)

1:00pm EDT

Gain Staging
Thursday October 10, 2024 1:00pm - 2:00pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Thursday October 10, 2024 1:00pm - 2:00pm EDT
1E05

1:30pm EDT

Enhancing and Optimizing Spatial Audio in Game Development: Best Practices and Innovations
Thursday October 10, 2024 1:30pm - 2:30pm EDT
This panel brings together leading industry experts to explore the critical role of spatial audio in enhancing the gaming experience. As games become more immersive, the demand for high-quality spatial audio has surged, making it an essential component for game developers. The discussion will delve into the intricacies of implementing spatial audio in both middleware and custom sound engines, addressing the challenges and solutions involved in achieving superior audio quality. Panelists will share insights on how to optimize spatial audio for various elements within a game, including sound effects, dialogue, and ambient sounds. Emphasis will be placed on customization techniques that ensure each audio component contributes to a cohesive and realistic auditory environment. A key focus of the panel will be on making spatial audio sound good across different types of games and game engines, highlighting the unique considerations and strategies for each scenario. Attendees will gain valuable knowledge on the latest advancements in spatial audio technology and practical tips for integrating these innovations into their game development workflows. Whether you are an audio engineer, game developer, or sound designer, this panel will provide you with the tools and strategies needed to elevate the audio experience in your games.
Speakers
KS

Kaushik Sunder

VP of Engineering, Embody
KB

Kevin Boettger

Spatial Audio Mastering Engineer, Embody
GD

Gordon Durity

Head of Audio Content, Electronic Arts
Thursday October 10, 2024 1:30pm - 2:30pm EDT
1E08

1:30pm EDT

They Might Be Giants - AES Masters Talk Audio History
Thursday October 10, 2024 1:30pm - 2:30pm EDT
Using the AES Oral Historical Videos, AES Legends, such as Phil Ramone, Les Paul, Bob Ludwig, Ray Dolby and others (some known and others not so widely known) with the assistance of Jim Anderson, will tell the inside and partly forgotten story of how modern recording techniques and facilities developed as told by the people who were there using the AES Oral Histories edited by the presenter.
Speakers
avatar for Jim Anderson

Jim Anderson

Producer/Engineer, Anderson Audio New York
Producer/Engineer
Thursday October 10, 2024 1:30pm - 2:30pm EDT
1E15

1:45pm EDT

From research to market: Entrepreneurial journeys in audio technology
Thursday October 10, 2024 1:45pm - 2:45pm EDT
Embarking on the journey from innovative research to launching a successful company in the audio technology sector can be both exciting and daunting. This workshop aims to demystify the entrepreneurial process for AES engineers and professionals who aspire to transform their research into a recognized product and a thriving business. By featuring a panel of seasoned entrepreneurs who have successfully navigated this path, attendees will gain invaluable insights and practical guidance on how to start and grow their own ventures in the audio industry.

Workshop Overview:
This workshop will provide a comprehensive roadmap for audio engineers and technologists considering the path of entrepreneurship. It will feature detailed discussions on the following key topics:

1. Identifying market opportunities:
- Understanding market needs and gaps in the audio technology sector.
- Conducting market research and validating your product idea.

2. Building a strong foundation:
- Transitioning from a research idea to a viable product.
- Establishing a solid business plan and securing initial funding.
- Protecting intellectual property through patents and trademarks.

3. From prototype to production:
- Developing and refining prototypes.
- Navigating the challenges of manufacturing and scaling up production.
- Ensuring quality control and meeting industry standards.

4. Go-to-market strategies:
- Crafting an effective marketing and branding strategy.
- Leveraging digital platforms and social media for product launch.
- Building partnerships and distribution channels.

5. Sustaining growth and innovation:
- Fostering a culture of continuous innovation and adaptation.
- Managing growth and scaling operations.
- Securing follow-on funding and exploring exit strategies.

Interactive elements:
The workshop will emphasize interaction and engagement, providing ample opportunities for attendees to pose questions and share their own experiences. Each panelist will deliver a brief presentation, followed by a moderated discussion and an open Q&A session.

Target audience:
This workshop is tailored for audio engineers, researchers, and technologists who have developed innovative audio technologies and are considering starting their own companies. It suits those with a basic understanding of audio technology and a keen interest in entrepreneurship.
Speakers
avatar for Tomasz Żernicki

Tomasz Żernicki

co-founder, my3DAudio
Tomasz Zernicki is co-founder and former CEO of Zylia (www.zylia.co), an innovative company that provides tools for 3D audio recording and music production.Additionally, he is a founder of my3DAudio Ventures, whose goal is to scale audio companies that reach the MVP phase and want... Read More →
avatar for Kapil Jain

Kapil Jain

CEO, Embody
We are Embody, and we believe in the power of AI to push the boundaries of immersive sound experience in gaming and entertainment.We are a team of data scientists, audio engineers, musicians and gamers who are building an ecosystem of tools and technologies for Immersive entertainment... Read More →
avatar for Edward Wersocki

Edward Wersocki

Senior Software Engineer, Everbridge
Edward is a Senior Software Engineer at Everbridge, specializing in the development of large-scale distributed systems. His research is centered on improving music perception for cochlear implant and hearing aid users, as well as advancing the integration of accessibility features... Read More →
avatar for Rebekah Wilson

Rebekah Wilson

CEO, Source Elements
Since 2005 Source Elements helps connect creative people around the world to collaborate in real-time production, approval, and delivery of professional quality media, for the media industry. Source Elements is the makers of Source-Connect: the industry standard in the film and television... Read More →
Thursday October 10, 2024 1:45pm - 2:45pm EDT
1E07

1:45pm EDT

Marcela Zorro & Florian Camerer: 3D for Two
Thursday October 10, 2024 1:45pm - 2:45pm EDT
Distinct differences between music and atmosphere recordings are even more noticeable in 3D audio than in stereo. In this 3D for Two, Marcela Zorro and Florian Camerer take turns when playing high resolution recording examples of music and outdoor atmosphere, explaining also microphone and production techniques from the two sides.

We listen and discuss how each method comes across in this room, touching also upon topics such as microphone arrays, location recording, mixing, use of LFE; and the current distribution formats available for 3D audio.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Florian Camerer

Florian Camerer

Senior Sound Engineer, ORF - Austrian TV
Florian is a post-production sound engineer for TV with more than three decades under his belt working for ORF - the Austrian public broadcaster. He started out in documentaries (location sound, sound design, mixing), recording and mixing ORF's first surround sound documentary in... Read More →
avatar for Marcela Zorro

Marcela Zorro

Universidad de los Andes
Marcela is an audio engineer specializing in the recording and production of classical music. She currently teaches in the audio production emphasis of the Music Department at Universidad de los Andes in Bogotá, Colombia. Marcela has been an active member of the AES since 2004.  She... Read More →
Thursday October 10, 2024 1:45pm - 2:45pm EDT
3D06

2:00pm EDT

Preserve, Protect, Profit - Reshaping How Artists Can Take Control of Their Own Legacies Through Technology
Thursday October 10, 2024 2:00pm - 2:45pm EDT
The evolution of the music industry, and how artists have traditionally approached their legacy, has taken a shift as more artists look to capitalize on the opportunity to digitize and monetize their legacies. This panel will dive into the role technology and other artificial intelligence advances the way musicians preserve, monetize, and authenticate their estate’s valuation. Hear from panelists who work in the music industry and media preservation space about the new age approach to ensuring the protection of iconic and cultural legacy assets for generations to come.  
Moderators
avatar for Meg Travis

Meg Travis

Director - Global Head of Marketing & Communications, Iron Mountain media and Services
Meg Travis leads industry and client engagement across North America, the United Kingdom and Europe. Her focus on global brand development has played an integral role in spearheading the company’s initiatives to bring media preservation and archiving to the forefront of the conversation... Read More →
Speakers
avatar for Terry Lewis

Terry Lewis

Five-time Grammy Award winners Jimmy Jam & Terry Lewis are world renowned songwriters, record producers, musicians, entrepreneurs, and the most influential and successful songwriting and production duo in modern music history. For over 40 years Jam & Lewis’ accomplishments have... Read More →
avatar for Francis Porter

Francis Porter

Co-Founder and COO, ANALOGr
Francis drives innovation at ANALOGr Authentication. As an award-winning documentarian, he draws on his experience to capture factual content that elicits the story value inherent in artist assets. As a technology product manager, he has led the platform to maintain and safeguard... Read More →
avatar for Thomas Scriven

Thomas Scriven

CEO, ANALOGr
Thomas Scriven is the CEO of ANALOGr and a trusted authority when establishing authentication, valuation, & monetization for artist owned assets. Previously, Thomas held various creative roles in the music business as an A&R, marketing manager, and artist manage.
avatar for Jimmy Jam

Jimmy Jam

Five-time Grammy Award winners Jimmy Jam & Terry Lewis are world renowned songwriters, record producers, musicians, entrepreneurs, and the most influential and successful songwriting and production duo in modern music history. For over 40 years Jam & Lewis’ accomplishments have... Read More →
Thursday October 10, 2024 2:00pm - 2:45pm EDT
3D04 (AES and AIMS Alliance Program Room)

2:00pm EDT

Mental Health in the Studio
Thursday October 10, 2024 2:00pm - 3:00pm EDT
Our program fosters candid conversations among studio professionals within a supportive environment. It examines the intricate relationship between mental health and client/artist dynamics, focusing on methods to empower artists during times of emotional strain. Participants will explore strategies for preserving personal equilibrium and managing professional boundaries, along with learning coping mechanisms for navigating high-pressure situations. Led by JC Losada "Mr Sonic" and Maria Rice, with special guest(s) TBA, the discussion draws upon experiences in studio environments, educational contexts, and clinical practice.
Speakers
Thursday October 10, 2024 2:00pm - 3:00pm EDT
1E09

2:00pm EDT

AI in Electronic Instrument Design
Thursday October 10, 2024 2:00pm - 3:00pm EDT
As applications of artificial intelligence and machine learning become prevalent across the music technology industry, this panel will examine the ways that these technologies are influencing the design of new electronic instruments. As part of the discussion we’ll looking at current instruments as well and the potential for new designs.
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
avatar for Akito van Troyer

Akito van Troyer

Associate Professor, Berklee College of Music
avatar for Dan Gonzalez

Dan Gonzalez

Principal Product Manager, iZotope & Native Instruments
VZ

Victor Zappi

Northeastern University
Thursday October 10, 2024 2:00pm - 3:00pm EDT
1E16

2:00pm EDT

Psychoacoustics and Practices of Immersive Sound Recording
Thursday October 10, 2024 2:00pm - 3:00pm EDT
In immersive sound recording with a microphone array, height channels are crucial for delivering realistic and enveloping auditory experiences. To maximize the benefits of these added channels and effectively capture height information, a solid understanding of the psychoacoustic principles of vertical stereophonic perception is essential. This tutorial/demo session will provide an overview of the key psychoacoustic principles that recording engineers should consider when creating immersive recordings with a microphone array. Various microphone array techniques and practical workflows for 3D sound capture will be introduced, along with a discussion of their advantages and disadvantages. The session will include demonstrations of various 7.1.4 recordings.
Speakers
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
Thursday October 10, 2024 2:00pm - 3:00pm EDT
1E06

2:20pm EDT

2nd order Boom Microphone
Thursday October 10, 2024 2:20pm - 2:40pm EDT
Boom microphones are mostly used in noisy environments and equipped with directional microphones to enhance speech pickup from talkers and suppress noise from surroundings. Theory suggests that using a 2nd order directional microphone array, that is simply a pair of two directional microphones, would greatly increase the amount of far-field noise rejection, with negligibly little degradation of the near-field pickup. We conducted a series of laboratory measurements to validate the theory and assess the feasibility of 2nd order boom microphone applications. The laboratory measurements were designed using Knowles Electronics Manikin for Acoustic Research (KEMAR) with two speakers to play cockpit noise, one being placed on axis of the KEMAR front and one being set to side. A mouth simulator inside KEMAR played artificial voice. The measurements were done at two distances, 6 and 25 mm, from the KEMAR mouth with a 1st order boom microphone, that is a single directional boom microphone, and a 2nd order boom microphone. Spice simulations were also conducted to support the experimental findings. Both the measurements and Spice simulations proved the theory was true until the 2nd order boom microphone was placed near KEMAR. Then, reflections off the KEMAR head degraded the performance of the 2nd order microphone while improving that of the 1st order microphone. The net result shows that the 2nd order microphone is not superior than the 1st order microphone. This article describes details of the theory, our experimental measurements, and the findings.
Moderators
PG

Paul Geluso

Director of the Music Technolo, New York University
Speakers
avatar for Jeong Nyeon Kim

Jeong Nyeon Kim

Senior Electro-Acoustic Application Engineer, Knowles Electrics
Authors
avatar for Jeong Nyeon Kim

Jeong Nyeon Kim

Senior Electro-Acoustic Application Engineer, Knowles Electrics
Thursday October 10, 2024 2:20pm - 2:40pm EDT
1E03

2:30pm EDT

Live Sound Q&A
Thursday October 10, 2024 2:30pm - 3:30pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Thursday October 10, 2024 2:30pm - 3:30pm EDT
1E05

2:40pm EDT

Improved Analogue-to-Digital Converter for High-quality Audio
Thursday October 10, 2024 2:40pm - 3:00pm EDT
A high-oversampled, low-bit modulator typical of modern audio ADCs needs a downsampler to provide PCM audio at sampling rates ranging from 44.1kHz to 768kHz. Traditionally, a multistage downsampler requantizes at each stage, raising questions about audio transparency. We present a decimator design in which there is no requantization other than a single dithered quantization when necessary to produce a final audio output of finite precision such as 24 bits. All processing is minimum-phase and concordant with the principles introduced in [1] which optimize for a specific compact impulse response and minimal (zero) modulation noise. [2]
Moderators
PG

Paul Geluso

Director of the Music Technolo, New York University
Speakers Authors
Thursday October 10, 2024 2:40pm - 3:00pm EDT
1E03

2:45pm EDT

Modern Production Workflows for Audio & Video
Thursday October 10, 2024 2:45pm - 3:45pm EDT
This workshop, featuring audio editors Sean Garnhart and Dan Brennan with video editor Gary Levy from Soundtrack Group, will delve into the dynamic interplay between video and audio professionals. Participants will hear about technical challenges and solutions from industry experts in real world film and TV environments from both a video and an audio perspective and how their workflows are evolving in tandem with new technologies and integrations.
Speakers
SG

Sean Garnhart

Sean Garnhart is a freelance supervising sound editor, sound designer and re-recording mixer who’s had the pleasure of working with M. Night Shyamalan, Spike Lee, Tim Burton, and the Coen Brothers. Some of his feature credits include OLD, Beauty & The Beast, The King of Staten Island... Read More →
DB

Dan Brennan

Dan Brennan is a highly regarded Re-Recording & Sound Supervisor with a career spanning over 20 years in the audio post-production industry. Dan's work is well-recognized in both film and television, with notable credits such as The Diplomat, Genius: Aretha, George and Tammy, Sorry... Read More →
GL

Gary Levy

Gary Levy is currently editing THE DIPLOMAT for Netflix. Television credits include TOKYO VICE, KALEIDOSCOPE, THE GOOD FIGHT, THE SINNER. NURSE JACKIE and THE OFFICE.​
Thursday October 10, 2024 2:45pm - 3:45pm EDT
1E15

2:45pm EDT

Get Smart! - Everything you wanted to know about game audio education but were afraid to ask!
Thursday October 10, 2024 2:45pm - 4:15pm EDT
Game Audio education programs have taken root and sprouted up all over the world. Game audio education is a hot topic. What are some of the latest training programs out there? What are the pros and cons of a degree program versus just surfing YouTube? I am already a teacher, how can I start a game audio program at my current school? Good questions! This panel brings together sound artists and entrepreneurs from some of the top private instructional institutions and teachers from some growing programs to discuss the latest and greatest educational models in audio for interactive media. Attendees will get a fantastic overview of what is being offered inside and outside of the traditional education system. This is a must for students and teachers alike, who are trying to navigate the waters and steer a path toward programs that are right for them in the shifting tides of audio for games and interactive media.
Speakers
avatar for Steven Horowitz

Steven Horowitz

Executive director of Technology and Applied Composition (TAC) program at SFCM, SFCM
Critically acclaimed as "One of the foremost figures in the field" Composer Steven Horowitz is the executive director of Technology and Applied Composition (TAC) program at SFCM, a position he took over in January, 2024.Horowitz comes to SFCM after 23 years as Audio Director at Nickelodeon... Read More →
avatar for Dafna Naphtali

Dafna Naphtali

Sound Artist / Composer
Dafna Naphtali is a sound-artist, vocalist, electronic musician and guitarist. As performer and composer of experimental, contemporary classical and improvised music since the mid-90’s, she creates custom Max/MSP programming for sound-processing of voice and other instruments. “luminary... Read More →
avatar for Alistair Hirst

Alistair Hirst

Sr. Game Dev Relations Mgr, Dolby
Alistair Hirst has shipped over 44 games across 11 platforms, doing sound design, music composition, audio programming and integration. He launched the Need for Speed franchise as Audio Director during his 10 years at Electronic Arts.  He co-founded OMNI Audio, a game audio company... Read More →
MD

Michele Darling

Chair, Electronic Production and Design, Berklee College of Music
Thursday October 10, 2024 2:45pm - 4:15pm EDT
1E08

3:00pm EDT

The sound of storytelling: the role of sound design and music in the ’drama’ genre of film
Thursday October 10, 2024 3:00pm - 3:20pm EDT
The integration of sound effects and music plays a central role in shaping the audience's emotional engagement and narrative comprehension in film. The 'drama' genre in film is primarily concerned with depicting human emotions and human narrative-based storytelling. Ten scenes were analysed, with participants in three groups exposed to six combinations of audio and visual stimuli. Participants reported salient sounds and their interpretations, focusing on context and emotional responses. One hypothesis is that effective sound design blurs the line between music and sound effects; another is that music conveys more emotion while sound effects enhance immersion. The results showed that 63\% of participants found the score more relevant to the context. The evaluation highlights that music alone emphasizes certain emotions more, while sound effects alone create moderate variability between emotion and sound identification.
Moderators
PG

Paul Geluso

Director of the Music Technolo, New York University
Speakers Authors
Thursday October 10, 2024 3:00pm - 3:20pm EDT
1E03

3:00pm EDT

State of the Music Business
Thursday October 10, 2024 3:00pm - 4:30pm EDT
State of the Music Business

In Depth Conversation on the State of the Music Business - Streaming,
Royalties and AI

Panelists:

Scott Rubin, Reach Music Publishing
Music Publisher
https://www.reachmusic.com/

Judy Tint, Attorney, Consultant, Producer, Professor/NYU
https://steinhardt.nyu.edu/people/judy-tint

Hank Shocklee, Artist, Musician, Producer, A&R, Content Creator

Moderator: Angela Piva. Co Chair Education Committee
Speakers
avatar for Judy Tint

Judy Tint

Judy Tint is an attorney, professor, and musician with over 30 years of experiencerepresenting artists, songwriters, producers, executives, and more. A graduate ofRutgers College and Columbia La School, she is a Clinical Associate Professor in theNYU Music Business Program, and served... Read More →
avatar for Hank Shocklee

Hank Shocklee

President, Shocklee Entertainment
Hank Shocklee is an audio technology innovator & futurist / music producer / sound designer who isalways inspired by pushing the limits of technology, creativity and problem solving. Best known as thearchitect of Public Enemy's "wall of noise" sound, he is a Rock and Roll Hall of... Read More →
avatar for Scott Rubin

Scott Rubin

Scott Rubin is a 27 year veteran of the music industry, with an emphasis in music publishing; spending allof his career with Reach Music Publishing (www.reachmusic.com). Since joining Reach Music in 1999,he has been an integral part of building their full service, independent music... Read More →
avatar for Angela Piva

Angela Piva

Angela Piva, Audio Pro/Audio Professor, highly skilled in all aspects of music & audio production, recording, mixing and mastering with over 35 years of professional audio engineering experience and accolades. Known as an innovator in sound technology, and for contributing to the... Read More →
Thursday October 10, 2024 3:00pm - 4:30pm EDT
1E07

3:00pm EDT

Making of Mean Girls 2024
Thursday October 10, 2024 3:00pm - 4:30pm EDT
Deep dive into the making of Mean Girls featuring Lewis Goldstein, Atmos Mixer Rerecording Engineer and Hanan Rubinstein the lead music producer. Recorded at Daxxit Sound Studios (Fort Lee, NJ) and The DiMenna Center for Classical music (NYC). Score was mixed at Daxxit sound studios, the music was mixed at Larrabee Studios (Los Angeles) and immersive theatrical Atmos was mixed at Parabolic Studios (NYC) by Lew Goldstein.

Moderated by: Roey Shamir

AES NY Section originally presented a similar panel on May 5th this year , Unfortunately the music supervisor Jack Grabon and one of the soundtrack mixers  Chris Galland can’t be with us, but we will share some of what they said at the original panel. Photo is from that event promo









Speakers
avatar for Roey Shamir

Roey Shamir

Mixmaster, INFX PRODUCTIONS INC.
• Freelance Audio Engineer, Mixer and Producer • Audio Engineer Mixer, Producer, Remixer for extensive major label records and filmproductions (Mix to Picture) • Mixed and Recorded music / albums for numerous RIAA multi platinum and Grammynominated... Read More →
avatar for Lewis Goldstein

Lewis Goldstein

Atmos Mixer Rerecording Engineer, Owner Parabolic Studios
avatar for Hanan Rubenstein

Hanan Rubenstein

Lead Music Producer and Score Mixer
Hanan is an in demand sound engineer, producer, vocal coach and a true multi-instrumentalist. From the stage to the studio he is known for his work with Alicia Keys, Adam Blackstone/BBE, Jeff Richmond and J.Cole. His other credits include a bevy of industry leading artists such as... Read More →

Thursday October 10, 2024 3:00pm - 4:30pm EDT
Stage

3:15pm EDT

DEI Town Hall
Thursday October 10, 2024 3:15pm - 4:15pm EDT
Reports from the DEI committee and subcommittees will be given. The floor will then be opening for questions and discussion.
Speakers
avatar for Mary Mazurek

Mary Mazurek

Audio Educator/ Recording Engineer, University of Lethbridge
Audio Educator/ Recording Engineer
avatar for Jiayue Cecilia Wu

Jiayue Cecilia Wu

Assistant Professor, Graduate Program Director (MSRA), University of Colorado Denver
Originally from Beijing, Dr. Jiayue Cecilia Wu (AKA: 武小慈) is a scholar, composer, audio engineer, and multimedia technologist. Her work focuses on how technology can augment the healing power of music. She earned her Bachelor of Science degree in Design and Engineering in 2000... Read More →
Thursday October 10, 2024 3:15pm - 4:15pm EDT
1E16

3:15pm EDT

Analog to Atmos: Bob Marley
Thursday October 10, 2024 3:15pm - 4:15pm EDT
The journey of Bob Marley’s music from analog to atmos will be discussed. In 2022-23, Burnin’, Catch A Fire, Exodus and Kaya were mixed in atmos by Nick Rives with the Marley estate’s approval. Eliot Kissileff created the stems used for the mixes. Nick went on to work on the Marley feature film released in 2024. Chris Walch, UMG VP of Assets and Archives will give some background on how the original tapes are preserved, digitized and presented for various projects. The panel will give insight into the evolving processes for creating the stems and focus on Nick’s creation of the atmos mixes, giving the audience insight into the processes and evolution of the format.
Speakers
avatar for Eliot Kissileff

Eliot Kissileff

Owner, Music and Media Preservation Services
After graduating from New York University’s Tisch School of the Arts with a BFA in film production, Eliot Kissileff attended the Institute of Audio Research before landing a job at Electric Lady Studios in 1996. With the goal of learning how to make vintage-sounding recordings for... Read More →
Thursday October 10, 2024 3:15pm - 4:15pm EDT
1E06

3:15pm EDT

Mastering Music - Busting Myths
Thursday October 10, 2024 3:15pm - 4:15pm EDT
Mastering for music in stereo in 2024. What are the real objectives of this discipline today? What are the axioms of this practice? And what is a typical modern processing chain made of? When and why do I need to do A/B comparisons? What does QC stand for? What types of processors are normally used and how? When and how to apply dither to a music program? What format and resolution should the mastering engineer work at? At which format and resolution should the mixing engineer deliver the final work? How loud should the final mixes be and why doesn't it really matter? These and other issues will be addressed in this tutorial.
Speakers
avatar for Ezequiel Morfi

Ezequiel Morfi

Studio Manager (Buenos Aires), TITANIO STUDIOS
I'm a producer and engineer dedicated to recording, mixing and mastering music in stereo!
Thursday October 10, 2024 3:15pm - 4:15pm EDT
1E09

4:00pm EDT

’The Good, the Bad, and the Ugly’ - Best practices in Immersive audio production
Thursday October 10, 2024 4:00pm - 5:30pm EDT
Current music production has opened up a relatively new opportunity in the creation of 3D content, now easily available online and in various forms. Artists and labels are increasingly asking for 3D content to be delivered along with a standard stereo master. Many practitioners, however, are at a loss when it comes to how and what to deliver, as various streaming services implement their own binaural processes. Remixing older catalog content or even recent Stereo releases has proven to be problematic when working from stems that are badly prepared, or very costly when going back and remixing from scratch, using the original multitrack masters. The panel will speak to best tactics regarding the creation of stems to be delivered for use in immersive mixing, along with some examples of routing with sub-master processing, along with proper use of LFE and center channel, sides, rear and height channels. Other issues to be discussed include room simulation, loudness measurement and the various binaural processes. Questions from the audience will be a welcome part of the workshop.
Speakers
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
avatar for George Massenburg

George Massenburg

Associate Professor of Sound Recording, Massenburg Design Works
George Y. Massenburg is a Grammy award-winning recording engineer and inventor. Working principally in Baltimore, Los Angeles, Nashville, and Macon, Georgia, Massenburg is widely known for submitting a paper to the Audio Engineering Society in 1972 regarding the parametric equali... Read More →
avatar for Marc Theriault

Marc Theriault

Mastering Engineer, Le Lab Mastering
Thursday October 10, 2024 4:00pm - 5:30pm EDT
1E15

4:30pm EDT

The medium is the message: Using embedded metadata to integrate systems
Thursday October 10, 2024 4:30pm - 5:30pm EDT
Metadata is essential in audio files for archiving, discovering, and accessing content. But in current multiplatform environments, metadata management can become difficult and complex. One solution is to implement a monolithic, centralised system, but this is often impractical and overly rigid in dynamic production workflows, since different clients or stakeholders may prefer specific delivery systems. As a result, media managers often have to deal with disparate distribution systems such as file-sharing services, e-mail attachments, download URLs, proprietary APIs, etc. The result of many of these distribution processes is that the audio content and its descriptive metadata often become separated, making the content potentially far less usable.

Embedding metadata in your audio files allows for easier, "one prong" delivery of content, in which you can obtain more robustness as the audio travels downstream along your workflows and into your client's systems. Furthermore, embedding metadata can also simplify system integration from a two-way process to a one-way process, where the receiving system does not need to be aware of specific requirements from the transmitting system.

In this workshop we will explore how the New York Public Radio Archives is using existing metadata fields in archival WAVE files to describe, authenticate, and augment their metadata. Using free or very low-cost tools, alongside well established standards, we will use some of the principles behind W3C's Resource Description Framework (RDF) to choose embedded metadata that is robust, consistent, and surprisingly flexible.
Speakers
avatar for Marcos Sueiro Bal

Marcos Sueiro Bal

Archives Manager, New York Public Radio
Marcos Sueiro Bal is the Archives Manager at New York Public Radio. He is a member of the IASA and ARSC Technical Committees. He specializes in audio reformatting and in digital metadata.
Thursday October 10, 2024 4:30pm - 5:30pm EDT
1E16

4:30pm EDT

Marcus "Da Fingaz" Manderson Presents: Mixing Night With Ken Lewis LIVE
Thursday October 10, 2024 4:30pm - 5:30pm EDT
Join Marcus (@dafingaz) as he brings legendary mix engineer, producer, and songwriter Ken Lewis to the AES stage.

Ken Lewis has one of the most unique resumes in the music industry, with 114 gold records to back it up. He's worked with artists like BTS, Kanye West, Future, Bruno Mars, and Taylor Swift. There's likely a song he's touched that's impacted your life. Through Mixing Night Audio, Ken and his team aim to help the next generation of creatives by sharing his experience, tools, and techniques—no gatekeeping, just actionable insights from his 30 years of making hit records.

Mixing Night offers some of the best-designed plugins on the market. You won’t want to miss this session—there may even be a special guest or two joining!
Speakers
avatar for Marcus Manderson

Marcus Manderson

Music Composer, Producer, Da Fingaz
I'm a versatile composer and producer, creating music in the epic cinematic, hip hop, pop, R&B, and more genres.Highlights include composing for Disney theme parks, composing music for the film "Hemisphere" (streaming on Amazon Prime Video, Google Play, Tubi), and securing sync placements... Read More →
Thursday October 10, 2024 4:30pm - 5:30pm EDT
1E09

4:30pm EDT

Cross-talk-cancellation for next-generation spatial audio content
Thursday October 10, 2024 4:30pm - 5:30pm EDT
Spatial audio reproduction is predominantly achieved through surround speaker setups or headphones. Surround setups offer high levels of immersion but can be impractical to set up and are unable to reproduce near-field binaural cues, such as whispers.
Conversely, while headphones are convenient, they struggle with sound externalization.

Cross-talk cancellation is a technique for reproducing binaural sounds via speakers, capable of delivering precise 3D audio, including both distant sounds (externalization) and near-field sounds (whispers). Despite its development over five decades ago, its commercial adoption has been slow. However, recent market innovations are enhancing user experiences significantly.

This panel aims to investigate the viability of cross-talk cancellation as an alternative to headphones and surround setups for high-quality binaural 3D audio, particularly in video games. We will engage experts from the gaming industry, spatial audio reproduction sector, and academia to gather insights on the method and its potential applications.
Speakers
avatar for Marcos Simon

Marcos Simon

CTO, Audioscenic
KS

Kaushik Sunder

VP of Engineering, Embody
EC

Edgar Choueiri

Princeton University
avatar for Jago Reed-Jones

Jago Reed-Jones

R&D Engineer, Audioscenic
I am a Research & Development Engineer at Audioscenic, where we are developing the next generation of 3D Audio Technology using binaural audio over loudspeakers.In addition, I am finishing a PhD at Liverpool John Moores University looking at use of neural networks to achieve binaural... Read More →
Thursday October 10, 2024 4:30pm - 5:30pm EDT
1E08

4:30pm EDT

Tools for Impressive 3D Audio Productions
Thursday October 10, 2024 4:30pm - 5:30pm EDT
3D audio productions can be realized in different ways: in acoustic rooms with hand-played instruments or in the box, where the sound is only created during production. What they have in common are the psychoacoustic rules that apply to all audio productions. The artistic work goes hand in hand with the psychoacoustic layer, crucial for immersive productions' perceived result: content resolution, transparency, and auditory connection/separation.
A high content resolution can be achieved with many individual sounds and useful tools such as reverb plug-ins. In the first part of the workshop, Lasse Nipkow presents a strategy for achieving a high resolution in 3D audio with mobile recordings and various sound bodies, thus creating mixes.
On the playback side, we now have a dozen or more loudspeakers. Therefore, the tools we need for mixing must overcome new challenges that do not exist in stereo. In the second part of the contribution, Grammy-nominated sound engineer Ronald Prent and Lasse Nipkow present outstanding plug-ins explicitly developed for 3D audio in recent years. Using impressive sound examples from various genres, they show how they use these tools effectively for their mixes.
This second contribution is related to a first, previous contribution by Lasse Nipkow and Konrad Strauss: ‘Psychoacoustics for Immersive Productions’.
Speakers
avatar for Lasse Nipkow

Lasse Nipkow

CEO, Silent Work LLC
Since 2010, Lasse Nipkow has been a renowned keynote speaker in the field of 3D audio music production. His expertise spans from seminars to conferences, both online and offline, and has gained significant popularity. As one of the leading experts in Europe, he provides comprehensive... Read More →
avatar for Ronald Prent

Ronald Prent

Proper Prent Sound
Ronald Prent started his career at Wisseloord Studios in the Netherlands in 1980 as an inexperiencedassistant and has since established himself as one of the most accomplished and innovative recording &mix engineers in the world. In the early years of surround sound, Ronald was a... Read More →
Thursday October 10, 2024 4:30pm - 5:30pm EDT
1E06

4:45pm EDT

Loudspeaker design method focusing on impedance characteristics
Thursday October 10, 2024 4:45pm - 5:30pm EDT
The majority of loudspeaker design now relies on simulation software such as Klippel. This approach has been effective in improving various speaker characteristics. However, the overreliance on simulation software without a deep understanding of electro-acoustic transducers has led to a neglect of the fundamental principles of speakers. This has become a bottleneck in enhancing the sound quality of speakers.

This session re-focuses on impedance characteristics, which are no longer receiving sufficient attention today. It explains the sound quality issues associated with impedance fluctuations, methods for addressing them, and how to design speakers based on information derived from impedance characteristics.
Thursday October 10, 2024 4:45pm - 5:30pm EDT
1E07

6:00pm EDT

Tech Tour: Greene Space
Thursday October 10, 2024 6:00pm - 6:45pm EDT
All Tech Tours are full. Know Before You Go emails were sent to accepted Tech Tour registrants with information about the tour.  

Duration: 45 minutes
Capacity: 45 (Once capacity is reached, registrants will be added to a wait list)


Tour Description: The Greene Space is a public, street-level performance space that is under the auspices of New York Public Radio and shares its mission to make the mind more curious, the heart more open and the spirit more joyful. But while radio and podcasts are media made for millions, The Greene Space was built to change the world one in-person audience at a time, moving New Yorkers to the intersection of art and politics, leading courageous conversations and curating performances of intense beauty that are deeply rooted locally but relevant to audiences globally.

The space provides the same technical service for broadcast quality audio and video recording and live streaming that we use to power our own radio stations, podcasts, and concerts and accommodates an in-person audience of up to 250 people.  One unique technical aspect of the space is that it's built on an analog-first design centered on a state-of-the-art API Vision audio console.  The acoustic construction and isolation challenges of the space were not insignificant, being located only a few meters above a major subway line and looking out through a mostly-glass facade to one of the heaviest traffic congestion points in the city and we will explain that as well as the technology of the space.

 
Arrival Instructions: Come to 44 Charlton St.  NY, NY 10013
Thursday October 10, 2024 6:00pm - 6:45pm EDT
Offsite
 
Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.