Loading…
AES Show 2024 NY has ended
Exhibits+ badges provide access to the ADAM Audio Immersive Room, the Genelec Immersive Room, Tech Tours, and the presentations on the Main Stage.

All Access badges provide access to all content in the Program (Tech Tours still require registration)

View the Exhibit Floor Plan.
strong> [clear filter]
arrow_back View All Dates
Wednesday, October 9
 

8:00am EDT

Attendee Registration
Wednesday October 9, 2024 8:00am - 6:00pm EDT
Wednesday October 9, 2024 8:00am - 6:00pm EDT
Crystal Palace Hall

9:00am EDT

Student Recording Competition Category 1
Wednesday October 9, 2024 9:00am - 10:00am EDT
Traditional Acoustic Recording finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
avatar for Martha DeFrancisco

Martha DeFrancisco

Associate Professor, McGill University
Martha De Francisco is Associate Professor of Sound Recording at McGill University in Montreal, Canada. A native of Colombia, she has established a prominent international career for over thirty years as a record producer/engineer for classical music, frequently invited to produce... Read More →
avatar for Kseniya Kawko

Kseniya Kawko

Tonmeister, msm-studios & Jazz Club Unterfahrt
Kseniya Kawko is a producer and recording engineer specialized in classical music and jazz. She holds Master of Music degrees from two world-renowned audio programs: Sound Recording, McGill University (Montréal, Canada) and Musikregie / Tonmeister, Hochschule für Musik Detmold (Germany... Read More →
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E07

9:00am EDT

New Directions in Modular Synthesizer Design
Wednesday October 9, 2024 9:00am - 10:00am EDT
With the explosive growth of modular instruments in the Eurorack format, synthesizer modules have become more sophisticated pointing to new directions in synthesis capabilities as well as real time performance control. This panel will consider some new developments in the field from leading manufacturers.
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
avatar for Leon Gruenbaum

Leon Gruenbaum

Consultant, Eventide
Hello. I am here to present the Misha, Eventide’s Misha Eurorack instrument and sequencer, for which I served as a consultant
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E09

9:00am EDT

Sound Design for Multi-Participant Extended Reality Experiences
Wednesday October 9, 2024 9:00am - 10:00am EDT
This workshop explores the unique challenges and opportunities in sound design and technical implementation for multi-participant extended reality (XR) experiences. Participants will learn techniques for creating immersive and interactive spatial audio to enhance presence and facilitate communication and collaboration in multi-participant XR environments. Through demonstrations and an expert-led panel discussion, attendees will gain insights into creating auditory experiences in XR.
Speakers
avatar for Agnieszka Roginska

Agnieszka Roginska

Professor, New York University
Agnieszka Roginska is a Professor of Music Technology at New York University. She conducts research in the simulation and applications of immersive and 3D audio including the capture, analysis and synthesis of auditory environments, auditory displays and applications in augmented... Read More →
avatar for Yi Wu

Yi Wu

PhD Student, New York University
avatar for Parichat Songmuang

Parichat Songmuang

Studio Manager/PhD Student, New York University
Parichat Songmuang graduated from New York University with her Master of Music degree in Music Technology at New York University and Advanced Certificate in Tonmeister Studies. As an undergraduate, she studied for her Bachelor of Science in Electronics Media and Film with a concentration... Read More →
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E08

9:00am EDT

Unlocking New Dimensions: Producing Music in Immersive Audio
Wednesday October 9, 2024 9:00am - 10:00am EDT
The field of audio production is always evolving. Now with immersive audio formats becoming more and more prominent, we should have a closer look at what possibilities come with it from a technical but most importantly from an artistic and musical standpoint.
In our Workshop, "Unlocking New Dimensions: Producing Music in Immersive Audio," we demonstrate how immersive audio formats can bring an artist's vision to life and how the storytelling in the music benefits from them.
In order to truly change the way people listen to music and provide an immersive experience, we must transform how we write and produce music, using immersive formats not just as a technical advancement but as a medium to create new art.
In this session, we will explore the entire production process, from recording to the final mix, and master with a focus on how one can create a dynamic and engaging listening experience with immersive formats like Dolby Atmos. We believe that immersive audio is more than just a technical upgrade—it's a new creative canvas. Our goal is to show how, by fully leveraging a format like Dolby Atmos, artists and producers can create soundscapes that envelop the listener and add new dimensions to the storytelling of music.

Philosophy

Artists often feel disconnected from the immersive production process. They rarely can give input on how their music is mixed in this format, leading to results that may not fully align with their artistic vision. At High Tide, we prioritize artist involvement, ensuring they are an integral part of the process. We believe that their input is crucial for creating an immersive experience that truly represents their vision. We will share insights and examples from our collaborations with artists like Amistat, an acoustic folk duo, and Tinush, an electronic music producer known for his attention to detail. These case studies will illustrate how our method fosters creativity and produces superior immersive audio experiences.

New workflows need new tools

A significant pain point in current immersive productions is the tendency to use only a few stems, which often limits the immersive potential. This often happens because the process of exporting individual tracks and preparing a mixing session can be time-consuming and labor-intensive. We will address these challenges in our presentation. We have developed innovative scripts and workflows that streamline this process, allowing us to work with all available tracks without the typical hassle. This approach not only enhances the quality of the final mix but also retains the intricate details and nuances of the original recordings.
Our workshop is designed to be interactive, with opportunities for attendees to ask questions throughout. We will provide real-world insights into our ProTools sessions, giving participants a detailed look at our Dolby Atmos mixing process. By walking through the entire workflow, from recording with Dolby Atmos in mind to the final mix, attendees will gain a comprehensive understanding of the steps involved and the benefits of this approach to create an engaging and immersive listening experience.
Speakers
avatar for Lennart Damann

Lennart Damann

Founder / Engineer, High Tide - Immersive Audio
avatar for Benedikt Ernst

Benedikt Ernst

High Tide - Immersive Audio
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E06

9:00am EDT

Groups, Matrices, and DCAS
Wednesday October 9, 2024 9:00am - 10:00am EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Wednesday October 9, 2024 9:00am - 10:00am EDT
1E05

9:00am EDT

Digital archiving and file management 101
Wednesday October 9, 2024 9:00am - 10:30am EDT
This hands-on tutorial will be an overview of digital archiving and file management concepts and methodologies. It will focus on best practices for file naming, folder structure and DAW session creation, preparation of DAW sessions and associated assets for long-term archiving, an overview of strategies and software solutions for backup and long-term archiving, and an survey of industry standards for digital archiving and deliverables. It will include demonstrations of archiving and backup software and asset preparation.

While this tutorial will be student-focused, anyone who wants to learn more about digital archiving and file management is welcome to attend.
Speakers
avatar for Konrad Strauss

Konrad Strauss

Professor, Indiana University Jacobs School of Music
Konrad Strauss is a Professor of Music in the Department of Audio Engineering and Sound Production at Indiana University’s Jacobs School of Music. He served as department chair and director of Recording Arts from 2001 to 2022. Prior to joining the faculty of IU, he worked as an... Read More →
Wednesday October 9, 2024 9:00am - 10:30am EDT
1E15

9:00am EDT

Applications of Artificial Intelligence and Machine Learning in Audio Quality Models, Part II
Wednesday October 9, 2024 9:00am - 10:30am EDT
The aim of this workshop is to provide participants with hands-on expertise in utilizing machine learning for audio quality modeling.

This workshop is a follow-up from the initial edition presented at the 156th AES Convention. The workshop expands on its first edition by including the participation of new panelists working at cutting-edge technology with multimedia industry leaders. Particularly, very recent advancements on deep learning for auditory perception will be presented, as well as a new open audio quality dataset for informing auditory models in audio quality metrics.

Machine learning techniques have been instrumental in understanding audio quality perception, revealing hidden relationships in subject response data to enhance device and algorithm development. Moreover, accurate quality models can be used in AI systems to predict audio quality -- a critical part of customer experience -- in situations where using human subjects is costly (e.g., in day-to-day product development) or impractical such as in audio transmission network monitoring and informing deep learning audio algorithms.

The complex nature of quality perception requires users and developers of these models to possess specific domain knowledge that extends beyond the general machine learning set of skills. This expertise includes experimental design in the domain of subjective audio quality assessment, data collection, data augmentation and filtering, in addition to model design and cross-validation.

The workshop aims to shed light on historical approaches to addressing these challenges by marrying aspects of machine learning and auditory perception knowledge. Furthermore, it will provide insights into state-of-the-art techniques and current related research topics in data-driven quality perception modelling.

The main goal of the workshop is to offer its attendees —- whether users or developers —- the necessary tools to assess the suitability of existing ML-based quality models for their use case.
Speakers
avatar for Pablo Delgado

Pablo Delgado

Fraunhofer IIS
Pablo Delgado is part of the scientific staff of the Advanced Audio Research Group at the Fraunhofer Institute for Integrated Circuits (IIS) in Erlangen, Germany. He specializes in psychoacoustics applied to audio and speech coding, as well as machine learning applications in audio... Read More →
avatar for Jan Skoglund

Jan Skoglund

Google
Jan Skoglund leads a team at Google in San Francisco, CA, developing speech and audio signal processing components for capture, real-time communication, storage, and rendering. These components have been deployed in Google software products such as Meet and hardware products such... Read More →
avatar for Phill Williams

Phill Williams

Audio Algorithms, Netflix
Phill is member of Netflix's Audio Algorithms team, working on all aspects of the audio delivery toolchain to get the best possible sound to everybody who watches Netflix, all over the world, all of the time.Prior to working at Netflix, Phill worked at Dolby Laboratories, as a contributor... Read More →
avatar for Sascha Dick

Sascha Dick

Sascha Dick received his Dipl.-Ing. degree in Information and Communication Technologies from the Friedrich Alexander University (FAU) of Erlangen-Nuremberg, Germany in 2011 with a thesis on an improved psychoacoustic model for spatial audio coding, and joined the Fraunhofer Institute... Read More →
Wednesday October 9, 2024 9:00am - 10:30am EDT
1E16

9:30am EDT

Acoustic modeling and designing of emergency sound systems in road tunnels
Wednesday October 9, 2024 9:30am - 9:50am EDT
Road tunnels are subject to strong regulations and high life-safety standards. Requirements importantly include the speech intelligibility of the emergency sound system. However, a road tunnel is an acoustically challenging environment due to extreme reverberation, high noise floors, and its tube-like geometry. Designing an adequate sound system is a challenge and requires extensive acoustic simulation. This article summarizes recent design work on several major road tunnel projects and gives practical guidelines for the successful completion of similar projects. The project includes several tunnels, each of several kilometers’ length with one to five lanes, transitions and sections, having a total length of 33 km. For each tunnel, first a working acoustic model had to be developed before the sound system itself could be designed and optimized. On-site measurements were conducted to establish data for background noise including jet fans and various traffic situations. Critical environmental parameters were measured and reverberation times were recorded using large balloon bursts. Sprayed concrete, road surface, as well as other finishes were modeled or estimated based on publicly available data for absorption and scattering characteristics. To establish the geometrical model, each tunnel was subdivided into manageable segments of roughly 1-2 km in length based on theoretical considerations. After calibrating the model, the sound system was designed as a large number of loudspeaker lines evenly distributed along the tunnel. Level and delay alignments as well as filter adjustments were applied to achieve the required average STI of 0.48. Validation by measurement showed good correlation with modeling results.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Authors
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
avatar for Tim Kuschel

Tim Kuschel

Acoustic Consultant, GUZ BOX design + audio
Experienced Acoustic Consultant with a demonstrated history of working in the architecture & planning industry. Skilled in architectural documentation, audio system design, acoustics and extensive experience using AFMG's acoustic modelling software EASE, Tim provides professional... Read More →
Wednesday October 9, 2024 9:30am - 9:50am EDT
1E03

9:50am EDT

Sound immission modeling of open-air sound systems
Wednesday October 9, 2024 9:50am - 10:10am EDT
Noise emission into the neighborhood is often a major concern when designing the configuration of an open-air sound system. In order for events to be approved, advance studies have to show that expected immission levels comply with given regulations and requirements of local authorities. For this purpose, certified engineering offices use dedicated software tools for modeling environmental noise propagation. However, predicting the radiation of modern sound systems is different from classical noise sources, such as trains or industrial plants. Sound systems and their directional patterns are modeled in electro-acoustic simulation software that can be fairly precise but that typically does not address environmental issues.
This paper proposes to use a simple data exchange format that can act as an open interface between sound system modeling tools and noise immission software. It is shown that most immission studies are conducted at points in the far field of the sound system. Far-field directivity data for the sound system is therefore a suitable solution if it is accompanied by a corresponding absolute level reference. The proposed approach has not only the advantage of being accurate for the given application but also involves low computational costs and is fully compliant with the existing framework of outdoor noise modeling standards. Concerns related to documentation and to the protection of proprietary signal processing settings are resolved as well. The proposed approach was validated by measurements at a number of outdoor concerts. Results are shown to be practically accurate within the given limits of uncertainty.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Authors
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Wednesday October 9, 2024 9:50am - 10:10am EDT
1E03

10:00am EDT

MoIP Security: From Quick and Dirty to Rock Solid
Wednesday October 9, 2024 10:00am - 10:30am EDT
Building on previous "101" presentations, this session focuses on advanced techniques for securing Media over IP (MoIP) networks. Attendees will learn to configure switches and media gateways, such as the Prodigy from DirectOut, for a secure setup. Emphasizing proactive problem prevention, the session covers basic to advanced security measures, including network segmentation, firewalls, and access restriction. Real-world case studies and future trends in MoIP security will be discussed. This presentation aims to transform your MoIP network from functional to fortified, ensuring both efficiency and security.
Speakers
avatar for Nicolas Sturmel

Nicolas Sturmel

Directout GmbH
Wednesday October 9, 2024 10:00am - 10:30am EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

10:00am EDT

Student Recording Competition Category 2
Wednesday October 9, 2024 10:00am - 10:45am EDT
Traditional Studio Recording finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
avatar for Andrew Schneider

Andrew Schneider

Recorderer / Mixerer / Producerer / Studio Ownerer, ACRE Audio
Andrew Schneider is a Brooklyn based Producer, Mixer and Engineer. He has a strong passion for working with independent artists to create unique recordings. American, Canadian and Swedish Grammy nominated, Andrew's production work includes recordings for labels Ipecac, Thrilljockey... Read More →
Wednesday October 9, 2024 10:00am - 10:45am EDT
1E07

10:00am EDT

Fania Records at 60
Wednesday October 9, 2024 10:00am - 11:00am EDT
Fania Records is often called the Motown of Latin music. With a who's who amongst salsa royalty, they hold a catalog of some of the greatest talent in Latin Music. From their humble beginnings selling cds out of cars, to Yankee Stadium and then Live in Zaire for the Rumble in the Jungle, Fania reached the top. Now at 60, Fania records is re-mastering much of their catalog. Join Grammy Winning Engineer JC "Mr. Sonic" Losada as he moderates a discussing with one of the re-mastering engineers Kevin Gray to hear more about the tech behind handling the Fania catalog.
Wednesday October 9, 2024 10:00am - 11:00am EDT
Stage

10:00am EDT

Technical Committee Meeting: Automotive Audio
Wednesday October 9, 2024 10:00am - 11:00am EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.

This Meeting will take place online and in-person. To Join online click this link: https://us02web.zoom.us/j/84593160017?pwd=IOfBmz7NYIaPf0ePwFNkSFWtOGbFWA.1
Wednesday October 9, 2024 10:00am - 11:00am EDT
2D02/03

10:00am EDT

Exhibit Hall
Wednesday October 9, 2024 10:00am - 6:00pm EDT
Step into the heart of innovation at the Audio Engineering Society’s Annual Conference Exhibit Hall. This dynamic space brings together leading companies and cutting-edge technologies from across the audio engineering industry. Attendees will have the opportunity to explore the latest advancements in audio equipment, software, and services, engage with industry experts, and discover new solutions to enhance their projects. Whether you’re looking to network with professionals, gain insights from live demonstrations, or simply stay ahead of the curve, the Exhibit Hall is the place to be. Don’t miss this chance to immerse yourself in the future of audio engineering!
Wednesday October 9, 2024 10:00am - 6:00pm EDT
Exhibit Hall

10:10am EDT

Virtual Acoustics Technology in the Recording Studio-A System Update
Wednesday October 9, 2024 10:10am - 10:30am EDT
This paper describes ongoing efforts toward optimizing the Virtual Acoustics Technology (VAT) system installed in the Immersive Media Lab at McGill University. Following the integration of the CAVIAR cancelling auralizer for feedback suppression, this current iteration of the active acoustics system is able to flexibly support the creation of virtual environments via the convolution of Spatial Room Impulse Responses (SRIRs) with real-time microphone signals. While the system has been successfully used for both recordings and live performances, we have nevertheless been looking to improve upon its “stability from feedback” and “natural sound quality,” two significant attributes of active acoustics systems [1]. We have implemented new software controls and microphone input methods to increase our ratio of gain before feedback, while additionally repositioning and adding loudspeakers to the system to generate a more even room coverage. Following these additions, we continue to evaluate the space through objective measurements and feedback from musicians and listeners.
[1] M. A. Poletti, “Active Acoustic Systems for the Control of Room Acoustics,” in Proceedings of the International Symposium on Room Acoustics, Melbourne, Australia, Aug. 2010.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Authors
avatar for Kathleen Ying-Ying Zhang

Kathleen Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Wednesday October 9, 2024 10:10am - 10:30am EDT
1E03

10:15am EDT

The Soundsystem as Survival and Resistance: A Black, Afro-Asian, and Indigenous Audio History
Wednesday October 9, 2024 10:15am - 11:00am EDT
Short Abstract:
This presentation explores Black, Indigenous, Afro-Asian, and Afro-Indigenous cross-cultural collaboration leading to influential innovations in sonic creative technology. Centering on the soundsystem as a form of self-determination and as a contemporary art form rooted in survival and resistance, the presentation aims to illuminate meaningful connections between ancestral and state-of-the-art sound technologies through the lens of critical decolonization theory. Highlighting innovations ranging from parametric EQ, radio, and the sound system as both cultural and analog technologies, the presentation shows how these could only exist through mutual influence and the building of nations and kinship across cultures bound by mutual experiences of oppression. These artistic traditions in music, gathering, and engineering emerged and continue to act as dynamic tools of survivance within Black and globally Indigenous communities. By skillfully using these tools, practitioners of these electronic and sonic art forms destabilize racialized boundaries and empower the collective ownership of cultural narratives—a fearless pursuit reflecting a shared vision of Black and Indigenous self-determination, liberation, and futurity for kin past, present, and future.

Extension:
The presentation’s audiovisual materials will comprise two aspects: archival materials from music and resistance movements in the featured locations and contemporary artworks dealing with identity politics and protest in connection with music. Featured locations will include, but are not limited to: Indian Country, USA & Canada; Kingston, Jamaica; Notting Hill, UK; and Detroit, MI. Timeline of technological innovation will center audio advancements in amplification and transmission within the post-transistor era (1947 - present), with a historical focus primarily from 1947 to 1996. Featured BIPOC artists and innovators in electrical engineering in the sonic arts will include and are not limited to: Hedley Jones, Tom Wong, and Don Lewis. Featured entrepreneurial artists in audio engineering, radio, recording, music and storytelling will include, but are not limited to: Vy Higgenson, Kemistry & Storm, Jeff Mills, K-Hand, and Patricia Chin. Featured Black, Asian and Indigenous owned radio stations and music labels that function as both mutual aid and community spaces will include, but are not limited to: Bvlbancha Liberation Radio, Cool Runnings Music, KTNN, VP Records, and the Sound System Futures Programme.

The objective of this session is to provide a brief overview mapping sound histories to the BIPOC innovators of origin with the aim of sparking generative discussion on Black and Indigenous creative technologies as both a cultural and computational history valuable and critical to the healing of community and dismantling of hegemonic societal structures. The session is structured largely in response to a critical lack of representation regarding BIPOC engineering history and legacy, and approaches this topic as a celebration of cross-cultural collaboration in a historical paradigm that often centers cross-cultural interactions as conflict-oriented. At the presentation’s conclusion participants will be encouraged to reflect on their own cross-cultural legacies as sources of innovation, as well as reflect on the unactivated dreams they may hold from their ancestors and how they might activate these dreams through sound, collective action, and community gathering.
Speakers
Wednesday October 9, 2024 10:15am - 11:00am EDT
1E08

10:15am EDT

KEYQUALIZATION New EQ Technique and KEY-QAlpha Product Demonstration by Roey Shamir
Wednesday October 9, 2024 10:15am - 11:15am EDT
Building on the industry standard that saw the development of Parametric Equalization as the first step in providing the audio engineer with complete control of spectrum modification, this presentation concerns the application of a new update, built on Parametric equalization technique to professional audio control. Invented in 1998 by Roey Shamir, now is a product in its alpha stage of development. The technique and product will be fully demonstrated by the inventor and well known mix producer (Roey Shamir).
The co developer of the hardware Useful Arts Audio / Peter Swann will also discuss matters of the hardware while its being developed. Peter  built his first console in 1982, and has been working ever since to create signal processing devices that don’t just measure well on the bench, but translate sound to compliment the perceptual mechanisms of the human ear and brain. In 2017, he launched Useful Arts Audio, which manufactures specialty front-end gear designed to accentuate musical performances by taking advantage of the natural emphasis the brain places on dynamics and harmonic content. Mr. Swann creates original designs aimed at translating incoming musical signals into compelling listening experiences.

Only the proof of concept Software will be demonstrated the hardware will not be ready for AES NY Convention Fall 2024.

There will be a well known song played in immersive 7.1.4 with examples of the new technique in the elements of the immersive mix.
Speakers
avatar for Peter Swann

Peter Swann

Peter Swann built his first console in 1982, and has been working ever since to create signal processing devices that don’t just measure well on the bench, but translate sound to compliment the perceptual mechanisms of the human ear and brain.  In 2017, he launched Useful Arts... Read More →
avatar for Roey Shamir

Roey Shamir

Mixmaster, INFX PRODUCTIONS INC.
• Freelance Audio Engineer, Mixer and Producer • Audio Engineer Mixer, Producer, Remixer for extensive major label records and filmproductions (Mix to Picture) • Mixed and Recorded music / albums for numerous RIAA multi platinum and Grammynominated... Read More →
Wednesday October 9, 2024 10:15am - 11:15am EDT
1E06

10:20am EDT

ADM-OSC 1.0
Wednesday October 9, 2024 10:20am - 10:40am EDT
Spatial and immersive audio has become increasingly mainstream, presented in concert halls and more recently through music streaming services. There is a diverse ecosystem of hardware and software controllers and renderers in both live and studio settings that would benefit from a standardized communication protocol. In 2019 a group of industry stakeholders began designing ADM-OSC to fill this need. ADM-OSC is a standard for transmitting metadata for object-based audio by implementing a namespace in parallel with the Audio Definition Model (ADM), a metadata standard developed in the broadcast industry. Open Sound Control (OSC) is a well-established data transport protocol developed for flexible and accurate communication of real-time performance data. By leveraging these open standards, we have created a lightweight specification that can be easily implemented in audio software, plugins, game engines, consoles, and controllers.
ADM-OSC has reached a level of maturity over multiple implementations be ready for an official 1.0 release. This paper will discuss the design of ADM-OSC 1.0 and how it was developed to facilitate interoperability for a range of stakeholders and use cases. The core address space for position data is described, as well as extensions for live control data. We conclude with an overview of future ADM-OSC development, including next steps in bringing together ideas and discussion from multiple industry partners.
Moderators Speakers
avatar for Michael Zbyszynski

Michael Zbyszynski

Software Development Engineer, L-Acoustics
Michael Zbyszyński is musician, researcher, teacher and developer in the field of contemporary electroacoustic music. He is currently part of the Creative Technologies R&D group at L-Acoustics. As a musician, his work spans from brass bands to symphony orchestras, including composition... Read More →
Authors
avatar for Michael Zbyszynski

Michael Zbyszynski

Software Development Engineer, L-Acoustics
Michael Zbyszyński is musician, researcher, teacher and developer in the field of contemporary electroacoustic music. He is currently part of the Creative Technologies R&D group at L-Acoustics. As a musician, his work spans from brass bands to symphony orchestras, including composition... Read More →
avatar for Hugo Larin

Hugo Larin

Senior Mgr. Business Development | FLUX:: GPLM, Harman International
Hugo Larin is a key collaborator to the FLUX: SPAT Revolution project and has deep roots in audio mixing, design and operation, as well as in networked control and data distribution. He leads the FLUX:: business development at HARMAN. His recent involvements and interests include object-based spatial audio mixing workflows, interoperability... Read More →
avatar for Lucas Zwicker

Lucas Zwicker

Senior Director, Workflow and Integration, CTO Office, Lawo AG
Lucas joined Lawo in 2014, having previously worked as a freelancer in the live sound and entertainment industry for several years. He holds a degree in event technology and a Bachelor of Engineering in electrical engineering and information technology from the University of Applied... Read More →
avatar for Mathieu Delquignies

Mathieu Delquignies

Education & Application Support France, d&b audiotechnik
Mathieu holds a Bachelors's degree in applied physics from Paris 7 University and Master's degree in sound engineering from ENS Louis Lumière in 2003. He has years of diverse freelance mixing and system designer experiences internationally, as well as loudspeakers, amplifiers, dsp... Read More →
Wednesday October 9, 2024 10:20am - 10:40am EDT
1E04

10:30am EDT

A General Overview of Methods for Generating Room Impulse Responses
Wednesday October 9, 2024 10:30am - 10:50am EDT
The utilization of room impulse responses has proven valuable for both the acoustic assessment of indoor environments and music production. Various techniques have been devised over time to capture these responses. Although algorithmic solutions have been in existence since the 1960s for generating synthetic reverberation in real time, they continue to be computationally demanding and in general lack the accuracy in comparison to measured authentic Room Impulse Responses (RIR). In recent times, machine learning has found application in diverse fields, including acoustics, leading to the development of techniques for generating RIRs. This paper provides a general overview, of approaches and methods for generating RIRs, categorized into algorithmic and machine learning techniques, with a particular emphasis on the latter. Discussion covers the acoustical attributes of rooms relevant to perceptual testing and methodologies for comparing RIRs. An examination of disparities between captured and generated RIRs is included to better delineate the key acoustic properties characterizing a room. The paper is designed to offer a general overview for those interested in RIR generation for music production purposes, with future work considerations also explored.
Moderators
avatar for Ying-Ying Zhang

Ying-Ying Zhang

PhD Candidate, McGill University
YIng-Ying Zhang is a music technology researcher and sound engineer. She is currently a PhD candidate at McGill University in the Sound Recording program where her research focuses on musician-centered virtual acoustic applications in recording environments. She received her Masters... Read More →
Speakers
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
Authors
avatar for Mihai-Vlad Baran

Mihai-Vlad Baran

McGill University
avatar for Richard King

Richard King

Professor, McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
Wednesday October 9, 2024 10:30am - 10:50am EDT
1E03

10:30am EDT

Revolutionizing Media Workflows: A modern approach to audio processing in a dynamic environment
Wednesday October 9, 2024 10:30am - 11:00am EDT
This presentation explores advanced cloud-technology-based audio processing, highlighting platform-agnostic solutions. It emphasizes the integration of SMPTE ST2110, NDI, and SRT for seamless, high-quality media transport. Key topics include the benefits of scalability, real-time processing capabilities, and integration across diverse systems, offering a comprehensive overview of modern, efficient and dynamic workflows in media production and distribution.
Speakers
avatar for Lucas Zwicker

Lucas Zwicker

Senior Director, Workflow and Integration, CTO Office, Lawo AG
Lucas joined Lawo in 2014, having previously worked as a freelancer in the live sound and entertainment industry for several years. He holds a degree in event technology and a Bachelor of Engineering in electrical engineering and information technology from the University of Applied... Read More →
Wednesday October 9, 2024 10:30am - 11:00am EDT
AES and AIMS Alliance Program Room
  Sponsor Session

10:30am EDT

Active Dynamics
Wednesday October 9, 2024 10:30am - 11:30am EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.

The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Wednesday October 9, 2024 10:30am - 11:30am EDT
1E05

10:40am EDT

Analysis of Ultra Wideband Wireless Audio Transceivers for High-Resolution Audio Transmission
Wednesday October 9, 2024 10:40am - 11:00am EDT
Modern Ultra Wide-Band Wireless (UWB) transceiver radio systems enhance digital audio wireless transmission by eliminating the need for audio data compression required by narrowband technologies such as Bluetooth. UWB systems, characterized by their high bandwidth, bypass compression computation delays, enabling audio data transmission from transmitter to receiver in under 10 milliseconds. This paper presents an analytical study of audio signals transmitted using a contemporary UWB transceiver system. The analysis confirms that UWB technology can deliver high-resolution (96kHz/24-bit) audio that is free from artifacts and comparable in quality to a wired digital link. This study underscores the potential of UWB systems to revolutionize wireless audio transmission by maintaining integrity and reducing latency, aligning with the rigorous demands of high-fidelity audio applications.
Moderators Speakers
avatar for Jeff Anderson

Jeff Anderson

SPARK Microsystems
Authors
avatar for Jeff Anderson

Jeff Anderson

SPARK Microsystems
Wednesday October 9, 2024 10:40am - 11:00am EDT
1E04

10:45am EDT

De-Mixing: Should we just because we can?
Wednesday October 9, 2024 10:45am - 11:30am EDT
Multi-Grammy Award winners, Michael Graves (Osiris Studio) and Cheryl Pawelski (Omnivore Recordings) will lead a discussion exploring the intersection of historical/catalog recordings and technology.

We’re at a point in technological developments where tech tools can be used to fix flawed media, restore poor recordings or de-mix to remix sources.

We will share questions we ask ourselves when we consider employing new technology and discuss the responsibility we feel toward the art, the artist and the artists’ intent.

To help aid the discussion, audio samples will be played by:

Chris Bell (I Am The Cosmos)
Hank Williams (The Garden Spot Programs, 1950)
Art Pepper (The Complete Maiden Voyage Recordings)
Janis Joplin & Jorma Kaukonen (The Legendary Typewriter Tape, 1964)
Various Artists (Written In Their Soul: The Stax Songwriter Demos)
Speakers
avatar for Michael Graves

Michael Graves

Mastering Engineer | Audio Restoration Specialist, Osiris Studio
Michael Graves is a multiple GRAMMY-Award winning mastering engineer. During his career, Graves has worked with clients worldwide and has gained a reputation as one of the leading audio restoration and re-mastering specialists working today, earning 5 GRAMMY Awards and 14 nominations... Read More →
avatar for Cheryl Pawelski

Cheryl Pawelski

Co-Founder/Co-Owner, Omnivore Recordings
Cheryl Pawelski is a four-time Grammy Award-winning producer who has for years been entrusted with preserving, curating, and championing some of music’s greatest legacies. Before co-founding Omnivore Entertainment Group, she held positions at Rhino, Concord, and EMI-Capitol Records... Read More →
Wednesday October 9, 2024 10:45am - 11:30am EDT
1E09

10:45am EDT

BBC Pop Hub - Moving Europe's largest radio station
Wednesday October 9, 2024 10:45am - 11:30am EDT
BBC Radio 2 is Europe’s largest radio station. In this session Jamie Laundon, Solution Lead at BBC Technology & Media Operations, will guide us through the move into the new, state-of-the-art "Pop Hub" studios on 8th floor of BBC Broadcasting House in London.

The project was delivered to extremely tight timescales and with no impact on our listeners. These studios have been designed as dynamic, collaborative spaces, with the aim to inspire our teams to produce their best work, and to reflect an energy and buzz that will be transmitted to our audiences.

The studios have been designed to be simple to use yet also feature complex workflows that permit studios to used in pairs and give additional assistance to presenters and musicians as required.

In this session you’ll learn about managing design decisions when working on an existing site, and the use of realtime media networks, AES67 and Dante to join infrastructure together. As well as audio equipment selection, we will also cover furniture design and some interesting challenges around accessibility, sight lines and camera angles. We will also cover the tools used to visualise radio, and create well-lit branded spaces that suit any of the BBC’s pop music networks. We will also talk about the collaboration and skill sharing between our audio, technology and operations teams to help deliver against very tight timelines.
Speakers
Wednesday October 9, 2024 10:45am - 11:30am EDT
1E16

10:45am EDT

How Audio Engineering can assist in Neurodivergent Hearing Loss Prevention
Wednesday October 9, 2024 10:45am - 11:30am EDT
Within live and studio sound engineering, the use of compression, normalization and parametric EQ is employed to perfect the sound of musicians we work with. These effects can impact the tone and timbre of an audio sample or even eliminate unwanted noise. Even the resistance of the electrical equipment used to aid us in our work as audio engineers has an impact on the quality of our end product. Many audio engineers and audio companies have used our understanding of the technology within our industry and applied it to the use of hearing mechanisms such as in-ear monitors or ear plugs.

Around 3-9% of adults struggle with some form of hearing disability or hearing loss. While the science and studies of audiology have helped us find ways to assist those disabled in this way, hearing disabilities tend to be comorbid with neurological or psychological disorders that are harder to manage when combined. PTSD, dementia, autism, as well as several other disorders severely impact the way our brains interact with auditory and kinesthetic stimuli that make current hearing aid technology more of a discomfort than a help to patients with these comorbidities.

After multiple observations of audio engineering techniques and the equipment we use, peer reviewed studying of clientele with these comorbidities, and applications of techniques to hearing sciences, we wish to display some of those applications and report our findings so as to make the world more accessible to those with hearing disabilities, neurological/psychological disorders, and those with both.
Wednesday October 9, 2024 10:45am - 11:30am EDT
1E15

10:45am EDT

Student Recording Competition Category 3
Wednesday October 9, 2024 10:45am - 11:45am EDT
Modern Studio Recording finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
MW

Marek Walaszek

General, Addicted To Music
Wednesday October 9, 2024 10:45am - 11:45am EDT
1E07

11:00am EDT

Reimagining Delay Effects: Integrating Generative AI for Creative Control
Wednesday October 9, 2024 11:00am - 11:20am EDT
This paper presents a novel generative delay effect that utilizes generative AI to create unique variations of a melody with each new echo. Unlike traditional delay effects, where repetitions are identical to the original input, this effect generates variations in pitch and rhythm, enhancing creative possibilities for artists. The significance of this innovation lies in addressing artists' concerns about generative AI potentially replacing their roles. By integrating generative AI into the creative process, artists retain control and collaborate with the technology, rather than being supplanted by it. The paper outlines the processing methodology, which involves training a Long Short-Term Memory (LSTM) neural network on a dataset of publicly available music. The network generates output melodies based on input characteristics, employing a specialized notation language for music. Additionally, the implementation of this machine learning model within a delay plugin's architecture is discussed, focusing on parameters such as buffer length and tail length. The integration of the model into the broader plugin framework highlights the practical aspects of utilizing generative AI in audio effects. The paper also explores the feasibility of deploying this technology on microcontrollers for use in instruments and effects pedals. By leveraging low-power AI libraries, this advanced functionality can be achieved with minimal storage requirements, demonstrating the efficiency and versatility of the approach. Finally, a demonstration of an early version of the generative delay effect will be presented.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers Authors
Wednesday October 9, 2024 11:00am - 11:20am EDT
1E03

11:00am EDT

Migrating to compute for live media production processing – the impact on audio
Wednesday October 9, 2024 11:00am - 11:30am EDT
The industry is on an sure and steady move to using compute for the complete chain of live media production processing. This evolution from dedicated hardware products provides significant flexibility yet comes with challenges relating to the precision of behavior needed to process broadcast audio and video. This presentation will outline the techniques being used for the transfer and processing of each media flow and examine the optimizations needed to ensure evolving systems are both feature-capable and fully usable.
Speakers
Wednesday October 9, 2024 11:00am - 11:30am EDT
AES and AIMS Alliance Program Room
  Sponsor Session

11:00am EDT

Ulrike & Jim: Enveloping Masterclass
Wednesday October 9, 2024 11:00am - 12:00pm EDT
Ulrike Schwarz and Jim Anderson discuss recordings from their extraordinary catalogues, including high resolution listening examples in formats 5.1, 9.1 and 7.1.4.

This masterclass series, featuring remarkable recording artists, is a chance to hear surround and 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Ulrike Schwarz

Ulrike Schwarz

Engineer/Producer, Co-Founder, Anderson Audio New York
Engineer/Producer, Co-Founder
avatar for Jim Anderson

Jim Anderson

Producer/Engineer, Anderson Audio New York
Producer/Engineer
Wednesday October 9, 2024 11:00am - 12:00pm EDT
3D06

11:00am EDT

Technical Committee Meeting: Acoustics and Sound Reinforcement
Wednesday October 9, 2024 11:00am - 12:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.
Wednesday October 9, 2024 11:00am - 12:00pm EDT
2D02/03

11:15am EDT

Immersive Voice and Audio Services (IVAS) Codec – A deeper look into the new 3GPP Standard for Immersive Communication
Wednesday October 9, 2024 11:15am - 12:45pm EDT
3GPP, the worldwide partnership project among seven major regional telecommunications standard development organizations responsible for mobile communication standards, has recently completed the standardization of a new codec intended for Immersive Voice and Audio Services (IVAS) in 5G systems. The new IVAS codec is a feature of 3GPP Release 18 (5G-Advanced) and enables completely new service scenarios by providing capabilities for interactive stereo and immersive audio communications, content sharing and distribution. The addressed service scenarios include conversational voice with stereo and immersive telephony/conferencing, XR (VR/AR/MR) communications, i.e., XR telephony and conferencing, and live and non-live streaming of user-generated immersive and XR content.
The IVAS Codec is an extension of the 3GPP EVS codec. Beyond support for mono voice and audio content, IVAS also brings support for coding and rendering of stereo, multichannel-based audio, scene-based audio (Ambisonics), object-based audio and metadata-assisted spatial audio – a new parametric audio format designed for direct spatial audio pick-up from smartphones. The performance of the codec was evaluated in the course of the 3GPP standardization in 23 experiments carried out by two independent laboratories each.
The workshop will introduce the IVAS Codec starting with a brief overview of standardization process and state of the art. The codec architecture, supported audio formats, and main technology blocks behind the supported coding and rendering options will be presented and discussed in detail together with test results and capabilities of the new codec in terms of compression efficiency and audio quality. In addition, the panelists will discuss new service scenarios for immersive audio communication and other mobile use-cases. If possible, immersive audio demonstrations will be carried out during the workshop. The panelists will collect and address questions and other input from the audience.
Speakers
avatar for Adriana Vasilache

Adriana Vasilache

Nokia Technologies
AF

Andrea Felice Genovese

Andrea Genovese is a Senior Research Engineer at Qualcomm
Wednesday October 9, 2024 11:15am - 12:45pm EDT
1E08

11:20am EDT

Rediscovering Xenakis: Decoding the Free Stochastic Music Program's Output File
Wednesday October 9, 2024 11:20am - 11:40am EDT
Composer Iannis Xenakis (1922-2001), in his book Formalized Music (1963), includes a piece of Fortran IV computer code which produces a composition of stochastically generated music. This early example of indeterminate music made by a computer was in fulfillment of his goal of creating a musical composition of minimum constraints by letting the computer stochastically generate parameters. Stochasticity is used on all levels of the composition, macro (note density and length of movements) and micro (length of notes and pitches). This paper carefully analyzes the output composition and the variations the program might produce. Efforts are then made to convert these compositions into MIDI format in order to promote their preservation and increase their accessibility. The preservation of the Free Stochastic Music Program is beneficial to understand one of the first examples of indeterminate computer music, with similar techniques being found in modern day music creation software.
Moderators Speakers Authors
Wednesday October 9, 2024 11:20am - 11:40am EDT
1E04

11:20am EDT

Acoustic Characteristics of Parasaurolophus Crest: Experimental Results from a simplified anatomical model
Wednesday October 9, 2024 11:20am - 11:40am EDT
This study presents a revised acoustic model of the Parasaurolophus crest, incorporating both the main airway and lateral diverticulum, based on previous anatomical models and recent findings. A physical device, as a simplified model of the crest, was constructed using a coupled piping system, and frequency sweeps were conducted to investigate its resonance behavior. Data were collected using a minimally invasive microphone, with a control group consisting of a simple open pipe for comparison. The results show that the frequency response of the experimental model aligns with that of the control pipe at many frequencies, but notable shifts and peak-splitting behavior were observed, suggesting a more active role of the lateral diverticulum in shaping the acoustic response than previously thought. These findings challenge earlier closed-pipe approaches, indicating that complex interactions between the main airway and lateral diverticulum generate additional resonant frequencies absent in the control pipe. The study provides empirical data that offer new insights into the resonance characteristics of the Parasaurolophus crest and contribute to understanding its auditory range, particularly for low-frequency sounds.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers Authors
Wednesday October 9, 2024 11:20am - 11:40am EDT
1E03

11:30am EDT

Good Enough Is Not Enough
Wednesday October 9, 2024 11:30am - 12:00pm EDT
For media or content delivery, “good enough” is not enough anymore. Just because signal distribution is over the internet doesn’t mean audiences will tolerate jitter or latency -- and they shouldn’t have to. New IP protocols, including SRT, NDI and the newest IPMX, are melding broadcast workflows with “traditional” AV technology. Live streaming has matured into a reliable and secure AV transport method, and it’s now possible to cost-effectively produce 1080P or 4K HDR content from a conference room camera, and then mix and integrate that with a broadcast-quality stream. Also to be covered: why IPMX will emerge as the new AVoIP standard.
Speakers
KH

Kevin Henneman

KMH Integration
Wednesday October 9, 2024 11:30am - 12:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

11:30am EDT

New Trends in Fully Virtual Design and Prototyping of Speaker Drivers
Wednesday October 9, 2024 11:30am - 12:30pm EDT
The audio industry is increasingly focused on fully virtual prototyping of devices, designs, and measurement setups. New modeling trends include performing detailed nonlinear distortion analysis, incorporating frequency-dependent properties in time-domain studies, creating lumped components for real-time system analysis, and virtually developing and testing innovative materials. In this talk, we will explore two examples to discuss the emerging modeling techniques and strategies that enable these advanced analyses and address challenges in speaker driver design.
Speakers
JH

Jinlan Huang

Applications Engineer, COMSOL, Inc.
Wednesday October 9, 2024 11:30am - 12:30pm EDT
Stage
  Sponsor Session

11:40am EDT

Unveiling the Impact of K-Pop Song Lyrics on Adolescent Self-Harm Rates: A Pilot Study of Gender Disparities
Wednesday October 9, 2024 11:40am - 12:00pm EDT
This study investigates the phenomenon of increased emergency room visits by teenagers following the broadcast of ‘High School Rapper 2’, which features self-harm in its lyrics. The rate of emergency room visits for self-harm among adolescents aged 10 to 24 notably increased during and shortly after the program. Specifically, in the 10-14 age group, visits per 100,000 population tripled from 0.9 to 3.1. For those aged 15-19, the rate rose from 5.7 to 10.8, and in the 20-24 age group, it increased from 7.3 to 11.0. This study aims to clarify the relationship between lyrics and self-harm rates among adolescents. We analyzed the lyrics of the top 20 songs performed on Melon, a popular music streaming platform for teenagers, from April to October 2018. Using Python's KoNLPy and Kkma libraries for tokenization, part-of-speech tagging, and stop-word filtering, we identified the top 50 frequently appearing words and narrowed them down to the top 20 contextually significant words. A correlation analysis (Pearson R) with "Emergency Department Visits due to Self-Harm" data revealed that the words 'sway', 'think', 'freedom' and 'today' had a statistically significant correlation (p < 0.05). Additionally, we found that males’ self-harm tendency was less influenced by the broadcast compared to females. This research suggests a computational approach to understanding the influence of music lyrics on adolescent self-harm behavior. This pilot study demonstrated correlations between specific words in K-pop lyrics and increased adolescent self-harm rates, with notable gender differences.
Moderators Speakers Authors
avatar for Sungjoon Kim

Sungjoon Kim

Research Intern, Korea Advanced Institute of Science and Technology
Wednesday October 9, 2024 11:40am - 12:00pm EDT
1E04

11:40am EDT

Interpreting user-generated audio from war zones
Wednesday October 9, 2024 11:40am - 12:00pm EDT
Increasingly, civilian inhabitants and combatants in conflict areas use their mobile phones to record video and audio of armed attacks. These user-generated recordings (UGRs) often provide the only source of immediate information about armed conflicts because access by professional journalists is highly restricted. Audio forensic analysis of these UGRs can help document the circumstances and aftermath of war zone incidents, but consumer off-the-shelf recording devices are not designed for battlefield circumstances and sound levels, nor do the battlefield circumstances provide clear, noise-free audio. Moreover, as with any user-generated material that generally does not have a documented chain-of-custody, there are forensic concerns about authenticity, misinformation, and propaganda that must be considered. In this paper we present several case studies of UGRs from armed conflict areas and describe several methods to assess the quality and integrity of the recorded audio. We also include several recommendations for amateurs who make UGRs so that the recorded material is more easily authenticated and corroborated. Audio and video examples are presented.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Authors
avatar for Rob Maher

Rob Maher

Professor, Montana State University
Audio digital signal processing, audio forensics, music analysis and synthesis.
Wednesday October 9, 2024 11:40am - 12:00pm EDT
1E03

11:45am EDT

Be a Leader at AES
Wednesday October 9, 2024 11:45am - 12:45pm EDT
Are you interested in a leadership position within AES? Whether it's a role in your local section as a chair or secretary, a committee, or our Board of Governors, join our panel of experienced AES leaders to find out what is involved and whether a role in leadership is right for you.







Speakers
avatar for David Bowles

David Bowles

Owner, Swineshead Productions, LLC
David v.R Bowles formed Swineshead Productions, LLC as a classical recording production company in 1995. His recordings have been GRAMMY- and JUNO-nominated and critically acclaimed worldwide. His releases in 3D Dolby Atmos can be found on Avie, OutHere Music (Delos) and Navona labels.Mr... Read More →
avatar for Leslie Gaston-Bird

Leslie Gaston-Bird

Owner, Mix Messiah Productions
Leslie Gaston-Bird (AMPS, MPSE) is author of the book "Women in Audio", part of the AES Presents series and published by Focal Press (Routledge). She is a voting member of the Recording Academy (The Grammys®) and its P&E (Producers and Engineers) Wing. Currently, she is a freelance... Read More →
avatar for Marcela Zorro

Marcela Zorro

Universidad de los Andes
Marcela is an audio engineer specializing in the recording and production of classical music. She currently teaches in the audio production emphasis of the Music Department at Universidad de los Andes in Bogotá, Colombia. Marcela has been an active member of the AES since 2004.  She... Read More →
avatar for Gary Gottlieb

Gary Gottlieb

AES President-Elect, Mendocino College
President-Elect, Co-Chair of the Events Coordination Committee, Chair of the Conference Policy Committee, and former Vice President of the Eastern Region, US and Canada; AES Fellow, Engineer, Author, Educator and Guest Speaker Gary Gottlieb refers to himself as a music generalist... Read More →
avatar for Angela Piva

Angela Piva

Angela Piva, Audio Pro/Audio Professor, highly skilled in all aspects of music & audio production, recording, mixing and mastering with over 35 years of professional audio engineering experience and accolades. Known as an innovator in sound technology, and for contributing to the... Read More →
Wednesday October 9, 2024 11:45am - 12:45pm EDT
1E15

11:45am EDT

Navigating the Legal Landscape of AI in Music and Tech
Wednesday October 9, 2024 11:45am - 12:45pm EDT
Artificial intelligence (AI) and machine learning (ML) is revolutionizing the music industry, from creation to distribution to consumption. These technologies, and the rapid growth, raise complex legal issues that parties should be aware of to ethically foster innovation while protecting legal rights. This panel will explore the legal challenges and implications of AI in music and music tech, providing insights from industry experts, legal professionals, and technologists, including discussion of topics on intellectual property rights, data privacy, the regulatory landscape, and how to navigate the complex field. Join us for an insightful panel that delves into the intersection of AI, music, and law, offering practical guidance and perspectives on protecting rights, legally and ethically sourcing datasets, and navigating this rapidly evolving landscape.

Topics:
1. Intellectual Property Rights: Understanding copyright issues related to AI-generated music. Was the ML model dataset trained on copyrighted works? Did the company clear the rights to train the model on the material used? Who owns the rights to a song composed by an AI? To what extent, if any? What are the implications for sampling and remixing AI-generated content? How AI impacts the traditional models of licensing and royalties. What changes are necessary to ensure fair compensation for creators?
2. Data Privacy and Security: Addressing concerns about data used in training AI models for music tech. What are the best practices for protecting artists' and users' data? Is data collected and stored properly? Are proper company policies in place?
3. Regulatory Landscape: Reviewing current laws and proposed regulations affecting AI in the music industry. How are different jurisdictions approaching these issues, and what can we expect in the future? Discussing the EU AI Act and other regulations on the horizon.
4. Ethical and Legal Accountability: Exploring the responsibilities of AI developers and users in the music industry. How can we ensure ethical use of AI technologies in music creation and distribution?
Speakers
avatar for Heather Rafter

Heather Rafter

Co-founder & Principal, RafterMarsh
Heather Dembert Rafter has been providing legal and business development services to the audio, music technology, and digital media industries for over twenty-five years. As principal counsel at RafterMarsh US, she leads the RM team in providing sophisticated corporate and IP advice... Read More →
DG

Denis Geokdag

CEO, Zynaptiq GmbH
avatar for Philipp Lengeling

Philipp Lengeling

Senior Counsel, RafterMarsh
Philipp G. Lengeling, Mag. iur., LL.M. (New York), Esq. is an attorney based in New York (U.S.A.) and Hamburg (Germany), who is heading the New York and Hamburg based offices for RafterMarsh, a transatlantic boutique law firm (California, New York, U.K., Germany).
avatar for Zack Klein

Zack Klein

Associate Counsel, RafterMarsh
Zack is an attorney licensed in New York and New Jersey, and is an associate with RafterMarsh US with focuses in the fields of tech, music law, intellectual property, licensing, as well as work in Web3. Zack has degrees in music technology and electrical engineering and his background... Read More →
Wednesday October 9, 2024 11:45am - 12:45pm EDT
1E09

11:45am EDT

Fix The Mix: Building a Thriving Career Behind the Board
Wednesday October 9, 2024 11:45am - 12:45pm EDT
Presented by We Are Moving The Needle, this panel brings together industry leaders from A&R, management, and artist relations to discuss building an impactful career behind the board. These experienced leaders will share the inside scoop on strategies to build your profile as a producer and engineer from navigating rates to defining your brand to attracting a manager. Whether you're an aspiring engineer or a seasoned pro, this conversation offers valuable insights, practical tips, and a chance to connect with a vibrant community of music industry leaders.
Speakers
LW

Lysee Webb

Founder, Van Pelt Management
avatar for Lee Dannay

Lee Dannay

VP of A&R, Thirty Tigers
Lee Dannay's 20 plus year career includes VP/A&R positions at Epic and Columbia Records, VP A&R in music publishing at Warner/Chappell Music, and Senior Music Producer at America’s Got Talent. She has signed and developed artists ranging from John Mayer, Shawn Mullins, Five For... Read More →
avatar for Samantha Sichel

Samantha Sichel

Head of Social Product & Digital Innovation, Live Nation
Live Nation Entertainment; Senior Manager of Marketing Solutions/Digital Business Development Samantha Sichel leads Digital Marketing Solutions for Live Nation Entertainment. This includes working with the Live Nation Sales and Account Management teams to conceptualize and budget... Read More →
Wednesday October 9, 2024 11:45am - 12:45pm EDT
1E16

12:00pm EDT

Experimental analysis of a car loudspeaker model based on imposed vibration velocity: effect of membrane discretization
Wednesday October 9, 2024 12:00pm - 12:20pm EDT
Nowadays, the research about the improvement of the interior sound quality of road vehicles is a relevant task. The cabin is an acoustically challenging environment due to the complex geometry, the different acoustic properties of the materials of cabin components and the presence of audio systems based on multiple loudspeaker units. This paper aims at presenting a simplified modelling approach designed to introduce the boundary condition imposed by a loudspeaker to the cabin system in the context of virtual acoustic analysis. The proposed model is discussed and compared with experimental measurements obtained from a test-case loudspeaker.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Speakers Authors
Wednesday October 9, 2024 12:00pm - 12:20pm EDT
1E03

12:00pm EDT

How do AES67 and SMPTE ST 2110 relate to IPMX
Wednesday October 9, 2024 12:00pm - 12:30pm EDT
Originally, the definition of IPMX started from the proven grounds of AES67 and SMPTE ST 2110. While AES67 and ST 2110 were build around the interoperability requirements of the broadcast realm, it became immediately apparent that for application to the world of ProAV enhanced functionality would be required, while other speciciations indispensible to the broadcast world would be too complex or rigid for ProAV application - namely the very strict timing and synchronization requirements only made possible by the use of PTP. During development of the IPMX specifications within the VSF group it became clear that certain constraints and enhancements have to be made with respect to AES67 and ST 2110 to enable wider adoption in the ProAV domain. IPMX is now taking shape with several companies having announced product availability. While the larger focus of IPMX interoperability is on the video side of things, audio is certainly a very important aspect. And with over 5000 AES67- or ST 2110-compatible audio devices in the market, it would certainly be a massive adoption accelerator if these products could interoperate in an IPMX environment. In this presentation Andreas will discuss how the addditional constraints and enhancements of the IPMX specifications affect interoperability with AES67 and ST 2110 and how these devices can be used in IPMX environments.
Speakers
Wednesday October 9, 2024 12:00pm - 12:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

12:00pm EDT

Student Recording Competition Category 4
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
Sound for Visual Media finalists will make presentations, we will hear their submissions, and the judges will give their comments.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Kia Eshghi

Kia Eshghi

Lecturer of Music Recording Technology, LaGuardia Community College
Director - AES Student Recording CompetitionVice Chair - AES Education CommitteeLecturer of Music Recording Technology at CUNY LaGuardia Community College in the Long Island City, Queens neighborhood of New York City.
avatar for Mary Mazurek

Mary Mazurek

Audio Educator/ Recording Engineer, University of Lethbridge
Audio Educator/ Recording Engineer
avatar for Jamie Baker

Jamie Baker

Professor of Sound Design, Hoboken Post LLC
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
1E06

12:00pm EDT

Implementing WHO Hearing Health Guidelines in Small Venues & Houses of Worship
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
This workshop falls under the category of an AES Tutorial. This workshop outlines the tested method implemented in a small house of worship, and aims to demystify WHO standards as well as communicate an effective solution to implement hearing health monitoring.

Implementing affordable and efficient hearing health monitoring in small music venues is crucial to align with the World Health Organization (WHO) guidelines aimed at preventing hearing damage due to prolonged exposure to high sound levels, which have become much more standardized in recent years. This tutorial will address the unique challenges faced by small music venues, especially houses of worship, which often operate under tighter budgets and where user-friendly permanent solutions are key to long term success. The WHO guidelines emphasize regular monitoring of sound levels, provision of adequate hearing protection, and educating both patrons and staff about the risks of hearing loss and the measures that can be taken to mitigate these risks. This workshop will provide a detailed, practical, and cost-effective roadmap for small venues to implement these recommendations without compromising on the quality of the musical experience. We will walk through interpreting the WHO and AES hearing health standards for live venues, and making sense of large technical documents to communicate to volunteers and newer employees, as well as outline a real world tested solution to implementing hearing health monitoring. Another area of focus will be a brief explanation of how human hearing works, and how it can be damaged. The tutorial will demonstrate how to collect and interpret data, and practical everyman solutions to adapt to hearing health challenges as they arise. Another critical aspect of the tutorial will be on hearing protection strategies. Small music venues can provide disposable earplugs to patrons and staff at a minimal cost or invest in reusable earplugs that offer better sound quality while protecting hearing. The tutorial will discuss the different types of earplugs available, their cost, and how to encourage their use among patrons without detracting from the concert experience. Education is a cornerstone of the WHO guidelines, and this tutorial will explore effective ways to raise awareness about hearing health. This includes training staff to recognize the signs of excessive sound exposure and understand the importance of hearing protection, as well as developing clear and engaging signage and informational materials for patrons. By the end of the workshop, attendees will have a comprehensive understanding of how to implement a hearing health monitoring system in their venues that is both practical and affordable, aligned with the WHO guidelines, and tailored to the unique needs of small music venues. They will leave equipped with actionable strategies, practical tools, and a network of peers committed to promoting hearing health in the music community. This tutorial not only aims to protect the hearing of patrons and staff but also to enhance the overall quality of the live music experience by fostering a safer and more aware listening environment.
Speakers
TS

Tommy Spurgeon

Physics Student & Undergraduate Researcher, University of South Carolina
Wednesday October 9, 2024 12:00pm - 12:45pm EDT
1E07

12:00pm EDT

Technical Committee Meeting: Coding of Audio Signals
Wednesday October 9, 2024 12:00pm - 1:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.

This Meeting will take place online and in-person. To Join online click this link: https://us02web.zoom.us/j/82216397103?pwd=L0VhaUlBK3JjcTJwai9jdjV6VUh2Zz09
Wednesday October 9, 2024 12:00pm - 1:00pm EDT
2D02/03

12:20pm EDT

A novel derivative-based approach for the automatic detection of time-reversed audio in the MPAI/IEEE-CAE ARP international standard
Wednesday October 9, 2024 12:20pm - 12:50pm EDT
The Moving Picture, Audio and Data Coding by Artificial Intelligence (MPAI) Context-based Audio Enhancement (CAE) Audio Recording Preservation (ARP) standard provides the technical specifications for a comprehensive framework for digitizing and preserving analog audio, specifically focusing on documents recorded on open-reel tapes. This paper introduces a novel, envelope derivative-based method incorporated within the ARP standard to detect reverse audio sections during the digitization process. The primary objective of this method is to automatically identify segments of audio recorded in reverse. Leveraging advanced derivative-based signal processing algorithms, the system enhances its capability to detect and reverse such sections, thereby reducing errors during analog-to-digital (A/D) conversion. This feature not only aids in identifying and correcting digitization errors but also improves the efficiency of large-scale audio document digitization projects. The system's performance has been evaluated using a diverse dataset encompassing various musical genres and digitized tapes, demonstrating its effectiveness across different types of audio content.
Moderators
avatar for Marina Bosi

Marina Bosi

Stanford University
Marina Bosi,  AES Past President, is a founding Director of the Moving Picture, Audio, and Data Coding by Artificial Intelligence (MPAI) and the Chair of the Context-based Audio Enhancement (MPAI-CAE) Development Group and IEEE SA CAE WG.  Dr. Bosi has served the Society as President... Read More →
Authors
Wednesday October 9, 2024 12:20pm - 12:50pm EDT
1E03

12:45pm EDT

Grammy Producers & Engineers' Wing: Songs That Changed My Life Vol. II
Wednesday October 9, 2024 12:45pm - 1:45pm EDT
Some songs are hits, some we just love, and some have changed our lives. Our panelists break down the DNA of their favorite tracks and explain what moved them, what grabbed them, and why these songs left a life-long impression. Back by popular demand, this presentation is guaranteed to make you feel good about being in the recording business.
Speakers
avatar for Maureen Droney

Maureen Droney

Vice President, The Recording Academy Producers & Engineering Wing
Maureen Droney is Vice President of the Recording Academy (GRAMMY Awards) Producers & Engineers Wing. A former recording engineer herself, Maureen has worked on GRAMMY-winning recordings for artists including Aretha Franklin, Whitney Houston and Santana—as well as on numerous other... Read More →
Wednesday October 9, 2024 12:45pm - 1:45pm EDT
Stage

1:00pm EDT

AES67 and ST 2110-30 Connection Management in AES70
Wednesday October 9, 2024 1:00pm - 1:30pm EDT
This presentation will describe the features of AES70-21, the recently published AES70 standard for management and monitoring of AES67 and ST 2110-30 audio stream connections.
Speakers
avatar for Jeff Berryman

Jeff Berryman

Senior Scientist, Bosch Communications
Jeff Berryman is Senior Scientist at Bosch Communications Systems.  Over the last 50 years, he has had a dual career in audio and information technology.  His audio experience is in professional loudspeaker and sound reinforcement systems design, focusing on concert touring and... Read More →
Wednesday October 9, 2024 1:00pm - 1:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

1:00pm EDT

Book Signing: Women in Audio by Leslie Gaston-Bird
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
Join us at the AES membership booth for an exclusive book signing event with AES President, Dr Leslie Gaston-Bird, the author of Women in Audio. This book features almost 100 profiles and stories of audio engineers who are women and have achieved success throughout the history of the trade. Beginning with a historical view, the book covers the achievements of women in various audio professions and then focuses on organizations that support and train women and girls in the industry. What follows are eight chapters divided by discipline, highlighting accomplished women in various audio fields: radio; sound for film and television; music recording and electronic music; hardware and software design; acoustics; live sound and sound for theater; education; audio for games, virtual reality, augmented reality, and mixed reality, as well as immersive sound.

Women in Audio - 1st Edition - Leslie Gaston-Bird - Routledge Book
Speakers
avatar for Leslie Gaston-Bird

Leslie Gaston-Bird

Owner, Mix Messiah Productions
Leslie Gaston-Bird (AMPS, MPSE) is author of the book "Women in Audio", part of the AES Presents series and published by Focal Press (Routledge). She is a voting member of the Recording Academy (The Grammys®) and its P&E (Producers and Engineers) Wing. Currently, she is a freelance... Read More →
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
AES Membership Booth

1:00pm EDT

Best Practices for Creating, Mixing, and Producing in Immersive Audio for Music, Film, TV, Podcasting and More
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
Join Focusrite Group’s Lee Whitmore in an interactive panel conversation and listening session with 19-time GRAMMY and Latin GRAMMY winner Rafa Sardina, Dolby’s Tom McAndrew, and ADAM Audio’s Mark Cummins. Listen to Dolby Atmos Music mixes by Rafa Sardina, and chat with industry experts about best practices for creating, mixing, and playback of immersive audio for music, film, TV, gaming, audiobooks and podcasting, as well as considerations for your studio and workflows.

Also held on 10/10 at 12:45pm
Speakers
MC

Mark Cummins

Head of Product Marketing, ADAM Audio
avatar for Tom McAndrew

Tom McAndrew

Senior Technical Manager – Content Relations, Dolby Laboratories
avatar for Rafa Sardina

Rafa Sardina

Rafa Sardina | Afterhours Studios / Abbey Road Institute Los Angeles
Los Angeles based 19-Time Grammy® & Latin Grammy® Winner Producer/Mixer/Engineer Rafa Sardina has built an eclectic and impressive client list over the years: Stevie Wonder, Camila Cabello, D’Angelo, Dr. Dre, Rosalía, John Legend, Elvis Costello & The Roots, Lady Gaga, Michael... Read More →
avatar for Dr. Lee Whitmore

Dr. Lee Whitmore

VP, Education and B2B, Americas, Focusrite Group
Dr. Lee Whitmore is a thought leader in music, audio, creative digital media, and education. He’s the Vice President for business-to-business (B2B), including education, pro, post, and government, at Focusrite Group, which includes FocusriteNovationSequential and Oberh... Read More →
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
1E06

1:00pm EDT

Technical Committee Meeting: Spatial Audio
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.

This Meeting will take place online and in-person. To Join online click this link: https://us02web.zoom.us/j/81754581433?pwd=hET9ewp9peRTCVsLGGAmKSuirkNh6J.1
Wednesday October 9, 2024 1:00pm - 2:00pm EDT
2D02/03

1:00pm EDT

dLive Certification
Wednesday October 9, 2024 1:00pm - 4:00pm EDT
In addition to long-form dLive Certification classes, training sessions will cover a variety of topics relevant to live sound engineers, including: Mixing Monitors from FOH; Vocal Processing; Groups, Matrices and DCAs; Active Dynamics; and Gain Staging.
 
The training will be led by industry veterans Michael Bangs and Jake Hartsfield. Bangs, whose career includes experience as a monitor engineer and production manager, has worked with A-list artists, including Aerosmith, Katy Perry, Tom Petty, Lynyrd Skynyrd and Kid Rock. Hartsfield is a seasoned live sound engineer, having mixed for artists like Vulfpeck, Ben Rector, Fearless Flyers, and more.

Signup link: https://zfrmz.com/DmSlX5gyZCfjrJUHa6bV
Wednesday October 9, 2024 1:00pm - 4:00pm EDT
1E05

1:30pm EDT

Catena Control Plane Standardization Update
Wednesday October 9, 2024 1:30pm - 2:00pm EDT
We will provide an update on progress within SMPTE on the Catena initiative. The goal of this initiative is to create a standardized, open source control plane for cloud, on-prem and hybrid environments.
Speakers
CL

Chris Lennon

Ross Video
Wednesday October 9, 2024 1:30pm - 2:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

2:00pm EDT

Auditory Envelopment and Affective Touch Hypothesis
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
Anticipation and pleasure in response to music listening can lead to dopamine release in the human striatal system, a neural midbrain reward and motivation cluster. The sensation of auditory envelopment, however, may also in itself have a stimulating effect. Theoretical reason and circumstantial evidence are given why this could be the case, thereby possibly constituting an auditory complement to a newly discovered and studied percept, affective touch, originating from C-tactile fibres in the skin, stimulated certain ways. In a pilot test, abstract sounds were used to determine the audibility of low frequency inter-aural fluctuation. Naïve subjects aged between 6 and 96 years were all sensitive to the conditions tested and asked to characterize the stimuli in their own words. Based on these results, controlling low frequency inter-aural fluctuation in listeners should be a priority when recording, mixing, distributing and reproducing audio.
Moderators Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
Authors
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
1E03

2:00pm EDT

Convergence - The vision of one AV network
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
The landscape of Media Networks has become fragmented with standards and solutions which are partly incompatible. This often forces to implement multiple parallel networks into ProAV systems which causes huge cost and efforts and leaves the customer with very inflexible systems. In contrast, the promise and vision of media networking has been the opposite: To bring data and media traffic into one network structure with opportunities for flexible connections and interoperability. The session shows the dimension of the problem and possible paths for solutions. The session will explain the idea of ‘convergence’ with examples from actual productions.
Speakers
avatar for Henning Kaltheuner

Henning Kaltheuner

Owner, Independent consultant
Independent consultant for the ProAV industry. After > 25y of practical Live Sound Engineering, Product Management and Product Design, my main experience and focus of my consultancy is Strategic Consulting and Market Research. I am expert in customer motivations and user needs, brand... Read More →
Wednesday October 9, 2024 2:00pm - 2:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

2:00pm EDT

Acoustical Simulation and Calculation Techniques for Small Spaces
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
This presentation provides an insightful overview of statistical acoustic calculation techniques and geometrical acoustic modeling. It addresses the unique challenges associated with small to medium-sized spaces and delves into the intricacies of contemporary ray tracing software tools. Attendees will also learn about the latest advancements of using these tools in immersive audio environments. Additionally, the presentation will showcase comparative low-frequency analysis calculations obtained using a variety of tools, including boundary element methods (BEM), finite element
methods (FEM) and others. The session includes a series of real-world case studies to illustrate these concepts in practical applications.
Speakers
avatar for Peter D'Antonio

Peter D'Antonio

Director of Research, REDIacoustics
Dr. Peter D’Antonio is a pioneering sound diffusion expert. He received his B.S. from St. John’s University in 1963 and his Ph.D. from the Polytechnic Institute of Brooklyn, in 1967. During his scientific career as a diffraction physicist, Dr. D’Antonio was a member of the Laboratory... Read More →
avatar for Dirk Noy

Dirk Noy

Managing Director Europe, WSDG
BioDirk Noy has a Master of Science (MSc) Diploma in Experimental Solid State Physics from the University of Basel, Switzerland and is a graduate from Full Sail University, Orlando, USA, where he was one of John Storyk’s students.Since joining the consulting firm Walters-Storyk... Read More →
avatar for Stefan Feistel

Stefan Feistel

Managing Director/Partner/Co-Founder, AFMG
Stefan Feistel is Managing Director/Partner/Co-Founder of AFMG, Berlin Germany. He is an expert in acoustical simulation and calculation techniques and applications.
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
1E08

2:00pm EDT

You're a Gear Geek - Just Like Us! Staging Gear in Trade Shows, Print Media and Online
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
Drawing on research to published in our forthcoming book, Gear: Cultures of Audio and Music Technologies (The MIT Press), this workshop focuses on gear in trade show, print media, and online fora. Samantha Bennett and Eliot Bates first introduce 'gear', hardware professional audio recording technologies, and 'gear cultures', those being 'milieux where sociability is centred around audio technology objects’ (Bates and Bennett, 2022). With special focus on Part III of our book, this workshop examines the myriad ways we see gear 'staged' - in events including AES and NAMM, in diverse print media, including the Sweetwater catalogue and Sound on Sound, and online in fora including Gearspace.

So much of the fetishistic and technostalgic qualities associated with gear today exceed the materials and design; and it is precisely these fetish ideologies that become central to gear sociability. So how does gear become social, and where do gear cultures gather? Since the mid 1990s, we have seen technological objects transformed into gear when staged within these trade show / media / online milieux. In these spaces, gear is gassed to the point where it attains fetish status. When staged, gear is sometimes framed in sex and war metaphor (Lakoff and Johnson, 1980), and heavily draws on heritage, canon, and iconicity to amplify and maintain gear discourses. In all these trade show / print media / online fora, we see the erasure of women’s labour in order to maintain hegemonic masculinities that gear cultures rely upon. In this workshop, Samantha and Eliot use a range of examples - from online gear lovers gassing over gear components, to Foxy Stardust selling us a DAW - to show how gear is called upon to do extensive work in structuring social relations, and how these gear-centric social formations produce gear cultures.
Speakers
avatar for Samantha Bennett

Samantha Bennett

The Australian National University
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
1E16

2:00pm EDT

Soundgirls Presents Best Skills for Live Sound
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
Join SoundGirls as we explore the different skill sets required for various production environments, from clubs to arena and stadium tours, as well as corporate events. We'll discuss what it takes to become the ideal candidate in each setting and how to stay relevant as a freelance live sound engineer.

In addition, we'll cover the importance of education in audio engineering, focusing on career development, networking, certifications, and the value of continuing education.

The panel will include members from SoundGirls.


----

Barbara Adams is an audio engineer and educator with 30 years of experience in the music industry. She specializes in live sound and production management. Her strong and varied experience also includes recording and artist management. She has worked in Philadelphia’s top music venues, toured and worked festivals as FOH and monitor engineer, and educated and mentored the next generation since 2010.

Barbara is an Associate Teaching Professor at Rowan University and a vital part of their music industry program. In addition to teaching classes in recording technology and live sound, she helps students find opportunities to gain valuable experience both in and out of the university community.

Barbara advocates for creating more diversity in this industry while she inspires the next generation of techs to follow their dreams in the music industry. She is a member of the Audio Engineering Society, Recording Academy, and the Philadelphia chapter president of Soundgirls.


Carolyn Slothour is a freelance live sound engineer based in New York with 10 years of experience in the audio industry with a specialization in FOH, monitor, and systems engineering. She has recently worked with artists such as Matchbox Twenty, Rob Thomas and Michael Bublé, and has mixed at festivals including Download, When We Were Young, Corona Capital, Slam Dunk and Lollapalooza. In addition to engineering, she develops curriculum and teaches live sound engineering courses for a non-profit organization whose mission is to improve accessibility and diversity within the audio industry. When not working, Carolyn enjoys making electronic music and spinning fire.


Kitzy is an audio engineer with experience in live sound, recording, post-production, and podcasting. As a touring FOH engineer, she has worked with artists like Maude Latour, Devon Again, Leanna Firestone, and Bartees Strange. In addition to her work as a touring engineer, Kitzy owns a small audio production & rental company in Philadelphia.

Kitzy was the co-host of the independent music podcast Left of the Dial, as well as the engineer behind the board for Left of the Dial's live in-studio sessions. She is also one of the co-hosts for the upcoming relaunch of the SoundGirls podcast.

Kitzy strives to be an advocate for people who are gender non-conforming, neurodivergent, and/or disabled in the live sound and production industry. When she isn't working, Kitzy enjoys fixing and riding motorcycles, playing with her housemate's dog, and working on her own music.
Wednesday October 9, 2024 2:00pm - 3:00pm EDT
1E09

2:00pm EDT

Is Audio AI Changing Audio Education?
Wednesday October 9, 2024 2:00pm - 3:15pm EDT
Quickly developing, it's often difficult to keep up with the latest in AI, and how AI based products are permeating into the industry. What AI based products are in use? Are they changing what we need to teach? What do we think will be coming next? Join us as a panel of educators and professionals (connected to AI in the industry) discuss these things, and answer your questions about and the state of AI in the industry. This panel is presented by the Technical Committee on Machine Learning and Artificla Intelligence and the Education Committee.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
avatar for Brecht De Man

Brecht De Man

Head of Research, PXL University of Applied Sciences and Arts
Brecht is an audio engineer with a broad background comprising research, software development, management and creative practice. He holds a PhD from the Centre for Digital Music at Queen Mary University of London on the topic of intelligent software tools for music production, and... Read More →
MD

Michele Darling

Chair, Electronic Production and Design, Berklee College of Music
Wednesday October 9, 2024 2:00pm - 3:15pm EDT
1E07

2:00pm EDT

Challenges and Opportunities of Binaural Audio
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
This workshop session will bring together leading sound engineers to explore the challenges and opportunities in binaural audio. The session will begin with an overview of the various binaural rendering methods currently used in the industry. The panelists will share their experiences with immersive audio, focusing on the intricacies of creating high-quality binaural mixes. They will discuss the philosophies behind binaural delivery, such as realistic mix room experiences versus creative purposes, as well as monitoring and quality control methods, and the perceived quality issues in current binaural rendering techniques around tonal fidelity and externalisation. The discussion will cover various applications of binaural audio, including music production, film, and broadcasting/streaming.
Speakers
avatar for Hyunkook Lee

Hyunkook Lee

Professor, Applied Psychoacoustics Lab, University of Huddersfield
Professor
avatar for Vicki Melchior

Vicki Melchior

Consultant, Audio DSP and Software
avatar for Bob Katz

Bob Katz

Digital Domain, Inc.
Bob Katz, workshop co-chair, has been an audio engineer since 1971 and AES member since 1975. He’s played the Bb Clarinet since age 10. Currently President and mastering engineer at Digital Domain, he is an AES Fellow, engineered 3 Grammy-winning albums and numerous nominees. He... Read More →
avatar for Morten Lindberg

Morten Lindberg

Producer and Engineer, 2L (Lindberg Lyd, Norway)
Recording Producer and Balance Engineer with 43 GRAMMY-nominations, 35 of these in craft categories Best Engineered Album, Best Surround Sound Album, Best Immersive Audio Album and Producer of the Year. Founder and CEO of the record label 2L. Grammy Award-winner 2020.
avatar for Justin Gray

Justin Gray

Chief Engineer, Synthesis Sound
Justin Gray is an award-winning producer, mixing engineer and mastering engineer based in Toronto. Justin is a leader in the field of immersive audio mixing & mastering, where he works with artists, producers and labels worldwide to bring their music to life in the Dolby Atmos and... Read More →
avatar for Andrew Scheps

Andrew Scheps

Owner, Tonequake Records
Andrew Scheps has worked with some of the biggest bands in the world: Green Day, Red Hot Chili Peppers, Weezer, Audioslave, Black Sabbath, Metallica, Linkin Park, Hozier, Kaleo and U2. He’s worked with legends such as Johnny Cash, Neil Diamond and Iggy Pop, as well as indie artists... Read More →
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
1E15

2:00pm EDT

The Technical Legacy of Dr. Robert Moog
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
This year marks the 60th anniversary of Dr. Robert Moog's AES paper, "Voltage-Controlled Electronic Music Modules" (paper number 346), as well as the debut of the Moog modular synthesizer. This along with a subsequent paper, "A Voltage-Controlled Low-Pass High-Pass Filter for Audio Signal Processing" (paper number 413) presented in 1965, form the basis for generations of electronic instruments that followed. Throughout his career, Bob Moog worked closely with artists to solve technical challenges that led to musically satisfying instruments. In this session we'll reflect on the development of these designs as well as the impact they have had on the industry from some of those who worked with Moog as well as designers who have built on this technical foundation.

This panel will be followed by a book signing at the AES Booth.

A Brief Testimonial on the Occasion of the 60thAnniversary of Bob Moog’s Historic AES Presentation on his Prototype Voltage-Controlled Modular System
By Albert Glinsky, author of Switched On: Bob Moog and the Synthesizer Revolution, Oxford University Press, 2022

Sixty years ago this week, Robert Arthur Moog, a shy, unassuming young man of 30, not generally known to anyone outside the world of theremin kits, sat at his little AES display at the Barbizon-Plaza Hotel in Manhattan, his table draped with his wife’s Indian print bedspread. Sitting atop it was his self-described “Abominatron”—four handmade modules with sheet metal panels hastily spray-painted, with jacks and knobs identified only by paper labels stuck on with rubber cement. He was surrounded on all sides by the titans of the audio electronics industry: Ampex, Scully, 3M. It was intimidating to a fault; he sat, hoping a few passersby would care to don a pair of earphones to sample the aural cherry pies or pumpkins of his State Fair-like presentation—a pleasant curiosity at best, he figured. Then something unexpected happened. As he famously recalled, a prominent visitor approached and uttered the words, “I’d like a couple of these and a couple of these.” He was in business, and he hadn’t planned on anything like that. History was made right then and there.


During this same 16th Annual AES convention week, on Wednesday, October 14th, at 9:30 in the morning, a session titled, MUSIC AND ELECTRONICS was chaired by the venerable Harald Bode. The second presenter of the morning, a certain Robert A. Moog of the R. A. Moog Company of Trumansburg NY, strode to the podium and read a 9-page typed manuscript titled, “Voltage-Controlled Electronic Music Modules.” Footnotes at the bottom of his last page credited a few predecessors on whose giants’ shoulders he felt he was standing—figures like Myron Schaeffer, Hugh LeCaine, and Bode himself. And the crowd was formidable: Le Caine was there, Bode, of course, and Harry F. Olson and Herbert Belar, inventors of the already classic RCA Mark I and Mark II synthesizers.


After the convention, Bob began selling custom variations on his prototype modules, and the rest, of course became history. But what some may not realize—something that really struck me as Bob’s biographer—is the influence, not just of the instruments themselves, but of Bob’s generosity in sharing the “recipe,” as I like to call it, for this groundbreaking system that would revolutionize music and sound art. Less than a year after his AES presentation, Bob published a detailed 7-page article in the July 1965 issue of the Journal of the Audio Engineering Society, replete with generous block diagrams, schematics, and photos. In the proprietary world of our current technology, this selfless, open-source act is astounding. Not that others hadn’t done this sort of thing before, and Bob had certainly come of age in the world of hobbyist magazines, DIY projects, and the brotherhood of tinkerers. But with the commercial potential of Bob’s system, it fairly gave the whole game away. Bob had inadvertently presented his own giant’s shoulders to the world, upon which so many others would stand—some eventually acknowledging the pedestal on which they held their own inventions, some not. But then, Bob’s passion was sheer invention, and often, as we know, to the detriment of profit and the big picture. But that was Bob: generosity of spirit, and the delight of scattering his seeds widely, even if some of it occasionally took root in competitors’ pastures. It’s not a stretch to say, and I say it with full confidence and conviction based on the facts: Buchla, ARP, EMS, and numerous others that followed—none of them would have been possible without Bob’s published recipe. The timeline bears it out, and some have openly admitted it.


The seminal genius of Bob’s work (and there are so many aspects to it), might be summed up by two statements he made in his AES talk that October morning in 1964. After citing several early historical attempts at producing music electronically, he explained, “The failure of any of these instruments to attain widespread success can be attributed at least partially, to the limitation on the degree of control which the performer is able to exercise over the instrument’s sounds.” He concluded his paper by saying, “In particular, the setting up of prototype experimental musical instruments, and the remote-control processing of conventional audio signals are ideal applications for the voltage-controlled modules.” Ideal indeed!! Amen Bob. And what a legacy you’ve left us. Thank you, and congratulations on such a rich 60-year legacy!!
Speakers
avatar for Michael Bierylo

Michael Bierylo

Chair Emeritus, Electronic Production and Design, Berklee College of Music
Chair Emeritus, Electronic Production and Design
MM

Michelle Moog-Koussa

Executive Director, Bob Moog Foundation
Wednesday October 9, 2024 2:00pm - 3:30pm EDT
Stage

2:00pm EDT

A study on the relative accuracy and robustness of the convolutional recurrent neural network based approach to binaural sound source localisation.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Binaural sound source localization is the task of finding the location of a sound source using binaural audio as affected by the head-related transfer functions (HRTFs) of a binaural array. The most common approach to this is to train a convolutional neural network directly on the magnitude and phase of the binaural audio. Recurrent layers can then also be introduced to allow for consideration of the temporal context of the binaural data, as to create a convolutional recurrent neural network (CRNN).
This work compares the relative performance of this approach for speech localization on the horizontal plane using four different CRNN models based on different types of recurrent layers; Conv-GRU, Conv-BiGRU, Conv-LSTM, and Conv-BiLSTM, as well as a baseline system of a more conventional CNN with no recurrent layers. These systems were trained and tested on datasets of binaural audio created by convolution of speech samples with BRIRs of 120 rooms, for 50 azimuthal directions. Additive noise created from additional sound sources on the horizontal plane were also added to the signal.
Results show a clear preference for use of CRNN over CNN, with overall localization error and front-back confusion being reduced, with it additionally being seen that such systems are less effected by increasing reverb time and reduced signal to noise ratio. Comparing the recurrent layers also reveals that LSTM based layers see the best overall localisation performance, while layers with bidirectionality are more robust, and so overall finding a preference for Conv-BiLSTM for the task.
Speakers
avatar for Jago T. Reed-Jones

Jago T. Reed-Jones

Research & Development Engineer, Audioscenic
I am a Research & Development Engineer at Audioscenic, where we are bringing spatial audio to people's homes using binaural audio over loudspeakers. In addition, I am finishing a PhD at Liverpool John Moores University looking at use of neural networks to achieve binaural sound source... Read More →
Authors
avatar for Jago T. Reed-Jones

Jago T. Reed-Jones

Research & Development Engineer, Audioscenic
I am a Research & Development Engineer at Audioscenic, where we are bringing spatial audio to people's homes using binaural audio over loudspeakers. In addition, I am finishing a PhD at Liverpool John Moores University looking at use of neural networks to achieve binaural sound source... Read More →
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Authoring Inter-Compatible Flexible Audio for Mass Personalization
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
The popularity of internet-based services for delivering media, i.e., audio and video on demand, creates an opportunity to offer personalized media experiences to audiences. A significant proportion of users are experiencing reduced enjoyment, or facing inaccessibility, due to combinations of different impairments, languages, devices and connectivity, and preferences. Audio currently leads video in the maturity of object-based production and distribution systems; thus, we sought to leverage existing object-based audio tools and frameworks to explore creation and delivery of personalized versions. In this practice-based investigation of personalization affordances, an immersive children’s fantasy radio drama was re-authored within five "dimensions of personalization", motivated by an analysis of under-served audiences, to enable us to develop an understanding of what and how to author personalized media. The dimensions were Duration, Language, Style, Device and Clarity. Our authoring approaches were designed to result in alternative versions that are inherently inter-compatible. The ability to combine them exploits the properties of object-based audio and creates the potential for mass personalization, even with a modest number of options within each of the dimensions. The result of this investigation is a structured set of adaptations, based around a common narrative. Future work will develop automation and smart network integration to trial the delivery of such personalized experiences at scale, and thereby validate the benefits that such forms of media personalization can bring.
Speakers
avatar for Craig Cieciura

Craig Cieciura

Research Fellow, University of Surrey
Craig graduated from the Music and Sound Recording (Tonmeister) course at The University of Surrey in 2016. He then completed his PhD at the same institution in 2022. His PhD topic concerned reproduction of object-based audio in the domestic environment using combinations of installed... Read More →
Authors
avatar for Craig Cieciura

Craig Cieciura

Research Fellow, University of Surrey
Craig graduated from the Music and Sound Recording (Tonmeister) course at The University of Surrey in 2016. He then completed his PhD at the same institution in 2022. His PhD topic concerned reproduction of object-based audio in the domestic environment using combinations of installed... Read More →
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Automatic Corrective EQ for Measurement Microphone Emulation
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
This study investigates automatic corrective equalisation (EQ) to adjust the frequency response of microphones with non-flat responses to match that of a flat frequency response measurement microphone. Non-flat responses in microphones can cause significant colouration, necessitating correction for accurate sound capture, particularly in spectral analysis scenarios. To address this, 10 non-flat microphones were profiled in an anechoic chamber, and a 1/3 octave digital graphic equaliser (GEQ) was employed to align their spectra with that of an industry-standard reference microphone. The system's performance was evaluated using acoustic guitar recordings, measuring spectral similarity between Pre- and Post-Corrected recordings and the reference microphone with objective metrics. Results demonstrated improvements in spectral similarity across the metrics, confirming the method's effectiveness in correcting frequency response irregularities. However, limitations such as the inability to account for proximity effects suggest the need for further refinement and validation in diverse acoustic environments.
Speakers
avatar for Matthew Cheshire

Matthew Cheshire

Lecturer in Audio Engineering / Researcher, Birmingham City University
Authors
avatar for Matthew Cheshire

Matthew Cheshire

Lecturer in Audio Engineering / Researcher, Birmingham City University
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Corelink Audio: The Development of A JUCE-based Networked Music Performance Solution
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Existing Networked Music Performance (NMP) solutions offer high-fidelity audio with minimal latency but often lack the versatility needed for multi-plugin configurations and multi-modal integrations, while ensuring ease of use and future development. This paper presents a novel NMP solution called Corelink Audio which aims to address these limitations. Corelink Audio is built on the Corelink network framework, a data agnostic communication protocol, and JUCE, an audio plugin development framework. Corelink audio can be used flexibly as an Audio Unit (AU) or Virtual Studio Technology (VST) plugin inside different host environments and can be easily integrated with other audio or non-audio data streams. Users can also extend the Corelink Audio codebase to tailor it to their own specific needs using the available JUCE and Corelink documentation. This paper details the architecture and technical specifics of the software. The performance of the system, including latency measurements and audio artifacts under varying conditions, is also evaluated and discussed.
Speakers
avatar for Zack Nguyen

Zack Nguyen

Software Developer, NYU IT
Authors
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Do We Need a Source Separation for Automated Subtitling on Media Contents?
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
In this paper, we investigate the efficacy of a speech and music source separation technique on an automated subtitling system under different signal-to-noise ratios (SNRs). To this end, we compare the generated subtitle errors by measuring the word error rate (WER) with and without source separation applied to speech in music. Experiments are first performed on a dataset by mixing speech from the LibriSpeech dataset with music from the MUSDB18-HQ dataset. Accordingly, it is revealed that when the SNR is below 5 dB, using separated speech yields the lowest subtitle error. By contrast, when the SNR exceeds 5 dB, using the mixture audio shows the lowest subtitle error. On the basis of these findings, we propose an automated subtitling system that dynamically chooses between using mixture audio or separated speech to generate subtitles. The system utilizes the estimated SNR as a threshold to decide whether to apply source separation, achieving the lowest average WER under various SNR conditions.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Miniaturized Full-Range MEMS Speaker for In-Ear Headphones
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
This paper reports a highly miniaturized loudspeaker for in-ear applications. It is manufactured using MEMS (micro-electromechanical systems) technology, a technology that is well known e.g. from MEMS microphones and accelerometers. The speaker is based on a mechanically open design, featuring a perfectly decoupled piezoelectric bending actuator. To provide good acoustic sealing, the actuator is surrounded by an acoustic shield. Measurements performed on loudspeaker prototypes attached to an artificial ear simulator showed that despite the small size of only 2.4 x 2.4 mm², sound pressure levels of 105 dB can be achieved within the full audio range. Moreover, low total harmonic distortion (THD) of less than 0.4 % at 90 dB and 1 kHz have been observed.
Speakers
avatar for Fabian Stoppel

Fabian Stoppel

Head Acoustic Systems and Microactuators, Fraunhofer Institute for Silicon Technology
Fabian Stoppel is an expert in the field of micro-electro-mechanical systems (MEMS) and holds a PhD degree in electrical engineering. Since 2010 he is with the Fraunhofer Institute for Silicon Technology (ISIT), Germany, where he is heading the group for acoustic micro systems. Dr... Read More →
Authors
avatar for Fabian Stoppel

Fabian Stoppel

Head Acoustic Systems and Microactuators, Fraunhofer Institute for Silicon Technology
Fabian Stoppel is an expert in the field of micro-electro-mechanical systems (MEMS) and holds a PhD degree in electrical engineering. Since 2010 he is with the Fraunhofer Institute for Silicon Technology (ISIT), Germany, where he is heading the group for acoustic micro systems. Dr... Read More →
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Music Auto-tagging in the long tail: A few-shot approach
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
In the realm of digital music, using tags to efficiently organize and retrieve music from extensive databases is crucial for music catalog owners. Human tagging by experts is labor-intensive but mostly accurate, whereas automatic tagging through supervised learning has approached satisfying accuracy but is restricted to a predefined set of training tags. Few-shot learning offers a viable solution to expand beyond this small set of predefined tags by enabling models to learn from only a few human-provided examples to understand tag meanings and subsequently apply these tags autonomously. We propose to integrate few-shot learning methodology into multi-label music auto-tagging by using features from pre-trained models as inputs to a lightweight linear classifier, also known as a linear probe. We investigate different popular pre-trained features, as well as different few-shot parametrizations with varying numbers of classes and samples per class. Our experiments demonstrate that a simple model with pre-trained features can achieve performance close to state-of-the-art models while using significantly less training data, such as 20 samples per tag. Additionally, our linear probe performs competitively with leading models when trained on the entire training dataset. The results show that this transfer learning-based few-shot approach could effectively address the issue of automatically assigning long-tail tags with only limited labeled data.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

On-Device Automatic Speech Remastering Solution in Real Time
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
With the development of AI technology, there are many attempts to provide new experiences to users by applying AI technology to various multimedia devices. Most of these technologies are provided through server-based AI models due to the large model size. In particular, most of the audio AI technologies are applied through apps and it is server-based in offline AI models. However, there is no doubt that AI technology which can be implemented in real time is important and attractive for streaming service devices such as TVs. This paper introduces an on-device automatic speech remastering solution. The automatic speech remastering solution indicates extracting speech in real-time from the on-device and automatically adjusts the speech level considering the current background sound and volume level of the device. In addition, the automatic speech normalization technique that reduces the variance in speech level for each content is applied. The proposed solution provides users with a high understanding and immersion in the contents by automatically improving the delivery of speech and normalizing speech levels without manually controlling the volume level. There are three key points in this paper. The first is a deep learning speech extraction model that can be implemented in real-time on TV devices, the second is an optimized implementation method using the DSP and NPU, and last is audio signal processing for the speech remastering to improve speech intelligibility.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:00pm EDT

Real-time Recognition of Speech Emotion for Human-robot Interaction
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
In this paper, we propose a novel method for real-time speech emotion recognition (SER) tailored for human-robot interaction. Traditional SER techniques, which analyze entire utterances, often struggle in real-time scenarios due to their high latency. To overcome this challenge, the proposed method breaks down speech into short, overlapping segments and uses a soft voting mechanism to aggregate emotion probabilities in real time. The proposed real-time method is applied to an SER model comprising the pre-trained wav2vec 2.0 and a convolutional network for feature extraction and emotion classification, respectively. The performance of the proposed method was evaluated on the KEMDy19 dataset, a Korean emotion dataset focusing on four key emotions: anger, happiness, neutrality, and sadness. Consequently, applying the real-time method, which processed each segment with a duration of 0.5 or 3.0 seconds, resulted in relative reduction of unweighted accuracy by 10.61% or 5.08%, respectively, compared to the method that processed entire utterances. However, the real-time factor (RTF) was significantly improved.
Wednesday October 9, 2024 2:00pm - 4:00pm EDT
Poster

2:15pm EDT

Mixing for Dolby Atmos Music
Wednesday October 9, 2024 2:15pm - 3:15pm EDT
Joseph J. Chudyk is a multi-platinum mixer and educator with significant experience mixing albums in immersive audio. In addition to teaching at Nazareth College in Rochester, NY, some of his recent notable Atmos mixes include work with Elle King and Miranda Lambert, and Mitchell Tenpenny. In this session Joe will play some of his Dolby Atmos Music mixes, talking about his experience, sharing tips and best practices, and more, with ample opportunities for attendees to ask questions.
Speakers
JC

Joe Chudyk

Joseph J. Chudyk is a multi-platinum mixer having worked with Kane Brown, Mitchell Tenpenny, Teddy Swims, Ghost, IDK, Newsboys, MxPx, Nathaniel Rateliff, Joe Pug, The National Reserve with members from the likes of Gregory Alan Isakov, Jeff Beck, Polar Bear Club, Air Dubai, OneRepublic... Read More →
Wednesday October 9, 2024 2:15pm - 3:15pm EDT
1E06

2:30pm EDT

A framework for high spatial-density auditory data displays using Matlab and Reaper
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
This research aimed to develop a software framework to study and optimize mapping strategies for complex data presented with auditory displays with high spatial resolution, such as wave-field synthesis and higher-order ambisonics systems. Our wave field synthesis system, the Collaborative-Research Augmented Immersive Virtual Environment Laboratory (CRAIVE-Lab), has a 128 full-range loudspeaker system along the circumference of the lab. We decided to use available software music synthesizers because they are built for excellent sound quality, and much knowledge exists on how to program analog synthesizers and other common methods for a desired sound output. At the scale of 128 channels, feeding 128 synthesizers with complex data was not practical for us for initial explorations because of computational resources and the complexity of data flow infrastructure. The proposed framework was programmed in Matlab, using Weather data from the NOAA database for initial exploration. Data is processed from 128 weather stations from East to West in the US spatially aligned with latitude. A MIDI script, in sequential order for all 128 channels, is compiled from converted weather parameters like temperature, precipitation amount, humidity, and wind speed. The MIDI file is then imported into Reaper to render a single sound file using software synthesizers that are operated with the MIDI file control data instructions. The rendered file is automatically cut into the 128 channels in Matlab and reimported into Reaper for audio playback.
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
1E03

2:30pm EDT

Converged AV Networking
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
The 'Converged AV Networking' presentation goes deeper into how modern AV protocols can harmoniously operate on a unified network infrastructure. It examines the complexities and challenges associated with clock synchronization, queuing mechanisms, and specific protocol requirements. Key topics covered include AES67 for audio over IP, ST2110 for professional media over managed IP networks, AVB/Milan for audio video bridging, PtPv2 for precise time protocol, gPtP for general precision time protocol, and QoS strategies like credit-based and strict priority queuing.
Speakers
avatar for Bart Swinnen

Bart Swinnen

CEO, Luminex Network Intelligence
Bart Swinnen, founder and CEO of Luminex Network Intelligence.Bart is for more than 2 decades closely involved in making AV Networking easy in the professional lighting, audio and video industry.With Luminex he provides a range of network switch products optimized for converged AV... Read More →
Wednesday October 9, 2024 2:30pm - 3:00pm EDT
3D04 (AES and AIMS Alliance Program Room)

3:00pm EDT

Immersive Voice and Audio Services (IVAS) Codec – The New 3GPP Standard for Immersive Communication
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
The recently standardized 3GPP codec for Immersive Voice and Audio Services (IVAS) is the first fully immersive communication codec designed for 5G mobile systems. The IVAS codec is an extension of the mono 3GPP EVS codec and offers additional support for coding and rendering of stereo, multi-channel, scene-based audio (Ambisonics), objects and metadata-assisted spatial audio. The IVAS codec enables completely new service scenarios with interactive stereo and immersive audio in communication, content sharing and distribution. This paper provides an overview of the underlying architecture and new audio coding and rendering technologies. Listening test results show the performance of the new codec in terms of compression efficiency and audio quality.
Moderators Speakers Authors
avatar for Adriana Vasilache

Adriana Vasilache

Nokia Technologies
avatar for Andrea Genovese

Andrea Genovese

Research, Qualcomm
Andrea Genovese is a Senior Researcher Engineer at Qualcomm Technologies Inc. working in Multimedia R&D. Andrea specializes in spatial audio and psychoacoustics, acoustic simulations, networked immersive distributed audio, and signal processing for environmental awareness. In 2023... Read More →
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
1E03

3:00pm EDT

Enabling Next Generation Audio - SMPTE ST 2110-41 with Serialized ADM Audio Metadata
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
One of achievements of the SMPTE ST 2110 suite of standards is to bring flexibility and agility to modern IP media systems. However, until now the format of the audio has been mostly static. With the recent publication of SMPTE ST 2110-41 and ST 2127-2, metadata can now be carried alongside the audio allowing adaptation to multiple audio formats. This enables more personalized and accessible services and brings operational benefits. The presentation will explain the opportunities of this transition and present a Paris Olympics case-study that used the first commercial products.
Speakers
avatar for James Cowdery

James Cowdery

Dolby Laboratories
Wednesday October 9, 2024 3:00pm - 3:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

3:00pm EDT

Technical Committee Meeting: Fiber Optics for Audio
Wednesday October 9, 2024 3:00pm - 4:00pm EDT
The AES Technical Committees (TC) lead the Society's involvement in technical issues, and become a visible resource of knowledge and expertise. This repository of knowledge represents the Society to the outside world, and acts to benefit its members within the Society. Each TC specializes in a specific area of audio.
Wednesday October 9, 2024 3:00pm - 4:00pm EDT
2D02/03

3:15pm EDT

Remote Live Events for Broadcast-music/spoken word
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
With upwards of fifty remotes each year, NY Public Radio broadcasts a great variety of events from the very simple to large scale events that can involve a many different challenges, including connectivity issues and internet issues with getting audio back to the studios for live broadcasts, dealing with microphone techniques for the greatest orchestras in the world at Carnegie Hall to dealing with the elements with live broadcasts in Central Park or other outdoor locations around NYC, dealing with cooperating with p.a.'s and unions at various sites, to broadcasts at a politician's residence. Each of these kinds of events might entail unique technical and personnel challenges, but there's also a lot of common sense approaches to working with each of these challenges, as well as applying new technologies as they come along to simplify things, and hopefully make things easier over the years to deal with these challenges.

We will discuss in detail two recent events, "This Land" a live broadcast of Classical, jazz, and Americana from Brooklyn Bridge Park and World Orchestra Week (WOW!) a Youth orchestra festival from Carnegie Hall, contrasting indoor vs outdoor and amplified vs unamplified.
Speakers
avatar for Edward Haber

Edward Haber

For 43 years I was an engineer at WNYC, then New York Public Radio (encompassing WNYC, WQXR, and NJPR); and for the last 36 of those years in charge of WNYC and latterly WQXR's remote recording, including recording the best orchestras in the world as part of WQXR's Carnegie Live radio... Read More →
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
1E08

3:15pm EDT

Gender Equity Townhall
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
The Annenberg Inclusion Initiative released a study demonstrating the dismal gender balance that continues to plague our industry. While there has been much positive change in the past few decades, there is still a lot of work to be done. Join representatives of the leading gender equity organizations including Women's Audio Mission, We Are Moving the Needle, She Is the Music, Soundgirls, and We Make Noise in a discussion on our common goals to change the gender imbalance in the industry moderated by Singer, Songwriter and Producer Eve Horne, founder of We Are the Unheard.
Speakers
avatar for Jeanne Montalvo

Jeanne Montalvo

Engineer/Producer, Self
Jeanne Montalvo is a Grammy-nominated audio engineer and award-winning radio producer. She was selected amongst thousands of applicants as the 2018 EQL resident at Spotify Studios and Electric Lady Studios in New York City, assisting in the recording process for artists like John... Read More →
avatar for Terri Winston

Terri Winston

Executive Director, Women's Audio Mission
Women's Audio Mission (WAM) is dedicated to closing the chronic gender gap in the music, recording and technology industries. Less than 5% of the people creating and shaping the sounds that make up the soundtrack of our lives are women and gender-expansive people. WAM is “changing... Read More →
EB

Erin Barra

MAMMA BARRA LLC; Artist/Manager If Alicia Keys and Nelly Furtado collided with Daft Punk and Esthero in the vast circuitry of an Akai APC 40, you'd probably end up with an experience as rich and soulful as Erin Barra. Joining the likes of MNDR and Tokimonsta, Erin is part of a... Read More →
avatar for Ann Mincieli

Ann Mincieli

Producer, Jungle City Studios
Ann Mincieli, longtime engineer and studio coordinator for Alicia Keys, has channelled her talents into making Jungle a technical prowess featuring the very best in vintage and modern technology. With over twenty years of experience in the music business, Ann Mincieli has travelled... Read More →
avatar for Eve Horne

Eve Horne

Peak Music
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
1E09

3:15pm EDT

Kseniya Kawko: Enveloping Masterclass
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
The masterclass consists of two parts, the first dedicated to jazz. Kseniya details 3D studio productions and live recordings from a German club “Unterfahrt”, featuring jazz legends among large ensembles and stunning new bands. The second part presents new msm-studios 3D tracks, including from the latest Boris Blank and Peter Gabriel albums.

This masterclass series, featuring remarkable recording artists, is a chance to hear 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for Kseniya Kawko

Kseniya Kawko

Tonmeister, msm-studios & Jazz Club Unterfahrt
Kseniya Kawko is a producer and recording engineer specialized in classical music and jazz. She holds Master of Music degrees from two world-renowned audio programs: Sound Recording, McGill University (Montréal, Canada) and Musikregie / Tonmeister, Hochschule für Musik Detmold (Germany... Read More →
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
3D06

3:15pm EDT

Implementing WHO safe listening standards: Insights from the Serendipity Arts Festival
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
This tutorial presents a detailed examination of the first known large-scale implementation of the WHO Global Standard for Safe Listening Venues and Events at the 2023 Serendipity Arts Festival, India's largest multidisciplinary arts event. The case study highlights the methods used to monitor and manage sound levels, the design and deployment of sound systems, and the provision of personal hearing protection, training and information. The session will delve into the practical challenges encountered, the strategies employed to adhere to the WHO Global Standard, and the outcomes of these efforts. Attendees will gain an understanding of the complexities involved in applying safe listening principles in a live event context and the implications for future large-scale events.
Speakers
avatar for Adam Hill

Adam Hill

Associate Professor of Electroacoustics, University of Derby
Adam Hill is an Associate Professor of Electroacoustics at the University of Derby where he leads the Electro-Acoustics Research Lab (EARLab) and runs the MSc Audio Engineering program. He received a Ph.D. from the University of Essex, an M.Sc. in Acoustics and Music Technology from... Read More →
Wednesday October 9, 2024 3:15pm - 4:15pm EDT
1E16

3:30pm EDT

Book Signing: The Technical Legacy of Dr. Robert Moog
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
A book signing of Switched On: Bob Moog and the Synthesizer Revolution by Albert Glinsky will be held following "The Technical Legacy of Dr. Robert Moog" panel on the main stage. Present for the signing will be Larry Fast, legendary synthesist, Steve Dunnington, Vice President of Engineering, Moog Music, Erik Norlander, Director of Keyboard Development for the inMusic Family of Brands, and Michelle Moog-Koussa, Executive Director of the Bob Moog Foundation. Copies of the biography will be for sale at the signing.
Speakers
MM

Michelle Moog-Koussa

Executive Director, Bob Moog Foundation
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
AES Membership Booth

3:30pm EDT

Enhancing Spatial Post-Filters through Non-Linear Combinations
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
This paper introduces a method to enhance the spatial selectivity of spatial post-filters estimated with first-order directional signals. The approach involves applying non-linear transformations on two different spatial post-filters and combining them with weights found by convex optimization of the resulting directivity patterns. The estimation of the post-filters is carried out similarly to the Cross Pattern Coherence (CroPaC) algorithm. The performance of the proposed method is evaluated in a two- and three-speaker scenario with different reverberation times and angular distances of the interfering speaker. The signal-to-interference, signal-to-distortion, and signal-to-artifact ratios are used for evaluation. The results show that the proposed method can improve the spatial selectivity of the post-filter estimated with first-order beampatterns. Using first-order patterns only, it even achieves better spatial separation than the original CroPaC post-filter estimated using first- and second-order signals.
Moderators Speakers
avatar for Stefan Wirler

Stefan Wirler

Aalto University
I'm a PhD student at the Aalto Acoustics Labs, in the Group of Ville Pulkki. I started my PhD studies in 2020. I received my Masters degree in Electrical Engineering from the Friedrich-Alexander University Erlangen-Nuremberg (FAU). My Master thesis was condicted at the AudioLabs in... Read More →
Authors
avatar for Stefan Wirler

Stefan Wirler

Aalto University
I'm a PhD student at the Aalto Acoustics Labs, in the Group of Ville Pulkki. I started my PhD studies in 2020. I received my Masters degree in Electrical Engineering from the Friedrich-Alexander University Erlangen-Nuremberg (FAU). My Master thesis was condicted at the AudioLabs in... Read More →
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
1E03

3:30pm EDT

Audio for Remote Live Production
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
More and more of live production is being done in a location-agnostic manner, with the processing, storage and control remote from the acquisition location. Whilst the video transport requirements often gather the headlines, the audio requirements can be far more complex. This presentation will unpack the technologies used to connect and transport audio, examine the dependencies with video and the different systems requirements in terms of timing references and how to have as much resilience as possible.
Speakers
Wednesday October 9, 2024 3:30pm - 4:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

3:30pm EDT

Newly Discovered Jazz Recordings at New York Public Library
Wednesday October 9, 2024 3:30pm - 4:15pm EDT
Last year saw the release of the Grammy-nominated album “Evenings at the Village Gate – John Coltrane with Eric Dolphy” on Verve. It was the result of many years of intensive detective work by Jessica Wood, Assistant Curator at New York Public Library’s Music & Recorded Sound Division. A long-missing cache of tapes from Richard Striker’s Institute of Sound, recorded by legendary New York engineer Richard Alderson, had been discovered and located by Wood.

Wood, Alderson, and NYPL Audio Preservation Engineer Jeff Willens then went through hundreds of tapes. Among them were unreleased studio and/or live recordings by Bob Dylan, John Coltrane, Nina Simone, Art Blakey, Thelonious Monk, and many other jazz and classical artists, as well as performers as diverse as Tom Paxton, Moondog, and poet Hugh Romney (aka Wavy Gravy).

In this talk, we will explain how this collection of vital, previously unknown recorded documents of the 1960s New York music scene was “lost,” located, and reclaimed by NYPL. We will also tell the story about how two of these unheard reels were brought to the public, filling some gaps in the Coltrane canon. And we will also share why a maverick NYC engineer at the top of his game, left the city and his invaluable tapes behind.

We will be playing samples from the Striker Collection during the talk and inviting attendees to help us “crowdsource” some of the unidentified recordings therein.
Speakers
avatar for Jeff Willens

Jeff Willens

Media Preservation Engineer, New York Public Library
avatar for Jessica Wood

Jessica Wood

Assistant Curator, Music Division and Rodgers & Hammerstein Archives of Recorded Sound, New York Public Library
Jessica is the Assistant Curator for the Music Division as well as the Rodgers & Hammerstein Archives of Recorded Sound at the New York Public Library for the Performing Arts. She previously worked as a project archivist at the Library of Congress' Music Division and as a sound recordings... Read More →
Wednesday October 9, 2024 3:30pm - 4:15pm EDT
1E07

3:30pm EDT

Student Mix Critiques
Wednesday October 9, 2024 3:30pm - 4:30pm EDT
These non-competitive sessions are an opportunity for AES student members at all stages in their studies to receive feedback on their mixes from a panel of industry professionals – live and in-person. Join us to hear mixes by other students, and get tips, tricks, and advice to push your skills to the next level! Advance submission procedures for each event will be posted at www.aesstudents.org. Very limited sign up will be available on-site. Maybe one of your mixes can be featured!
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
Wednesday October 9, 2024 3:30pm - 4:30pm EDT
1E06

4:00pm EDT

The Impact of Height Microphone Layer Position on Perceived Realism of Organ Recording Reproduction
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
For on-site immersive recordings, height microphones are often placed carefully to avoid a distorted or unrealistic image, with many established immersive microphone arrays placing the height microphones 1.5 m or less above the horizontal layer. However, with an instrument so acoustically symbiotic with its space as the pipe organ, the impact of non-coincident height microphone placement has not previously been explored in-depth. Despite this, the pipe organ's radiation characteristics may benefit from non-coincident height microphone placement, providing subjectively improved tone color without sacrificing perceived realism. Subjective listening tests were conducted comparing a pipe organ recording with coincident and non-coincident height microphone positions. The findings of this case study conclude that non-coincident height microphone placement does not significantly impact perceived realism of the immersive organ recording.
Moderators Speakers
avatar for Jessica Luo

Jessica Luo

Graduate Student, New York University
Authors
avatar for Jessica Luo

Jessica Luo

Graduate Student, New York University
avatar for Garrett Treanor

Garrett Treanor

New York University
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
1E03

4:00pm EDT

Realtime distance collaboration in the performing arts
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
Is it possible for musicians to play together over the Internet in real time? The answer is yes. It is possible up to a distance of 3000 km. When using MVTP (Modular Video Transmisson Platform) technology based on the use of optimized JPEG XS codec running on a powerful FPGA. The presentation describes the workflow and user case of using Ultra low latency CESNET MVTP technology in a real environment. During a shared musical performance between two or three locations or during remote teaching or workshops at a music university.
Speakers
PK

Petr Krkavec

DISK Systems
Wednesday October 9, 2024 4:00pm - 4:30pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

4:00pm EDT

Designing an Immersive Audio Production Studio - an in depth Study of Rue Boyer, Paris
Wednesday October 9, 2024 4:00pm - 5:30pm EDT
With consumers able to experience immersive audio more easily and cost-effectively than ever, and both Spotify as well as Apple Music platforms offering playout facilities for immersive content, demand is seemingly continuing to accelerate. The presentation offers a discussion of acoustical and architectural requirements and challenges as well as the relevant standards to adhere to. As a case study, the presentation will review Paris' newest studio, including the design strategies as well as advanced predictive acoustic analysis involved in creating the studio's design. It will also the
studio's unique elevator assisted “disappearing console” strategy for its Studio A Control Room changing into a multi-purpose live room and classroom.
Speakers
avatar for Sergio Molho

Sergio Molho

Partner, Director of Business Development, WSDG Walters-Storyk Design Group
Partner, Director of Business Development
avatar for John Storyk

John Storyk

Founding Partner, WSDG
Since graduating from Princeton as an architect in 1968, and co-founding WSDG (Walters-Storyk Design Group) with his wife, Beth Walters in 1986, John Storyk has been involved with the design of over 3500 recording studios, corporate conference rooms, performance venues, broadcast... Read More →
avatar for Dirk Noy

Dirk Noy

Managing Director Europe, WSDG
BioDirk Noy has a Master of Science (MSc) Diploma in Experimental Solid State Physics from the University of Basel, Switzerland and is a graduate from Full Sail University, Orlando, USA, where he was one of John Storyk’s students.Since joining the consulting firm Walters-Storyk... Read More →
Wednesday October 9, 2024 4:00pm - 5:30pm EDT
1E15

4:15pm EDT

Immediate Family Film at AES NY 2024
Wednesday October 9, 2024 4:15pm - 5:15pm EDT
The Immediate Family (the modern iteration of a legendary studio ensemble known as “The Section”) is a unique group of iconic musicians composed of four of the most recorded, respected, and sought-after players in modern music: Danny Kortchmar (guitar and vocals), Waddy Wachtel (guitar and vocals), Leland Sklar (bass), Russ Kunkel (drums) and the addition of prominent touring, session guitarist and songwriter Steve Postell (guitar and vocals). Frequent collaborators both in the studio and on stage, their work can be heard on albums from Jackson Browne, James Taylor, Linda Ronstadt, Stevie Nicks, Keith Richards, Warren Zevon, Graham Nash, Neil Young, David Crosby, and many, many more. Kortchmar, Sklar, and Kunkel have worked together since the early ‘70s and made up three-quarters of The Section, best known for both their studio and live work in support of some of the top-selling singer/songwriters and solo singers in the history of music, as well as their acclaimed instrumental albums. (By the mid-70s, they were later joined by Wachtel). The film reveals the machinery behind the booming era of the singer-songwriter, when the talents of these four musicians were in furious demand. Collectively, these musicians helped define the sound of a generation. The band is managed by Industry veterans, Fred Croshal of Croshal Entertainment Group and David Helfant of Arpeggio Entertainment, who also served as Executive Producers of the band’s Documentary. Join band members Danny Kortchmar and Waddy Wachtel, and record producers Russ Titelman and Niko Bolas, and acclaimed filmmaker Denny Tedesco to discuss this award- winning documentary.
Speakers
avatar for Danny Kortchmar

Danny Kortchmar

Danny Kortchmar is a guitarist, songwriter, and producer, renowned for his work with some of the biggest names in rock and pop music. Kortchmar became a prominent figure in the 1970s as part of the West Coast music scene. He has collaborated extensively with artists like James Taylor... Read More →
avatar for Denny Tedesco

Denny Tedesco

Acclaimed filmmaker Denny Tedesco (director of The Wrecking Crew and now the full-length documentary Immediate Family), grew up in Los Angeles. In 1996, Denny set out to tell the story of his terminally ill father, legendary session guitarist Tommy Tedesco.  “The Wrecking Crew... Read More →
avatar for Niko Bolas

Niko Bolas

Niko Bolas is an award-winning music producer and recording engineer with credits that include work with such artists as LeAnn Rimes, The Mavericks, Prince, Neil Young, Tyler Bank, Boz Scaggs, Sheryl Crow, Toto, Los Lobos, Jewel, OneRepublic, Johnny Cash, and Keith Urban.He is also... Read More →
avatar for Russ Titelman

Russ Titelman

Russ Titelman is an acclaimed American record producer, known for his versatile work across rock, pop, and jazz genres. Titelman began his career working with Phil Spector as a guitarist and singer in the early 60’s. Then, after being signed to Don Kirchner's publishing company... Read More →

Wednesday October 9, 2024 4:15pm - 5:15pm EDT
Stage

4:30pm EDT

Spatial Matrix Synthesis
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
Spatial Matrix synthesis is presented in this paper. This modulation synthesis technique creates acoustic velocity fields from acoustic pressure signals by using spatial transformation matrices, thus generating complete sound fields for spatial audio. The analysis presented here focuses on orthogonal rotation matrices in both two and three dimensions and compares the results in each scenario with other sound modulation synthesis methods, including amplitude and frequency modulation. As an alternative method for spatial sound synthesis that exclusively modifies the acoustic velocity vector through effects comparable to those created by both amplitude and frequency modulations, Spatial Matrix synthesis is argued to generate inherently spatial sounds, giving this method the potential to become a new musical instrument for spatial music.
Moderators Speakers
avatar for Timothy Schmele

Timothy Schmele

Researcher, Eurecat
Researcher in audio, audio technology, immersive audio, sonification and composition practices. Composer of electroacoustic music.
Authors
avatar for Timothy Schmele

Timothy Schmele

Researcher, Eurecat
Researcher in audio, audio technology, immersive audio, sonification and composition practices. Composer of electroacoustic music.
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
1E03

4:30pm EDT

IPMX Site Open “By Invitation Only” to Developers, Integrators, and End Users in New York City
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
“The Gateway” is an initiative by the Swedish-American Chamber of Commerce in the heart of Manhattan, New York on the 29th floor of the 900 Third Avenue tower. The facility offers event spaces and conference rooms for members, but it is also the locus of a constantly evolving panoply of latest technologies from the Pro AV and IT world assembled and managed by HarperVision Associates.In addition to cutting edge interoperability demos and first looks at brand new products, the Gateway NYC and HarperVision also host key AV/IT events such as Collaboration Week by the IMCCA (Interactive Multimedia & Collaborative Communications Alliance).Greg Harper, President of HarperVision Associates and CIO of the Gateway, has also recently led an initiative involving IPMX, a suite of open specifications and standards that build upon SMPTE ST 2110, NMOS, and AES67. This includes a live interoperability demonstration consisting of several brands including: converters, presentation switchers, routing software, and more.This growing interop site is a gathering place to witness working products, conduct events and presentations, and to showcase IPMX in an enterprise AV/IT setting. Businesses, government users, equipment vendors and developers, integrators, and educators alike are encouraged to communicate with Greg about site visits, events, and other initiatives.
Speakers
avatar for Greg Harper

Greg Harper

President, HarperVision
Wednesday October 9, 2024 4:30pm - 5:00pm EDT
3D04 (AES and AIMS Alliance Program Room)
  Sponsor Session

4:30pm EDT

Ask Me Anything! (About Starting Your Career)
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
Aimed at students, novices, and anyone thinking of transitioning into working in the audio industry as either a professional or hobbyist/amateur. What best prepared our panel of professionals for entry into their career? We will discuss skills, knowledge, work habits, obstacles, working conditions, goals and realities, and whatever else YOU want us to discuss. Join us for a relaxed and informal discussion featuring professionals from different sectors of the industry (including music production, live sound, broadcasting, and research/development, and maybe more!) – and bring your questions – a good portion of the discussion will be guided by audience interests.
Speakers
avatar for Ian Corbett

Ian Corbett

Coordinator & Professor, Audio Engineering & Music Technology, Kansas City Kansas Community College
Dr. Ian Corbett is the Coordinator and Professor of Audio Engineering and Music Technology at Kansas City Kansas Community College. He also owns and operates off-beat-open-hats LLC, providing live sound, recording, and audio production services to clients in the Kansas City area... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E16

4:30pm EDT

7 Non Traditional Production Methods w/ QuestionATL
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
Some engineers and producers take advantage of a few of the customizations and optimizations of their machines (laptops, desktops, mobile recorders, tablets, etc.) However, unlocking the deeper settings of your production machine can lead to greater efficiency and productivity with cutting-edge applications and technologies. QuestionATL will take attendees through a variety of production optimizations–including mobile devices when working remotely with artists who need to provide high-resolution content without having access to a studio or professional-grade gear.
Speakers
avatar for Question Atl

Question Atl

QuestionATL Music
QuestionATL is a blind Rap Artist & Producer. He is self-taught on several instruments, began freestyling at 5 and making beats at 12. He has won over 20 beat battles and produced for several artists. QuestionATL released his debut project, The Dream, on all streaming platforms as... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E08

4:30pm EDT

Crafting Modern Hip Hop/Rap: A Hands-On Guide
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
This 45-minute workshop is designed for music production students eager to dive into the world of modern hip hop and rap. With hard-hitting beats, intricate rhythms, and innovative production techniques, hip hop/rap continues to dominate the music scene. Our session will cover essential techniques and creative tips to help you produce tracks that capture the essence of contemporary hip hop and rap.

Workshop Overview

Title: Crafting Modern Hip Hop/Rap: A Hands-On Guide

Duration: 45 minutes

Target Audience: Music production students and aspiring producers

Prerequisites: Basic knowledge of DAWs (Digital Audio Workstations) and audio production

Objectives

By the end of this workshop, participants will:

Understand the key elements of modern hip hop and rap music.
Learn how to create hard-hitting beats and basslines.
Explore techniques for crafting catchy hooks and verses.
Gain insights into mixing techniques to achieve a polished hip hop/rap sound.

Workshop Content

Introduction to Modern Hip Hop/Rap (5 minutes)
Brief history and evolution of the genre
Key characteristics: Beats, basslines, and lyrical flow
Influential artists and iconic tracks

Creating Hard-Hitting Beats (10 minutes)
Overview of drum machines and sample packs
Layering and sequencing drum sounds
Practical demo: Designing a punchy beat

Crafting Catchy Hooks and Verses (10 minutes)
Importance of hooks in hip hop and rap
Techniques for writing memorable verses and choruses
Practical demo: Building a hook and verse from scratch

Basslines and 808s (10 minutes)
Essential bass patterns and 808 techniques
Using sub-bass and distortion for impact
Practical demo: Programming a powerful bassline

Mixing for Hip Hop/Rap (5 minutes)
Key principles of mixing for clarity and impact
Using effects to enhance the vibe: Reverb, delay, and saturation
Practical demo: Quick mixing tips for a hip hop/rap track

Q&A and Wrap-Up (5 minutes)
Recap of key points
Q&A session for participants
Additional resources and next steps
Speakers
avatar for Tyron Williams

Tyron Williams

Music Educator, Certified Banga Productions LLC
Greeting everyone. My name is Tyron Williams, better known as TWiZ, an accomplished GRAMMY nominated Music Educator, record producer, audio engineer, and content creator hailing from Savannah, Georgia. Currently, I live in the NYC area extending my expertise to aspiring young artists... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E07

4:30pm EDT

George Massenburg: Enveloping Masterclass
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
George plays high resolution stereo, 5.1 and 3D recordings from his fabulous back catalogue, commenting on production techniques, dynamics processing and the most appropriate listening format considering each track.

This masterclass series, featuring remarkable recording artists, is a chance to hear 3D audio at its best; as we discuss factors of production, distribution and reproduction that makes it worth the effort. Thomas exemplifies the underrated qualia, auditory envelopment (AE); and we evaluate how robustly AE latent in the content may be heard across this particular listening room.

Seats are limited to keep playback variation at bay, and the session is concluded with Q&A. If you attend multiple masterclasses, consider choosing different seats each time.
Speakers
avatar for Thomas Lund

Thomas Lund

Senior Technologist, Genelec Oy
Thomas Lund has authored papers on human perception, spatialisation, loudness, sound exposure and true-peak level. He is researcher at Genelec, and convenor of a working group on hearing health under the European Commission. Out of a medical background, Thomas previously served in... Read More →
avatar for George Massenburg

George Massenburg

Associate Professor of Sound Recording, Massenburg Design Works
George Y. Massenburg is a Grammy award-winning recording engineer and inventor. Working principally in Baltimore, Los Angeles, Nashville, and Macon, Georgia, Massenburg is widely known for submitting a paper to the Audio Engineering Society in 1972 regarding the parametric equali... Read More →
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
3D06

4:30pm EDT

Vocal Production with Simone Torres
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
Join Simone as she dives into the art of vocal production, sharing her unique process and insights. Discover the psychology behind the craft, learn how she creates safe spaces for her artists, and get your questions answered in a live Q&A session!

_____

Simone Torres is a Diamond, Grammy Nominated, and Latin Grammy honored engineer, vocal producer, and artist from New York. She has worked with artists such as Becky G, Chlöe Bailey, Sofia Carson, Normani, Anitta, Cardi B, Camila Cabello, Monsta X, Jessie J, Billy Porter, The Backstreet Boys, Dua Lipa, and more.

Some notable accomplishments include being honored by the Latin Recording Academy as one of their Leading Ladies of Entertainment (2023), Vocal Producing Sofia Carson's "Slow Dance" with BTS' Jimin, Becky G’s “Coming Your Way,” and “La Nena,” Normani’s “Motivation” and Engineering Cardi B’s 11x Platinum “I Like It” and 4x Platinum “Be Careful.”

IG: @iamsimonetorres
Speakers
Wednesday October 9, 2024 4:30pm - 5:30pm EDT
1E09

4:45pm EDT

Composing and Producing Music for Immersive Audio
Wednesday October 9, 2024 4:45pm - 5:30pm EDT
Justin Gray is a leader in Immersive Audio music production and has presented at numerous AES events on the topic. As an early adopter of object-based technology for music production, Justin has mixed and mastered over 2000 songs in immersive formats (Dolby Atmos and Sony 360).

In 2020, Justin began composing new music for a project specifically conceived for immersive production. This project was recorded over 40 days in a world-class studio in Toronto, Canada, and included over 40 musicians from across the globe. The music draws influence from a diverse range of musical traditions and was composed, orchestrated, recorded, edited, mixed, and mastered with immersive intent at each stage.

Justin has drawn upon his broad musical experience as a composer, arranger, performer, recording engineer, producer, and engineer to bring this musical project to life. He is inspired to share the music, philosophy, and immersive production process with the AES community.

As Immersive Audio continues to expand its scope in the global music community, Justin believes that the next step is for artists to integrate immersive design into their creative practice. This can begin at the initial composition stage, and carry through each stage of a musical production.

In this workshop, Justin will share examples from the project in Dolby Atmos. He will also explore composing and orchestrating music with immersive intent, a rarely discussed topic in this field. The presentation will also include an in-depth look into immersive production techniques, including contemporary immersive recording techniques and immersive mixing philosophy/aesthetics.

Each performance in the project was filmed at a cinematic level, further supporting the learning experience.

Justin also rigorously documented the technical aspects of the project and will have additional visual content to support the presentation.

This AES event will be one of the first public previews of the music. The full album will be released in various cities worldwide in Dolby Atmos movie theatres in early 2025.



Speakers
avatar for Justin Gray

Justin Gray

Chief Engineer, Synthesis Sound
Justin Gray is an award-winning producer, mixing engineer and mastering engineer based in Toronto. Justin is a leader in the field of immersive audio mixing & mastering, where he works with artists, producers and labels worldwide to bring their music to life in the Dolby Atmos and... Read More →
Wednesday October 9, 2024 4:45pm - 5:30pm EDT
1E06

6:00pm EDT

Richard Heyser Memorial Lecture: Doohickii Digitali Redux
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
Anthony Agnello and Richard Factor have over 100 combined years as members of the AES. They have been fortunate to witness and, in some cases, perpetrate the evolution of audio engineering from electro-mechanical to digital. In May of 1979, they published an article in db Magazine, “Doohickii Digitali”, that concluded:

"… we have summarized some of the possibilities of digital signal processing (and completely ignored digital recording, a probably more important subject).  While we don’t have the “all digital” studio yet, it can be reasonably foreseen that before too long, all the components will be available to make one, so that the only analog elements remaining will be the microphone and speaker.  Since these are the primary limitations of the recording process at present, we anxiously await the development of a standard digital I/O bus for the human element."


In this lecture, they will pick up not only where they left off, but also before they began.
Speakers
avatar for Anthony Agnello

Anthony Agnello

Managing Director, Eventide
Tony Agnello was born in Brooklyn, graduated from Brooklyn Technical High School in 1966, received the BSEE from City College of NY in 1971, the MSEE from the City University of NY in 1974 followed by post graduate studies in Digital Signal Processing at Brooklyn’s Polytechnical... Read More →
avatar for Richard Factor

Richard Factor

Chairman / Prime Fossil, Eventide
Richard Factor was born in 1945 and missed being a baby boomer by only weeks.  He lived and was schooled in Manhattan and the Bronx, NYC, until he moved to New Jersey at age 40 and then to Sedona, Arizona in his 60s.  He obtained a degree in Economics but preferred to pursue broadcasting... Read More →
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
1E08

6:00pm EDT

Tech Tour: Lincoln Center - David Rubenstein Atrium & David Geffen Hall
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
All Tech Tours are full. Know Before You Go emails were sent to accepted Tech Tour registrants with information about the tour.  

Duration: 1.5 hours
Capacity: 20 (Once capacity is reached, registrants will be added to a wait list

Tour Description: Join Programming Senior Manager, Godfrey Palaia, and Technical Director, Rudy R. Rodriguez, for an exclusive tour of two iconic spaces at Lincoln Center. Godfrey will guide participants through the vibrant David Rubinstein Atrium, showcasing its unique programming and design. Rudy will then lead the group to David Geffen Hall, offering insights into the cutting-edge technology and production infrastructure that power its world-class performances.
Wednesday October 9, 2024 6:00pm - 7:30pm EDT
Offsite

7:00pm EDT

DEI Mixer at Tailor Public House
Wednesday October 9, 2024 7:00pm - 9:00pm EDT
Experience the vibrant convergence of diversity, equity, and inclusion within the audio engineering community at the AES Show 2024 DEI Mixer.

Join us for an engaging evening of networking, insightful conversations, and a celebration of our shared commitment to creating a more inclusive soundscape. Connect with fellow professionals, share your experiences, and explore the diverse perspectives that shape our industry. Together, we are harmonizing the future of audio engineering.

We are excited to have We Are Moving the Needle as a co-organizer for this event, reinforcing our focus on gender equity in the music industry. 

Additionally, we are honored to have long-time DEI advocate Rebekah Wilson, CEO of Source Elements, and AES President Leslie Gaston-Bird delivering a powerful opening to set the tone for the evening.


Light refreshments and beverages will be provided. This event will be held offsite at
Tailor Public House. Sponsored by: Source Elements 
Wednesday October 9, 2024 7:00pm - 9:00pm EDT
Offsite
 
Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.
Filtered by Date -