Skip to main content
School of Electronic Engineering and Computer Science

Smart EQ: Personalizing Audio with Context-aware AI using Listener Preferences and Psychological Factors

Annual stipend: UKRI stipend rate (plus London wage).

Start date: September 2024 (Monday 16th if not visa delay).

Supervisors: Dr. Charalampos Saitis, Dr. George Fazekas in collaboration with Yamaha.

Abstract: There is evidence that production techniques like equalisation (EQ) can significantly influence a listener's emotional response to music. In professional music production, precise EQ adjustments are used to evoke specific emotional effects. However, in consumer audio, users often struggle to achieve the desired sound quality, leading to a diminished music-listening experience and a pervasive sense of frustration. Ideal EQ settings are highly personal and context-dependent, with no universal solution. Many users find it challenging to achieve the sound they desire due to diverse individual preferences and situational contexts. Some prioritise content, while others focus on sound quality. Psychological factors such as introversion versus extroversion and individual sensitivity also influence preferred EQ settings. These preferences fluctuate with mood, environment, and specific listening contexts, adding another layer of complexity.

The goal of this PhD is to create a model predicting ideal EQ settings for individuals based on their unique preferences and contextual factors. This involves conducting experiments with a diverse group of listeners to gather data on EQ preferences across various contexts. Semantic descriptors related to EQ settings will be collected via crowdsourcing, along with psychological traits and contextual information. Using this rich dataset, a predictive model will be developed with state-of-the-art machine learning techniques, incorporating insights from psychological and contextual factors to enhance accuracy. Validation will be an iterative process involving rigorous testing and refinement. Collaboration with Yamaha will aim to integrate the model into a consumer audio system capable of real-time adaptive EQ adjustments, followed by user testing to evaluate and refine the system. Research by Dourou (2022) found that listeners with low arousal levels prefer EQ settings that boost lower frequencies, aligning with observations in music production and highlighting the need for personalised EQ. Stables et al. (2016) provide a framework for understanding terms and processes to achieve desired timbral effects, informing context-aware EQ development. Further inspiration comes from multimodal models for music and language (Manco et al., 2022) and crowdsourcing for semantic descriptors (Cartwright and Pardo, 2014) as well as other works listed in the references below.

The ideal candidate will have an interest in music technology, music emotion research, signal processing, machine learning as well as the latest deep learning techniques. Basic understanding of music theory is useful but not essential.

How to Apply - https://aim.qmul.ac.uk/apply/

Primary supervisor: Dr George Fazekas and Dr Charalampos Saitis (C4DM) Email: george.fazekas@qmul.ac.uk; c.saitis@qmul.ac.uk

AIM CDT Website: UKRI Centre for Doctoral Training in Artificial Intelligence and Music (qmul.ac.uk)

C4DM Website: Home Page (qmul.ac.uk)

Back to top