State-of-the-art headphone-based spatial audio rendering techniques rely on the use of Head-Related Transfer Functions (HRTFs) which are peculiar filters that capture the acoustic effects of the human head. HRTFs allow accurate simulation of the signal that arrives at the eardrum as a function of the sound source’s spatial position. However, obtaining personal HRTFs normally requires dedicated hardware, a controlled environment, and long acquisition time. Within the Sound of Vision project we are developing a novel framework for low-cost individual HRTF determination based on the study of the human outer ear. HRTFs for different ear shapes will be acquired in a controlled semi-anechoic laboratory environment including a brand new custom built HRTF measurement system (see demo time-lapse video) and a KEMAR mannequin for acoustic research with interchangeable ears. The relation between HRTFs and morphologic data – acquired through a laser scanner – will be established through machine learning techniques. The aim is to use the measured data to develop a methodology for determining the individual HRTF directly from a 3D representation of the user’s head.