@article {ALVAREZ-MARTíNEZ:2024:0736-2935:6411, title = "HRTF individualization using ear images from 3D meshes", journal = "INTER-NOISE and NOISE-CON Congress and Conference Proceedings", parent_itemid = "infobike://ince/incecp", publishercode ="ince", year = "2024", volume = "270", number = "5", publication date ="2024-10-04T00:00:00", pages = "6411-6421", itemtype = "ARTICLE", issn = "0736-2935", url = "https://ince.publisher.ingentaconnect.com/content/ince/incecp/2024/00000270/00000005/art00045", doi = "doi:10.3397/IN_2024_3741", author = "ALVAREZ-MART{\’ı}NEZ, Ariel and L{\’o}PEZ MONFORT, Jos{\’e} Javier", abstract = "The head-related transfer function (HRTF) describes how a human receives sound from different directions in space. It is unique to each listener. The best way to obtain an individualized HRTF is through direct measurement, but it requires sophisticated laboratory equipment and long measurement times. This work proposes an approach for HRTF individualization employing ear images from 3D meshes using a deep neural network and spherical harmonics transform (SHT). The method relies on the HUTUBS dataset, including 3D meshes and HRTFs. The model uses ear images to predict a low-dimensional representation of the HRTF. Initially, ten images of the right ear of each 3D mesh are taken in different positions. The model consists of two main parts. The first is a convolutional neural network (CNN), which is used to extract features from the ear images. The second part learns from the feature map obtained and predict the spherical harmonic coefficients. Finally, the individualized HRTF is obtained through inverse SHT. The performance of the method is evaluated by computing the log-spectral distortion (LSD) between the measured HRTF and the predicted one. The results show favorable LSD values compared to other models addressing the same problem.", }