Neurally Controlled Graphic Equalizer
Loading...
Access rights
openAccess
URL
Journal Title
Journal ISSN
Volume Title
A1 Alkuperäisartikkeli tieteellisessä aikakauslehdessä
This publication is imported from Aalto University research portal.
View publication in the Research portal (opens in new window)
View/Open full text file from the Research portal (opens in new window)
View publication in the Research portal (opens in new window)
View/Open full text file from the Research portal (opens in new window)
Authors
Date
2019-12
Major/Subject
Mcode
Degree programme
Language
en
Pages
2140-2149
Series
IEEE/ACM Transactions on Audio, Speech, and Language Processing, Volume 27, issue 12
Abstract
This paper describes a neural network based method to simplify the design of a graphic equalizer without sacrificing the accuracy of approximation. The key idea is to train a neural network to predict the mapping from target gains to the optimized band filter gains at specified center frequencies. The prediction is implemented with a feedforward neural network having a hidden layer with 20 neurons in the case of the ten-octave graphic equalizer. The band filter coefficients can then be quickly and easily computed using closed-form formulas. This work turns, for the first time, the accurate graphic equalization design into a feedforward calculation without matrix inversion or iterations. The filter gain control using the neural network reduces the computing time by 99.6% in comparison to the least-squares design method it is imitating and contributes an approximation error of less than 0.1 dB. The resulting neurally controlled graphic equalizer will be highly useful in various audio and music processing applications, which require time-varying equalization.Description
Keywords
Gain, Equalizers, Graphics, Bandwidth, Speech processing, Optimization, Audio systems, equalizers, feedforward neural networks, IIR filters, supervised learning
Other note
Citation
Välimäki, V & Rämö, J 2019, ' Neurally Controlled Graphic Equalizer ', IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 27, no. 12, pp. 2140-2149 . https://doi.org/10.1109/TASLP.2019.2935809