|Title:||Sound Signal Invariant DAE Neural Network-Based Quantizer Architecture of Audio/Speech Coder Using the Matching Pursuit Algorithm|
|Authors:||Avramov, V. V.|
Petrovsky, A. A.
|Keywords:||публикации ученых;Stepwise activation function;Audio/speech coding;Quantization;Neural network|
|Citation:||Avramov, V. Sound Signal Invariant DAE Neural Network-Based Quantizer Architecture of Audio/Speech Coder Using the Matching Pursuit Algorithm / V. Avramov, V. Herasimovich, A. Petrovsky // Advances in Neural Networks – ISNN 2018. Lecture Notes in Computer Science. – 2018. – Vol. 10878. – P. 511 – 520. – DOI: 10.1007/978-3-319-92537-0_59.|
|Abstract:||The paper is devoted to the quantization algorithm development based on the neural networks framework. This research is considered in the context of the scalable real-time audio/speech coder based on the perceptually adaptive matching pursuit algorithm. The encoder parameterizes the input sound signal frame with some amount of real numbers that are need to be compactly represented in binary form, i.e. quantized. The neural network quantization approach gives great opportunity for such a goal because the data quantized in whole vector but not in separate form and it can effectively use correlations between each element of the input coded vector. Deep autoencoder (DAE) neural network-based architecture for the quantization part of the encoding algorithm is shown. Its structure and learning features are described. Conducted experiments points out the big compression ratio with high reconstructed signal quality of the developed audio/speech coder quantization scheme.|
|Appears in Collections:||Публикации в зарубежных изданиях|
|Avramov_Sound.pdf||84.71 kB||Adobe PDF||View/Open|
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.