Search
Now showing items 1-10 of 10
Modeling plate and spring reverberation using a DSP-informed deep neural network
Plate and spring reverberators are electromechanical systems first used and researched as means to substitute real room reverberation. Nowadays they are often used in music production for aesthetic reasons due to their ...
End-to-End Probabilistic Inference for Nonstationary Audio Analysis
A typical audio signal processing pipeline includes multiple disjoint analysis stages, including calculation of a time-frequency representation followed by spectrogram-based feature analysis. We show how time-frequency ...
Data-Efficient Weakly Supervised Learning for Low-Resource Audio Event Detection Using Deep Learning
We propose a method to perform audio event detection under the common constraint that only limited training data are available. In training a deep learning system to perform audio event detection, two practical problems ...
Estimating & Mitigating the Impact of Acoustic Environments on Machine-to-Machine Signalling
The advance of technology for transmitting Data-over-Sound in various IoT and telecommunication applications has led to the concept of machine-to-machine over-the-air acoustic signalling. Reverberation can have a detrimental ...
Spectral Visibility Graphs: Application to Similarity of Harmonic Signals
Graph theory is emerging as a new source of tools for time series analysis. One promising method is to transform a signal into its visibility graph, a representation which captures many interesting aspects of the signal. ...
Musical Features for Automatic Music Transcription Evaluation
(Transactions of the International Society for Music Information Retrieval (TISMIR, 2021)
This technical report gives a detailed, formal description of the features introduced in the paper: Adrien Ycart, Lele Liu, Emmanouil Benetos and Marcus T. Pearce. "Investigating the Perceptual Validity of Evaluation Metrics ...
Alzheimer's Dementia Recognition Using Acoustic, Lexical, Disfluency and Speech Pause Features Robust to Noisy Inputs
We present two multimodal fusion-based deep learning models that consume ASR transcribed speech and acoustic data simultaneously to classify whether a speaker in a structured diagnostic task has Alzheimer's Disease and to ...
Joint Scattering for Automatic Chick Call Recognition
(2021)
Animal vocalisations contain important information about health, emotional state, and behaviour, thus can be potentially used for animal welfare monitoring. Motivated by the spectro-temporal patterns of chick calls in the ...
The CORSMAL benchmark for the prediction of the properties of containers
(2021-07-27)
The contactless estimation of the weight of a container and the amount of its content manipulated by a person are key pre-requisites for safe human-to-robot handovers. However, opaqueness and transparencies of the container ...