Browsing Electronic Engineering and Computer Science by Author "Ma, Y"
Now showing items 1-16 of 16
-
Antenna Classification using Gaussian Mixture Models (GMM) and Machine Learning
Ma, Y; Hao, Y (2020-07-09) -
Autonomous Compressive-Sensing-Augmented Spectrum Sensing
Zhang, X; Ma, Y; Gao, Y; Zhang, W (2018-08) -
ChatMusician: Understanding and Generating Music Intrinsically with LLM
Yuan, R; Lin, H; Wang, Y; Tian, Z; Wu, S; Shen, T; Zhang, G; Wu, Y; Liu, C; Zhou, Z (2024-08-11)While Large Language Models (LLMs) demonstrate impressive capabilities in text generation, we find that their ability has yet to be generalized to music, humanity’s creative language. We introduce ChatMusician, an open-source ... -
Deep learning framework for subject-independent emotion detection using wireless signals.
Khan, AN; Ihalage, AA; Ma, Y; Liu, B; Liu, Y; Hao, Y (PLoS, 2021-02)Emotion states recognition using wireless signals is an emerging area of research that has an impact on neuroscientific studies of human behaviour and well-being monitoring. Currently, standoff emotion detection is mostly ... -
Distributed Compressive Sensing Augmented Wideband Spectrum Sharing for Cognitive IoT
Zhang, X; Ma, Y; Qi, H; Gao, Y; Xie, Z; Xie, Z; Zhang, M; Wang, X; Wei, G; Li, Z (2018-08) -
DMRN+17: Digital Music Research Network One-day Workshop 2022
Miller, J; Lewis, D; Guo, Z; Li, Y; Ma, Y; Vahidi, C; Boon, H; Wolstanholme, L; Gil Panal, JM; Hayes, B (Centre for Digital Music - C4DM, 2022-12-20)DMRN+17: Digital Music Research Network One-day Workshop 2022 Queen Mary University of London - Tuesday 20th December 2022. The Digital Music Research Network (DMRN) aims to promote research in the area of Digital Music, ... -
Large-Scale Pretrained Model for Self-Supervised Music Audio Representation Learning
Li, Y; Yuan, R; Zhang, G; Ma, Y; Lin, C; Chen, X; Ragni, A; Yin, H; Hu, Z; He, H (2022-12-20)Self-supervised learning technique is an under-explored topic for music audio due to the challenge of designing an appropriate training paradigm. We hence propose MAP-MERT, a large-scale music audio pre-trained model for ... -
LyricWhiz: Robust Multilingual Lyrics Transcription by Whispering to ChatGPT
Zhuo, L; Yuan, R; Pan, J; Ma, Y; Li, Y; Zhang, G; Liu, S; Dannenberg, R; Fu, J; Lin, C (International Society for Music Information Retrieval Conference (ISMIR), 2023-11-05)We introduce LyricWhiz, a robust, multilingual, and zero-shot automatic lyrics transcription method achieving state-of-the-art performance on various lyrics transcription datasets, even in challenging genres such as rock ... -
MARBLE: Music Audio Representation Benchmark for Universal Evaluation
Yuan, R; Ma, Y; Li, Y; Zhang, G; Chen, X; Yin, H; Zhuo, L; Liu, Y; Huang, J; Tian, Z (37th Conference on Neural Information Processing Systems (NeurIPS), 2023)In the era of extensive intersection between art and Artificial Intelligence (AI), such as image generation and fiction co-creation, AI for music remains relatively nascent, particularly in music understanding. This is ... -
MERT: Acoustic Music Understanding Model with Large-Scale Self-supervised Training
Li, Y; Yuan, R; Zhang, G; Ma, Y; Chen, X; Yin, H; Xiao, C; Lin, C; Ragni, A; Benetos, E (2024-05-07)Self-supervised learning (SSL) has recently emerged as a promising paradigm for training generalisable models on large-scale data in the fields of vision, text, and speech. Although SSL has been proven effective in speech ... -
MERTech: instrument playing technique detection using self-supervised pretrained model with multi-task finetuning
Li, D; Ma, Y; Wei, W; KONG, Q; Wu, Y; Che, M; Xia, F; Benetos, E; Li, W; IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP) (IEEE, 2024-04-14)Instrument playing techniques (IPTs) constitute a pivotal component of musical expression. However, the development of automatic IPT detection methods suffers from limited labeled data and inherent class imbalance issues. ... -
MusiLingo: bridging music and text with pre-trained language models for music captioning and query response
Deng, Z; Ma, Y; Liu, Y; Guo, R; Zhang, G; Chen, W; Huang, W; Benetos, E; 2024 Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL 2024) (2024-06-16)Large Language Models (LLMs) have shown immense potential in multimodal applications, yet the convergence of textual and musical domains remains not well-explored. To address this gap, we present MusiLingo, a novel system ... -
MusiLingo: Bridging Music and Text with Pre-trained Language Models for Music Captioning and Query Response.
Deng, Z; Ma, Y; Liu, Y; Guo, R; Zhang, G; Chen, W; Huang, W; Benetos, E (2024) -
On the effectiveness of speech self-supervised learning for music
Ma, Y; Yuan, R; Li, Y; Zhang, G; Chen, X; Yin, H; Lin, C; Benetos, E; Ragni, A; Gyenge, N (International Society for Music Information Retrieval Conference (ISMIR), 2023-11-05)Self-supervised learning (SSL) has shown promising results in various speech and natural language processing applications. However, its efficacy in music information retrieval (MIR) still remains largely unexplored. While ... -
Optimization of High-resolution and Ambiguity-free Sparse Planar Array Geometry for Automotive MIMO Radar
Huan, M; Liang, J; Ma, Y; Liu, W; Wu, Y; Zeng, Y (Institute of Electrical and Electronics Engineers (IEEE), 2024)The next-generation 4D imaging automotive radar is characterized by high angular resolution, unambiguous detection, low latency, low cost, and small size. This study provides an enhanced analysis of the angular ambiguity ... -
Sparsity Independent Sub-Nyquist Rate Wideband Spectrum Sensing on Real-Time TV White Space
Ma, Y; Gao, Y; Cavallaro, A; Parini, CG; Zhang, W; Liang, Y-C (2017-10)