# 李學翰's RD Page
## Singing Voice Separation
### [Multi-Scale multi-band densenets for audio source separation](https://arxiv.org/pdf/1706.09588.pdf)
Takahashi, Naoya, and Yuki Mitsufuji. "Multi-scale multi-band densenets for audio source separation." In 2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), pp. 21-25. IEEE, 2017.
### [Towards Real-Time Single-Channel Singing-Voice Separation with Pruned Multi-Scaled Densenets](https://ieeexplore.ieee.org/document/9053542)
M. Huber, G. Schindler, C. Schörkhuber, W. Roth, F. Pernkopf and H. Fröning, "Towards Real-Time Single-Channel Singing-Voice Separation with Pruned Multi-Scaled Densenets," ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 2020, pp. 806-810, doi: 10.1109/ICASSP40776.2020.9053542.
### [Mmdenselstm: An Efficient Combination of Convolutional and Recurrent Neural Networks for Audio Source Separation](https://arxiv.org/pdf/1805.02410.pdf)
N. Takahashi, N. Goswami and Y. Mitsufuji, "Mmdenselstm: An Efficient Combination of Convolutional and Recurrent Neural Networks for Audio Source Separation," 2018 16th International Workshop on Acoustic Signal Enhancement (IWAENC), Tokyo, Japan, 2018, pp. 106-110, doi: 10.1109/IWAENC.2018.8521383.
### [Hybrid spectrogram and waveform source separation](https://arxiv.org/pdf/2111.03600.pdf)
Défossez, Alexandre. "Hybrid spectrogram and waveform source separation." arXiv preprint arXiv:2111.03600 (2021).
### [Hybrid Transformers for Music Source Separation](https://arxiv.org/pdf/2211.08553.pdf)
Rouard, Simon, Francisco Massa, and Alexandre Défossez. "Hybrid transformers for music source separation." In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1-5. IEEE, 2023.
### [Kuielab-mdx-net: A two-stream neural network for music demixing](https://arxiv.org/pdf/2111.12203.pdf)
Kim, Minseok, Woosung Choi, Jaehwa Chung, Daewon Lee, and Soonyoung Jung. "Kuielab-mdx-net: A two-stream neural network for music demixing." arXiv preprint arXiv:2111.12203 (2021)
### [D3net: Densely connected multidilated densenet for music source separation](https://arxiv.org/pdf/2010.01733.pdf)
Takahashi, Naoya, and Yuki Mitsufuji. "D3net: Densely connected multidilated densenet for music source separation." arXiv preprint arXiv:2010.01733 (2020).
### [Decoupling Magnitude and Phase Estimation with Deep ResUNet for Music Source Separation](https://arxiv.org/pdf/2109.05418.pdf)
Qiuqiang Kong, , Yin Cao, Haohe Liu, Keunwoo Choi, and Yuxuan Wang. "Decoupling Magnitude and Phase Estimation with Deep ResUNet for Music Source Separation." (2021).
### [On Training Targets for Supervised Speech Separation](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4293540/pdf/nihms642326.pdf)
Y. Wang, A. Narayanan and D. Wang, "On Training Targets for Supervised Speech Separation," in IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 12, pp. 1849-1858, Dec. 2014, doi: 10.1109/TASLP.2014.2352935.
### [Supervised Speech Separation Based on Deep Learning: An Overview](https://web.cse.ohio-state.edu/~wang.77/papers/Wang-Chen.taslp18.pdf)
D. Wang and J. Chen, "Supervised Speech Separation Based on Deep Learning: An Overview," in IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 26, no. 10, pp. 1702-1726, Oct. 2018, doi: 10.1109/TASLP.2018.2842159.
### [Phase-sensitive and recognition-boosted speech separation using deep recurrent neural networks](https://www.jonathanleroux.org/pdf/Erdogan2015ICASSP04.pdf)
H. Erdogan, J. R. Hershey, S. Watanabe and J. Le Roux, "Phase-sensitive and recognition-boosted speech separation using deep recurrent neural networks," 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 2015, pp. 708-712, doi: 10.1109/ICASSP.2015.7178061.
### [使用 U-Net 及其壓縮版本來進行歌聲分離](https://www.airitilibrary.com/Publication/alDetailedMesh?docid=U0001-2408202115100900)
王俞禮. "使用 U-Net 及其壓縮版本來進行歌聲分離." (2021): 1-68.
### [改良 U-Net 對歌曲人聲分離效果](https://www.airitilibrary.com/Publication/alDetailedMesh?docid=U0001-2007202015240300)
黃翔宇. "改良 U-Net 對歌曲人聲分離效果." (2020): 1-50.
## Dataset
### [MUSDB18 - a corpus for music separation](https://hal.inria.fr/hal-02190845/document)
Zafar Rafii, Antoine Liutkus, Fabian-Robert Stöter, Stylianos Ioannis Mimilakis, Rachel Bittner. MUSDB18 - a corpus for music separation. 2017.
## Augmentation
### [Improving music source separation based on deep neural networks through data augmentation and network blending](https://ieeexplore.ieee.org/document/7952158)
S. Uhlich et al., "Improving music source separation based on deep neural networks through data augmentation and network blending," 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 2017, pp. 261-265, doi: 10.1109/ICASSP.2017.7952158.
### [Improved singing voice separation with chromagram-based pitch-aware remixing](https://arxiv.org/pdf/2203.15092.pdf)
S. Yuan et al., "Improved Singing Voice Separation with Chromagram-Based Pitch-Aware Remixing," ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, Singapore, 2022, pp. 111-115, doi: 10.1109/ICASSP43922.2022.9747612.
## Metric
### [Performance measurement in blind audio source separation](https://hal.inria.fr/inria-00544230/document)
E. Vincent, R. Gribonval and C. Fevotte, "Performance measurement in blind audio source separation," in IEEE Transactions on Audio, Speech, and Language Processing, vol. 14, no. 4, pp. 1462-1469, July 2006, doi: 10.1109/TSA.2005.858005.
## Chord Estimation
### [MIREX 2018 submission: A structural chord representation for automatic large-vocabulary chord transcription](https://www.music-ir.org/mirex/abstracts/2018/JLCX1.pdf)
Jiang, Junyan, Ke Chen, Wei Li, and Guangyu Xia. "MIREX 2018 submission: A structural chord representation for automatic large-vocabulary chord transcription." Proceedings of the Music Information Retrieval Evaluation eXchange (2018).
### [Symbolic Representation of Musical Chords: A Proposed Syntax for Text Annotations](https://archives.ismir.net/ismir2005/paper/000080.pdf)
Harte, Christopher, Mark B. Sandler, Samer A. Abdallah, and Emilia Gómez. "Symbolic Representation of Musical Chords: A Proposed Syntax for Text Annotations." In ISMIR, vol. 5, pp. 66-71. 2005.
## Genre Classification
### [On the complexity analysis and visualization of musical information](https://www.mdpi.com/1099-4300/21/7/669)
Lopes, António M., and J. A. Tenreiro Machado. "On the complexity analysis and visualization of musical information." Entropy 21, no. 7 (2019): 669.