$\require{mediawiki-texvc}$

연합인증

연합인증 가입 기관의 연구자들은 소속기관의 인증정보(ID와 암호)를 이용해 다른 대학, 연구기관, 서비스 공급자의 다양한 온라인 자원과 연구 데이터를 이용할 수 있습니다.

이는 여행자가 자국에서 발행 받은 여권으로 세계 각국을 자유롭게 여행할 수 있는 것과 같습니다.

연합인증으로 이용이 가능한 서비스는 NTIS, DataON, Edison, Kafe, Webinar 등이 있습니다.

한번의 인증절차만으로 연합인증 가입 서비스에 추가 로그인 없이 이용이 가능합니다.

다만, 연합인증을 위해서는 최초 1회만 인증 절차가 필요합니다. (회원이 아닐 경우 회원 가입이 필요합니다.)

연합인증 절차는 다음과 같습니다.

최초이용시에는
ScienceON에 로그인 → 연합인증 서비스 접속 → 로그인 (본인 확인 또는 회원가입) → 서비스 이용

그 이후에는
ScienceON 로그인 → 연합인증 서비스 접속 → 서비스 이용

연합인증을 활용하시면 KISTI가 제공하는 다양한 서비스를 편리하게 이용하실 수 있습니다.

Abstract AI-Helper 아이콘AI-Helper

Recent developments in the field of separation of mixed signals into music/voice components have attracted the attention of many researchers. Recently, iterative kernel back-fitting, also known as kernel additive modeling, was proposed to achieve good results for music/voice separation. To obtain mi...

주제어

AI 본문요약
AI-Helper 아이콘 AI-Helper

* AI 자동 식별 결과로 적합하지 않은 문장이 있을 수 있으니, 이용에 유의하시기 바랍니다.

제안 방법

  • From the complex spectrogram X of the input music signal, each complex spectrogram, SV, SH, and SP, for the vocal, harmonic, and percussive components is estimated by each generalized WbE, GV, GH, and GP, of decomposed spectral amplitude by singular value decomposition (SVD) for the vocal, harmonic, and percussive components, respectively. The WbE estimation gain, Gj, for each source j (= 0, 1, 2, … , J) is explained in detail in Algorithm 2.
  • In this paper, a generalized weighted β-order MMSE estimation (WbE) method based on kernel back-fitting (KBF) was proposed and evaluated for the separation of mixed signals into music/voice components.
  • In this paper, an advanced music/voice separation method is proposed, in which WbE and KBF are combined for improvement of the separation performance.
  • The proposed estimation method takes full advantage of both a generalized weighted β-order spectral amplitude estimator and an SVD-based subspace decomposition.

대상 데이터

  • For the first experiment, 150 full-length song tracks [23] were used (50 songs from the ccMixter database containing many different musical genres, 50 songs from a self-recording studio music database, and 50 songs from the MIR-1 K database), where all singing voices and music accompaniments were recorded separately. All of the song data were stored in PCM format with mono, 16-bit depth, and 44.
  • For the second performance comparison, the proposed algorithm, SVD-WbE-KAM, was compared with REPETSIM [26], RPCA [27], and SVD-GW-KAM. To evaluate the separation of background music and singing voice, 40 fulllength song tracks [24] were used (20 songs from the ccMixter database containing many different musical genres, and 20 songs from the MIR-1 K database). Figures 1 and 2 show boxplots of the SDR for the vocals and the music accompaniment, respectively.
본문요약 정보가 도움이 되었나요?

참고문헌 (27)

  1. Z. Rafii and B. Pardo, "REpeating Pattern Extraction Technique (REPET): A Simple Method for Music/Voice Separation," IEEE Trans. Audio, Speech, Language Process., vol. 21, no. 1, Jan. 2013, pp. 73-84. 

  2. N.C. Maddage, C. Xu, and Y. Wang, "Singer Identification Based on Vocal and Instrumental Models," Proc. Int. Conf. Pattern Recogn., Cambridge, UK, Aug. 23-26, 2004, pp. 375-378. 

  3. M. Ryynanen and A. Klapuri, "Transcription of the Singing Melody in Polyphonic Music," Int. Conf. Music Inf. Retrieval, Victoria, Canada, Oct. 8-12, 2006, pp. 222-227. 

  4. S. Marchand et al., "DReaM: A Novel System for Joint Source Separation and Multi-track Coding," 133rd AES Conv., San Francisco, CA, USA, Oct. 26-29, 2012. 

  5. J. Nikunen, T. Virtanen, and M. Vilermo, "Multichannel Audio Upmixing Based on Non-negative Tensor Factorization Representation," IEEE Workshop Appl. Signal Process. Audio Acoust., New Paltz, NY, USA, Oct. 16-19, 2011, pp. 33-36. 

  6. U. Simsekli, Y.K. Yilmaz, and A.T. Cemgil, "Score Guided Audio Restoration via Generalized Coupled Tensor Factorisation," IEEE Int. Conf. Acoust., Speech Signal Process., Kyoto, Japan, Mar. 25-30, 2012, pp. 5369-5372. 

  7. J.L. Durrieu, B. David, and G. Richard, "A Musically Motivated Mid-level Representation for Pitch Estimation and Musical Audio Source Separation," IEEE J. Sel. Topics Signal Process., vol. 5, no. 6, Oct. 2011, pp. 1180-1191. 

  8. C.L. Hsu and J.S.R. Jang, "On the Improvement of Singing Voice Separation for Monaural Recordings Using the MIR-1K Dataset," IEEE Trans. Audio, Speech, Language Process., vol. 18, no. 2, Feb. 2010, pp. 310-319. 

  9. T. Virtanen, A. Mesaros, and M. Ryynanen, "Combining Pitch-Based Inference and Non-negative Spectrogram Factorization in Separating Vocals from Polyphonic Music," ISCA Tutorial Res. Workshop Statistical Perceptual Audition, Brisbane, Australia, Sept. 21, 2008, pp. 17-22. 

  10. A. Liutkus et al., "Kernel Additive Models for Source Separation," IEEE Trans. Signal Process., vol. 62, no. 16, Aug. 2014, pp. 4298-4310. 

  11. D. Fitzgerald, "Harmonic/Percussive Separation Using Median Filtering," Int. Conf. Digital Audio Effects, Graz, Austria, Sept. 6-10, 2010, pp. 1-4. 

  12. Z. Rafii and B. Pardo, "A Simple Music/Voice Separation Method Based on the Extraction of the Repeating Musical Structure," IEEE Int. Conf. Acoust., Speech Signal Process., Prague, Czech Republic, May 22-27, 2011, pp. 221-224. 

  13. A. Liutkus et al., "Adaptive Filtering for Music/Voice Separation Exploiting the Repeating Musical Structure," IEEE Int. Conf. Acoust., Speech Signal Process., Kyoto, Japan, Mar. 25-30, 2012, pp. 53-56. 

  14. Z. Rafii and B. Pardo, "Music/Voice Separation Using the Similarity Matrix," Int. Conf. Music Inf. Retrieval, Porto, Portugal, Oct. 8-12, 2012, pp. 583-588. 

  15. O. Yilmaz and S. Rickard, "Blind Separation of Speech Mixtures via Time-Frequency Masking," IEEE Trans. Signal Process., vol. 52, no. 7, July 2004, pp. 1830-1847. 

  16. Y. Ephraim and D. Malah, "Speech Enhancement Using a Minimum Mean-Square Error Short-Time Spectral Amplitude Estimator," IEEE Trans. Acoust., Speech, Signal Process., vol. 32, no. 6, Dec. 1984, pp. 1109-1121. 

  17. E. Plourde and B. Champagne, "Auditory-Based Spectral Amplitude Estimators for Speech Enhancement," IEEE Trans. Audio, Speech, Language Process., vol. 16, no. 8, Nov. 2008, pp. 1614-1623. 

  18. C.H. You, S.N. Koh, and S. Rahardja, " ${\beta}$ -Order MMSE Spectral Amplitude Estimation for Speech Enhancement," IEEE Trans. Speech, Audio Process., vol. 13, no. 4, July. 2005, pp. 475-486. 

  19. F. Deng, F. Bao, and C.-C. Bao, "Speech Enhancement Using Generalized ${\beta}$ -Order Spectral Amplitude Estimator," Speech Commun., vol. 59, Apr. 2014, pp. 55-68. 

  20. C.H. You, S.N. Koh, and S. Rahardja, "Masking-Based ${\beta}$ -Order MMSE Speech Enhancement," Speech Commun., vol. 48, no. 1, Jan. 2006, pp. 57-70. 

  21. C.H. You, S.N. Koh, and S. Rahardja, "Improved Adaptive ${\beta}$ - Order MMSE Speech Enhancement," APSIPA Ann. Summit Conf., Sapporo, Japan, Oct. 4-7, 2009, pp. 797-800. 

  22. D.D. Greenwood, "A Cochlear Frequency-Position Function for Several Species-29 Years Later," J. Acoust. Soc. America, vol. 87, no. 6, July 1990, pp. 2592-2605. 

  23. Multimedia Technology Laboratory homepage, Accessed Nov. 20, 2015. http://imsp.kw.ac.kr/Research.html 

  24. E. Vincent, R. Gribonval, and C. Fevotte, "Performance Measurement in Blind Audio Source Separation," IEEE Trans. Audio, Speech, Language Process., vol. 14, no. 4, July 2006, pp. 1462-1469. 

  25. R.C. Hendriks et al., "Minimum Mean-Square Error Amplitude Estimators for Speech Enhancement under the Generalized Gamma Distribution," Int. Workshop Acoust. Echo Noise Contr., Paris, France, Sept. 12-14, 2006, pp. 1-4. 

  26. Z. Rafii, A. Liutkus, and B. Pardo, "REPET for Background/Foreground Separation in Audio," in Blind Source Separation: Advances in Theory, Algorithms and Appl., Berlin, Germany: Springer, 2014, pp. 395-411. 

  27. P.S. Huang et al., "Singing-Voice Separation from Monaural Recordings Using Robust Principal Component Analysis," IEEE Int. Conf. Acoust., Speech Signal Process., Kyoto, Japan, Mar. 25-30, 2012, pp. 57-60. 

저자의 다른 논문 :

LOADING...

관련 콘텐츠

오픈액세스(OA) 유형

BRONZE

출판사/학술단체 등이 한시적으로 특별한 프로모션 또는 일정기간 경과 후 접근을 허용하여, 출판사/학술단체 등의 사이트에서 이용 가능한 논문

유발과제정보 저작권 관리 안내
섹션별 컨텐츠 바로가기

AI-Helper ※ AI-Helper는 오픈소스 모델을 사용합니다.

AI-Helper 아이콘
AI-Helper
안녕하세요, AI-Helper입니다. 좌측 "선택된 텍스트"에서 텍스트를 선택하여 요약, 번역, 용어설명을 실행하세요.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.

선택된 텍스트

맨위로