최소 단어 이상 선택하여야 합니다.
최대 10 단어까지만 선택 가능합니다.
다음과 같은 기능을 한번의 로그인으로 사용 할 수 있습니다.
NTIS 바로가기응용통계연구 = The Korean journal of applied statistics, v.35 no.4, 2022년, pp.469 - 484
문혜인 (단국대학교 대학원 응용통계학과) , 손원 (단국대학교 대학원 응용통계학과)
Since a large text corpus contains hundred-thousand unique words, text data is one of the typical large-dimensional data. Therefore, various feature selection methods have been proposed for dimension reduction. Feature selection methods can improve the prediction accuracy. In addition, with reduced ...
Bird S, Klein E, and Loper E (2009). Natural language processing with Python: analyzing text with the natural language toolkit. O'Reilly.
Forman G (2003). An extensive empirical study of feature selection metrics for text classification. Journal of Machine Learning Research, 3, 1289-1305.
Kou G, Yang P, Peng Y, Xiao F, Chen Y, and Alsaadi FE (2020). Evaluation of feature selection methods for text classification with small datasets using multiple criteria decision-making methods. Applied Soft Computing, 86, 105836.
Manning C and Schutze H (1999). Foundations of statistical natural language processing. MIT press.
Mladenic D and Grobelnik M (1999). Feature selection for unbalanced class distribution and naive bayes. In Proceedings of the 16th International Conference on Machine Learning (ICML). 258-267.
Rao JNK and Scott AJ (1987). On simple adjustments to chi-square tests with sample survey data. The annals of statistics, 15(1), 385-397.
SonW(2020). Skewness of chi-square statistic for imbalanced text data. Journal of the Korean Data & Information Science Society, 31(5), 807-821.
Yang Y and Pedersen JO (1997). A comparative study on feature selection in text categorization. In Proceedings of the 14th International Conference on Machine Learning (ICML). 412-420.
*원문 PDF 파일 및 링크정보가 존재하지 않을 경우 KISTI DDS 시스템에서 제공하는 원문복사서비스를 사용할 수 있습니다.
오픈액세스 학술지에 출판된 논문
※ AI-Helper는 부적절한 답변을 할 수 있습니다.