RISS 학술연구정보서비스

검색
다국어 입력

http://chineseinput.net/에서 pinyin(병음)방식으로 중국어를 변환할 수 있습니다.

변환된 중국어를 복사하여 사용하시면 됩니다.

예시)
  • 中文 을 입력하시려면 zhongwen을 입력하시고 space를누르시면됩니다.
  • 北京 을 입력하시려면 beijing을 입력하시고 space를 누르시면 됩니다.
닫기
    인기검색어 순위 펼치기

    RISS 인기검색어

      KCI우수등재

      문서 분류를 위한 신경망 모델에 적합한 텍스트 전처리와 워드 임베딩의 조합

      한글로보기

      https://www.riss.kr/link?id=A105451690

      • 0

        상세조회
      • 0

        다운로드
      서지정보 열기
      • 내보내기
      • 내책장담기
      • 공유하기
      • 오류접수

      부가정보

      다국어 초록 (Multilingual Abstract)

      Neural networks with word embedding have recently used for document classification. Researchers concentrate on designing new architecture or optimizing model parameters to increase performance. However, most recent studies have overlooked text preproc...

      Neural networks with word embedding have recently used for document classification. Researchers concentrate on designing new architecture or optimizing model parameters to increase performance. However, most recent studies have overlooked text preprocessing and word embedding, in that the description of text preprocessing used is insufficient, and a certain pretrained word embedding model is mostly used without any plausible reasons. Our paper shows that finding a suitable combination of text preprocessing and word embedding can be one of the important factors required to enhance the performance. We conducted experiments on AG’s News dataset to compare those possible combinations, and zero/random padding, and presence or absence of fine-tuning. We used pretrained word embedding models such as skip-gram, GloVe, and fastText. For diversity, we also use an average of multiple pretrained embeddings (Average), randomly initialized embedding (Random), task data-trained skip-gram (AGNews-Skip). In addition, we used three advanced neural networks for the sake of generality. Experimental results based on OOV (Out Of Vocabulary) word statistics suggest the necessity of those comparisons and a suitable combination of text preprocessing and word embedding.

      더보기

      참고문헌 (Reference)

      1 "word2vec: tool fot computing continuous distributed representations of words"

      2 "fastText: wiki word vectors"

      3 Conneau, Alexis, "Very deep convolutional networks for natural language processing"

      4 Lai, Siwei, "Recurrent Convolutional Neural Networks for Text Classification" 333 : 2015

      5 Camacho-Collados, Jose, "On the Role of Text Preprocessing in Neural Network Architectures: An Evaluation Study on Text Categorization and Sentiment Analysis"

      6 Chen, Huimin, "Neural sentiment classification with user and product attention" 2016

      7 Bahdanau, Dzmitry, "Neural machine translation by jointly learning to align and translate"

      8 Loper, Edward, "NLTK: The natural language toolkit" Association for Computational Linguistics 2002

      9 Lei, Tao, "Molding cnns for text: non-linear, non-consecutive convolutions"

      10 Hochreiter, Sepp, "Long short-term memory" 9 (9): 1735-1780, 1997

      1 "word2vec: tool fot computing continuous distributed representations of words"

      2 "fastText: wiki word vectors"

      3 Conneau, Alexis, "Very deep convolutional networks for natural language processing"

      4 Lai, Siwei, "Recurrent Convolutional Neural Networks for Text Classification" 333 : 2015

      5 Camacho-Collados, Jose, "On the Role of Text Preprocessing in Neural Network Architectures: An Evaluation Study on Text Categorization and Sentiment Analysis"

      6 Chen, Huimin, "Neural sentiment classification with user and product attention" 2016

      7 Bahdanau, Dzmitry, "Neural machine translation by jointly learning to align and translate"

      8 Loper, Edward, "NLTK: The natural language toolkit" Association for Computational Linguistics 2002

      9 Lei, Tao, "Molding cnns for text: non-linear, non-consecutive convolutions"

      10 Hochreiter, Sepp, "Long short-term memory" 9 (9): 1735-1780, 1997

      11 Wen, Ying, "Learning text representation using recurrent convolutional neural network with highway layers"

      12 Yang, Zichao, "Hierarchical attention networks for document classification" 2016

      13 Kowsari, Kamran, "Hdltex: Hierarchical deep learning for text classification"

      14 Pennington, Jeffrey, "Glove: Global vectors for word representation" 2014

      15 "GloVe: Global Vectors for Word Representation"

      16 Raffel, Colin, "Feed-forward networks with attention can solve some long-term memory problems"

      17 Bojanowski, Piotr, "Enriching word vectors with subword information"

      18 Xiao, Yijun, "Efficient character-level document classification by combining convolution and recurrent layers"

      19 Tang, Duyu, "Document modeling with gated recurrent neural network for sentiment classification" 2015

      20 Mikolov, Tomas, "Distributed representations of words and phrases and their compositionality, Advances in neural information processing systems" 2013

      21 Johnson, Rie, "Deep pyramid convolutional neural networks for text categorization" 1 : 2017

      22 Kim, Yoon, "Convolutional neural networks for sentence classification"

      23 Zhang, Xiang, "Character-level convolutional networks for text classification;Advances in neural information processing systems" 2015

      24 Zhang, Ye, "A sensitivity analysis of (and practitioners' guide to) convolutional neural networks for sentence classification"

      25 Zhou, Chunting, "A C-LSTM neural network for text classification"

      더보기

      분석정보

      View

      상세정보조회

      0

      Usage

      원문다운로드

      0

      대출신청

      0

      복사신청

      0

      EDDS신청

      0

      동일 주제 내 활용도 TOP

      더보기

      주제

      연도별 연구동향

      연도별 활용동향

      연관논문

      연구자 네트워크맵

      공동연구자 (7)

      유사연구자 (20) 활용도상위20명

      인용정보 인용지수 설명보기

      학술지 이력

      학술지 이력
      연월일 이력구분 이력상세 등재구분
      2021 평가예정 계속평가 신청대상 (등재유지)
      2016-01-01 평가 우수등재학술지 선정 (계속평가)
      2015-01-01 평가 등재학술지 유지 (등재유지) KCI등재
      2002-01-01 평가 학술지 통합 (등재유지) KCI등재
      더보기

      학술지 인용정보

      학술지 인용정보
      기준연도 WOS-KCI 통합IF(2년) KCIF(2년) KCIF(3년)
      2016 0.19 0.19 0.19
      KCIF(4년) KCIF(5년) 중심성지수(3년) 즉시성지수
      0.2 0.18 0.373 0.07
      더보기

      이 자료와 함께 이용한 RISS 자료

      나만을 위한 추천자료

      해외이동버튼