graph deep learning/#6 Graph Recurrent Networks
6.4 Sentence LSTM
yuuuun
2020. 11. 22. 14:54
반응형
- Text Encoding을 향상시키기 위하여 S-LSTM이 제안되었음
- text를 graph로 변환한 뒤, representation을 학습하여 Graph-LSTM을 용이하게 함.
- S-LSTM model regards each word as a node in the graph and it adds a supernode.
- For each node, the word node could aggregate information from its adjacent words as well as the supernode.
- The supernode could aggregate information from all of the word nodes as well as itself.
- NLP
- Hidden states of words can be used to solve the word-level tasks such as sequence labeling, POS tagging and so on.
반응형