000 04396nam a22005775i 4500
001 978-981-99-1600-9
003 DE-He213
005 20240507154521.0
007 cr nn 008mamaa
008 230823s2023 si | s |||| 0|eng d
020 _a9789819916009
_9978-981-99-1600-9
024 7 _a10.1007/978-981-99-1600-9
_2doi
050 4 _aQA76.9.N38
072 7 _aUYQL
_2bicssc
072 7 _aCOM073000
_2bisacsh
072 7 _aUYQL
_2thema
082 0 4 _a006.35
_223
245 1 0 _aRepresentation Learning for Natural Language Processing
_h[electronic resource] /
_cedited by Zhiyuan Liu, Yankai Lin, Maosong Sun.
250 _a2nd ed. 2023.
264 1 _aSingapore :
_bSpringer Nature Singapore :
_bImprint: Springer,
_c2023.
300 _aXX, 521 p. 169 illus. in color.
_bonline resource.
336 _atext
_btxt
_2rdacontent
337 _acomputer
_bc
_2rdamedia
338 _aonline resource
_bcr
_2rdacarrier
347 _atext file
_bPDF
_2rda
505 0 _aChapter 1. Representation Learning and NLP -- Chapter 2. Word Representation -- Chapter 3. Compositional Semantics -- Chapter 4. Sentence Representation -- Chapter 5. Document Representation -- Chapter 6. Sememe Knowledge Representation -- Chapter 7. World Knowledge Representation -- Chapter 8. Network Representation -- Chapter 9. Cross-Modal Representation -- Chapter 10. Resources -- Chapter 11. Outlook.
506 0 _aOpen Access
520 _aThis book provides an overview of the recent advances in representation learning theory, algorithms, and applications for natural language processing (NLP), ranging from word embeddings to pre-trained language models. It is divided into four parts. Part I presents the representation learning techniques for multiple language entries, including words, sentences and documents, as well as pre-training techniques. Part II then introduces the related representation techniques to NLP, including graphs, cross-modal entries, and robustness. Part III then introduces the representation techniques for the knowledge that are closely related to NLP, including entity-based world knowledge, sememe-based linguistic knowledge, legal domain knowledge and biomedical domain knowledge. Lastly, Part IV discusses the remaining challenges and future research directions. The theories and algorithms of representation learning presented can also benefit other related domains such as machine learning, social network analysis, semantic Web, information retrieval, data mining and computational biology. This book is intended for advanced undergraduate and graduate students, post-doctoral fellows, researchers, lecturers, and industrial engineers, as well as anyone interested in representation learning and natural language processing. As compared to the first edition, the second edition (1) provides a more detailed introduction to representation learning in Chapter 1; (2) adds four new chapters to introduce pre-trained language models, robust representation learning, legal knowledge representation learning and biomedical knowledge representation learning; (3) updates recent advances in representation learning in all chapters; and (4) corrects some errors in the first edition. The new contents will be approximately 50%+ compared to the first edition. This is an open access book.
650 0 _aNatural language processing (Computer science).
650 0 _aComputational linguistics.
650 0 _aArtificial intelligence.
650 0 _aData mining.
650 1 4 _aNatural Language Processing (NLP).
650 2 4 _aComputational Linguistics.
650 2 4 _aArtificial Intelligence.
650 2 4 _aData Mining and Knowledge Discovery.
700 1 _aLiu, Zhiyuan.
_eeditor.
_4edt
_4http://id.loc.gov/vocabulary/relators/edt
700 1 _aLin, Yankai.
_eeditor.
_4edt
_4http://id.loc.gov/vocabulary/relators/edt
700 1 _aSun, Maosong.
_eeditor.
_4edt
_4http://id.loc.gov/vocabulary/relators/edt
710 2 _aSpringerLink (Online service)
773 0 _tSpringer Nature eBook
776 0 8 _iPrinted edition:
_z9789819915996
776 0 8 _iPrinted edition:
_z9789819916016
776 0 8 _iPrinted edition:
_z9789819916023
856 4 0 _uhttps://doi.org/10.1007/978-981-99-1600-9
912 _aZDB-2-SCS
912 _aZDB-2-SXCS
912 _aZDB-2-SOB
999 _c37464
_d37464