- Natural Language Processing with TensorFlow
- Thushan Ganegedara
- 216字
- 2021-06-25 21:28:24
Chapter 4. Advanced Word2vec
In Chapter 3, Word2vec – Learning Word Embeddings, we introduced you to Word2vec, the basics of learning word embeddings, and the two common Word2vec algorithms: skip-gram and CBOW. In this chapter, we will discuss several topics related to Word2vec, focusing on these two algorithms and extensions.
First, we will explore how the original skip-gram algorithm was implemented and how it compares to its more modern variant, which we used in Chapter 3, Word2vec – Learning Word Embeddings. We will examine the differences between skip-gram and CBOW and look at the behavior of the loss over time of the two approaches. We will also discuss which method works better, using both our observation and the available literature.
We will discuss several extensions to the existing Word2vec methods that boost performance. These extensions include using more effective sampling techniques to sample negative examples for negative sampling and ignoring uninformative words in the learning process, among others. You will also learn a novel word embedding learning technique known as Global Vectors (GloVe) and the specific advantages that GloVe has over skip-gram and CBOW.
Finally, you will learn how to use Word2vec to solve a real-world problem: document classification. We will see this with a simple trick of obtaining document embeddings from word embeddings.
- Python自然語言處理實戰:核心技術與算法
- Java加密與解密的藝術(第2版)
- Raspberry Pi for Secret Agents(Third Edition)
- Haxe Game Development Essentials
- Spring Boot進階:原理、實戰與面試題分析
- Instant Ext.NET Application Development
- 智能搜索和推薦系統:原理、算法與應用
- Python自然語言理解:自然語言理解系統開發與應用實戰
- Python Web自動化測試設計與實現
- 深度學習入門:基于Python的理論與實現
- JSP應用與開發技術(第3版)
- Python程序設計:基礎與實踐
- Spring Boot 2+Thymeleaf企業應用實戰
- Mastering Unity 2017 Game Development with C#(Second Edition)
- Java程序性能優化實戰