- Natural Language Processing with TensorFlow
- Thushan Ganegedara
- 216字
- 2021-06-25 21:28:24
Chapter 4. Advanced Word2vec
In Chapter 3, Word2vec – Learning Word Embeddings, we introduced you to Word2vec, the basics of learning word embeddings, and the two common Word2vec algorithms: skip-gram and CBOW. In this chapter, we will discuss several topics related to Word2vec, focusing on these two algorithms and extensions.
First, we will explore how the original skip-gram algorithm was implemented and how it compares to its more modern variant, which we used in Chapter 3, Word2vec – Learning Word Embeddings. We will examine the differences between skip-gram and CBOW and look at the behavior of the loss over time of the two approaches. We will also discuss which method works better, using both our observation and the available literature.
We will discuss several extensions to the existing Word2vec methods that boost performance. These extensions include using more effective sampling techniques to sample negative examples for negative sampling and ignoring uninformative words in the learning process, among others. You will also learn a novel word embedding learning technique known as Global Vectors (GloVe) and the specific advantages that GloVe has over skip-gram and CBOW.
Finally, you will learn how to use Word2vec to solve a real-world problem: document classification. We will see this with a simple trick of obtaining document embeddings from word embeddings.
- Learning Cython Programming
- AWS Serverless架構:使用AWS從傳統部署方式向Serverless架構遷移
- R語言游戲數據分析與挖掘
- 焊接機器人系統操作、編程與維護
- Apache Spark 2.x for Java Developers
- Natural Language Processing with Java and LingPipe Cookbook
- Extreme C
- Web前端應用開發技術
- CRYENGINE Game Development Blueprints
- 創意UI:Photoshop玩轉APP設計
- C#面向對象程序設計(第2版)
- 嵌入式Linux C語言程序設計基礎教程
- Java EE項目應用開發
- IBM DB2 9.7 Advanced Application Developer Cookbook
- HTML5+CSS+JavaScript深入學習實錄