Learning Word Embedding
Published:Oct 15, 2017 00:00
•1 min read
•Lil'Log
Analysis
The article provides a concise introduction to word embeddings, specifically focusing on the need to convert text into numerical representations for machine learning. It highlights one-hot encoding as a basic method. The explanation is clear and suitable for a beginner audience.
Key Takeaways
Reference
“One of the simplest transformation approaches is to do a one-hot encoding in which each distinct word stands for one dimension of the resulting vector and a binary value indicates whether the word presents (1) or not (0).”