來源:深度學習與NLP
文章來源:微信公衆号 資料派THU
本文收集整理了一批基于Tensorflow實作的深度學習/機器學習的深度NLP模型。
收集整理了一批基于Tensorflow實作的深度學習/機器學習的深度NLP模型。
基于Tensorflow的自然語言處理模型,為自然語言處理問題收集機器學習和Tensorflow深度學習模型,100%Jupeyter NoteBooks且内部代碼極為簡潔。
資源整理自網絡,源位址:
https://github.com/huseinzol05目錄
- Text classification
- Chatbot
- Neural Machine Translation
- Embedded
- Entity-Tagging
- POS-Tagging
- Dependency-Parser
- Question-Answers
- Supervised Summarization
- Unsupervised Summarization
- Stemming
- Generator
- Language detection
- OCR (optical character recognition)
- Speech to Text
- Text to Speech
- Text Similarity
- Miscellaneous
- Attention
目标
原始的實作稍微有點複雜,對于初學者來說有點難。是以我嘗試将其中大部分内容簡化,同時,還有很多論文的内容亟待實作,一步一步來。
内容
文本分類:
連結:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/text-classification- Basic cell RNN
- Bidirectional RNN
- LSTM cell RNN
- GRU cell RNN
- LSTM RNN + Conv2D
- K-max Conv1d
- LSTM RNN + Conv1D + Highway
- LSTM RNN with Attention
- Neural Turing Machine
- Seq2Seq
- Bidirectional Transformers
- Dynamic Memory Network
- Residual Network using Atrous CNN + Bahdanau Attention
- Transformer-XL
完整清單包含(66 notebooks)
聊天機器人:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/chatbot- Seq2Seq-manual
- Seq2Seq-API Greedy
- Bidirectional Seq2Seq-manual
- Bidirectional Seq2Seq-API Greedy
- Bidirectional Seq2Seq-manual + backward Bahdanau + forward Luong
- Bidirectional Seq2Seq-API + backward Bahdanau + forward Luong + Stack Bahdanau Luong Attention + Beam Decoder
- Bytenet
- Capsule layers + LSTM Seq2Seq-API + Luong Attention + Beam Decoder
- End-to-End Memory Network
- Attention is All you need
- Transformer-XL + LSTM
- GPT-2 + LSTM
完整清單包含(51 notebooks)
機器翻譯(英語到越南語):
https://github.com/huseinzol05/NLP-ModelsTensorflow/tree/master/neural-machine-translation完整清單包含(49 notebooks)
詞向量:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/embedded- Word Vector using CBOW sample softmax
- Word Vector using CBOW noise contrastive estimation
- Word Vector using skipgram sample softmax
- Word Vector using skipgram noise contrastive estimation
- Lda2Vec Tensorflow
- Supervised Embedded
- Triplet-loss + LSTM
- LSTM Auto-Encoder
- Batch-All Triplet-loss LSTM
- Fast-text
- ELMO (biLM)
詞性标注:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/pos-tagging- Bidirectional RNN + Bahdanau Attention + CRF
- Bidirectional RNN + Luong Attention + CRF
- Bidirectional RNN + CRF
實體識别:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/entity-tagging- Char Ngrams + Bidirectional RNN + Bahdanau Attention + CRF
- Char Ngrams + Residual Network + Bahdanau Attention + CRF
依存分析:
https://github.com/huseinzol05/NLP-ModelsTensorflow/tree/master/dependency-parser- Residual Network + Bahdanau Attention + CRF
- Residual Network + Bahdanau Attention + Char Embedded + CRF
問答:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/question-answer- End-to-End Memory Network + Basic cell
- End-to-End Memory Network + GRU cell
- End-to-End Memory Network + LSTM cell
詞幹抽取:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/stemming- LSTM + Seq2Seq + Beam
- GRU + Seq2Seq + Beam
- LSTM + BiRNN + Seq2Seq + Beam
- GRU + BiRNN + Seq2Seq + Beam
- DNC + Seq2Seq + Greedy
有監督摘要抽取:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/summarization- LSTM Seq2Seq using topic modelling
- LSTM Seq2Seq + Luong Attention using topic modelling
- LSTM Seq2Seq + Beam Decoder using topic modelling
- LSTM Bidirectional + Luong Attention + Beam Decoder using topic modelling
- LSTM Seq2Seq + Luong Attention + Pointer Generator
無監督摘要抽取:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/unsupervised-summarization- Skip-thought Vector (unsupervised)
- Residual Network using Atrous CNN (unsupervised)
- Residual Network using Atrous CNN + Bahdanau Attention (unsupervised)
**OCR (字元識别):
**
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/ocr- CNN + LSTM RNN
語音識别:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/speech-to-text- Tacotron
- Bidirectional RNN + Greedy CTC
- Bidirectional RNN + Beam CTC
- Seq2Seq + Bahdanau Attention + Beam CTC
- Seq2Seq + Luong Attention + Beam CTC
- Bidirectional RNN + Attention + Beam CTC
- Wavenet
語音合成:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/text-to-speech- Seq2Seq + Luong Attention
- Seq2Seq + Bahdanau Attention
生成器:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/generator- Character-wise RNN + LSTM
- Character-wise RNN + Beam search
- Character-wise RNN + LSTM + Embedding
- Word-wise RNN + LSTM
- Word-wise RNN + LSTM + Embedding
- Character-wise + Seq2Seq + GRU
- Word-wise + Seq2Seq + GRU
- Character-wise RNN + LSTM + Bahdanau Attention
- Character-wise RNN + LSTM + Luong Attention
語言檢測:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/language-detection- Fast-text Char N-Grams
文本相似性:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/text-similarity- Character wise similarity + LSTM + Bidirectional
- Word wise similarity + LSTM + Bidirectional
- Character wise similarity Triplet loss + LSTM
- Word wise similarity Triplet loss + LSTM
注意力機制:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/attention- Bahdanau
- Luong
- Hierarchical
- Additive
- Soft
- Attention-over-Attention
- Bahdanau API
- Luong API
其他:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/misc- Attention heatmap on Bahdanau Attention
- Attention heatmap on Luong Attention
非深度學習:
https://github.com/huseinzol05/NLP-Models-Tensorflow/tree/master/not-deep-learning- Markov chatbot
- Decomposition summarization (3 notebooks)