NLP-Models-Tensorflow, Gathers machine learning and tensorflow deep learning models for NLP problems, code simplify inside Jupyter Notebooks 100%. I will attached github repositories for models that I not implemented from scratch, basically I copy, paste and fix those code for deprecated issues.
nlp machine-learning embedded deep-learning chatbot language-detection lstm summarization attention speech-to-text neural-machine-translation optical-character-recognition pos-tagging lstm-seq2seq-tf dnc-seq2seq luong-apiThe code uses PyTorch https://pytorch.org. Note that the original experiments were done using torch-autograd, we have so far validated that CIFAR-10 experiments are exactly reproducible in PyTorch, and are in process of doing so for ImageNet (results are very slightly worse in PyTorch, due to hyperparameters). This section describes how to get the results in the table 1 of the paper.
pytorch knowledge-distillation attention deep-learningTensorflow implementation of attention mechanism for text classification tasks. Inspired by "Hierarchical Attention Networks for Document Classification", Zichao Yang et al. (http://www.aclweb.org/anthology/N16-1174).
attention tensorflow rnn text-classification sentiment-analysisThis is a Tensorflow implementation of Spatial Transformer Networks by Max Jaderberg, Karen Simonyan, Andrew Zisserman and Koray Kavukcuoglu, accompanying by two-part blog tutorial series. Spatial Transformer Networks (STN) is a differentiable module that can be inserted anywhere in ConvNet architecture to increase its geometric invariance. It effectively gives the network the ability to spatially transform feature maps at no extra data or supervision cost.
spatial-transformer-network tensorflow stn attention affine-transformation convnetLong-range arena is an effort toward systematic evaluation of efficient transformer models. The project aims at establishing benchmark tasks/dtasets using which we can evaluate transformer-based models in a systematic way, by assessing their generalization power, computational efficiency, memory foot-print, etc. Long-range arena also implements different variants of Transformer models in JAX, using Flax.
nlp deep-learning transformers attention flax jaxCode for genereting data, training and testing. There are two tasks: ordering single numbers, or sums or numbers (the _sums suffix). The training scripts save model weights each epoch. When started, they attempt to load appropriate weights - corresponding to the hidden size and the chosen number of steps in a sequence.
pointer-networks seq2seq attention lstmThis is HAN version of sentiment classification, with pre-trained character-level embedding, and used RAN instead of GRU.
deep-learning rnn tensorflow attention document chinese nlpThis is a implementation of Recurrent Discounted Attention unit that extends Tensorflow's RNNCell, RDA is builds on the RWA by additionally allowing the discounting of the past.
tensorflow attention rnn machine-learning deep-learningUpdate: Joint Model for law cases prediction is released. run python HAN_train.py to train the model for predict accusation, relevant articles and term of imprisonment.
ai law accusation relevant-articles hierarchical-attention-network crime fasttext attention text-categorization textcnn text-classificationThis is a PyTorch implementation of Recurrent Models of Visual Attention by Volodymyr Mnih, Nicolas Heess, Alex Graves and Koray Kavukcuoglu. The Recurrent Attention Model (RAM) is a recurrent neural network that processes inputs sequentially, attending to different locations within the image one at a time, and incrementally combining information from these fixations to build up a dynamic internal representation of the image.
pytorch attention recurrent-attention-model recurrent-models ramThis is an unofficial Tensorflow implementation of Attend, Infear, Repeat (AIR), as presented in the following paper: S. M. Ali Eslami et. al., Attend, Infer, Repeat: Fast Scene Understanding with Generative Models. I describe the implementation and the issues I run into while working on it in this blog post.
tensorflow vae neural-networks attention-mechanism generative-model computer-vision computer-graphics rnn attention attend-infer-repeatThis project is modified version of https://github.com/jlindsey15/RAM. The critical problem of last implemetnation is that the location network cannot learn because of tf.stop_gradient implementation so that they got just '94% accuracy'. It seems relatively bad compared to the result of paper. If 'tf.stop_gradient' was commented, the classification result was very bad. The reason I think is that the problem is originated from sharing the gradient flow through location, core, glimpse network. Through gradient sharing, gradients of classification part are corrupted by gradients of reinforcement part so that classification result become very bad. (If someone want to share gradient, the weighted loss should be needed. please refer https://arxiv.org/pdf/1412.7755.pdf) According to their post research, 'Multiple Object Recognition with Visual Attention' (https://arxiv.org/pdf/1412.7755.pdf) they softly separate location network and others through multi-layer RNN. From this, I assume that sharing the gradient through whole network is not a good idea so separate them, and finally got a good result. In summary, the learning stretegy is as follow. location network, baseline network : learn with gradients of reinforcement learning only.
deep-learning ram attention reinforcement-learning recurrent-neural-networksThis toolkit provides the voice activity detection (VAD) code and our recorded dataset. J. Kim and M. Hahn, "Voice Activity Detection Using an Adaptive Context Attention Model," in IEEE Signal Processing Letters, vol. PP, no. 99, pp. 1-1.
vad dnn lstm bdnn acam attention speech data voice-detection speech-recognitionAn implementation of the Show, Attend and Tell (Xu, Kelvin et. al., 2016) paper in TensorFlow, for the OpenAI Im2LaTeX suggested problem. The crux of the model is contained in cnn_enc_gru_dec_attn.py that uses the embedding attention decoder from TensorFlow to attend on the output of the CNN.
deep-learning attention-mechanism attention show-and-tell show-attend-and-tell tensorflowCurrently working on replicating the Deep Attentive Reader in Keras. This model has been designed in the same way as described in the paper Teaching Machines to Read and Comprehend. So far, the results are consistantly improving - once 10% Accuracy has been achieved I intend to move over to an AWS instance to reduce the training time.
deep-learning keras nlp-machine-learning nlp embeddings attentionThis repository is the implementation of Bilinear Attention Networks for the visual question answering task. Our single model achieved 70.35 and an ensemble of 15 models achieved 71.84 (Test-standard, VQA 2.0). For the detail, please refer to our technical report. This repository is based on and inspired by @hengyuan-hu's work. We sincerely thank for their sharing of the codes.
visual-question-answering attention bilinear-pooling pytorch-implmentionThis JavaScript/WebGL library detects if the user is looking at the screen or not. It is very robust to all lighting conditions and lightweight (only 150KB gzipped for the main script and the neural network JSON model). It is great for playing a video only if the user is watching it. In the paths /integrationDemo*, there are several integration examples. You can host them through a static HTTPS server.
face detection attention webgl deep-learning glance watching webcam video face-tracking face-detection real-time library lightweight
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.