the purpose of this repository is to explore text classification methods in NLP with deep learning. sentence similarity project has been released you can check it if you like.
classification nlp fasttext textcnn textrnn tensorflow multi-label multi-class attention-mechanism text-classification convolutional-neural-networks sentence-classification memory-networksI tried to implement the idea in Attention Is All You Need. They authors claimed that their model, the Transformer, outperformed the state-of-the-art one in machine translation with only attention, no CNNs, no RNNs. How cool it is! At the end of the paper, they promise they will make their code available soon, but apparently it is not so yet. I have two goals with this project. One is I wanted to have a full understanding of the paper. Often it's hard for me to have a good grasp before writing some code for it. Another is to share my code with people who are interested in this model before the official code is unveiled. I got a BLEU score of 17.14. (Recollect I trained with a small dataset, limited vocabulary) Some of the evaluation results are as follows. Details are available in the results folder.
attention-mechanism translation attention-is-all-you-need implementationUpdate (December 2, 2016) TensorFlow implementation of Show, Attend and Tell: Neural Image Caption Generation with Visual Attention which introduces an attention based image caption generator. The model changes its attention to the relevant part of the image while it generates each word.First, clone this repo and pycocoevalcap in same directory.
tensorflow image-captioning show-attend-and-tell attention-mechanism mscoco-image-datasetFelix Hieber, Tobias Domhan, Michael Denkowski, David Vilar, Artem Sokolov, Ann Clifton and Matt Post (2017): Sockeye: A Toolkit for Neural Machine Translation. In eprint arXiv:cs-CL/1712.05690.If you are interested in collaborating or have any questions, please submit a pull request or issue. You can also send questions to sockeye-dev-at-amazon-dot-com.
deep-learning deep-neural-networks mxnet machine-learning machine-translation neural-machine-translation encoder-decoder attention-mechanism sequence-to-sequence sequence-to-sequence-models sockeye attention-is-all-you-need attention-alignment-visualization attention-model seq2seq convolutional-neural-networks translationYou can click the following links for detailed installation instructions. PubMed Paper Reading Dataset This dataset gathers 14,857 entities, 133 relations, and entities corresponding tokenized text from PubMed. It contains 875,698 training pairs, 109,462 development pairs, and 109,462 test pairs.
nlp text-generation pytorch generation datasets natural-language-generation attention-mechanism memory-networks end-to-end-learning paper-generationNeural Machine Translation with Keras (Theano and Tensorflow). for obtaining the required packages for running this library.
neural-machine-translation keras deep-learning sequence-to-sequence theano machine-learning nmt machine-translation lstm-networks gru tensorflow attention-mechanism web-demo transformer attention-is-all-you-need attention-model attention-seq2seqFinally, execute_cora.py puts all of the above together and may be used to execute a full training run on Cora. An experimental sparse version is also available, working only when the batch size is equal to 1. The sparse model may be found at models/sp_gat.py.
graph-attention-networks attention-mechanism self-attention tensorflow neural-networksChainer-based Python implementation of Transformer, an attention-based seq2seq model without convolution and recurrence. If you want to see the architecture, please see net.py. See "Attention Is All You Need", Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, arxiv, 2017.
chainer neural-network deep-learning deep-neural-networks attention-mechanism googleadaptive attention model. tensorflow implementation of knowing when to look: adaptive attention via visual sentinel for image captioning.
image-captioning tensorflow knowing-when-to-look attention-mechanism visual-sentinelEfficient implementation of structured sparsity inducing attention mechanisms: fusedmax, oscarmax and sparsemax. Currently available for pytorch v0.2. Requires python (3.6, 3.5, or 2.7), cython, numpy, scipy, scikit-learn, and lightning.
attention-mechanism attention-mechanisms fused-lasso deep-learning deeplearning deep-neural-networks sparsity sparse segmentationYou can download the model here. You can download the model here.
nlp deep-learning attention-mechanism summarization pointer-networks pytorch-implmention seq2seq-attnJoint model for intent detection and slot filling based on attention, input alignment and knowledge. with ability to detect whether a input sentence is a noise input or meanfuling input by combine feature from domain detection, intent detection and slot filling.
intent-detection slot-filling joint-models cnn attention-mechanism bi-directional knowledge-representationTo train the model require GPU with 12GB Memory, if you do not have GPU, you can directly use the pretrained model for inference. This code is written in Lua and requires Torch. The preprocssinng code is in Python, and you need to install NLTK if you want to use NLTK to tokenize the caption.
image-captioning torch attention-mechanismSimpleDNN is a machine learning lightweight open-source library written in Kotlin whose purpose is to support the development of feed-forward and recurrent Artificial Neural Networks.
neural-network kotlin machine-learning recurrent-neural-networks feedforward-neural-network attention-mechanism natural-language-processing artificial-intelligenceA fixed-size, go-back-k recurrent attention module on an RNN so as to have linear short-term memory by the means of attention. The LARNN model can be easily used inside a loop on the cell state just like any other RNN. The cell state keeps the k last states for its multi-head attention mechanism. The LARNN is derived from the Long Short-Term Memory (LSTM) cell. The LARNN introduces attention on the state's past values up to a certain range, limited by a time window k to keep the forward processing linear in time in terms sequence length (time steps).
rnn lstm attention-mechanism attention-model attention-is-all-you-need recurrent-neural-networks pytorchThis is an unofficial Tensorflow implementation of Attend, Infear, Repeat (AIR), as presented in the following paper: S. M. Ali Eslami et. al., Attend, Infer, Repeat: Fast Scene Understanding with Generative Models. I describe the implementation and the issues I run into while working on it in this blog post.
tensorflow vae neural-networks attention-mechanism generative-model computer-vision computer-graphics rnn attention attend-infer-repeatA. R. Kosiorek, A. Bewley, I. Posner, "Hierarchical Attentive Recurrent Tracking", NIPS 2017. The notebook scripts/demo.ipynb contains a demo, which shows how to evaluate tracker on an arbitrary image sequence. By default, it runs on images located in imgs folder and uses a pretrained model. Before running the demo please download AlexNet weights first (described in the Training section).
neural-nets object-tracking attention-mechanism rnn tensorflow kitti-datasetAn implementation of the Show, Attend and Tell (Xu, Kelvin et. al., 2016) paper in TensorFlow, for the OpenAI Im2LaTeX suggested problem. The crux of the model is contained in cnn_enc_gru_dec_attn.py that uses the embedding attention decoder from TensorFlow to attend on the output of the CNN.
deep-learning attention-mechanism attention show-and-tell show-attend-and-tell tensorflowThis repository contains the code for building a system similar to the one from the work Video Description using Bidirectional Recurrent Neural Networks, presented at the International Conference of Artificial Neural Networks (ICANN'16). With this module, you can replicate our experiments and easily deploy new models. ABiViRNet is built upon our fork of Keras framework (version 1.2) and tested for the Theano backend. See data_engine/README.md for detailed information.
video-captioning deep-learning keras theano attention-mechanism lstmNetwork pruning is aimed at imposing sparsity in a neural network architecture by increasing the portion of zero-valued weights for reducing its size energy efficiency consideration and increasing evaluation speed. In most of the conducted research efforts, the sparsity is enforced for network pruning without any attention to the internal network characteristics such as unbalanced outputs of the neurons or more specifically the distribution of the weights and outputs of the neurons. That may cause severe accuracy drop due to uncontrolled sparsity. In this work, we propose an attention mechanism that simultaneously controls the sparsity intensity and supervised network pruning by keeping important information bottlenecks of the network to be active. On CIFAR-10, the proposed method outperforms the best baseline method by 6% and reduced the accuracy drop by 2.6× at the same level of sparsity. Please refer to Official TensorFLow installation guideline for further details considering your specific system architecture.
sparsity deep-learning convolutional-neural-networks attention-mechanism
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.