The model can be composed of an LSTM or a Quasi-Recurrent Neural Network (QRNN) which is two or more times faster than the cuDNN LSTM in this setup while achieving equivalent or better accuracy. The codebase is now PyTorch 0.4 compatible for most use cases (a big shoutout to https://github.com/shawntan for a fairly comprehensive PR https://github.com/salesforce/awd-lstm-lm/pull/43). Mild readjustments to hyperparameters may be necessary to obtain quoted performance. If you desire exact reproducibility (or wish to run on PyTorch 0.3 or lower), we suggest using an older commit of this repository. We are still working on pointer, finetune and generate functionalities.
lstm pytorch language-model sgd qrnnProvides an implementation of today's most used tokenizers, with a focus on performance and versatility.
nlp natural-language-processing transformers gpt language-model bert natural-language-understanding🤗 Transformers provides thousands of pretrained models to perform tasks on texts such as classification, information extraction, question answering, summarization, translation, text generation and more in over 100 languages. Its aim is to make cutting-edge NLP easier to use for everyone. 🤗 Transformers provides APIs to quickly download and use those pretrained models on a given text, fine-tune them on your own datasets and then share them with the community on our model hub. At the same time, each python module defining an architecture is fully standalone and can be modified to enable quick research experiments.
nlp natural-language-processing tensorflow pytorch transformer speech-recognition seq2seq flax gpt pretrained-models language-models natural-language-generation nlp-library language-model bert natural-language-understanding jax xlnet pytorch-transformers model-hubA Machine Learning library written in pure Go designed to support relevant neural architectures in Natural Language Processing. spaGO is self-contained, in that it uses its own lightweight computational graph framework for both training and inference, easy to understand from start to finish.
nlp machine-learning natural-language-processing deep-learning neural-network automatic-differentiation artificial-intelligence recurrent-networks lstm computation-graph question-answering bart automatic-translation deeplearning language-model bert transformer-architecture bert-as-service named-entities-recognitionHaystack is an end-to-end framework that enables you to build powerful and production-ready pipelines for different search use cases. Whether you want to perform Question Answering or semantic document search, you can use the State-of-the-Art NLP models in Haystack to provide unique search experiences and allow your users to query in natural language. Haystack is built in a modular fashion so that you can combine the best technology from other open-source projects like Huggingface's Transformers, Elasticsearch, or Milvus.
search nlp search-engine elasticsearch information-retrieval pytorch question-answering summarization transfer-learning ann language-model semantic-search squad bert dpr retriever neural-search natural-languageLingvo is a framework for building neural networks in Tensorflow, particularly sequence models. A list of publications using Lingvo can be found here.
nlp research translation tensorflow machine-translation speech distributed tts speech-synthesis mnist speech-recognition lm seq2seq speech-to-text gpu-computing language-model asrThis package provides spaCy components and architectures to use transformer models via Hugging Face's transformers in spaCy. The result is convenient access to state-of-the-art transformer architectures, such as BERT, GPT-2, XLNet, etc. This release requires spaCy v3. For the previous version of this library, see the v0.6.x branch.
nlp machine-learning natural-language-processing google pytorch spacy openai transfer-learning language-model bert natural-language-understanding spacy-pipeline spacy-extension pytorch-model gpt-2 huggingface xlnetThis is a PyTorch implementation of the TensorFlow code provided with OpenAI's paper "Improving Language Understanding by Generative Pre-Training" by Alec Radford, Karthik Narasimhan, Tim Salimans and Ilya Sutskever. This implementation comprises a script to load in the PyTorch model the weights pre-trained by the authors with the TensorFlow implementation.
neural-networks pytorch openai language-model transformerWe uploaded the preprocessed PubMed texts that were used to pre-train the BlueBERT models. This repository provides codes and models of BlueBERT, pre-trained on PubMed abstracts and clinical notes (MIMIC-III). Please refer to our paper Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets for more details.
natural-language-processing pubmed mimic-iii language-model bert pubmed-abstracts bert-modelpackage lingo provides the data structures and algorithms required for natural language processing.Specifically, it provides a POS Tagger (lingo/pos), a Dependency Parser (lingo/dep), and a basic tokenizer (lingo/lexer) for English. It also provides data structures for holding corpuses (lingo/corpus), and treebanks (lingo/treebank).
natural-language-processing nlp nlp-library nlp-parsing nlp-dependency-parsing nlp-machine-learning language-model part-of-speech-tagger part-of-speech inflection conll-ugetlang provides fast natural language detection in Go.
nlp natural-language language-modelThis paper tries to utilize the diversity of word meaning to train the Deep Neural Network. In the language modeling (prediction of the word sequence), we want to express the diversity of word meaning. For example, when predicting the word next to "Banana is delicious ___", the answer is "fruit", but "sweets", "food" is also ok. But ordinary one-hot vector teaching is not suitable to achieve it. Because any similar words ignored, but the exact answer word.
language-model machine-learning natural-language-processingJupyter notebooks for Chainer hands-on
chainer deep-learning jupyter-notebook dqn word2vec rnn language-model cnn convolutional-neural-networksRNN(LSTM, GRU) in Theano with mini-batch training; character-level language models in Theano
rnn lstm gru language-modelThis is a character RNN-based language model in PyTorch. Code are based on examples in https://github.com/pytorch/examples/tree/master/word_language_model. It can handle any Unicode corpus. All configurations and hyper paramters are centerized in a JSON file (hps/penn.json is an example for PTB). See the example for what are specified.
deep-learning language-model pytorchImportant: Please note that these scripts form in no way a complete application ready for end-user consumption. However, if you are a developer interested in natural language processing you may find some of them useful. Contributions, patches and pull requests are very welcome. At the time of this writing, the scripts here are focused on building the English and German VoxForge models. However, there is no reason why they couldn't be used to build other language models as well, feel free to contribute support for those.
kaldi speech-corpora voxforge sequitur lexicon cmu-sphinx language-model asr speech-recognitionZeroth is an open source project for Korean speech recognition implemented using the Kaldi toolkit. This project was developed as part of Atlas’s (https://www.goodatlas.com) Language AI platform, which enables enterprises to add intelligence to their B2C communications.
kaldi zeroth korean asr language-model speech-recognition audio-data lexicon open-source fastcampus data-augmentation tdnnThis package adds kenLM support as a spaCy 2.0 extension. Train a kenLM language model first (or use the test model from test.arpa).
spacy spacy-nlp nlp kenlm language-model spacy-extensionThis is a token-stream based parser for VHDL-2008.
vhdl language-model python-3 parser
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.