Displaying 1 to 20 from 84 results

brain.js is a library of Neural Networks written in JavaScript. 💡 Note: This is a continuation of the harthur/brain repository (which is not maintained anymore). For more details, check out this issue.

neural-network brain recurrent-neural-networks easy-to-use api web nodejs browser convolutional-neural-networks node stream ai artificial-intelligence brainjs brain.js feed-forward classifier neural network neural-networks machine-learning synapse recurrent long-short-term-memory gated-recurrent-unit rnn lstm gruIntel MKL-DNN repository migrated to https://github.com/intel/mkl-dnn. The old address will continue to be available and will redirect to the new repo. Please update your links. Intel(R) Math Kernel Library for Deep Neural Networks (Intel(R) MKL-DNN) is an open source performance library for deep learning applications. The library accelerates deep learning applications and framework on Intel(R) architecture. Intel(R) MKL-DNN contains vectorized and threaded building blocks which you can use to implement deep neural networks (DNN) with C and C++ interfaces.

intel mkl-dnn deep-learning deep-neural-networks cnn rnn lstm c-plus-plus intel-architecture xeon xeon-phi atom core simd sse42 avx2 avx512 avx512-vnni performanceAiLearning: 机器学习 - MachineLearning - ML、深度学习 - DeepLearning - DL、自然语言处理 NLP

fp-growth apriori mahchine-leaning naivebayes svm adaboost kmeans svd pca logistic regression recommendedsystem sklearn scikit-learn nlp deeplearning dnn lstm rnnThis chapter intends to introduce the main objects and concepts in TensorFlow. We also introduce how to access the data for the rest of the book and provide additional resources for learning about TensorFlow. After we have established the basic objects and methods in TensorFlow, we now want to establish the components that make up TensorFlow algorithms. We start by introducing computational graphs, and then move to loss functions and back propagation. We end with creating a simple classifier and then show an example of evaluating regression and classification algorithms.

tensorflow tensorflow-cookbook linear-regression neural-network tensorflow-algorithms rnn cnn svm nlp packtpub machine-learning tensorboard classification regression kmeans-clustering genetic-algorithm odeCompared to a classical approach, using a Recurrent Neural Networks (RNN) with Long Short-Term Memory cells (LSTMs) require no or almost no feature engineering. Data can be fed directly into the neural network who acts like a black box, modeling the problem correctly. Other research on the activity recognition dataset can use a big amount of feature engineering, which is rather a signal processing approach combined with classical data science techniques. The approach here is rather very simple in terms of how much was the data preprocessed. Let's use Google's neat Deep Learning library, TensorFlow, demonstrating the usage of an LSTM, a type of Artificial Neural Network that can process sequential data / time series.

machine-learning deep-learning lstm human-activity-recognition neural-network rnn recurrent-neural-networks tensorflow activity-recognitionAn ai powered automatically generats poems in Chinese. 很久以来，我们都想让机器自己创作诗歌，当无数作家、编辑还没有抬起笔时，AI已经完成了数千篇文章。现在，这里是第一步....

tensorflow poetry lstm rnnMulti-layer Recurrent Neural Networks (LSTM, RNN) for word-level language models in Python using TensorFlow. Mostly reused code from https://github.com/sherjilozair/char-rnn-tensorflow which was inspired from Andrej Karpathy's char-rnn.

rnn tensorflow rnn-tensorflow lstmIn these tutorials for pyTorch, we will build our first Neural Network and try to build some advanced Neural Network architectures developed recent years. Thanks for liufuyang's notebook files which is a great contribution to this tutorial.

neural-network pytorch-tutorial batch-normalization cnn rnn autoencoder pytorch regression classification batch tutorial dropout dqn reinforcement-learning gan generative-adversarial-network machine-learningIn these tutorials, we will build our first Neural Network and try to build some advanced Neural Network architectures developed recent years. All methods mentioned below have their video and text tutorial in Chinese. Visit 莫烦 Python for more.

tensorflow tensorflow-tutorials gan generative-adversarial-network rnn cnn classification regression autoencoder deep-q-network dqn machine-learning tutorial dropout neural-network텐서플로우를 기초부터 응용까지 단계별로 연습할 수 있는 소스 코드를 제공합니다. 텐서플로우 공식 사이트에서 제공하는 안내서의 대부분의 내용을 다루고 있으며, 공식 사이트에서 제공하는 소스 코드보다는 훨씬 간략하게 작성하였으므로 쉽게 개념을 익힐 수 있을 것 입니다. 또한, 모든 주석은 한글로(!) 되어 있습니다.

neural-network tensorflow mnist autoencoder rnn deep-learning tutorial chatbot seq2seq dqn word2vec cnn gan inception深度学习入门课、资深课、特色课、学术案例、产业实践案例、深度学习知识百科及面试题库The course, case and knowledge of Deep Learning and AI

nlp video reinforcement-learning detection cnn transformer gan dqn classification rnn sarsa segmentation recommender-system bert pose dssm tinybert dynabertWe would like to thank all participants for their submissions and we hope that this challenge provided insights to practitioners and researchers about the challenges in semantic code search and motivated new research. We would like to encourage everyone to continue using the dataset and the human evaluations, which we now provide publicly. Please, see below for details, specifically the Evaluation section. No new submissions to the challenge will be accepted.

nlp data-science data machine-learning natural-language-processing deep-learning tensorflow ml cnn open-data neural-networks rnn datasets representation-learning nlp-machine-learning bert programming-language-theory self-attention machine-learning-on-source-codeThis repository contains the iPython notebook and training data to accompany the O'Reilly tutorial on sentiment analysis with LSTMs in Tensorflow. See the original tutorial to run this code in a pre-built environment on O'Reilly's servers with cell-by-cell guidance, or run these files on your own machine. There is also another file called Pre-Trained LSTM.ipynb which allows you to input your own text, and see the output of the trained network. Before running the notebook, you'll first need to download all data we'll be using. This data is located in the models.tar.gz and training_data.tar.gz tarballs. We will extract these into the same directory as Oriole LSTM.ipynb. As always, the first step is to clone the repository.

sentiment-analysis tensorflow lstm rnnTensorflow implementation of attention mechanism for text classification tasks. Inspired by "Hierarchical Attention Networks for Document Classification", Zichao Yang et al. (http://www.aclweb.org/anthology/N16-1174).

attention tensorflow rnn text-classification sentiment-analysisChatbot in 200 lines of code using TensorLayer

tensorlayer tensorflow chatbot rnn lstm bot nlp chat corpusA vanilla sequence to sequence model presented in https://arxiv.org/abs/1409.3215, https://arxiv.org/abs/1406.1078 consits of using a recurrent neural network such as an LSTM (http://dl.acm.org/citation.cfm?id=1246450) or GRU (https://arxiv.org/abs/1412.3555) to encode a sequence of words or characters in a source language into a fixed length vector representation and then deocoding from that representation using another RNN in the target language. An extension of sequence to sequence models that incorporate an attention mechanism was presented in https://arxiv.org/abs/1409.0473 that uses information from the RNN hidden states in the source language at each time step in the deocder RNN. This attention mechanism significantly improves performance on tasks like machine translation. A few variants of the attention model for the task of machine translation have been presented in https://arxiv.org/abs/1508.04025.

pytorch seq2seq deep-learning rnnIt is assumed that you already have training and test data. The data is made from many examples (I'm using 684K examples), each example is made from the text from the start of the article, which I call description (or desc), and the text of the original headline (or head). The texts should be already tokenized and the tokens separated by spaces. Once you have the data ready save it in a python pickle file as a tuple: (heads, descs, keywords) were heads is a list of all the head strings, descs is a list of all the article strings in the same order and length as heads. I ignore the keywrods information so you can place None.

keras nlp generation summarization rnnCS224n: Natural Language Processing with Deep Learning Assignments Winter, 2017

cs224n deep-learning natural-language-processing word2vec rnn tensorflow
We have large collection of open source products. Follow the tags from
Tag Cloud >>

Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
**Add Projects.**