This is the official code repository for Machine Learning with TensorFlow. Get started with machine learning using TensorFlow, Google's latest and greatest machine learning library.
tensorflow machine-learning regression convolutional-neural-networks logistic-regression book reinforcement-learning autoencoder linear-regression classification clusteringIn these tutorials for pyTorch, we will build our first Neural Network and try to build some advanced Neural Network architectures developed recent years. Thanks for liufuyang's notebook files which is a great contribution to this tutorial.
neural-network pytorch-tutorial batch-normalization cnn rnn autoencoder pytorch regression classification batch tutorial dropout dqn reinforcement-learning gan generative-adversarial-network machine-learningIn these tutorials, we will build our first Neural Network and try to build some advanced Neural Network architectures developed recent years. All methods mentioned below have their video and text tutorial in Chinese. Visit 莫烦 Python for more.
tensorflow tensorflow-tutorials gan generative-adversarial-network rnn cnn classification regression autoencoder deep-q-network dqn machine-learning tutorial dropout neural-networkThis repo contains the source code in my personal column (https://zhuanlan.zhihu.com/zhaoyeyu), implemented using Python 3.6. Including Natural Language Processing and Computer Vision projects, such as text generation, machine translation, deep convolution GAN and other actual combat code.
deep-learning tensorflow-examples convolutional-neural-networks recurrent-neural-networks autoencoder gan style-transfer natural-language-processing machine-translationSome examples require MNIST dataset for training and testing. Don't worry, this dataset will automatically be downloaded when running examples (with input_data.py). MNIST is a database of handwritten digits, for a quick description of that dataset, you can check this notebook.
recurrent-neural-networks convolutional-neural-networks deep-learning-tutorial tensorflow tensorlayer keras deep-reinforcement-learning tensorflow-tutorials deep-learning machine-learning notebook autoencoder multi-layer-perceptron reinforcement-learning tflearn neural-networks neural-network neural-machine-translation nlp cnn텐서플로우를 기초부터 응용까지 단계별로 연습할 수 있는 소스 코드를 제공합니다. 텐서플로우 공식 사이트에서 제공하는 안내서의 대부분의 내용을 다루고 있으며, 공식 사이트에서 제공하는 소스 코드보다는 훨씬 간략하게 작성하였으므로 쉽게 개념을 익힐 수 있을 것 입니다. 또한, 모든 주석은 한글로(!) 되어 있습니다.
neural-network tensorflow mnist autoencoder rnn deep-learning tutorial chatbot seq2seq dqn word2vec cnn gan inceptionNiftyNet is a consortium of research organisations (BMEIS -- School of Biomedical Engineering and Imaging Sciences, King's College London; WEISS -- Wellcome EPSRC Centre for Interventional and Surgical Sciences, UCL; CMIC -- Centre for Medical Image Computing, UCL; HIG -- High-dimensional Imaging Group, UCL), where BMEIS acts as the consortium lead. NiftyNet is not intended for clinical use.
tensorflow distributed ml neural-network python2 python3 pip deep-neural-networks deep-learning convolutional-neural-networks medical-imaging medical-image-computing medical-image-processing medical-images segmentation gan autoencoder medical-image-analysis image-guided-therapyWelcome to my GitHub repo. I am a Data Scientist and I code in R, Python and Wolfram Mathematica. Here you will find some Machine Learning, Deep Learning, Natural Language Processing and Artificial Intelligence models I developed.
anomaly-detection deep-learning autoencoder keras keras-models denoising-autoencoders generative-adversarial-network glove keras-layer word2vec nlp natural-language-processing sentiment-analysis opencv segnet resnet-50 variational-autoencoder t-sne svm-classifier latent-dirichlet-allocationImportant Notes: PyOD contains some neural network based models, e.g., AutoEncoders, which are implemented in keras. However, PyOD would NOT install keras and tensorflow automatically. This would reduce the risk of damaging your local installations. You are responsible for installing keras and tensorflow if you want to use neural net based models. An instruction is provided here. Anomaly detection resources, e.g., courses, books, papers and videos.
outlier-detection anomaly-detection outlier-ensembles outliers anomaly machine-learning data-mining unsupervised-learning python2 python3 fraud-detection autoencoder neural-networks deep-learningDifferent models can be chosen using th main.lua -model <modelName>. The denoising criterion can be used to replace the standard (autoencoder) reconstruction criterion by using the denoising flag. For example, a denoising AAE (DAAE) [10] can be set up using th main.lua -model AAE -denoising. The corruption process is additive Gaussian noise *~ N(0, 0.5)*.
autoencoder deep-learningDeep learning is one of the most popular domains in the artificial intelligence (AI) space, which allows you to develop multi-layered models of varying complexities. This book is designed to help you grasp things, from basic deep learning algorithms to the more advanced algorithms. The book is designed in a way that first you will understand the algorithm intuitively, once you have a basic understanding of the algorithms, then you will master the underlying math behind them effortlessly and then you will learn how to implement them using TensorFlow step by step. The book covers almost all the state of the art deep learning algorithms. First, you will get a good understanding of the fundamentals of neural networks and several variants of gradient descent algorithms. Later, you will explore RNN, Bidirectional RNN, LSTM, GRU, seq2seq, CNN, capsule nets and more. Then, you will master GAN and various types of GANs and several different autoencoders.
tensorflow word-embeddings gru autoencoder gans doc2vec skip-thoughts adagrad cyclegan deep-learning-mathematics capsule-network few-shot-learning quick-thought deep-learning-scratch nadam deep-learning-math lstm-math cnn-math rnn-derivation contractive-autonencodersThe project was created as part of the Creative Applications of Deep Learning with TensorFlow (CADL) Kadenze course's final assignment. It is an experimental attempt to transfer artistic style learned from a series of paintings "live" onto a video sequence by fitting a variational autoencoder with 512 codes to both paintings and video frames, isolating the mean feature-space embeddings and modifying the video's embeddings to be closer to those of the paintings. Because the general visual quality of the VAE's decoded output is relatively low, a convolutional post-processing network based on residual convolutions was trained with the purpose of making the resulting image less similar to the VAE's generated output and more similar to the original input images. The basic idea was to have an upsampling network here, but it quickly turned out to be a very naive idea at this point of development. Instead, it now downsizes the input, learns filters in a residual network and then samples back up to the input frame size; I would have liked to perform convolutions directly on the input, but memory limitations prevented the usage of a useful amount of feature maps.
tensorflow deep-learning neural-network variational-inference autoencoder vae vaegan generative-art generative-adversarial-network experiment cadl kadenze online-courseKeras implementation of Structural Deep Network Embedding, KDD 2016
keras network-embedding autoencoder visualization link-predictionDisclaimer: VAE coming soon... The last activation of the decoder layer, the loss function, and the normalization scheme used on the training data are crucial for obtaining good reconstructions and preventing exploding negative losses.
autoencoder pytorch variational-autoencoderTensorbag is a collection of tensorflow tutorial on different Deep Learning and Machine Learning algorithms. The tutorials are organised as jupyter notebooks and require tensorflow >= 1.5. There is a subset of notebooks identified with the tag [quiz] that directly ask to the reader to complete part of the code. In the same folder there is always a complementary notebook with the complete solution.
cifar-10 generative-adversarial-networks mnist lenet-5 convolutional-neural-networks resnet-18 notebook deep-learning tensorflow resnet tfrecord-format cifar-100 tensorflow-tutorials kmeans-clustering perceptron autoencoder tutorialThis repository contains a pre-trained Split-Brain Autoencoder network. The network achieves state-of-the-art results on several large-scale unsupervised representation learning benchmarks. This code requires a working installation of Caffe. For guidelines and help with installation of Caffe, consult the installation guide and Caffe users group.
unsupervised-learning autoencoder caffeUsing Autoencoders for classification as unsupervised machine learning algorithms with Deep Learning. Project in Unsupervised Classification With Autoencoder.ipynb file.
deep-learning machine-learning autoencoderThis repository is under the MIT license. See the LICENSE file for detail. We need TensorFlow (version>=1.4).
point-cloud deep-learning pointnet autoencoderWelcome to my GitHub repo. I am a Data Scientist and I code in R, Python and Wolfram Mathematica. Here you will find some Machine Learning, Deep Learning, Natural Language Processing and Artificial Intelligence models I developed.
r python3 python-3 mathematica lasagne theano theano-models autoencoder face-recognition natural-language-processing nlp nlp-machine-learning deep-learning keras lstm lstm-neural-networks timeseries time-series-analysis word2vecThis repository contains the H2O presentation for Trevor Hastie and Rob Tibshirani's Statistical Learning and Data Mining IV course in Washington, DC on October 19, 2016.
statistical-learning deep-learning tutorial h2o machine-learning r autoencoder
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.