- 15

This repository implements the paper, Neural Discrete Representation Learning (VQ-VAE) in Tensorflow. ⚠️ This is not an official implementation, and might have some glitch (,or a major defect).

https://github.com/hiwonjoon/tf-vqvaeTags | tensorflow vae cifar10 mnist generative-model |

Implementation | Jupyter Notebook |

License | Public |

Platform |

Collection of generative models, e.g. GAN, VAE in Pytorch and Tensorflow. Also present here are RBM and Helmholtz Machine. Generated samples will be stored in GAN/{gan_model}/out (or VAE/{vae_model}/out, etc) directory during training.

vae gan pytorch tensorflow generative-model machine-learning rbm restricted-boltzmann-machineSimple Tensorflow implementation of "Densenet" using Cifar10, MNIST

densenet tensorflow densenet-tensorflowTrain Compositional Pattern Producing Network as a Generative Model, using Generative Adversarial Networks and Variational Autoencoder techniques to produce high resolution images. Run python train.py from the command line to train from scratch and experiment with different settings.

All pull requests are welcome, make sure to follow the contribution guidelines when you submit pull request.

tensorflow tensorflow-tutorials mnist-classification mnist machine-learning android tensorflow-models machine-learning-android tensorflow-android tensorflow-model mnist-model deep-learning deep-neural-networks deeplearning deep-learning-tutorialTensorflow implementation of Deep Convolutional Generative Adversarial Networks which is a stabilize Generative Adversarial Networks. The referenced torch code can be found here.

tensorflow dcgan gan generative-modelTensorflow implementation of BEGAN: Boundary Equilibrium Generative Adversarial Networks.

gan tensorflow celeba generative-model began googleSimple Tensorflow implementation of "Squeeze and Excitation Networks" using Cifar10 (ResNeXt, Inception-v4, Inception-resnet-v2)

senet inception-resnet inception densenet resnext tensorflowThis is a Tensorflow implementation of Conditional Image Generation with PixelCNN Decoders which introduces the Gated PixelCNN model based on PixelCNN architecture originally mentioned in Pixel Recurrent Neural Networks. The model can be conditioned on latent representation of labels or images to generate images accordingly. Images can also be modelled unconditionally. It can also act as a powerful decoder and can replace deconvolution (transposed convolution) in Autoencoders and GANs. A detailed summary of the paper can be found here. The gating accounts for remembering the context and model more complex interactions, like in LSTM. The network stack on the left is the Vertical stack that takes care of blind spots that occure while convolution due to the masking layer (Refer the Pixel RNN paper to know more about masking). Use of residual connection significantly improves the model performance.

deep-learning generative-algorithm paper convolution deepmind tensorflowA PyTorch-based package containing useful models for modern deep semi-supervised learning and deep generative models. Want to jump right into it? Look into the notebooks. 2018.04.17 - The Gumbel softmax notebook has been added to show how you can use discrete latent variables in VAEs. 2018.02.28 - The β-VAE notebook was added to show how VAEs can learn disentangled representations.

semi-supervised-learning pytorch generative-modelsCNN's with Noisy Labels - This notebook looks at a recent paper that discusses how convolutional neural networks that are trained on random labels (with some probability) are still able to acheive good accuracy on MNIST. I thought that the paper showed some eye-brow raising results, so I went ahead and tried it out for myself. It was pretty amazing to see that even when training a CNN with random labels 50% of the time, and the correct labels the other 50% of the time, the network was still able to get a 90+% accuracy. Character Level RNN (Work in Progress) - This notebook shows you how to train a character level RNN in Tensorflow. The idea was inspired by Andrej Karpathy's famous blog post and was based on this Keras implementation. In this notebook, you'll learn more about what the model is doing, and how you can input your own dataset, and train a model to generate similar looking text.

tensorflow deep-learning machine-learningTensorFlow is Google's machine learning runtime. It is implemented as C++ runtime, along with Python framework to support building a variety of models, especially neural networks for deep learning. It is interesting to be able to use TensorFlow in a node.js application using just JavaScript (or TypeScript if that's your preference). However, the Python functionality is vast (several ops, estimator implementations etc.) and continually expanding. Instead, it would be more practical to consider building Graphs and training models in Python, and then consuming those for runtime use-cases (like prediction or inference) in a pure node.js and Python-free deployment. This is what this node module enables.

tensorflow node-tensorflow nodejs machine-learning deep-learning npm-package tf tensor ml ai neural-networks neuralnetworks deeplearning model numerical-computation googleImplementation of the method described in our Arxiv paper. We present an autoencoder that leverages learned representations to better measure similarities in data space. By combining a variational autoencoder with a generative adversarial network we can use learned feature representations in the GAN discriminator as basis for the VAE reconstruction objective. Thereby, we replace element-wise errors with feature-wise errors to better capture the data distribution while offering invariance towards e.g. translation. We apply our method to images of faces and show that it outperforms VAEs with element-wise similarity measures in terms of visual fidelity. Moreover, we show that the method learns an embedding in which high-level abstract visual features (e.g. wearing glasses) can be modified using simple arithmetic.

This package is part of the Kadenze Academy program Creative Applications of Deep Learning w/ TensorFlow. from cadl import and then pressing tab to see the list of available modules.

deep-learning neural-network tutorial mooc gan vae vae-gan pixelcnn wavenet magenta nsynth tensorflow celeba cyclegan dcgan word2vec glove autoregressive conditional courseTensorFlow Implementation for learned compression of images using Generative Adversarial Networks. The method was developed by Agustsson et. al. in Generative Adversarial Networks for Extreme Learned Image Compression. The proposed idea is very interesting and their approach is well-described. Training is conducted with batch size 1 and reconstructed samples / tensorboard summaries will be periodically written every certain number of steps (default is 128). Checkpoints are saved every 10 epochs.

generative-adversarial-network image-compression tensorflow gan computer-visionThis is a PyTorch implementation of the TensorFlow code provided with OpenAI's paper "Improving Language Understanding by Generative Pre-Training" by Alec Radford, Karthik Narasimhan, Tim Salimans and Ilya Sutskever. This implementation comprises a script to load in the PyTorch model the weights pre-trained by the authors with the TensorFlow implementation.

neural-networks pytorch openai language-model transformerGAN Playground lets you play around with Generative Adversarial Networks right in your browser. Currently, it contains three built-in datasets: MNIST, Fashion MNIST, and CIFAR-10. GAN Playground provides you the ability to set your models' hyperparameters and build up your discriminator and generator layer-by-layer. You can observe the network learn in real time as the generator produces more and more realistic images, or more likely, gets stuck in failure modes such as mode collapse.

generative-adversarial-network deep-learning machine-learning neural-network ganTensorFlow implementation of Learning from Simulated and Unsupervised Images through Adversarial Training. Result of lambda=1.0 with optimizer=sgd after 8,000 steps.

tensorflow synthetic-images deep-learning apple generative-modelSome examples require MNIST dataset for training and testing. Don't worry, this dataset will automatically be downloaded when running examples (with input_data.py). MNIST is a database of handwritten digits, for a quick description of that dataset, you can check this notebook.

recurrent-neural-networks convolutional-neural-networks deep-learning-tutorial tensorflow tensorlayer keras deep-reinforcement-learning tensorflow-tutorials deep-learning machine-learning notebook autoencoder multi-layer-perceptron reinforcement-learning tflearn neural-networks neural-network neural-machine-translation nlp cnnA composable GAN API and CLI. Built for developers, researchers, and artists. HyperGAN is currently in open beta.

gan supervised-learning unsupervised-learning learning generative-adversarial-network generative-model artificial-intelligence machine-learning machine-learning-api tensorflow classification generator discriminatorMNIST 예제를 CNN 모델로 학습하는 코드를 조금 보강하고 정리해서 TensorFlow-MNIST 저장소에 올려두었습니다. summary 를 저장해서 TensorBoard 를 사용할 수 있게 하였으며, 모델을 생성하는 부분과 Trainer, Tester 를 분리하여 학습한 모델을 저장 후 따로 사용할 수 있도록 해 두었으니 참고 해 주세요.

We have large collection of open source products. Follow the tags from
Tag Cloud >>

Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
**Add Projects.**