- 33

This repository provides implementation of NeuroEvolution of Augmenting Topologies (NEAT) method written in Go language. The Neuroevolution (NE) is an artificial evolution of Neural Networks (NN) using genetic algorithms in order to find optimal NN parameters and topology. Neuroevolution of NN may assume search for optimal weights of connections between NN nodes as well as search for optimal topology of resulting NN. The NEAT method implemented in this work do search for both: optimal connections weights and topology for given task (number of NN nodes per layer and their interconnections).

https://github.com/yaricom/goNEAT

Unity Machine Learning Agents (ML-Agents) is an open-source Unity plugin that enables games and simulations to serve as environments for training intelligent agents. Agents can be trained using reinforcement learning, imitation learning, neuroevolution, or other machine learning methods through a simple-to-use Python API. We also provide implementations (based on TensorFlow) of state-of-the-art algorithms to enable game developers and hobbyists to easily train intelligent agents for 2D, 3D and VR/AR games. These trained agents can be used for multiple purposes, including controlling NPC behavior (in a variety of settings such as multi-agent and adversarial), automated testing of game builds and evaluating different game design decisions pre-release. ML-Agents is mutually beneficial for both game developers and AI researchers as it provides a central platform where advances in AI can be evaluated on Unity’s rich environments and then made accessible to the wider research and game developer communities. For more information, in addition to installation and usage instructions, see our documentation home. If you have used a version of ML-Agents prior to v0.3, we strongly recommend our guide on migrating to v0.3.

reinforcement-learning unity3d deep-learning unity deep-reinforcement-learning neural-networksMeta learning is an exciting research trend in machine learning, which enables a model to understand the learning process. Unlike other ML paradigms, with meta learning you can learn from small datasets faster. Hands-On Meta Learning with Python starts by explaining the fundamentals of meta learning and helps you understand the concept of learning to learn. You will delve into various one-shot learning algorithms, like siamese, prototypical, relation and memory-augmented networks by implementing them in TensorFlow and Keras. As you make your way through the book, you will dive into state-of-the-art meta learning algorithms such as MAML, Reptile, and CAML. You will then explore how to learn quickly with Meta-SGD and discover how you can perform unsupervised learning using meta learning with CACTUs. In the concluding chapters, you will work through recent trends in meta learning such as adversarial meta learning, task agnostic meta learning, and meta imitation learning.

reinforcement-learning tensorflow keras one-shot-learning reptile maml mann zero-shot-learning ntm shot-learning siamese-network relation-network metalearning few-shot-learning prototypical-networks meta-sgd matching-networks deep-meta-learning meta-imitation-learning prototypical-networkConvNetJS is a Javascript implementation of Neural networks, It currently supports Common Neural Network modules, Classification (SVM/Softmax) and Regression (L2) cost functions, A MagicNet class for fully automatic neural network learning (automatic hyperparameter search and cross-validatations), Ability to specify and train Convolutional Networks that process images, An experimental Reinforcement Learning module, based on Deep Q Learning.

artificial-intelligence neural-networks machine-learning deep-learningSome examples require MNIST dataset for training and testing. Don't worry, this dataset will automatically be downloaded when running examples (with input_data.py). MNIST is a database of handwritten digits, for a quick description of that dataset, you can check this notebook.

recurrent-neural-networks convolutional-neural-networks deep-learning-tutorial tensorflow tensorlayer keras deep-reinforcement-learning tensorflow-tutorials deep-learning machine-learning notebook autoencoder multi-layer-perceptron reinforcement-learning tflearn neural-networks neural-network neural-machine-translation nlp cnnTrending deep learning Github repositories can be found here. Hint: This will be updated regularly.

deep-learning deep-neural-networks deep-reinforcement-learning convolutional-neural-networks recurrent-neural-networks stargazers-count artificial-neural-networks artificial-intelligence machine-learning top-repositoriesIn these tutorials, we will demonstrate and visualize algorithms like Genetic Algorithm, Evolution Strategy, NEAT etc. All methods mentioned below have their video and text tutorial in Chinese. Visit èŽ«çƒ¦ Python for more.

evolutionary-algorithm genetic-algorithm neuroevolution microbial-genetic-algorithm travel-sale-problem evolution-strategy es reinforcement-learning neural-network microbial-ga neat neural-nets travel-sales-problem nes evolution-strategies openai distributed-es machine-learning tutorialRepository for the book Introduction to Artificial Neural Networks and Deep Learning: A Practical Guide with Applications in Python. Deep learning is not just the talk of the town among tech folks. Deep learning allows us to tackle complex problems, training artificial neural networks to recognize complex patterns for image and speech recognition. In this book, we'll continue where we left off in Python Machine Learning and implement deep learning algorithms in PyTorch.

deep-learning neural-network machine-learning tensorflow artificial-intelligence data-science pytorchkeras-rl implements some state-of-the art deep reinforcement learning algorithms in Python and seamlessly integrates with the deep learning library Keras. Just like Keras, it works with either Theano or TensorFlow, which means that you can train your algorithm efficiently either on CPU or GPU. Furthermore, keras-rl works with OpenAI Gym out of the box. This means that evaluating and playing around with different algorithms is easy. Of course you can extend keras-rl according to your own needs. You can use built-in Keras callbacks and metrics or define your own. Even more so, it is easy to implement your own environments and even algorithms by simply extending some simple abstract classes. In a nutshell: keras-rl makes it really easy to run state-of-the-art deep reinforcement learning algorithms, uses Keras and thus Theano or TensorFlow and was built with OpenAI Gym in mind.

keras tensorflow theano reinforcement-learning neural-networks machine-learningTensorLayer is a novel TensorFlow-based deep learning and reinforcement learning library designed for researchers and engineers. It provides a large collection of customizable neural layers / functions that are key to build real-world AI applications. TensorLayer is awarded the 2017 Best Open Source Software by the ACM Multimedia Society. Simplicity : TensorLayer lifts the low-level dataflow interface of TensorFlow to high-level layers / models. It is very easy to learn through the rich example codes contributed by a wide community.

tensorlayer deep-learning tensorflow machine-learning data-science neural-network reinforcement-learning artificial-intelligence gan a3c tensorflow-tutorials dqn object-detection chatbot tensorflow-tutorial imagenet googleThis repository contains material related to Udacity's Deep Reinforcement Learning Nanodegree program. The tutorials lead you through implementing various algorithms in reinforcement learning. All of the code is in PyTorch (v0.4) and Python 3.

deep-reinforcement-learning reinforcement-learning reinforcement-learning-algorithms neural-networks pytorch pytorch-rl ddpg dqn ppo dynamic-programming cross-entropy hill-climbing ml-agents openai-gym-solutions openai-gym rl-algorithmsCoach is a python reinforcement learning framework containing implementation of many state-of-the-art algorithms. It exposes a set of easy-to-use APIs for experimenting with new RL algorithms, and allows simple integration of new environments to solve. Basic RL components (algorithms, environments, neural network architectures, exploration policies, ...) are well decoupled, so that extending and reusing existing components is fairly painless.

reinforcement-learning deep-learning mxnet tensorflow openai-gym rl starcraft imitation-learning hierarchical-reinforcement-learning coach mujoco starcraft2 onnx roboschool carla starcraft2-ai distributed-reinforcement-learningRLzoo is a collection of the most practical reinforcement learning algorithms, frameworks and applications. It is implemented with Tensorflow 2.0 and API of neural network layers in TensorLayer 2, to provide a hands-on fast-developing approach for reinforcement learning practices and benchmarks. It supports basic toy-tests like OpenAI Gym and DeepMind Control Suite with very simple configurations. Moreover, RLzoo supports robot learning benchmark environment RLBench based on Vrep/Pyrep simulator. Other large-scale distributed training framework for more realistic scenarios with Unity 3D, Mujoco, Bullet Physics, etc, will be supported in the future. A Springer textbook is also provided, you can get the free PDF if your institute has Springer license. Different from RLzoo for simple usage with high-level APIs, we also have a RL tutorial that aims to make the reinforcement learning tutorial simple, transparent and straight-forward with low-level APIs, as this would not only benefits new learners of reinforcement learning, but also provide convenience for senior researchers to testify their new ideas quickly.

reinforcement-learning deep-learning tensorflow deep-reinforcement-learning tensorlayer reinforcement-learning-practicesCoach is a python reinforcement learning research framework containing implementation of many state-of-the-art algorithms. It exposes a set of easy-to-use APIs for experimenting with new RL algorithms, and allows simple integration of new environments to solve. Basic RL components (algorithms, environments, neural network architectures, exploration policies, ...) are well decoupled, so that extending and reusing existing components is fairly painless.

coach openai-gym reinforcement-learning tensorflow rl carla imitation-learning mujoco roboschool deep-learning hierarchical-reinforcement-learning starcraft starcraft2 starcraft2-aiCompared to a classical approach, using a Recurrent Neural Networks (RNN) with Long Short-Term Memory cells (LSTMs) require no or almost no feature engineering. Data can be fed directly into the neural network who acts like a black box, modeling the problem correctly. Other research on the activity recognition dataset can use a big amount of feature engineering, which is rather a signal processing approach combined with classical data science techniques. The approach here is rather very simple in terms of how much was the data preprocessed. Let's use Google's neat Deep Learning library, TensorFlow, demonstrating the usage of an LSTM, a type of Artificial Neural Network that can process sequential data / time series.

machine-learning deep-learning lstm human-activity-recognition neural-network rnn recurrent-neural-networks tensorflow activity-recognitionThe goal of the project is to provide machine learning for everyone, both technical and non-technical users. I needed a tool sometimes, which I can use to fast create a machine learning prototype. Whether to build some proof of concept or create a fast draft model to prove a point. I find myself often stuck at writing boilerplate code and/or thinking too much of how to start this.

data-science machine-learning automation neural-network scikit-learn sklearn machine-learning-algorithms artificial-intelligence neural-networks data-analysis machine-learning-library machinelearning preprocessing automl multilayer-perceptron-network scikitlearn-machine-learning multilayer-perceptron automl-api automl-algorithms automl-experimentsThe Accord.NET project provides machine learning, statistics, artificial intelligence, computer vision and image processing methods to .NET. It can be used on Microsoft Windows, Xamarin, Unity3D, Windows Store applications, Linux or mobile.

machine-learning framework c-sharp nuget visual-studio statistics unity3d neural-network support-vector-machines computer-vision image-processing ffmpegNetKet is an open-source project delivering cutting-edge methods for the study of many-body quantum systems with artificial neural networks and machine learning techniques. It is a Python library built on JAX. Netket supports MacOS and Linux. We reccomend to install NetKet using pip For instructions on how to install the latest stable/beta release of NetKet see the Getting Started section of our website.

quantum machine-learning-algorithms neural-networks restricted-boltzmann-machine rbm convolutional-neural-networks physics-simulation hamiltonian markov-chain-monte-carlo monte-carlo-methods exact-diagonalization variational-method variational-monte-carlo quantum-state-tomography complex-neural-networkReinforcement Learning with Python will help you to master basic reinforcement learning algorithms to the advanced deep reinforcement learning algorithms. The book starts with an introduction to Reinforcement Learning followed by OpenAI and Tensorflow. You will then explore various RL algorithms and concepts such as the Markov Decision Processes, Monte-Carlo methods, and dynamic programming, including value and policy iteration. This example-rich guide will introduce you to deep learning, covering various deep learning algorithms. You will then explore deep reinforcement learning in depth, which is a combination of deep learning and reinforcement learning. You will master various deep reinforcement learning algorithms such as DQN, Double DQN. Dueling DQN, DRQN, A3C, DDPG, TRPO, and PPO. You will also learn about recent advancements in reinforcement learning such as imagination augmented agents, learn from human preference, DQfD, HER and many more.

reinforcement-learning deep-reinforcement-learning sarsa q-learning policy-gradients deep-q-network deep-learning-algorithms asynchronous-advantage-actor-critic deep-deterministic-policy-gradient deep-recurrent-q-network double-dqn dueling-dqn hindsight-experience-replay drqn trpo ppo[1] Nicolas Schweighofer and Kenji Doya. Meta-learning in reinforcement learning. Neural Networks, 16(1):5–9, 2003. [2] Sepp Hochreiter, A Steven Younger, and Peter R Conwell. Learning to learn using gradient descent. In International Conference on Artificial Neural Networks, pages 87–94. Springer, 2001.

deep-learning meta-learning one-shot-learning learning-to-learn few-shot-learningWhile research in Deep Learning continues to improve the world, we use a bunch of tricks to implement algorithms with TensorLayer day to day. Here are a summary of the tricks to use TensorLayer. If you find a trick that is particularly useful in practice, please open a Pull Request to add it to the document. If we find it to be reasonable and verified, we will merge it in.

tensorlayer tensorflow deep-learning machine-learning data-science neural-network reinforcement-learning neural-networks tensorflow-tutorials tensorflow-models computer-vision tensorflow-framework tensorflow-library tflearn keras tensorboard nlp natural-language-processing lasagne tensorflow-experiments
We have large collection of open source products. Follow the tags from
Tag Cloud >>

Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
**Add Projects.**