It is written in Theano and Lasagne. It uses end-to-end trained embeddings of 5 different emotions to generate responses conditioned by a given emotion. The code is flexible and allows to condition a response by an arbitrary categorical variable defined for some samples in the training data. With CakeChat you can, for example, train your own persona-based neural conversational model[5] or create an emotional chatting machine without external memory[4].
https://cakechat.replika.aiTags | conversational-ai conversational-agents conversational-bots dialogue-agents dialogue-systems dialog-systems nlp deep-learning seq2seq seq2seq-chatbot seq2seq-model theano lasagne |
Implementation | Python |
License | Apache |
Platform | Windows Linux |
The Plato Research Dialogue System is a flexible framework that can be used to create, train, and evaluate conversational AI agents in various environments. It supports interactions through speech, text, or dialogue acts and each conversational agent can interact with data, human users, or other conversational agents (in a multi-agent setting). Every component of every agent can be trained independently online or offline and Plato provides an easy way of wrapping around virtually any existing model, as long as Plato's interface is adhered to. Plato wrote several dialogues between characters who argue on a topic by asking questions. Many of these dialogues feature Socrates including Socrates' trial. (Socrates was acquitted in a new trial held in Athens, Greece on May 25th 2012).
nlp machine-learning deep-learning conversational-ui dialogue-systems conversational-ai conversational-agentThis work tries to reproduce the results of A Neural Conversational Model (aka the Google chatbot). It uses a RNN (seq2seq model) for sentence predictions. It is done using python and TensorFlow. The loading corpus part of the program is inspired by the Torch neuralconvo from macournoyer.
chatbot deep-learning tensorflow seq2seqThis is an attempt at implementing Sequence to Sequence Learning with Neural Networks (seq2seq) and reproducing the results in A Neural Conversational Model (aka the Google chatbot). Human: What is the purpose of living? Machine: To live forever.
seq2seq torch machine-learning deep-learning neural-conversation-modelsStealth is a Ruby based framework for creating conversational (voice & chat) bots. It's design is inspired by Ruby on Rails's philosophy of convention over configuration. It has an MVC architecture with the slight caveat that views are aptly named replies. Stealth is extensible. All service integrations are split out into separate Ruby Gems. Things like analytics and natural language processing (NLP) can be added in as gems as well.
chatbot chatbot-framework voice bot bot-framework bots natural-language-processing stealth rails alexa-skill alexa-skills-kit facebook-messenger-bot conversational-ui conversational-bots conversational-agents conversational-aiLicensed under the Apache License, Version 2.0. Copyright 2019 Rasa Technologies GmbH. Copy of the license. A list of the Licenses of the dependencies of the project can be found at the bottom of the Libraries Summary.
nlp bot machine-learning bots botkit chatbot bot-framework machine-learning-library rasa chatbot-framework conversational-agents conversational-bots conversational-aiRasa is an open source machine learning framework to automate text-and voice-based conversations. With Rasa, you can build chatbots on Facebook, Slack, Microsoft Bot Framework, Rocket.Chat, Mattermost, Telegram etc. Rasa's primary purpose is to help you build contextual, layered conversations with lots of back-and-forth. To have a real conversation, you need to have some memory and build on things that were said earlier. Rasa lets you do that in a scalable way.
nlp machine-learning machine-learning-library bot bots botkit conversational-bots conversational-agents conversational-ai spacy mitie chatbot chatbots chatbots-framework bot-framework natural-language-processingImport key components to build HelloBot. Create skills as pre-defined responses for a user's input containing specific keywords. Every skill returns response and confidence.
bot nlp chatbot dialogue-systems question-answering chitchat slot-filling intent-classification entity-extraction named-entity-recognition keras tensorflow deep-learning deep-neural-networks intent-detection dialogue-agents dialogue-managerNote: the repository is not maintained. Feel free to PM me if you'd like to take up the maintainance. Build a general-purpose conversational chatbot based on a hot seq2seq approach implemented in tensorflow. Since it doesn't produce good results so far, also consider other implementations of seq2seq.
seq2seq chatbot tensorflowThe present repo contains the code accompanying the blog post 🦄 How to build a State-of-the-Art Conversational AI with Transfer Learning. This code is a clean and commented code base with training and testing scripts that can be used to train a dialog agent leveraging transfer Learning from an OpenAI GPT and GPT-2 Transformer language model.
nlp deep-learning dialog pytorch neural-networks chatbots transfer-learning gpt gpt-2Quickly add voice to your app with the Alan Platform. Create an in-app voice assistant to enable human-like conversations and provide a personalized voice experience for every user. A powerful web-based IDE where you can write, test and debug dialog scenarios for your voice assistant or chatbot.
machine-learning text-to-speech sdk chatbot voice voice-commands speech-recognition voice-control voice-assistant conversational-ai voice-ai alan-voice alan-ai alan-studio alan-web-sdk alan-sdk-web ai conversational nlp natural-language dialogs dialog-flow enterprise-ai voice-development add-voice-to-your-app voice-interfaceLinks to the implementations of neural conversational models for different frameworks. Contributions are welcomed. A dialog system that is able to express emotions in a text conversation. See online demo.
seq2seq chatbotNOTE: THE CODE IS UNDER DEVELOPMENT, PLEASE ALWAYS PULL THE LATEST VERSION FROM HERE. In recent years, sequence-to-sequence (seq2seq) models are used in a variety of tasks from machine translation, headline generation, text summarization, speech to text, to image caption generation. The underlying framework of all these models are usually a deep neural network which contains an encoder and decoder. The encoder processes the input data and a decoder receives the output of the encoder and generates the final output. Although simply using an encoder/decoder model would, most of the time, produce better result than traditional methods on the above-mentioned tasks, researchers proposed additional improvements over these sequence to sequence models, like using an attention-based model over the input, pointer-generation models, and self-attention models. However, all these seq2seq models suffer from two common problems: 1) exposure bias and 2) inconsistency between train/test measurement. Recently a completely fresh point of view emerged in solving these two problems in seq2seq models by using methods in Reinforcement Learning (RL). In these new researches, we try to look at the seq2seq problems from the RL point of view and we try to come up with a formulation that could combine the power of RL methods in decision-making and sequence to sequence models in remembering long memories. In this paper, we will summarize some of the most recent frameworks that combines concepts from RL world to the deep neural network area and explain how these two areas could benefit from each other in solving complex seq2seq tasks. In the end, we will provide insights on some of the problems of the current existing models and how we can improve them with better RL models. We also provide the source code for implementing most of the models that will be discussed in this paper on the complex task of abstractive text summarization.
reinforcement-learning actor-critic policy-gradient abstractive-text-summarization pointer-generator nlpThis repository is the home for a set of templates and solutions to help build conversational experiences using Azure Bot Service and Bot Framework. Things look a little different around here? Find out more in our Wiki page and here.
bot bot-framework azure-bot-service assistant conversation-experiences virtual-assistant conversation conversational-ui conversational-agents conversational-bots conversational-ai conversational-interfaces virtualassistant virtual va skills microsoftNLP-Models-Tensorflow, Gathers machine learning and tensorflow deep learning models for NLP problems, code simplify inside Jupyter Notebooks 100%. I will attached github repositories for models that I not implemented from scratch, basically I copy, paste and fix those code for deprecated issues.
nlp machine-learning embedded deep-learning chatbot language-detection lstm summarization attention speech-to-text neural-machine-translation optical-character-recognition pos-tagging lstm-seq2seq-tf dnc-seq2seq luong-apiHi! You have just found Seq2Seq. Seq2Seq is a sequence to sequence learning add-on for the python deep learning library Keras. Using Seq2Seq, you can build and train sequence-to-sequence neural network models in Keras. Such models are useful for machine translation, chatbots (see [4]), parsers, or whatever that comes to your mind. That's it! You have successfully compiled a minimal Seq2Seq model! Next, let's build a 6 layer deep Seq2Seq model (3 layers for encoding, 3 layers for decoding).
Stock-Prediction-Models, Gathers machine learning and deep learning models for Stock forecasting, included trading bots and simulations. I code LSTM Recurrent Neural Network and Simple signal rolling agent inside Tensorflow JS, you can try it here, huseinhouse.com/stock-forecasting-js, you can download any historical CSV and upload dynamically.
deep-learning monte-carlo trading-bot lstm stock-market stock-price-prediction seq2seq learning-agents stock-price-forecasting evolution-strategies lstm-sequence stock-prediction-models deep-learning-stock strategy-agent monte-carlo-markov-chainChatterBot is a machine-learning based conversational dialog engine build in Python which makes it possible to generate responses based on collections of known conversations. The language independent design of ChatterBot allows it to be trained to speak any language. An untrained instance of ChatterBot starts off with no knowledge of how to communicate. Each time a user enters a statement, the library saves the text that they entered and the text that the statement was in response to. As ChatterBot receives more input the number of responses that it can reply and the accuracy of each response in relation to the input statement increase. The program selects the closest matching response by searching for the closest matching known statement that matches the input, it then returns the most likely response to that statement based on how frequently each response is issued by the people the bot communicates with.
chatterbot machine-learning chatbot conversation language botThis is a project use seq2seq model to play couplets (对对联)。 This project is written with Tensorflow. You can try the demo at https://ai.binwang.me/couplet. You will need some data to run this program, the dataset can be downloaded from this project.
seq2seq deep-learning machine-learningMinimal Seq2Seq model with attention for neural machine translation in PyTorch. This implementation relies on torchtext to minimize dataset management and preprocessing parts.
seq2seq deep-learning machine-translationAn implementation of the End-to-End Task-Completion Neural Dialogue Systems and A User Simulator for Task-Completion Dialogues. This document describes how to run the simulation and different dialogue agents (rule-based, command line, reinforcement learning). More instructions to plug in your customized agents or user simulators are in the Recipe section of the paper.
user-simulator nlg dialogue-agents nlu end-to-end
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.