Jovo Framework - Build cross-platform voice applications for Amazon Alexa and Google Home

  •        134

Jovo is the first open source framework that lets you build voice apps for both Amazon Alexa and Google Assistant with only one code base. Besides cross-platform development, Jovo also offers a variety of integrations and easy prototyping capabilities.

https://github.com/jovotech/jovo-framework-nodejs
https://www.jovo.tech/

Tags
Implementation
License
Platform

   




Related Projects

sonus - :speech_balloon: /so.nus/ STT (speech to text) for Node with offline hotword detection


Sonus lets you quickly and easily add a VUI (Voice User Interface) to any hardware or software project. Just like Alexa, Google Now, and Siri, Sonus is always listening offline for a customizable hotword. Once that hotword is detected your speech is streamed to the cloud recognition service of your choice - then you get the results. Generally, running npm install should suffice. This module however, requires you to install SoX.

Google-Actions-Java-SDK - Unofficial Google Actions Java SDK - for Android engineers and all Java lovers


Official Google Actions SDK is written in Node.js. But in many situations voice interfaces like Google Home or Google Assistant will extend or replace mobile apps. If you are old fashioned Android engineer and the most of your code is already written in Java, why not reuse it and build voice extension to app on your own? And this is the main reason to build Google Actions Java SDK - enabling as much developers as possible to build their brilliant ideas for Google Assistant and Home. Currently this is just working proof of concept of Google Actions Java SDK. It means that there is no documentation, fixed interface, (not much) unit tests and many, many others.

avs-device-sdk - An SDK for commercial device makers to integrate Alexa directly into connected products


If you are updating from v1.3 or earlier to v1.6, you must update your AlexaClientSDKConfig.json to include a Notifications database. An updated sample is available in the quickstart guides for Ubuntu Linux, Raspberry Pi, macOS, and Generic Linux. The Alexa Voice Service (AVS) enables developers to integrate Alexa directly into their products, bringing the convenience of voice control to any connected device. AVS provides developers with access to a suite of resources to quickly and easily build Alexa-enabled products, including APIs, hardware development kits, software development kits, and documentation.

alexa-skills-kit-js - SDK and example code for building voice-enabled skills for the Amazon Echo.


This Node.js Alexa Skills Kit project has been deprecated. A new Alexa Skills Kit SDK for Node.js, along with updated skill samples, is available in our new home: https://github.com/alexa. If you still need access to the deprecated Node.js Alexa Skills Kit project, please go to deprecated branch.


geemusic - A bridge between Google Music and Amazon's Alexa


GeeMusic is an Alexa skill which bridges Google Music and Amazon's Alexa. It hopes to rescue all of those who want an Echo/Dot but don't want to switch off of Google Music or pay extra for an Amazon Music Unlimited subscription. This project is still in its early phases and subject to a bit of change, however it is functional and ready for use! The only catch is that you'll need to run it on your own server for the time being (ideally I'll eventually release this on the Alexa Skills marketplace, but there's a lot of work to do before then).

Abot - Digital Assistant Framework


Abot (pronounced Eh-Bot, like the Canadians) is a digital assistant framework that enables anyone to easily build a digital assistant similar to Apple's Siri, Microsoft's Cortana, Google Now, or Amazon Alexa. Further, Abot supports a human-aided training backend enabling anyone to build services like Facebook M. Abot is the first A.I. framework that aims to be available everywhere and— ultimately—to do everything.

alexa-avs-sample-app - 2018-01-25 - The AVS Java Sample App is maintenance mode


⚠️ Starting January 25, 2018, the AVS Java Sample App will be put into maintenance mode. To leverage the latest Alexa features, please use the AVS Device SDK C++ Sample App, which you can find here. To discuss any specific dependencies on the AVS Java Sample App, feel free to reach out to us here. This project provides a step-by-step walkthrough to help you build a hands-free Alexa Voice Service (AVS) prototype in 60 minutes, using wake word engines from Sensory or KITT.AI. Now, in addition to pushing a button to "start listening", you can now also just say the wake word "Alexa", much like the Amazon Echo. You can find step-by-step instructions to set up the hands-free prototype on Raspberry Pi, or follow the instructions to set up the push-to-talk only prototype on Linux, Mac, or Windows.

alexa-skills-kit-sdk-for-nodejs - The Alexa Skills Kit SDK for Node


Alexa SDK team is proud to present the new Alexa Node.js SDK -- the open-source Alexa Skill Development Kit built by developers for developers. Creating an Alexa skill using the Alexa Skill Kit, Node.js and AWS Lambda has become one of the most popular ways we see skills created today. The event-driven, non-blocking I/O model of Node.js is well-suited for an Alexa skill and Node.js is one of the largest ecosystems of open source libraries in the world. Plus, AWS Lambda is free for the first one million calls per month, which is sufficient for most Alexa skill developers. Also, when using AWS Lambda you don't need to manage any SSL certificates since the Alexa Skills Kit is a trusted trigger.

alexa-cookbook - A series of sample code projects to be used for educational purposes during Alexa hackathons and workshops, and as a reference for tutorials and blog posts


These folders contain short code samples that show you how to add features to an Alexa skill. Let's assume we are a chef, with an idea for a fancy new recipe. The recipe will require one or more ingredients to be combined to create the finished product. Alexa skills follow the same model. Basic ingredients can be added together to create a robust skill. This Cookbook is a desk reference for other Alexa skill design materials. Back to the github.com/alexa home page.

ask-alexa-pykit - A minimalist SDK for developing skills for the Amazon Echo's ASK - Alexa Skills Kit using Amazon Web Services's Python Lambda Functions


A minimalist framework for developing apps (skills) for the Amazon Echo's SDK: The Alexa Skills Kit (ASK). Let me know if you know any other projects that use or build on top of ask-alexa-pykit.

skill-sample-nodejs-fact - Build An Alexa Fact Skill


This Alexa sample skill is a template for a basic fact skill. Provided a list of interesting facts about a topic, Alexa will select a fact at random and tell it to the user when the skill is invoked. If you would like to see an example of this skill in action, you can enable the Gloucester Facts from the Alexa Skill Store.

Google Speech Recognition Example


Google Speech Recognition contains a working example of application that uses google speech recognition API. App contains all necessary dlls to record, decode and send your voice request to google service and recieve a text representation of what you've said. It's developed i...

alexa-home - Using Amazon Echo to control the home! http://linuxforhobbyists.com to learn to DIY


Welcome to Alexa Home! The goal of this project is to use your Amazon Echo* to control various home automation software. Interested in integrating a new module? Check out some of the code and send in a pull request! Glad to help out. The repository includes two components: a web scraper to get commands from Amazon Echo history (watir-login.rb) and a Sinatra server that takes those commands and, using various modules, triggers certain actions.

echo-sonos - Amazon Echo integration with Sonos


All of the pieces for an Amazon Echo (Alexa) <-> Sonos integration. echo-sonos supports Apple Music, Spotify, Deezer, Deezer Elite, Sonos playlists, Sonos favorites, SiriusXM, the local Sonos music library, and configurable node-sonos-http-api presets.

mycroft-skills - A repository for sharing and collaboration for third-party Mycroft skills development


The official home of Skills for the Mycroft ecosystem. These Skills are written by both the MycroftAI team and others within the Community. If you are building Skills, please ensure that you use the Meta Editor for your README.md file. The Skills list is generated from parsing the README.md files.

mycroft-core - Mycroft Core, the Mycroft Artificial Intelligence platform.


Mycroft is a hackable open source voice assistant. This script sets up dependencies and a virtualenv. If running in an environment besides Ubuntu/Debian, Arch or Fedora you may need to manually install packages as instructed by dev_setup.sh.

susi_desktop - Desktop Client for http://api.susi.ai


Susi AI is an intelligent Open Source personal assistant. It is capable of chat and voice interaction by using APIs to perform actions such as music playback, making to-do lists, setting alarms, streaming podcasts, playing audiobooks, and providing weather, traffic, and other real time information. Additional functionalities can be added as console services using external APIs. Susi AI is able to answer questions and depending on the context will ask for additional information in order to perform the desired outcome. The core of the assistant is the Susi AI server that holds the "intelligence" and "personality" of Susi AI. The Android and web applications make use of the APIs to access information from a hosted server. The app is based on electron hence make sure that you have electron installed, else simply run npm i -g electron.

voice-elements - :speaker: Web Component wrapper to the Web Speech API, that allows you to do voice recognition and speech synthesis using Polymer


Web Component wrapper to the Web Speech API, that allows you to do voice recognition (speech to text) and speech synthesis (text to speech) using Polymer. Or download as ZIP.