# Technology Overview

Mycroft is the name of a suite of software and hardware tools that use [natural language processing](https://en.wikipedia.org/wiki/Natural_language_processing) and [machine learning](https://en.wikipedia.org/wiki/Machine_learning) to provide an open source voice assistant.

{% embed url="<https://www.youtube.com/watch?v=m4L0QfzUeEI>" %}
Our vision for Mycroft
{% endembed %}

## Mycroft components

Mycroft is modular. Some components can be easily 'swapped out' for others:

* Wake Word detection
* Speech to Text (STT)
* Intent parser

### Wake Word detection

A Wake Word is a phrase you use to tell Mycroft you're about to issue a command. By default, this is `Hey Mycroft`, but you can configure your own Wake Word in your [Mycroft Home](https://home.mycroft.ai) account.

There are two technologies that Mycroft.AI currently uses for Wake Word detection:

* [PocketSphinx](https://github.com/cmusphinx/pocketsphinx): PocketSphinx is part of the broader [CMUSphinx package](https://cmusphinx.github.io/), developed by [Carnegie Mellon University](https://www.cmu.edu). PocketSphinx is a lightweight speech recognition engine, specifically tuned for handheld and mobile devices.

Because PocketSphinx is trained on English speech, your Wake Word currently needs to be an English word, like `Hello Mike`, `Hi there Mickey` or `Hey Mike`. Wake Words in other languages, like Spanish, French or German, won't work as well.

* [Precise](https://mycroft.ai/documentation/precise): Unlike PocketSphinx, which is based on Speech to Text technology, Precise is a neural network that is trained on audio data. It doesn't matter what *words* you want to use for your Wake Word. Instead, you train it on *sounds*. The downside is that Precise needs to be trained on your chosen Wake Word. Precise is the default Wake Word Listener for the "Hey Mycroft" wake word, PocketSphinx provides a fallback to this if Precise is unavailable.

### Speech to Text (STT)

Speech to Text (STT) software is used to take spoken words, and turn them into text phrases that can then be acted on.

We are working with Mozilla to build [DeepSpeech](https://github.com/mozilla/DeepSpeech). A fully open source STT engine, based on Baidu’s Deep Speech architecture and implemented with Google’s [TensorFlow](https://www.tensorflow.org/) framework.

DeepSpeech is not yet ready for production use and Mycroft currently uses [Google STT](https://cloud.google.com/speech/) as the default STT engine.

Mycroft also supports other STT engines that can be configured using the [Configuration Manager](https://mycroft-ai.gitbook.io/docs/using-mycroft-ai/customizations/config-manager):

* [IBM Watson Speech to Text](https://www.ibm.com/watson/services/speech-to-text/) (IBM API key required)
* [wit.ai Speech to Text](https://wit.ai/blog/2014/02/12/speech-api) (wit.ai API key required)

### Intent parser

An intent parser is software which identifies what the user's *intent* is based on their speech. An intent parser usually takes the output of a Speech to Text (STT) engine as an input.

For example, Julie Speaks the following to Mycroft: `Hey Mycroft, tell me about the weather`

Julie's *intent* is to find out about the weather (probably in her current location).

An intent parser can then match the *intent* with a suitable Skill to handle the *intent*.

* [Adapt intent parser](https://github.com/MycroftAI/adapt): Adapt is the default intent parser for all Mycroft platforms. Adapt was developed by Mycroft and is available under an open source license.
* [Padatious](https://github.com/MycroftAI/padatious): Padatious is a neural network based intent parser. Padatious is currently under active development by Mycroft and is available under an open source license. It is likely that some Mycroft platforms will switch to using Padatious in the future instead of Adapt.

### Text to Speech

Text to Speech (TTS) software takes written text, such as text files on a computer, and uses a *voice* to *speak* the text. Text to Speech can have different voices, depending on the TTS engine used.

* [Mimic](https://github.com/MycroftAI/mimic): Mycroft's default local text to speech (TTS) engine, based on CMU's Flite (Festival Lite)
* [Mimic2](https://github.com/MycroftAI/mimic2): Mycroft's own cloud based text to speech (TTS) engine, based on Tacotron providing a much better voice quality.

In your home.mycroft.ai account, you can select voices from these as well as

* [Google TTS](https://play.google.com/store/apps/details?id=com.google.android.tts): you need to choose which voice to use

even more tts engines are available but require manual configuration.

### Middleware

The Mycroft middleware has two components:

* [Mycroft Core](https://github.com/MycroftAI/mycroft-core): this code, written in Python, is the core software that provides the 'glue' between other modules. Mycroft Core is available under an Apache 2.0 open source license.
* [Mycroft Home and Mycroft API](https://home.mycroft.ai): this is the platform where data on Users and Devices is held. This platform provides abstraction services, such as storing API keys that are used to access third-party services to provide Skill functionality. The code for this platform is available under an AGPL 3.0 open source license.

### Mycroft Skills

[Mycroft Skills](https://github.com/MycroftAI/mycroft-skills) are like 'add-ons' or 'plugins' that provide additional functionality. Skills can be developed by Mycroft Developers, or by Community Developers, and vary in their functionality and maturity.

[Mycroft Skills Kit (MSK)](https://github.com/mycroftai/mycroft-skills-kit) is a Python-based utility that has been created to make it easier for Skill Authors to create, test and submit Skills to the [Skills Marketplace](https://market.mycroft.ai).

{% content-ref url="mycroft-skills-kit" %}
[mycroft-skills-kit](https://mycroft-ai.gitbook.io/docs/mycroft-technologies/mycroft-skills-kit)
{% endcontent-ref %}

[Mycroft Skills Manager (MSM)](https://github.com/mycroftai/mycroft-skills-manager) is a command line tool used to add, manage and remove Skills on any Mycroft installation.

### Devices and Enclosures

Mycroft is designed to run on many different platforms. Each dedicated platform is called a device, these include:

* **Mark 1** - our first reference hardware device using a dedicated software image.
* **Mark 2** - our latest reference hardware device using a dedicated software image.
* **Picroft** - any Raspberry Pi 3 or 4 that is running the Picroft software image.

The enclosure refers to the specific code that is required for that device. It might define unique functionality such as the eyes on the Mark 1, or a specific way of interacting with the hardware, such as controlling the volume levels at a hardware level via i2c.


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://mycroft-ai.gitbook.io/docs/mycroft-technologies/overview.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
