<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://openvoice-tech.net/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Florian</id>
	<title>Open Voice Technology Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://openvoice-tech.net/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Florian"/>
	<link rel="alternate" type="text/html" href="https://openvoice-tech.net/wiki/Special:Contributions/Florian"/>
	<updated>2026-04-16T21:31:05Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2389</id>
		<title>SEPIA</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2389"/>
		<updated>2022-05-13T08:50:03Z</updated>

		<summary type="html">&lt;p&gt;Florian: updated SEPIA version info&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Voice Assistant&lt;br /&gt;
| Url = https://github.com/SEPIA-Framework/&lt;br /&gt;
| License = MIT License&lt;br /&gt;
| InternetRequired = No, but services like weather and Wiki won&#039;t work&lt;br /&gt;
| SupportedLanguages = en, de + partial support for 15 more (beta)&lt;br /&gt;
| WakeWord = Hey SEPIA, customizable&lt;br /&gt;
| DefaultTTS = eSpeak, picoTTS, MBROLA, MaryTTS + compatible APIs&lt;br /&gt;
| SupportedOS = Windows, Linux, Mac OS X, Raspberry Pi&lt;br /&gt;
| CodeLanguage = Java (Assist-Server), HTML/Javascript (Client), Python (STT-Server)&lt;br /&gt;
| LatestVersion = 2.6.2&lt;br /&gt;
| LatestReleaseDate = 2022-05-08&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Introduction: S.E.P.I.A. Open Assistant Framework ==&lt;br /&gt;
[https://sepia-framework.github.io/ S.E.P.I.A.] is an acronym for: &#039;&#039;&#039;s&#039;&#039;&#039;elf-hosted, &#039;&#039;&#039;e&#039;&#039;&#039;xtendable, &#039;&#039;&#039;p&#039;&#039;&#039;ersonal, &#039;&#039;&#039;i&#039;&#039;&#039;ntelligent &#039;&#039;&#039;a&#039;&#039;&#039;ssistant. It is a modular, open-source framework equipped with all the required tools to build your own, full-fledged digital voice-assistant, including [[:Category:STT|speech recognition]] ([[:Category:STT|STT]]), [[:Category:Wake words|wake-word]] detection, [[:Category:TTS|text-to-speech]] ([[:Category:TTS|TTS]]), [[Natural language understanding|natural-language-understanding (NLU)]], dialog-management, SDK(s), a cross-platform client app and much more.&lt;br /&gt;
&lt;br /&gt;
The framework consists of several, highly customizable micro-services that work together seamlessly to form the SEPIA Open Assistant. It follows the &#039;&#039;&#039;client-server principle using a lightweight Java server and Elasticsearch DB as &amp;quot;brain&amp;quot;&#039;&#039;&#039; and a JavaScript based client that works for example as &#039;&#039;&#039;smart-speaker, smart-display or full-fledged digital assistant app&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
All components work on Linux, Windows and Mac and have been optimized to even &#039;&#039;&#039;run smoothly on a Raspberry Pi&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Out-of-the-box SEPIA currently has smart-services for: news, music (radio), timers, alarms, reminders, to-do and shopping lists, smart home (e.g. using open-source tools like openHAB), navigation, places, weather, Wikipedia, web-search, soccer-results (Bundesliga), a bit of small-talk and more. To build custom services there is a Java SDK and a code editor integrated into the SEPIA Control HUB web-app. The client can be extended with custom HTML widgets.&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-docs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Components==&lt;br /&gt;
Some of the core SEPIA framework components:&lt;br /&gt;
&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-assist-server SEPIA Assist-Server] - The &amp;quot;brain&amp;quot; of SEPIA responsible for: user-accounts, database integration, [[Natural language understanding|NLU]], dialog-management, open-source [[:Category:TTS|TTS]], smart-services (weather, navigation, alarms, news etc.), remote actions and more&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-assist-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-websocket-server-java SEPIA Chat-Server] - A [[wikipedia:WebSocket|WebSocket]] based Chat-Server that takes care of real-time, asynchronous user-to-user and user-to-assistant communication&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-websocket-server-java&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-html-client-app SEPIA Cross-Platform Client] - The primary SEPIA client based on HTML5 web technology that runs on all modern browsers (mobile and desktop) and communicates with other SEPIA components via HTTP or WebSocket connections&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-html-client-app&amp;lt;/ref&amp;gt;. It supports headless- and kiosk-mode via [https://github.com/bytemind-de/nodejs-client-extension-interface CLEXI server] to build independent smart-speaker or smart-display like devices&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-installation-and-setup/tree/master/sepia-client-installation&amp;lt;/ref&amp;gt;. The client is available as Android app as well via [https://play.google.com/store/apps/details?id=de.bytemind.sepia.app.web Play Store] or [https://github.com/SEPIA-Framework/sepia-installation-and-setup/releases direct download].&lt;br /&gt;
*[[SEPIA Speech-To-Text Server|SEPIA STT-Server]] - A WebSocket based, full-duplex Python server for real-time [[:Category:STT|automatic speech recognition]] (ASR) supporting [https://github.com/SEPIA-Framework/sepia-stt-server multiple open-source ASR engines]. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
There are many more SEPIA components and tools to discover on the [https://github.com/SEPIA-Framework official GitHub project page].&lt;br /&gt;
&lt;br /&gt;
==Installation==&lt;br /&gt;
The [https://github.com/SEPIA-Framework/sepia-docs official documentation page] has the most recent installation guides.&lt;br /&gt;
&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;br /&gt;
[[Category:Project]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2255</id>
		<title>SEPIA</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2255"/>
		<updated>2022-01-30T22:59:47Z</updated>

		<summary type="html">&lt;p&gt;Florian: Updated SEPIA release version&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Voice Assistant&lt;br /&gt;
| Url = https://github.com/SEPIA-Framework/&lt;br /&gt;
| License = MIT License&lt;br /&gt;
| InternetRequired = No, but services like weather and Wiki won&#039;t work&lt;br /&gt;
| SupportedLanguages = en, de + partial support for 15 more (beta)&lt;br /&gt;
| WakeWord = Hey SEPIA, customizable&lt;br /&gt;
| DefaultTTS = eSpeak, picoTTS, MBROLA, MaryTTS + compatible APIs&lt;br /&gt;
| SupportedOS = Windows, Linux, Mac OS X, Raspberry Pi&lt;br /&gt;
| CodeLanguage = Java (Assist-Server), HTML/Javascript (Client), Python (STT-Server)&lt;br /&gt;
| LatestVersion = 2.6.1&lt;br /&gt;
| LatestReleaseDate = 2022-01-30&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Introduction: S.E.P.I.A. Open Assistant Framework ==&lt;br /&gt;
[https://sepia-framework.github.io/ S.E.P.I.A.] is an acronym for: &#039;&#039;&#039;s&#039;&#039;&#039;elf-hosted, &#039;&#039;&#039;e&#039;&#039;&#039;xtendable, &#039;&#039;&#039;p&#039;&#039;&#039;ersonal, &#039;&#039;&#039;i&#039;&#039;&#039;ntelligent &#039;&#039;&#039;a&#039;&#039;&#039;ssistant. It is a modular, open-source framework equipped with all the required tools to build your own, full-fledged digital voice-assistant, including [[:Category:STT|speech recognition]] ([[:Category:STT|STT]]), [[:Category:Wake words|wake-word]] detection, [[:Category:TTS|text-to-speech]] ([[:Category:TTS|TTS]]), [[Natural language understanding|natural-language-understanding (NLU)]], dialog-management, SDK(s), a cross-platform client app and much more.&lt;br /&gt;
&lt;br /&gt;
The framework consists of several, highly customizable micro-services that work together seamlessly to form the SEPIA Open Assistant. It follows the &#039;&#039;&#039;client-server principle using a lightweight Java server and Elasticsearch DB as &amp;quot;brain&amp;quot;&#039;&#039;&#039; and a JavaScript based client that works for example as &#039;&#039;&#039;smart-speaker, smart-display or full-fledged digital assistant app&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
All components work on Linux, Windows and Mac and have been optimized to even &#039;&#039;&#039;run smoothly on a Raspberry Pi&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Out-of-the-box SEPIA currently has smart-services for: news, music (radio), timers, alarms, reminders, to-do and shopping lists, smart home (e.g. using open-source tools like openHAB), navigation, places, weather, Wikipedia, web-search, soccer-results (Bundesliga), a bit of small-talk and more. To build custom services there is a Java SDK and a code editor integrated into the SEPIA Control HUB web-app. The client can be extended with custom HTML widgets.&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-docs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Components==&lt;br /&gt;
Some of the core SEPIA framework components:&lt;br /&gt;
&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-assist-server SEPIA Assist-Server] - The &amp;quot;brain&amp;quot; of SEPIA responsible for: user-accounts, database integration, [[Natural language understanding|NLU]], dialog-management, open-source [[:Category:TTS|TTS]], smart-services (weather, navigation, alarms, news etc.), remote actions and more&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-assist-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-websocket-server-java SEPIA Chat-Server] - A [[wikipedia:WebSocket|WebSocket]] based Chat-Server that takes care of real-time, asynchronous user-to-user and user-to-assistant communication&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-websocket-server-java&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-html-client-app SEPIA Cross-Platform Client] - The primary SEPIA client based on HTML5 web technology that runs on all modern browsers (mobile and desktop) and communicates with other SEPIA components via HTTP or WebSocket connections&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-html-client-app&amp;lt;/ref&amp;gt;. It supports headless- and kiosk-mode via [https://github.com/bytemind-de/nodejs-client-extension-interface CLEXI server] to build independent smart-speaker or smart-display like devices&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-installation-and-setup/tree/master/sepia-client-installation&amp;lt;/ref&amp;gt;. The client is available as Android app as well via [https://play.google.com/store/apps/details?id=de.bytemind.sepia.app.web Play Store] or [https://github.com/SEPIA-Framework/sepia-installation-and-setup/releases direct download].&lt;br /&gt;
*[[SEPIA Speech-To-Text Server|SEPIA STT-Server]] - A WebSocket based, full-duplex Python server for real-time [[:Category:STT|automatic speech recognition]] (ASR) supporting [https://github.com/SEPIA-Framework/sepia-stt-server multiple open-source ASR engines]. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
There are many more SEPIA components and tools to discover on the [https://github.com/SEPIA-Framework official GitHub project page].&lt;br /&gt;
&lt;br /&gt;
==Installation==&lt;br /&gt;
The [https://github.com/SEPIA-Framework/sepia-docs official documentation page] has the most recent installation guides.&lt;br /&gt;
&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;br /&gt;
[[Category:Project]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Glossary_of_voice_tech&amp;diff=2254</id>
		<title>Glossary of voice tech</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Glossary_of_voice_tech&amp;diff=2254"/>
		<updated>2022-01-28T11:50:59Z</updated>

		<summary type="html">&lt;p&gt;Florian: added category &amp;quot;Glossary&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Open Voice Tech]]&lt;br /&gt;
&lt;br /&gt;
In the field of voice technology there are lots of buzzwords. Some are self explaining, other lead to confusion regularly. This list should be a glossary.&lt;br /&gt;
&lt;br /&gt;
==General terms==&lt;br /&gt;
&lt;br /&gt;
*[[:Category:Dataset|Dataset]]&lt;br /&gt;
*[[Research papers|Papers]] (&#039;&#039;research papers&#039;&#039;)&lt;br /&gt;
*[[Phonemes]]&lt;br /&gt;
*[[Model]]&lt;br /&gt;
*[[Checkpoint]]&lt;br /&gt;
*[[Repository]]&lt;br /&gt;
&lt;br /&gt;
==STT terms==&lt;br /&gt;
&lt;br /&gt;
*[[:Category:Wake words|Wake word]]&lt;br /&gt;
*[[Hotword]]&lt;br /&gt;
*[[Voice print]]&lt;br /&gt;
*[[Word error rate]] (&#039;&#039;WER&#039;&#039;)&lt;br /&gt;
*[[Diarization]]&lt;br /&gt;
*[[Barge-in]]&lt;br /&gt;
&lt;br /&gt;
==TTS terms==&lt;br /&gt;
&lt;br /&gt;
*&lt;br /&gt;
&lt;br /&gt;
==Voice assistant terms==&lt;br /&gt;
&lt;br /&gt;
*[[Conversational AI]]&lt;br /&gt;
*[[Natural language understanding]] (&#039;&#039;NLU&#039;&#039;)&lt;br /&gt;
*[[Intent]]&lt;br /&gt;
*[[Utterance]]&lt;br /&gt;
*[[Voiceonly]]&lt;br /&gt;
&lt;br /&gt;
==Machine learning==&lt;br /&gt;
&lt;br /&gt;
*[[Epoch]]&lt;br /&gt;
*[[Step]]&lt;br /&gt;
*[[Batch size]]&lt;br /&gt;
*[[Learning rate]]&lt;br /&gt;
*[[Inference]]&lt;br /&gt;
*[[Alignment]]&lt;br /&gt;
[[Category:Glossary]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Barge-in&amp;diff=2253</id>
		<title>Barge-in</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Barge-in&amp;diff=2253"/>
		<updated>2022-01-28T11:50:23Z</updated>

		<summary type="html">&lt;p&gt;Florian: removed category &amp;quot;Open Voice Assistants&amp;quot; since this is a vocabulary explanation. Created new category &amp;quot;Glossary&amp;quot; for it.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&amp;quot;Barge-in is a feature that allows callers to interrupt a prompt and provide their response before the prompt has finished playing&amp;quot;&#039;&#039; at pag. 24 of book &amp;quot;Voice User Interface Design&amp;quot; by James Giangola et al.&lt;br /&gt;
&lt;br /&gt;
In other words, barge-in is, in any voice interface system / [[:Category:Open Voice Assistants|voice assistant]], the user capability of interrupt/stop the assistant spoken (a [[:Category:TTS|text-to-speech]] synthetic voice play), to impose a new overriding user voice request to be processed asap.&lt;br /&gt;
[[Category:Glossary]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Mycroft_Precise&amp;diff=2252</id>
		<title>Mycroft Precise</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Mycroft_Precise&amp;diff=2252"/>
		<updated>2022-01-28T11:47:44Z</updated>

		<summary type="html">&lt;p&gt;Florian: Removed category &amp;quot;Open Voice Assistants&amp;quot; because Precise is a component of Mycroft&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Mycroft]]&lt;br /&gt;
[[Category:Wake words]]&lt;br /&gt;
&lt;br /&gt;
Mycroft Precise is a simple to use [[:Category:Wake words|wake word listener]] created by [[Mycroft]].&lt;br /&gt;
&lt;br /&gt;
Precise can be used to model custom wake words.  They have a good [https://github.com/MycroftAI/mycroft-precise/wiki/Training-your-own-wake-word#how-to-train-your-own-wake-word documentation page] on that, with additional tips and tricks from other [[Mycroft]] community members [https://github.com/el-tocino/localcroft/blob/master/precise/Precise.md here] and [https://github.com/sparky-vision/mycroft-precise-tips here].&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Porcupine&amp;diff=2251</id>
		<title>Porcupine</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Porcupine&amp;diff=2251"/>
		<updated>2022-01-28T11:39:58Z</updated>

		<summary type="html">&lt;p&gt;Florian: Minor text format updates&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Porcupine by Picovoice.ai &amp;lt;ref&amp;gt;https://picovoice.ai/platform/porcupine/&amp;lt;/ref&amp;gt; is a commercial wake-word engine with some open components licensed under Apache 2.0. The open-source parts change from version to version but usually include a SDK to integrate the engine and a number of wake-words that can be used freely like &amp;quot;Computer&amp;quot;, &amp;quot;Jarvis&amp;quot;, &amp;quot;Alexa&amp;quot;, &amp;quot;Hey Siri&amp;quot; etc..&lt;br /&gt;
&lt;br /&gt;
When it was first released to the public in March 2018 it quickly became the go-to solution for open-source projects due to its high accuracy, low resource requirements and good platform support though the selection of free wake-words was pretty limited at that time (Picovoice, Alexa, Raspberry and some more &amp;quot;exotic&amp;quot; words &amp;lt;ref&amp;gt;https://github.com/Picovoice/porcupine/tree/0507c5250b50d0a938b714ba604819d61fc9e602/resources/keyword_files&amp;lt;/ref&amp;gt;). &lt;br /&gt;
&lt;br /&gt;
The most unique aspect of Porcupine is the ability to create new wake-words from just one line of text, no audio recording, no training data! &amp;lt;ref&amp;gt;https://medium.com/@alirezakenarsarianhari/yet-another-wake-word-detection-engine-a2486d36d8d4&amp;lt;/ref&amp;gt;. Unfortunately this feature was only available (with some very limited exceptions) to commercial customers until Porcupine v2.0. Since v2.0 everyone can create their own wake-words via the Picovoice online console within the limited free-tier &amp;lt;ref&amp;gt;https://picovoice.ai/console/&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://picovoice.ai/blog/introducing-picovoices-free-tier/&amp;lt;/ref&amp;gt;. The drawback of version 2.0 is that the usage requires an API key and devices need to be activated by contacting the Picovoice server every now and then.&lt;br /&gt;
&lt;br /&gt;
== Features Summary ==&lt;br /&gt;
&lt;br /&gt;
* Commercial product with a free tier and some open parts. Access key and online activation since v2.0 (December 2021).&lt;br /&gt;
* At the time of writing (Jan. 2022) probably the best wake-word engine for open-source projects in terms of accuracy and platform support (empiric data, needs some real statistics).&lt;br /&gt;
* Available on basically any platform and in any programming language from x86 to ARM, from PCs to microcontrollers, from Android to Raspberry Pi, from Python to Javascript (WASM support).&lt;br /&gt;
*Popular wake-words available in free version (e.g. &amp;quot;Computer&amp;quot;, &amp;quot;Jarvis&amp;quot;, &amp;quot;Alexa&amp;quot; etc.). To avoid registration, access key management and online activation use Porcupine v1.9.&lt;br /&gt;
* Everyone can create custom wake-words with one line of text via web console (v2.0).&lt;br /&gt;
* Low resource consumption (works on Raspberry Pi Zero).&lt;br /&gt;
&lt;br /&gt;
== Links ==&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Porcupine&amp;diff=2250</id>
		<title>Porcupine</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Porcupine&amp;diff=2250"/>
		<updated>2022-01-28T11:36:59Z</updated>

		<summary type="html">&lt;p&gt;Florian: updated some info text&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Porcupine by Picovoice.ai &amp;lt;ref&amp;gt;https://picovoice.ai/platform/porcupine/&amp;lt;/ref&amp;gt; is a commercial wake-word engine with some open components licensed under Apache 2.0. The open-source parts change from version to version but usually include a SDK to integrate the engine and a number of wake-words that can be used freely like &amp;quot;Computer&amp;quot;, &amp;quot;Jarvis&amp;quot;, &amp;quot;Alexa&amp;quot;, &amp;quot;Hey Siri&amp;quot; etc..&lt;br /&gt;
&lt;br /&gt;
When it was first released to the public in March 2018 it quickly became the go-to solution for open-source projects due to its high accuracy, low resource requirements and good platform support though the selection of free wake-words was pretty limited at that time (Picovoice, Alexa, Raspberry and some more &amp;quot;exotic&amp;quot; words &amp;lt;ref&amp;gt;https://github.com/Picovoice/porcupine/tree/0507c5250b50d0a938b714ba604819d61fc9e602/resources/keyword_files&amp;lt;/ref&amp;gt;). &lt;br /&gt;
&lt;br /&gt;
The most unique aspect of Porcupine is the ability to create new wake-words from just one line of text, no audio recording, no training data! &amp;lt;ref&amp;gt;https://medium.com/@alirezakenarsarianhari/yet-another-wake-word-detection-engine-a2486d36d8d4&amp;lt;/ref&amp;gt;. Unfortunately this feature was only available (with some very limited exceptions) to commercial customers until Porcupine v2.0. Since v2.0 everyone can create their own wake-words via the Picovoice online console within the limited free-tier &amp;lt;ref&amp;gt;https://picovoice.ai/console/&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://picovoice.ai/blog/introducing-picovoices-free-tier/&amp;lt;/ref&amp;gt;. The drawback of version 2.0 is that usage requires an API key and devices need to be activated by contacting the Picovoice server every now and then.&lt;br /&gt;
&lt;br /&gt;
== Features Summary ==&lt;br /&gt;
&lt;br /&gt;
* Commercial product with a free tier and some open parts. Access key and online activation since v2.0 (December 2021).&lt;br /&gt;
* At the time of writing (Jan. 2022) probably the best wake-word engine for open-source projects in terms of accuracy and platform support (empiric data, needs some real statistics).&lt;br /&gt;
* Available on basically any platform and in any programming language from x86 to ARM, from PCs to microcontrollers, from Android to Raspberry Pi, from Python to Javascript (WASM support).&lt;br /&gt;
*Popular wake-words available in free version (e.g. &amp;quot;Computer&amp;quot;, &amp;quot;Jarvis&amp;quot;, &amp;quot;Alexa&amp;quot; etc.). To avoid registration access key management and online activation use Porcupine v1.9.&lt;br /&gt;
* Everyone can create custom wake-words with one line of text via web console (v2.0).&lt;br /&gt;
* Low resource consumption (works on Raspberry Pi Zero).&lt;br /&gt;
&lt;br /&gt;
== Links ==&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Porcupine&amp;diff=2249</id>
		<title>Porcupine</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Porcupine&amp;diff=2249"/>
		<updated>2022-01-28T11:33:37Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created Porcupine page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Porcupine by Picovoice.ai &amp;lt;ref&amp;gt;https://picovoice.ai/platform/porcupine/&amp;lt;/ref&amp;gt; is a commercial wake-word engine with some open components licensed under Apache 2.0. The open-source parts change from version to version but usually include a SDK to integrate the engine and a number of wake-words that can be used freely like &amp;quot;Computer&amp;quot;, &amp;quot;Jarvis&amp;quot;, &amp;quot;Alexa&amp;quot;, &amp;quot;Hey Siri&amp;quot; etc..&lt;br /&gt;
&lt;br /&gt;
When it was first released to the public in March 2018 it quickly became the go-to solution for open-source projects due to its high accuracy, low resource requirements and good platform support though the selection of free wake-words was pretty limited at that time (Picovoice, Alexa, Raspberry and some more &amp;quot;exotic&amp;quot; words &amp;lt;ref&amp;gt;https://github.com/Picovoice/porcupine/tree/0507c5250b50d0a938b714ba604819d61fc9e602/resources/keyword_files&amp;lt;/ref&amp;gt;). &lt;br /&gt;
&lt;br /&gt;
The most unique aspect of Porcupine is the ability to create new wake-words from just one line of text, no audio recording, no training data! &amp;lt;ref&amp;gt;https://medium.com/@alirezakenarsarianhari/yet-another-wake-word-detection-engine-a2486d36d8d4&amp;lt;/ref&amp;gt;. Unfortunately this feature was only available (with some very limited exceptions) to commercial customers until Porcupine v2.0. Since v2.0 everyone can create their own wake-words via the Picovoice online console within the limited free-tier &amp;lt;ref&amp;gt;https://picovoice.ai/console/&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://picovoice.ai/blog/introducing-picovoices-free-tier/&amp;lt;/ref&amp;gt;. The drawback of version 2.0 is that usage requires an API key and devices need to be activated by contacting the Picovoice server every now and then.&lt;br /&gt;
&lt;br /&gt;
== Features Summary ==&lt;br /&gt;
&lt;br /&gt;
* Commercial product with a free tier and some open parts. Access key and online activation since v2.0 (December 2021).&lt;br /&gt;
* At the time of writing (Jan. 2022) probably the best wake-word engine for open-source projects in terms of accuracy and platform support (empiric data, needs some real statistics).&lt;br /&gt;
* Available on basically any platform and in any programming language from x86 to ARM, from PCs to microcontrollers, from Android to Raspberry Pi, from Python to Javascript (WASM support).&lt;br /&gt;
* Custom wake-words can be created with one line of text&lt;br /&gt;
* Low resource consumption (works on Raspberry Pi Zero)&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2248</id>
		<title>SEPIA</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2248"/>
		<updated>2022-01-28T10:40:53Z</updated>

		<summary type="html">&lt;p&gt;Florian: Updated infobox&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Voice Assistant&lt;br /&gt;
| Url = https://github.com/SEPIA-Framework/&lt;br /&gt;
| License = MIT License&lt;br /&gt;
| InternetRequired = No, but services like weather and Wiki won&#039;t work&lt;br /&gt;
| SupportedLanguages = en, de + partial support for 15 more (beta)&lt;br /&gt;
| WakeWord = Hey SEPIA, customizable&lt;br /&gt;
| DefaultTTS = eSpeak, picoTTS, MBROLA, MaryTTS + compatible APIs&lt;br /&gt;
| SupportedOS = Windows, Linux, Mac OS X, Raspberry Pi&lt;br /&gt;
| CodeLanguage = Java (Assist-Server), HTML/Javascript (Client), Python (STT-Server)&lt;br /&gt;
| LatestVersion = 2.6.0&lt;br /&gt;
| LatestReleaseDate = 2021-10-14&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Introduction: S.E.P.I.A. Open Assistant Framework ==&lt;br /&gt;
[https://sepia-framework.github.io/ S.E.P.I.A.] is an acronym for: &#039;&#039;&#039;s&#039;&#039;&#039;elf-hosted, &#039;&#039;&#039;e&#039;&#039;&#039;xtendable, &#039;&#039;&#039;p&#039;&#039;&#039;ersonal, &#039;&#039;&#039;i&#039;&#039;&#039;ntelligent &#039;&#039;&#039;a&#039;&#039;&#039;ssistant. It is a modular, open-source framework equipped with all the required tools to build your own, full-fledged digital voice-assistant, including [[:Category:STT|speech recognition]] ([[:Category:STT|STT]]), [[:Category:Wake words|wake-word]] detection, [[:Category:TTS|text-to-speech]] ([[:Category:TTS|TTS]]), [[Natural language understanding|natural-language-understanding (NLU)]], dialog-management, SDK(s), a cross-platform client app and much more.&lt;br /&gt;
&lt;br /&gt;
The framework consists of several, highly customizable micro-services that work together seamlessly to form the SEPIA Open Assistant. It follows the &#039;&#039;&#039;client-server principle using a lightweight Java server and Elasticsearch DB as &amp;quot;brain&amp;quot;&#039;&#039;&#039; and a JavaScript based client that works for example as &#039;&#039;&#039;smart-speaker, smart-display or full-fledged digital assistant app&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
All components work on Linux, Windows and Mac and have been optimized to even &#039;&#039;&#039;run smoothly on a Raspberry Pi&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Out-of-the-box SEPIA currently has smart-services for: news, music (radio), timers, alarms, reminders, to-do and shopping lists, smart home (e.g. using open-source tools like openHAB), navigation, places, weather, Wikipedia, web-search, soccer-results (Bundesliga), a bit of small-talk and more. To build custom services there is a Java SDK and a code editor integrated into the SEPIA Control HUB web-app. The client can be extended with custom HTML widgets.&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-docs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Components==&lt;br /&gt;
Some of the core SEPIA framework components:&lt;br /&gt;
&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-assist-server SEPIA Assist-Server] - The &amp;quot;brain&amp;quot; of SEPIA responsible for: user-accounts, database integration, [[Natural language understanding|NLU]], dialog-management, open-source [[:Category:TTS|TTS]], smart-services (weather, navigation, alarms, news etc.), remote actions and more&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-assist-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-websocket-server-java SEPIA Chat-Server] - A [[wikipedia:WebSocket|WebSocket]] based Chat-Server that takes care of real-time, asynchronous user-to-user and user-to-assistant communication&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-websocket-server-java&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-html-client-app SEPIA Cross-Platform Client] - The primary SEPIA client based on HTML5 web technology that runs on all modern browsers (mobile and desktop) and communicates with other SEPIA components via HTTP or WebSocket connections&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-html-client-app&amp;lt;/ref&amp;gt;. It supports headless- and kiosk-mode via [https://github.com/bytemind-de/nodejs-client-extension-interface CLEXI server] to build independent smart-speaker or smart-display like devices&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-installation-and-setup/tree/master/sepia-client-installation&amp;lt;/ref&amp;gt;. The client is available as Android app as well via [https://play.google.com/store/apps/details?id=de.bytemind.sepia.app.web Play Store] or [https://github.com/SEPIA-Framework/sepia-installation-and-setup/releases direct download].&lt;br /&gt;
*[[SEPIA Speech-To-Text Server|SEPIA STT-Server]] - A WebSocket based, full-duplex Python server for real-time [[:Category:STT|automatic speech recognition]] (ASR) supporting [https://github.com/SEPIA-Framework/sepia-stt-server multiple open-source ASR engines]. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
There are many more SEPIA components and tools to discover on the [https://github.com/SEPIA-Framework official GitHub project page].&lt;br /&gt;
&lt;br /&gt;
==Installation==&lt;br /&gt;
The [https://github.com/SEPIA-Framework/sepia-docs official documentation page] has the most recent installation guides.&lt;br /&gt;
&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;br /&gt;
[[Category:Project]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Comparison_of_voice_assistants&amp;diff=2083</id>
		<title>Comparison of voice assistants</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Comparison_of_voice_assistants&amp;diff=2083"/>
		<updated>2021-12-18T14:11:45Z</updated>

		<summary type="html">&lt;p&gt;Florian: Added SEPIA info and removed license field duplicate&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;There are lots of (open) voice assistants out there. Maybe we can make a comparison list of which assistants exists, in which points they&#039;re equal and which aspects differ.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Comparison of available voice assistants&lt;br /&gt;
!&lt;br /&gt;
![[Mycroft]] AI&lt;br /&gt;
![[SEPIA]]&lt;br /&gt;
![[Rhasspy]]&lt;br /&gt;
![[Leon]]&lt;br /&gt;
![[Genie]]&lt;br /&gt;
|-&lt;br /&gt;
|Target group&lt;br /&gt;
|&lt;br /&gt;
|Makers, tinkerers, smart-home enthusiasts,&lt;br /&gt;
end-users via mobile app &lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|License&lt;br /&gt;
|&lt;br /&gt;
|MIT&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Requires internet access&lt;br /&gt;
|&lt;br /&gt;
|100% offline is possible, but services&lt;br /&gt;
like Wikipedia or news require internet&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Offline STT&lt;br /&gt;
|&lt;br /&gt;
|via SEPIA STT&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Offline TTS&lt;br /&gt;
|&lt;br /&gt;
|via SEPIA-Home&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|URL&lt;br /&gt;
|&lt;br /&gt;
|https://sepia-framework.github.io/&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Comment&lt;br /&gt;
|&lt;br /&gt;
|Highly customizable. Modules can be distributed to&lt;br /&gt;
multiple Raspberry Pis in home network.&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Linux&lt;br /&gt;
|&lt;br /&gt;
|Yes&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Windows&lt;br /&gt;
|&lt;br /&gt;
|Yes&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Mac OS X&lt;br /&gt;
|&lt;br /&gt;
|Yes&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Raspberry Pi&lt;br /&gt;
|&lt;br /&gt;
|Yes&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|Own hw product line&lt;br /&gt;
|&lt;br /&gt;
|No (maybe someday ^^)&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|...&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA_Speech-To-Text_Server&amp;diff=2007</id>
		<title>SEPIA Speech-To-Text Server</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA_Speech-To-Text_Server&amp;diff=2007"/>
		<updated>2021-12-03T10:33:34Z</updated>

		<summary type="html">&lt;p&gt;Florian: optimized format&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;SEPIA Speech-To-Text (STT) Server is a [[wikipedia:WebSocket|WebSocket]] based, full-duplex Python server for real-time automatic speech recognition (ASR) supporting multiple open-source ASR engines. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results.&lt;br /&gt;
&lt;br /&gt;
One goal of this project is to offer a standardized, secure, real-time interface for all the great open-source ASR tools out there. The server works on all major platforms including single-board devices like Raspberry Pi (4).&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Features ==&lt;br /&gt;
&lt;br /&gt;
* WebSocket server ([https://fastapi.tiangolo.com/ Python Fast-API]) that can receive audio streams and send transcribed text at the same time&lt;br /&gt;
* Modular architecture to support multiple ASR engines like [[Vosk]] (reference implementation), [[Coqui]], [[Deepspeech]], [[Scribosermo]] and more (under construction)&lt;br /&gt;
* Optional post processing of result (e.g. via [https://github.com/allo-media/text2num text2num] and custom modules)&lt;br /&gt;
* Standardized API for all engines and support for individual engine features ([[speaker identification]], grammar, confidence score, word timestamps, alternative results, etc.)&lt;br /&gt;
* On-the-fly server and engine configuration via HTTP REST API and WebSocket &#039;welcome&#039; event (including custom grammar, if supported by engine and model)&lt;br /&gt;
* User authentication via simple common token or individual tokens for multiple users&lt;br /&gt;
* Docker containers with support for all major platform architectures: x86 64Bit (amd64), ARM 32Bit (armv7l) and ARM 64Bit (aarch64)&lt;br /&gt;
* Fast enough to run even on Raspberry Pi 4 (2GB) in real-time (depending on engine and model configuration)&lt;br /&gt;
* Compatible to [[SEPIA]] Framework client (v0.24+)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:STT]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA_Speech-To-Text_Server&amp;diff=2006</id>
		<title>SEPIA Speech-To-Text Server</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA_Speech-To-Text_Server&amp;diff=2006"/>
		<updated>2021-12-03T10:33:11Z</updated>

		<summary type="html">&lt;p&gt;Florian: formatting&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;SEPIA Speech-To-Text (STT) Server is a [[wikipedia:WebSocket|WebSocket]] based, full-duplex Python server for real-time automatic speech recognition (ASR) supporting multiple open-source ASR engines. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results.&lt;br /&gt;
&lt;br /&gt;
One goal of this project is to offer a standardized, secure, real-time interface for all the great open-source ASR tools out there. The server works on all major platforms including single-board devices like Raspberry Pi (4).&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Features ==&lt;br /&gt;
&lt;br /&gt;
* WebSocket server ([https://fastapi.tiangolo.com/ Python Fast-API]) that can receive audio streams and send transcribed text at the same time&lt;br /&gt;
* Modular architecture to support multiple ASR engines like [[Vosk]] (reference implementation), [[Coqui]], [[Deepspeech]], [[Scribosermo]] and more (under construction)&lt;br /&gt;
* Optional post processing of result (e.g. via [https://github.com/allo-media/text2num text2num] and custom modules)&lt;br /&gt;
* Standardized API for all engines and support for individual engine features ([[speaker identification]], grammar, confidence score, word timestamps, alternative results, etc.)&lt;br /&gt;
* On-the-fly server and engine configuration via HTTP REST API and WebSocket &#039;welcome&#039; event (including custom grammar, if supported by engine and model)&lt;br /&gt;
* User authentication via simple common token or individual tokens for multiple users&lt;br /&gt;
* Docker containers with support for all major platform architectures: x86 64Bit (amd64), ARM 32Bit (armv7l) and ARM 64Bit (aarch64)&lt;br /&gt;
* Fast enough to run even on Raspberry Pi 4 (2GB) in real-time (depending on engine and model configuration)&lt;br /&gt;
* Compatible to [[SEPIA]] Framework client (v0.24+)&lt;br /&gt;
[[Category:STT]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Vosk&amp;diff=2005</id>
		<title>Vosk</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Vosk&amp;diff=2005"/>
		<updated>2021-12-03T10:32:10Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created Vosk page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[https://github.com/alphacep/vosk-api Vosk] is an open-source speech recognition toolkit by Alphacephei&amp;lt;ref&amp;gt;https://alphacephei.com/vosk/&amp;lt;/ref&amp;gt;. Key features are:&lt;br /&gt;
&lt;br /&gt;
# Supports 20+ languages and dialects - English, Indian English, German, French, Spanish, Portuguese, Chinese, Russian, Turkish, Vietnamese, Italian, Dutch, Catalan, Arabic, Greek, Farsi, Filipino, Ukrainian, Kazakh, Swedish. More to come.&lt;br /&gt;
# Works offline, even on lightweight devices - Raspberry Pi, Android, iOS&lt;br /&gt;
# Installs with simple &amp;lt;code&amp;gt;pip3 install vosk&amp;lt;/code&amp;gt;&lt;br /&gt;
# Portable per-language models are only 50Mb each, but there are much bigger server models available.&lt;br /&gt;
# Provides streaming API for the best user experience (unlike popular speech-recognition python packages)&lt;br /&gt;
# There are bindings for different programming languages, too - java/csharp/javascript etc.&lt;br /&gt;
# Allows quick reconfiguration of vocabulary for best accuracy.&lt;br /&gt;
# Supports speaker identification beside simple speech recognition.&lt;br /&gt;
&lt;br /&gt;
[[Category:STT]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA_Speech-To-Text_Server&amp;diff=2004</id>
		<title>SEPIA Speech-To-Text Server</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA_Speech-To-Text_Server&amp;diff=2004"/>
		<updated>2021-12-03T10:26:40Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created SEPIA STT Server page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;SEPIA Speech-To-Text (STT) Server is a [[wikipedia:WebSocket|WebSocket]] based, full-duplex Python server for real-time automatic speech recognition (ASR) supporting multiple open-source ASR engines. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results.&lt;br /&gt;
&lt;br /&gt;
One goal of this project is to offer a standardized, secure, real-time interface for all the great open-source ASR tools out there. The server works on all major platforms including single-board devices like Raspberry Pi (4).&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Features ==&lt;br /&gt;
&lt;br /&gt;
* WebSocket server ([https://fastapi.tiangolo.com/ Python Fast-API]) that can receive audio streams and send transcribed text at the same time&lt;br /&gt;
* Modular architecture to support multiple ASR engines like [[Vosk]] (reference implementation), [[Coqui]], [[Deepspeech]], [[Scribosermo]] and more (under construction)&lt;br /&gt;
* Optional post processing of result (e.g. via [https://github.com/allo-media/text2num text2num] and custom modules)&lt;br /&gt;
* Standardized API for all engines and support for individual engine features ([[speaker identification]], grammar, confidence score, word timestamps, alternative results, etc.)&lt;br /&gt;
* On-the-fly server and engine configuration via HTTP REST API and WebSocket &#039;welcome&#039; event (including custom grammar, if supported by engine and model)&lt;br /&gt;
* User authentication via simple common token or individual tokens for multiple users&lt;br /&gt;
* Docker containers with support for all major platform architectures: x86 64Bit (amd64), ARM 32Bit (armv7l) and ARM 64Bit (aarch64)&lt;br /&gt;
* Fast enough to run even on Raspberry Pi 4 (2GB) in real-time (depending on engine and model configuration)&lt;br /&gt;
* Compatible to [[SEPIA]] Framework client (v0.24+)&lt;br /&gt;
&lt;br /&gt;
[[Category:STT]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2003</id>
		<title>SEPIA</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=2003"/>
		<updated>2021-12-03T10:17:52Z</updated>

		<summary type="html">&lt;p&gt;Florian: Added components section. More references.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction: S.E.P.I.A. Open Assistant Framework ==&lt;br /&gt;
[https://sepia-framework.github.io/ S.E.P.I.A.] is an acronym for: &#039;&#039;&#039;s&#039;&#039;&#039;elf-hosted, &#039;&#039;&#039;e&#039;&#039;&#039;xtendable, &#039;&#039;&#039;p&#039;&#039;&#039;ersonal, &#039;&#039;&#039;i&#039;&#039;&#039;ntelligent &#039;&#039;&#039;a&#039;&#039;&#039;ssistant. It is a modular, open-source framework equipped with all the required tools to build your own, full-fledged digital voice-assistant, including [[:Category:STT|speech recognition]] ([[:Category:STT|STT]]), [[:Category:Wake words|wake-word]] detection, [[:Category:TTS|text-to-speech]] ([[:Category:TTS|TTS]]), [[Natural language understanding|natural-language-understanding (NLU)]], dialog-management, SDK(s), a cross-platform client app and much more.&lt;br /&gt;
&lt;br /&gt;
The framework consists of several, highly customizable micro-services that work together seamlessly to form the SEPIA Open Assistant. It follows the &#039;&#039;&#039;client-server principle using a lightweight Java server and Elasticsearch DB as &amp;quot;brain&amp;quot;&#039;&#039;&#039; and a JavaScript based client that works for example as &#039;&#039;&#039;smart-speaker, smart-display or full-fledged digital assistant app&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
All components work on Linux, Windows and Mac and have been optimized to even &#039;&#039;&#039;run smoothly on a Raspberry Pi&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Out-of-the-box SEPIA currently has smart-services for: news, music (radio), timers, alarms, reminders, to-do and shopping lists, smart home (e.g. using open-source tools like openHAB), navigation, places, weather, Wikipedia, web-search, soccer-results (Bundesliga), a bit of small-talk and more. To build custom services there is a Java SDK and a code editor integrated into the SEPIA Control HUB web-app. The client can be extended with custom HTML widgets.&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-docs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Components==&lt;br /&gt;
Some of the core SEPIA framework components:&lt;br /&gt;
&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-assist-server SEPIA Assist-Server] - The &amp;quot;brain&amp;quot; of SEPIA responsible for: user-accounts, database integration, [[Natural language understanding|NLU]], dialog-management, open-source [[:Category:TTS|TTS]], smart-services (weather, navigation, alarms, news etc.), remote actions and more&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-assist-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-websocket-server-java SEPIA Chat-Server] - A [[wikipedia:WebSocket|WebSocket]] based Chat-Server that takes care of real-time, asynchronous user-to-user and user-to-assistant communication&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-websocket-server-java&amp;lt;/ref&amp;gt;.&lt;br /&gt;
*[https://github.com/SEPIA-Framework/sepia-html-client-app SEPIA Cross-Platform Client] - The primary SEPIA client based on HTML5 web technology that runs on all modern browsers (mobile and desktop) and communicates with other SEPIA components via HTTP or WebSocket connections&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-html-client-app&amp;lt;/ref&amp;gt;. It supports headless- and kiosk-mode via [https://github.com/bytemind-de/nodejs-client-extension-interface CLEXI server] to build independent smart-speaker or smart-display like devices&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-installation-and-setup/tree/master/sepia-client-installation&amp;lt;/ref&amp;gt;. The client is available as Android app as well via [https://play.google.com/store/apps/details?id=de.bytemind.sepia.app.web Play Store] or [https://github.com/SEPIA-Framework/sepia-installation-and-setup/releases direct download].&lt;br /&gt;
*[[SEPIA Speech-To-Text Server|SEPIA STT-Server]] - A WebSocket based, full-duplex Python server for real-time [[:Category:STT|automatic speech recognition]] (ASR) supporting [https://github.com/SEPIA-Framework/sepia-stt-server multiple open-source ASR engines]. It can receive a stream of audio chunks via the secure WebSocket connection and return transcribed text almost immediately as partial and final results&amp;lt;ref&amp;gt;https://github.com/SEPIA-Framework/sepia-stt-server&amp;lt;/ref&amp;gt;.&lt;br /&gt;
There are many more SEPIA components and tools to discover on the [https://github.com/SEPIA-Framework official GitHub project page].&lt;br /&gt;
&lt;br /&gt;
==Installation==&lt;br /&gt;
The [https://github.com/SEPIA-Framework/sepia-docs official documentation page] has the most recent installation guides.&lt;br /&gt;
&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;br /&gt;
[[Category:Project]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=1998</id>
		<title>SEPIA</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=SEPIA&amp;diff=1998"/>
		<updated>2021-12-01T18:17:44Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created basic SEPIA page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction: S.E.P.I.A. Open Assistant Framework ==&lt;br /&gt;
S.E.P.I.A. is an acronym for: &#039;&#039;&#039;s&#039;&#039;&#039;elf-hosted, &#039;&#039;&#039;e&#039;&#039;&#039;xtendable, &#039;&#039;&#039;p&#039;&#039;&#039;ersonal, &#039;&#039;&#039;i&#039;&#039;&#039;ntelligent &#039;&#039;&#039;a&#039;&#039;&#039;ssistant. It is a modular, open-source framework equipped with all the required tools to build your own, full-fledged digital voice-assistant, including speech recognition (STT), wake-word detection, text-to-speech (TTS), natural-language-understanding, dialog-management, SDK(s), a cross-platform client app and much more.&lt;br /&gt;
&lt;br /&gt;
The framework consists of several, highly customizable micro-services that work together seamlessly to form the SEPIA Open Assistant. It follows the &#039;&#039;&#039;client-server principle using a lightweight Java server and Elasticsearch DB as &amp;quot;brain&amp;quot;&#039;&#039;&#039; and a Javascript based client that works for example as &#039;&#039;&#039;smart-speaker, smart-display or full-fledged digital assistant app&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
All components work on Linux, Windows and Mac and have been optimized to even &#039;&#039;&#039;run smoothly on a Raspberry Pi&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Out-of-the-box SEPIA currently has smart-services for: news, music (radio), timers, alarms, reminders, to-do and shopping lists, smart home (e.g. using open-source tools like openHAB), navigation, places, weather, Wikipedia, web-search, soccer-results (Bundesliga), a bit of small-talk and more. To build custom services there is a Java SDK and a code editor integrated into the SEPIA Control HUB web-app. The client can be extended with custom HTML widgets.&lt;br /&gt;
&lt;br /&gt;
== Components ==&lt;br /&gt;
Under construction&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Under construction&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1842</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1842"/>
		<updated>2021-11-13T13:34:00Z</updated>

		<summary type="html">&lt;p&gt;Florian: Added HTTP REST API note&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
*Pronunciation of times, dates, temperatures etc. can be very good. MaryTTS uses an extensive, handcrafted set of rules (and statistics) to handle this.&amp;lt;ref&amp;gt;http://mary.dfki.de/documentation/overview.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
*MaryTTS is server-based meaning it offers a HTTP REST API out-of-the-box to synthesize text&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
Voice models can be downloaded via script from inside the release version or using these links:&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/marytts/voice-bits1-hsmm/releases bits1-hsmm] - A female German hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-bits3-hsmm/releases/ bits3-hsmm] - A male German hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-cmu-slt-hsmm/ voice-cmu-slt-hsmm] (included) - A female US English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-cmu-bdl-hsmm/releases/ cmu-bdl-hsmm] - A male US English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-dfki-spike-hsmm/releases/ dfki-spike-hsmm] - A male British English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-dfki-prudence-hsmm/releases/ dfki-prudence-hsmm] - A female British English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-upmc-jessica-hsmm/releases/ upmc-jessica-hsmm] - A female French hidden semi-Markov model voice&lt;br /&gt;
* ...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11 (Debian: &#039;sudo apt-get install openjdk-11-jdk-headless&#039;)&lt;br /&gt;
* Download the release ZIP file: [https://github.com/marytts/marytts/releases v5.2 official] or [https://github.com/fquirin/marytts/releases v6.0 snpashot]&lt;br /&gt;
* Extract the ZIP and start the server (run scripts are in &#039;marytts\bin&#039;)&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
==== Voice: dfki-spike-hsmm en_GB ====&lt;br /&gt;
&lt;br /&gt;
* Test system: Raspberry Pi4 4GB&lt;br /&gt;
* Sentence: &amp;quot;Hello this is a test&amp;quot;&lt;br /&gt;
* Run-time: 0.33 s&lt;br /&gt;
* Real-time-factor: 0.19&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=OpenTTS&amp;diff=1841</id>
		<title>OpenTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=OpenTTS&amp;diff=1841"/>
		<updated>2021-11-13T13:31:34Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created OpenTTS draft&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is OpenTTS ==&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
https://github.com/synesthesiam/opentts&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=ESpeak&amp;diff=1840</id>
		<title>ESpeak</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=ESpeak&amp;diff=1840"/>
		<updated>2021-11-13T13:30:32Z</updated>

		<summary type="html">&lt;p&gt;Florian: Format tweaks&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What are eSpeak and eSpeak-NG? ==&lt;br /&gt;
eSpeak is a compact open source (GNU GPL license) software speech synthesizer for English and other languages, for Linux and Windows. It uses a &amp;quot;formant synthesis&amp;quot; method. This allows many languages to be provided in a small size. The speech is clear, and can be used at high speeds, but is not as natural or smooth as larger synthesizers which are based on human speech recordings.&lt;br /&gt;
&lt;br /&gt;
Originally known as &#039;&#039;&#039;speak&#039;&#039;&#039; and originally written for Acorn/RISC_OS computers starting in 1995, eSpeak is an enhancement and re-write, including a relaxation of the original memory and processing power constraints and comes with support for additional languages.&amp;lt;ref&amp;gt;http://espeak.sourceforge.net/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In 2010 Reece H. Dunn started maintaining a version of eSpeak on GitHub that was forked in late 2015 and renamed to &#039;&#039;&#039;eSpeak NG&#039;&#039;&#039;. The new eSpeak NG project is a significant departure from the eSpeak project with the intention of cleaning up the existing codebase, adding new features and improving the supported languages.&amp;lt;ref&amp;gt;https://github.com/espeak-ng/espeak-ng&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Advantages ==&lt;br /&gt;
Although it has been practically around since the dawn of time (or at least the personal computer) there are still good reasons to use eSpeak:&lt;br /&gt;
&lt;br /&gt;
* It is blazing fast, audio generation is almost instant even on SBCs like the Raspberry Pi Zero&lt;br /&gt;
* It is very small and memory consumption is negligible on modern systems (~5MB [source?])&lt;br /&gt;
* It is available on basically any platform either as command line program, shared library, SAPI5 version for Windows screen-readers, etc..&lt;br /&gt;
* Although the voice sounds rather robotic it has a certain charm, especially if you are building little robots ;-)&lt;br /&gt;
* It is available in dozens of languages&lt;br /&gt;
* It can translate text into phoneme codes and is often used to generate new dictionary entries for other systems (e.g. STT)&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Debian command line tool (espeak-ng recommended):&amp;lt;syntaxhighlight lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
sudo apt-get install -y espeak-ng espeak-ng-espeak&lt;br /&gt;
&amp;lt;/syntaxhighlight&amp;gt;NOTE: eSpeak is often found in other projects like [[OpenTTS]] or [https://github.com/SEPIA-Framework/sepia-assist-server/tree/master/Xtensions/TTS SEPIA Assist Server] and ready to use out-of-the-box.&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
eSpeak-NG (en):&lt;br /&gt;
&lt;br /&gt;
* Test system: Raspberry Pi4 4GB&lt;br /&gt;
* Sentence: &amp;quot;Hello this is a test&amp;quot;&lt;br /&gt;
* Run-time: 0.08 s&lt;br /&gt;
* Real-time-factor: 0.062&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=ESpeak&amp;diff=1839</id>
		<title>ESpeak</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=ESpeak&amp;diff=1839"/>
		<updated>2021-11-13T13:29:44Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created eSpeak page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What are eSpeak and eSpeak-NG? ==&lt;br /&gt;
eSpeak is a compact open source (GNU GPL license) software speech synthesizer for English and other languages, for Linux and Windows. It uses a &amp;quot;formant synthesis&amp;quot; method. This allows many languages to be provided in a small size. The speech is clear, and can be used at high speeds, but is not as natural or smooth as larger synthesizers which are based on human speech recordings.&lt;br /&gt;
&lt;br /&gt;
Originally known as &#039;&#039;&#039;speak&#039;&#039;&#039; and originally written for Acorn/RISC_OS computers starting in 1995, eSpeak is an enhancement and re-write, including a relaxation of the original memory and processing power constraints and comes with support for additional languages.&amp;lt;ref&amp;gt;http://espeak.sourceforge.net/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In 2010 Reece H. Dunn started maintaining a version of eSpeak on GitHub that was forked in late 2015 and renamed to &#039;&#039;&#039;eSpeak NG&#039;&#039;&#039;. The new eSpeak NG project is a significant departure from the eSpeak project with the intention of cleaning up the existing codebase, adding new features and improving the supported languages.&amp;lt;ref&amp;gt;https://github.com/espeak-ng/espeak-ng&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Advantages ==&lt;br /&gt;
Although it has been practically around since the dawn of time (or at least the personal computer) there are still good reasons to use eSpeak:&lt;br /&gt;
&lt;br /&gt;
* It is blazing fast, audio generation is almost instant even on SBCs like the Raspberry Pi Zero&lt;br /&gt;
* It is very small and memory consumption is negligible on modern systems (~5MB [source?])&lt;br /&gt;
* It is available on basically any platform either as command line program, shared library, SAPI5 version for Windows screen-readers, etc..&lt;br /&gt;
* Although the voice sounds rather robotic it has a certain charm, especially if you are building little robots ;-)&lt;br /&gt;
* It is available in dozens of languages&lt;br /&gt;
* It can translate text into phoneme codes and is often used to generate new dictionary entries for other systems (e.g. STT)&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Debian command line tool (espeak-ng recommended):&amp;lt;syntaxhighlight lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
sudo apt-get install -y espeak-ng espeak-ng-espeak&lt;br /&gt;
&amp;lt;/syntaxhighlight&amp;gt;NOTE: eSpeak is often found in other projects like [[OpenTTS]] or [https://github.com/SEPIA-Framework/sepia-assist-server/tree/master/Xtensions/TTS SEPIA Assist Server] and ready to use out-of-the-box.&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
eSpeak-NG (en):&lt;br /&gt;
&lt;br /&gt;
* Test system: Raspberry Pi4 4GB&lt;br /&gt;
* Sentence: &amp;quot;Hello this is a test&amp;quot;&lt;br /&gt;
* Run-time: 0.08 s&lt;br /&gt;
* Real-time-factor: 0.062&lt;br /&gt;
&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1838</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1838"/>
		<updated>2021-11-13T12:54:54Z</updated>

		<summary type="html">&lt;p&gt;Florian: Smaller format fix&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
*Pronunciation of times, dates, temperatures etc. can be very good. MaryTTS uses an extensive, handcrafted set of rules (and statistics) to handle this.&amp;lt;ref&amp;gt;http://mary.dfki.de/documentation/overview.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
Voice models can be downloaded via script from inside the release version or using these links:&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/marytts/voice-bits1-hsmm/releases bits1-hsmm] - A female German hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-bits3-hsmm/releases/ bits3-hsmm] - A male German hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-cmu-slt-hsmm/ voice-cmu-slt-hsmm] (included) - A female US English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-cmu-bdl-hsmm/releases/ cmu-bdl-hsmm] - A male US English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-dfki-spike-hsmm/releases/ dfki-spike-hsmm] - A male British English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-dfki-prudence-hsmm/releases/ dfki-prudence-hsmm] - A female British English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-upmc-jessica-hsmm/releases/ upmc-jessica-hsmm] - A female French hidden semi-Markov model voice&lt;br /&gt;
* ...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11 (Debian: &#039;sudo apt-get install openjdk-11-jdk-headless&#039;)&lt;br /&gt;
* Download the release ZIP file: [https://github.com/marytts/marytts/releases v5.2 official] or [https://github.com/fquirin/marytts/releases v6.0 snpashot]&lt;br /&gt;
* Extract the ZIP and start the server (run scripts are in &#039;marytts\bin&#039;)&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
==== Voice: dfki-spike-hsmm en_GB ====&lt;br /&gt;
&lt;br /&gt;
* Test system: Raspberry Pi4 4GB&lt;br /&gt;
* Sentence: &amp;quot;Hello this is a test&amp;quot;&lt;br /&gt;
* Run-time: 0.33 s&lt;br /&gt;
* Real-time-factor: 0.19&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1837</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1837"/>
		<updated>2021-11-13T12:45:40Z</updated>

		<summary type="html">&lt;p&gt;Florian: Model links, releases, more info&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
*Pronunciation of times, dates, temperatures etc. can be very good. MaryTTS uses an extensive, handcrafted set of rules (and statistics) to handle this.&amp;lt;ref&amp;gt;http://mary.dfki.de/documentation/overview.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
Voice models can be downloaded via script from inside the release version or using these links:&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/marytts/voice-bits1-hsmm/releases bits1-hsmm] - A female German hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-bits3-hsmm/releases/ bits3-hsmm] - A male German hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-cmu-slt-hsmm/ voice-cmu-slt-hsmm] (included) - A female US English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-cmu-bdl-hsmm/releases/ cmu-bdl-hsmm] - A male US English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-dfki-spike-hsmm/releases/ dfki-spike-hsmm] - A male British English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-dfki-prudence-hsmm/releases/ dfki-prudence-hsmm] - A female British English hidden semi-Markov model voice&lt;br /&gt;
* [https://github.com/marytts/voice-upmc-jessica-hsmm/releases/ upmc-jessica-hsmm] - A female French hidden semi-Markov model voice&lt;br /&gt;
* ...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11 (Debian: &#039;sudo apt-get install openjdk-11-jdk-headless&#039;)&lt;br /&gt;
* Download the release ZIP file: [https://github.com/marytts/marytts/releases v5.2 official] or [https://github.com/fquirin/marytts/releases v6.0 snpashot]&lt;br /&gt;
* Extract the ZIP and start the server (run scripts are in &#039;marytts\bin&#039;)&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
==== Mary-TTS - Voice: dfki-spike-hsmm en_GB ====&lt;br /&gt;
&lt;br /&gt;
* Test system: Raspberry Pi4 4GB&lt;br /&gt;
* Sentence: &amp;quot;Hello this is a test&amp;quot;&lt;br /&gt;
* Run-time: 0.33 s&lt;br /&gt;
* Real-time-factor: 0.19&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
[UNDER CONSTRUCTION]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1836</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1836"/>
		<updated>2021-11-13T12:14:47Z</updated>

		<summary type="html">&lt;p&gt;Florian: Performance metrik&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
*Pronunciation of times, dates, temperatures etc. can be very good. MaryTTS uses an extensive, handcrafted set of rules (and statistics) to handle this.&amp;lt;ref&amp;gt;http://mary.dfki.de/documentation/overview.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11&lt;br /&gt;
* Download the release ZIP file&lt;br /&gt;
* Extract and start the server&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&lt;br /&gt;
==== Mary-TTS - Voice: dfki-spike-hsmm en_GB ====&lt;br /&gt;
Test system: Raspberry Pi4 4GB&lt;br /&gt;
&lt;br /&gt;
Sentence: &amp;quot;Hello this is a test&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Run-time: 0.33 s&lt;br /&gt;
&lt;br /&gt;
Real-time-factor: 0.19&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1835</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1835"/>
		<updated>2021-11-13T12:12:43Z</updated>

		<summary type="html">&lt;p&gt;Florian: Added pronunciation advantage&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
*Pronunciation of times, dates, temperatures etc. can be very good. MaryTTS uses an extensive, handcrafted set of rules (and statistics) to handle this.&amp;lt;ref&amp;gt;http://mary.dfki.de/documentation/overview.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11&lt;br /&gt;
* Download the release ZIP file&lt;br /&gt;
* Extract and start the server&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1834</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1834"/>
		<updated>2021-11-13T12:06:02Z</updated>

		<summary type="html">&lt;p&gt;Florian: Set category&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11&lt;br /&gt;
* Download the release ZIP file&lt;br /&gt;
* Extract and start the server&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
[[Category:TTS]]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1833</id>
		<title>MaryTTS</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=MaryTTS&amp;diff=1833"/>
		<updated>2021-11-13T12:02:05Z</updated>

		<summary type="html">&lt;p&gt;Florian: Created MaryTTS page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== What is MaryTTS? ==&lt;br /&gt;
[http://mary.dfki.de/ Mary (&#039;&#039;&#039;M&#039;&#039;&#039;odular &#039;&#039;&#039;A&#039;&#039;&#039;rchitecture for &#039;&#039;&#039;R&#039;&#039;&#039;esearch in s&#039;&#039;&#039;Y&#039;&#039;&#039;ynthesis) Text-to-Speech] is an open-source (GNU LGPL license&amp;lt;ref&amp;gt;https://github.com/marytts/marytts/blob/master/LICENSE.md&amp;lt;/ref&amp;gt;), multilingual Text-to-Speech Synthesis platform &#039;&#039;&#039;written in Java&#039;&#039;&#039;. It was originally developed as a collaborative project of [http://www.dfki.de/web DFKI’s] Language Technology Lab and the [http://www.coli.uni-saarland.de/groups/WB/Phonetics/ Institute of Phonetics] at Saarland University, Germany. It is now maintained by the Multimodal Speech Processing Group in the [https://www.mmci.uni-saarland.de/ Cluster of Excellence MMCI] and DFKI.&amp;lt;ref&amp;gt;http://mary.dfki.de/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MaryTTS has been around for a very! long time. Version 3.0 even dates back to 2006, long before Deep Learning was a broadly known term and the last official release was version 5.2 in 2016&amp;lt;ref&amp;gt;http://mary.dfki.de/download/index.html&amp;lt;/ref&amp;gt;. The system uses [[wikipedia:Speech_synthesis#Unit_selection_synthesis|unit selection]] and [[wikipedia:Hidden_Markov_model|HMM]]-based techniques to build voices (today probably called AI, back then called statistics ^^). If you want to learn more about the architecture check out the [http://mary.dfki.de/documentation/index.html official documentation].&lt;br /&gt;
&lt;br /&gt;
There is still activity on the GitHub page and internally there has been some major code refactoring but it is currently unclear if there will ever be another release version. There has been an unofficial snapshot release for the SEPIA Framework which runs stable on Java 11 but should be considered experimental: [https://github.com/fquirin/marytts/releases MaryTTS 6.0 snapshot] ([https://hub.docker.com/r/sepia/marytts Docker container]).&lt;br /&gt;
&lt;br /&gt;
== Advantages of MaryTTS ==&lt;br /&gt;
MaryTTS has certain advantages compared to modern Deep Learning systems or classical, synthetic engines like [[eSpeak]]:&lt;br /&gt;
&lt;br /&gt;
* The quality of the voice depends strongly on the model but can be surprisingly good, not state-of-the-art but much better than a synthetic voice.&lt;br /&gt;
* Audio generation is very fast and ranges from 0.2 to 0.5 [[Real-time-factor|RTF]] on a Raspberry Pi 4 depending on the selected voice&amp;lt;ref&amp;gt;https://b07z.net/downloads/tts.html&amp;lt;/ref&amp;gt; meaning it can actually be used on edge devices.&lt;br /&gt;
* RAM consumption is moderate but you should probably reserve around 256-512 MB.&lt;br /&gt;
* Installation is super easy and it runs on Windows, Mac and Linux (every system that can install Java 8)&lt;br /&gt;
* Language support is very good: German, British and American English, French, Italian, Luxembourgish, Russian, Swedish, Telugu, Turkish and more&lt;br /&gt;
&lt;br /&gt;
== Models ==&lt;br /&gt;
[TBD]&lt;br /&gt;
&lt;br /&gt;
== Installation ==&lt;br /&gt;
Installation is super easy:&lt;br /&gt;
&lt;br /&gt;
* Install Java 8 or 11&lt;br /&gt;
* Download the release ZIP file&lt;br /&gt;
* Extract and start the server&lt;br /&gt;
&lt;br /&gt;
By default you can access the server in your browser via: http://localhost:59125/ &lt;br /&gt;
&lt;br /&gt;
In a production system you might want to run MaryTTS behind a [[wikipedia:Reverse_proxy|reverse proxy]] (like Nginx or Apache) to [https://github.com/SEPIA-Framework/sepia-assist-server/blob/master/Xtensions/TTS/marytts/INSTALL.md#solving-cors-problems avoid CORS issues].&lt;br /&gt;
&lt;br /&gt;
== Performance ==&lt;br /&gt;
[TBD]&lt;/div&gt;</summary>
		<author><name>Florian</name></author>
	</entry>
</feed>