<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://openvoice-tech.net/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Eltocino</id>
	<title>Open Voice Technology Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://openvoice-tech.net/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Eltocino"/>
	<link rel="alternate" type="text/html" href="https://openvoice-tech.net/wiki/Special:Contributions/Eltocino"/>
	<updated>2026-04-18T02:28:43Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Research_papers&amp;diff=1908</id>
		<title>Research papers</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Research_papers&amp;diff=1908"/>
		<updated>2021-11-17T06:51:29Z</updated>

		<summary type="html">&lt;p&gt;Eltocino: finding research papers for speech&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Most current research papers on speech can be found online.  &lt;br /&gt;
&lt;br /&gt;
[https://scholar.google.com/scholar Google Scholar] is a useful place that links out to many other sites.  &lt;br /&gt;
&lt;br /&gt;
[https://scholar.archive.org/ Internet Archive] Scholar section can also be used to search for papers.  &lt;br /&gt;
&lt;br /&gt;
[https://arxiv.org/corr Arxiv] has an entire subsection for computer research, and the papers are linked from the abstract page in available formats.  &lt;br /&gt;
&lt;br /&gt;
Conference pages are also a good place to find new and interesting papers you might otherwise miss, finding a title on interspeech&#039;s presentations and searching for it on arxiv usually nets you a copy to read.&lt;/div&gt;</summary>
		<author><name>Eltocino</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Building_voice_datasets&amp;diff=1907</id>
		<title>Building voice datasets</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Building_voice_datasets&amp;diff=1907"/>
		<updated>2021-11-17T06:38:58Z</updated>

		<summary type="html">&lt;p&gt;Eltocino: first submit&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Building a quality dataset for your STT purposes&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
There are several datasets you can build for STT purposes.  Wake word, fine-tuning, complete STT. Each has a separate end goal.  For all three, there seems to be several clear tips.  First, diversity is the key to improving general usability.  The more variance you have in the data, pitch, speed, phonemes, tones, accents, speakers, and such, the better the end models typically are.  Second, larger datasets tend to result in better usability, with the caveat of the first tip.  Third, recording quality can play a big role in how the end model turns out.  Transcription quality as well as audio quality matter.  It may be tempting to record noise and layer it over all your studio-quality clips, but this may not improve your end result.  Some wakeword tools promise usability after recording as few as one to three utterances.  These may be usable for a single user, in matching audio settings, but would likely not be useful as a general purpose wake word model.  &lt;br /&gt;
&lt;br /&gt;
Before you start gathering a single clip, read the docs for the tool you&#039;re using.  [https://stt.readthedocs.io/en/stable/playbook/DATA_FORMATTING.html Coqui STT] specifies 10-20 second clips, so a dataset with five minute long clips would not be a good choice.  Make sure you can get data in the formats that your tool needs.  For audio this is usually not a difficult problem, tools like [https://en.wikipedia.org/wiki/SoX SoX] or [https://www.ffmpeg.org/ FFMpeg] can convert pretty much any source to any type these days. &lt;br /&gt;
&lt;br /&gt;
 * How to get a more diverse set of data?  &lt;br /&gt;
&lt;br /&gt;
For wake words, this is mostly about the users who will be using the end models.  They should provide a good chunk of data to start with.  Expand by finding personalized variables that you&#039;re missing, ie, accents, pitches.  For this type of diversification, ask your friends, colleagues, enemies, the internet.  Pay a service to garner data for you (ie, mechanical turk, craig&#039;s list, etc).  Write a paper at college and collect samples from as many different students as you can.  Find users all over the world that have some level of interest and simply ask them to contribute.  It is possible to use TTS services to provide data, however, this has limitations and in some cases may reduce usability. &lt;br /&gt;
&lt;br /&gt;
For STT, you want both personal diversity (lots of different voices), as well as sentence, word, and phoneme diversity (lots of different sentences and words).  Having 1000 people read the same sentence to build a dataset might make a great model for that particular sentence.  If that&#039;s your end goal, then this might be a strategy to pursue.  For more general purpose models, gather as wide a set of phonemes as possible, typically in a ration mimicking the general language&#039;s usage.  Word diversity can be simply getting a dictionary and starting at aardvark and working through to zymurgy (for English speakers). Another tactic would be to utilize the top 1000 most commonly used words.  Putting both of those together, you should then find a source of sentences, or build a list of your own to record.  There are several open data sets you can also copy from (Common voice).  For domain-specific modeling, you will certainly want to focus on both word- and sentence-choice to be relevant to the domain you&#039;re trying to target.   Mozilla&#039;s Common Voice English set has over 40,000 speakers and more than 200,000 different words.  It&#039;s a massive sampling of the language with a huge range of voice types, and word and sentences.  Within those users there&#039;s also a massive range of recording quality from studio-level to buzzing, hissy computer mic clips.  &lt;br /&gt;
&lt;br /&gt;
* Getting more data&lt;br /&gt;
&lt;br /&gt;
Expanding the diversity of your dataset is a good start for this.  Still need more data?  Domain specific modelers will want to search industry resources for conference or work-area recordings that can be used.  General purpose users can incorporate existing datasets when they can.  Common Voice, Tedlium3, Librivox all have decent sized sets that can be used to supplement or build general datasets.  For fine-tuning, your needs will be more specific, but collecting the broadest, largest set of data you can will still leave you in a better position than having too little to effectively tune.  &lt;br /&gt;
&lt;br /&gt;
* Quality, or can you hear me now?&lt;br /&gt;
&lt;br /&gt;
Common Voice has a massive range of samples within it.  In addition to simply collecting sentences, users can also verify samples to confirm they&#039;re a match to the expected transcript. This has a two-fold benefit: sentences that don&#039;t match the transcript can be noted for exclusion, and the poorest quality samples that are unintelligible or have other audio quality problems can be noted for exclusion.&lt;/div&gt;</summary>
		<author><name>Eltocino</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Category:Wake_words&amp;diff=1903</id>
		<title>Category:Wake words</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Category:Wake_words&amp;diff=1903"/>
		<updated>2021-11-17T05:58:56Z</updated>

		<summary type="html">&lt;p&gt;Eltocino: Updated dataset line to be more generic, since stt datasets will have similar info.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:STT]]&lt;br /&gt;
[[Category:Open Voice Assistants]]&lt;br /&gt;
&lt;br /&gt;
Wake words, sometimes called key words, are a special category of Speech-To-Text. Wake words are used to &amp;quot;wake&amp;quot; a listening device and start its functions. In most cases these &amp;quot;wake words&amp;quot; are detected locally on devices while actual speech recognition is mostly done by internet cloud services. Mycroft defaults to &amp;quot;Hey, Mycroft&amp;quot; for its wake word, for instance. Some platforms allow for multiple wake words to be used. Coqui STT engine can even be configured as a wake word listener.    &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Wake word listeners&#039;&#039;&#039;: &lt;br /&gt;
&lt;br /&gt;
* [[Mycroft Precise]]&lt;br /&gt;
* [[Porcupine]]&lt;br /&gt;
* [[Snowboy]] &lt;br /&gt;
* [[Howl]]&lt;br /&gt;
* [[Coqui]] STT&lt;br /&gt;
* Google tensorflow lite speech recognition &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Customizing wake words&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What makes [[a good wake word]]?&lt;br /&gt;
* Building a quality dataset&lt;/div&gt;</summary>
		<author><name>Eltocino</name></author>
	</entry>
	<entry>
		<id>https://openvoice-tech.net/index.php?title=Recording_tipps&amp;diff=1902</id>
		<title>Recording tipps</title>
		<link rel="alternate" type="text/html" href="https://openvoice-tech.net/index.php?title=Recording_tipps&amp;diff=1902"/>
		<updated>2021-11-17T05:57:03Z</updated>

		<summary type="html">&lt;p&gt;Eltocino: added link to librivox recording tips.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Recording tipps]]&lt;br /&gt;
[[Category:Lessons learned]]&lt;br /&gt;
&lt;br /&gt;
When you plan to record a voice dataset to be used for a TTS model training you should check these tips and tricks:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Use a good microphone and a quiet recording room setup&#039;&#039;&#039; (no computer fans, air conditioning, ...)&lt;br /&gt;
* Use a text corpus with cleaned numbers/abbreviations and good phoneme coverage&lt;br /&gt;
* Read in a neutral style, but with a natural speech flow and do not swallow up letters&lt;br /&gt;
* Adjust tone and pitch with punctuation&lt;br /&gt;
* Use a constant recording speed&lt;br /&gt;
* Check your recordings regularly in high volume for background noise&lt;br /&gt;
* Take breaks regularly and do not record more than four hours a day&lt;br /&gt;
* Record error free&lt;br /&gt;
* Investing in a quality interface and mic can make a big difference in quality. A 24 bit 96khz interface with a large diaphragm condenser can be had for about $200 USD.&lt;br /&gt;
* Record at the highest quality level practical.  You can convert to lesser formats later, but you cannot up convert cleanly&lt;br /&gt;
* Review your work at regular intervals and compare with previous recording to ensure consistent quality&lt;br /&gt;
* Do not be afraid to ask for help! Getting feedback on your data early on can help prevent wasted effort.&lt;br /&gt;
*There&#039;s a wealth of information on the internet about recording.  For instance, https://wiki.librivox.org/index.php/Newbie_Guide_to_Recording from Librivox is a useful guide with numerous sub pages of information.  Some is audio-book specific, but the majority is useful for anyone recording voice.&lt;/div&gt;</summary>
		<author><name>Eltocino</name></author>
	</entry>
</feed>