Text To Speech (TTS) models have made great strides in quality over the last few years. Unfortunately, it's not currently possible to use these libraries without installing a large number of dependencies.
The txtai TextToSpeech pipeline has the following objectives:
- Fast performance both on CPU and GPU
- Ability to batch large text values and stream it through the model
- Minimal install footprint
- All dependencies must be Apache 2.0 compatible
This article will go through a set of text to speech generation examples.
Install dependencies
Install txtai
and all dependencies.
# Install txtai
pip install txtai[pipeline-audio,pipeline-data] onnxruntime-gpu librosa
# Install NLTK
import nltk
nltk.download('averaged_perceptron_tagger_eng')
Create a TextToSpeech instance
The TextToSpeech instance is the main entrypoint for generating speech from text. The pipeline is backed by models from the ESPnet project. ESPnet has a number of high quality TTS models available on the Hugging Face Hub.
This pipeline can use the following models on the Hugging Face Hub.
The default model is ljspeech-jets-onnx
. Each of the models above are ESPnet models exported to ONNX using espnet_onnx. More on that process can be found in the links above.
from txtai.pipeline import TextToSpeech
# Create text-to-speech model
tts = TextToSpeech()
Generate speech
The first example shows how to generate speech from text. Let's give it a try!
import librosa.display
import matplotlib.pyplot as plt
text = "Text To Speech models have made great strides in quality over the last few years."
# Generate raw waveform speech
speech, rate = tts(text), 22050
# Print waveplot
plt.figure(figsize=(15, 5))
plot = librosa.display.waveplot(speech[0], sr=speech[1])
The graph shows a plot of the audio. It clearly shows pauses between words and sentences as we would expect in spoken language. Now let's play the generated speech.
from IPython.display import Audio, display
import os
import soundfile as sf
def play(speech):
# Convert to MP3 to save space
sf.write("speech.wav", speech[0], speech[1])
!ffmpeg -i speech.wav -y -b:a 64 speech.mp3 2> /dev/null
# Play speech
display(Audio(filename="speech.mp3"))
play(speech)
Transcribe audio back to text
Next we'll use OpenAI Whisper to transcribe the generated audio back to text.
from txtai.pipeline import Transcription
# Transcribe files
transcribe = Transcription("openai/whisper-base")
# Print result
transcribe(speech, rate)
Text to speech models have made great strides in quality over the last few years.
And as expected, the transcription matches the original text.
Streaming speech generation
The TextToSpeech pipeline supports incrementally generating snippets of speech. This enables the pipeline to work with streaming LLM generation.
text = "This is streaming speech generation. It's designed to take output tokens from a streaming LLM. It returns snippets of speech.".split()
for speech, _ in tts(text, stream=True):
print(speech.shape)
(32768,)
(31488,)
(26368,)
Audio books
The TextToSpeech pipeline is designed to work with large blocks of text. It could be used to build audio for entire chapters of books.
In the next example below, we'll read the beginning of the book the Great Gatsby
. We'll load a new model that enables setting a speaker.
# Beginning of The Great Gatsby from Project Gutenberg
# https://www.gutenberg.org/ebooks/64317
text = """
In my younger and more vulnerable years my father gave me some advice
that I've been turning over in my mind ever since.
“Whenever you feel like criticizing anyone,” he told me, “just
remember that all the people in this world haven't had the advantages
that you've had.”
He didn't say any more, but we've always been unusually communicative
in a reserved way, and I understood that he meant a great deal more
than that.
"""
tts = TextToSpeech("neuml/vctk-vits-onnx")
speech = tts(text, speaker=3)
play(speech)
Text To Speech Workflow
In the last example, we'll cover building a text-to-speech workflow. This workflow is no different in that it connects multiple pipelines together, each of which are backed by machine learning models.
The workflow extracts text from a webpage, summarizes it and then generates audio of the summary.
summary:
path: sshleifer/distilbart-cnn-12-6
textractor:
join: true
lines: false
minlength: 100
paragraphs: true
sentences: false
texttospeech:
path: neuml/vctk-vits-onnx
workflow:
tts:
tasks:
- action: textractor
task: url
- action: summary
- action: texttospeech
args:
speaker: 15
from txtai.app import Application
app = Application("workflow.yml")
speech = list(app.workflow("tts", ["https://en.wikipedia.org/wiki/Natural_language_processing"]))[0]
play(speech)
Wrapping up
This article gave a brief introduction on text to speech models. The text to speech pipeline in txtai is designed to be easy to use and handles the most common text to speech tasks in English.
This work is made possible by the excellent advancements in text to speech modeling. ESPnet is a great project and should be checked out for more advanced and a wider range of use cases. This pipeline was also made possible by the great work from espnet_onnx in building a framework to export models to ONNX.
Looking forward to seeing what the community dreams up using this pipeline!
Top comments (0)