Latest Version: v0.0.0
Currently demo.
develop version: Python 3.11. Maybe 3.9 < Python < 3.13 is ok.
Installable Python kits, and information about using Python, are available at python.org.
The Model Context Protocol allows applications to provide context for LLMs in a standardized way, separating the concerns of providing context from the actual LLM interaction. This Python SDK implements the full MCP specification, making it easy to:
- Build MCP clients that can connect to any MCP server
- Create MCP servers that expose resources, prompts and tools
- Use standard transports like stdio and SSE
- Handle all MCP protocol messages and lifecycle events
We recommend using uv to manage your Python projects.
If you have uv-managed, Add MCP to your project dependencies:
uv add "mcp"Alternatively, for projects using pip for dependencies:
pip install "mcp"Currently using pyttsx3
Full documentation of the Library: https://pyttsx3.readthedocs.io/en/latest/
pip install pyttsx3If you get installation errors , make sure you first upgrade your wheel version using : pip install –upgrade wheel
If you are on a linux system and if the voice output is not working , then: Install espeak , ffmpeg and libespeak1 as shown below:
sudo apt update && sudo apt install espeak ffmpeg libespeak1pip install RealtimeSTTThis will install all the necessary dependencies, including a CPU support only version of PyTorch.
Although it is possible to run RealtimeSTT with a CPU installation only (use a small model like "tiny" or "base" in this case) you will get way better experience using CUDA (please scroll down).
Before installing RealtimeSTT please execute:
sudo apt-get update
sudo apt-get install python3-dev
sudo apt-get install portaudio19-devBefore installing RealtimeSTT please execute:
brew install portaudioTo upgrade your PyTorch installation to enable GPU support with CUDA, follow these instructions based on your specific CUDA version. This is useful if you wish to enhance the performance of RealtimeSTT with CUDA capabilities.
To update PyTorch and Torchaudio to support CUDA 11.8, use the following commands:
pip install torch==2.5.1+cu118 torchaudio==2.5.1 --index-url https://download.pytorch.org/whl/cu118To update PyTorch and Torchaudio to support CUDA 12.X, execute the following:
pip install torch==2.5.1+cu121 torchaudio==2.5.1 --index-url https://download.pytorch.org/whl/cu121Replace 2.5.1 with the version of PyTorch that matches your system and requirements.
Note: To check if your NVIDIA GPU supports CUDA, visit the official CUDA GPUs list.
If you didn't use CUDA models before, some additional steps might be needed one time before installation. These steps prepare the system for CUDA support and installation of the GPU-optimized installation. This is recommended for those who require better performance and have a compatible NVIDIA GPU. To use RealtimeSTT with GPU support via CUDA please also follow these steps:
-
Install NVIDIA CUDA Toolkit:
- select between CUDA 11.8 or CUDA 12.X Toolkit
- for 12.X visit NVIDIA CUDA Toolkit Archive and select latest version.
- for 11.8 visit NVIDIA CUDA Toolkit 11.8.
- Select operating system and version.
- Download and install the software.
- select between CUDA 11.8 or CUDA 12.X Toolkit
-
Install NVIDIA cuDNN:
- select between CUDA 11.8 or CUDA 12.X Toolkit
- for 12.X visit cuDNN Downloads.
- Select operating system and version.
- Download and install the software.
- for 11.8 visit NVIDIA cuDNN Archive.
- Click on "Download cuDNN v8.7.0 (November 28th, 2022), for CUDA 11.x".
- Download and install the software.
- for 12.X visit cuDNN Downloads.
- select between CUDA 11.8 or CUDA 12.X Toolkit
-
Install ffmpeg:
Note: Installation of ffmpeg might not actually be needed to operate RealtimeSTT *thanks to jgilbert2017 for pointing this out
You can download an installer for your OS from the ffmpeg Website.
Or use a package manager:
-
On Ubuntu or Debian:
sudo apt update && sudo apt install ffmpeg -
On Arch Linux:
sudo pacman -S ffmpeg
-
On MacOS using Homebrew (https://brew.sh/):
brew install ffmpeg
-
On Windows using Winget official documentation :
winget install Gyan.FFmpeg
-
On Windows using Chocolatey (https://chocolatey.org/):
choco install ffmpeg
-
On Windows using Scoop (https://scoop.sh/):
scoop install ffmpeg
-
Look here for information about how to train your own OpenWakeWord models. You can use a simple Google Colab notebook for a start or use a more detailed notebook that enables more customization (can produce high quality models, but requires more development experience).
You might need to use tf2onnx to convert tensorflow tflite models to onnx format:
pip install -U tf2onnx
python -m tf2onnx.convert --tflite my_model_filename.tflite --output my_model_filename.onnxQ: I encountered the following error: "Unable to load any of {libcudnn_ops.so.9.1.0, libcudnn_ops.so.9.1, libcudnn_ops.so.9, libcudnn_ops.so} Invalid handle. Cannot load symbol cudnnCreateTensorDescriptor." How do I fix this?
A: This issue arises from a mismatch between the version of ctranslate2 and cuDNN. The ctranslate2 library was updated to version 4.5.0, which uses cuDNN 9.2. There are two ways to resolve this issue:
- Downgrade
ctranslate2to version 4.4.0:pip install ctranslate2==4.4.0
- Upgrade cuDNN on your system to version 9.2 or above.
Contributions are always welcome!
Shoutout to Steven Linn for providing docker support.