Vikky is a personal assistant that helps developers to be more productive and efficient. Vikky is a bot that can be integrated with the developer's workspace and can be used to automate tasks, provide reminders, and help developers to stay focused. It is a physical bot that can be placed on the developer's desk and can be controlled using voice commands. Raspberry PI is used as the hardware platform for Vikky. Vikky's main focus is to help developers to be more productive and efficient by providing reminders, automating tasks, and helping developers to stay focused.
If you want to install the TTS Engine on your machine, please follow the steps below. This has only been tested on Linux, but I think it will work on Windows with small tweaks. If you are on Windows, I would recommend WSL with an Ubuntu image. Proper Windows and Mac support is in development.
- Install the
espeaksynthesizer according to the installation instructions for your operating system. - Install the required Python packages, e.g., by running
pip install -r requirements.txton Mac or Linux systems without an Nvidia GPU, andpip install -r requirements_cuda.txtif you have a modern Nvidia GPU. - Download the models:
- For voice recognition, we use Whisper.cpp
- Compile them yourself.
- To pull the code, from the Vikky directory use:
git submodule update --init --recursive - Move to the right subdirectory:
cd submodules/whisper.cpp - Compile for your system (see the Documentation), e.g.
- Raspberry PI with:
make libwhisper.so -jormake -j- sudo apt install ffmpeg
- No such file or directory as SDL.h => sudo apt install libsdl2-dev
- Mac with CoreML:
WHISPER_COREML=1 make -j
- Raspberry PI with:
- To pull the code, from the Vikky directory use:
- Compile them yourself.
- For the LLM, you have two option:
- Use a commercial API or install an inference backend yourself, such as Ollama:
- Find and install a backend with an OpenAI compatible API (most of them)
- Edit the glados_config.yaml
- update
completion_urlto the URL of your local server - for commercial APIs, add the
api_key - remove the LlamaServer configurations (make them null)
- update
- Use a commercial API or install an inference backend yourself, such as Ollama: