A Multimodal User Interface for HiperController Software
Hiperwall Controller Multimodal integrates hand tracking using a webcam and voice recognition to control various mouse actions. By detecting specific hand gestures and interpreting voice commands, users can interact with their computer in a more intuitive and natural way.
Before you begin, ensure you have met the following requirements:
- Python 3.7 or higher
- OS: Windows 10 or higher, macOS 14, Linux
- Clone the repository:
git clone https://github.com/sion99/multimodal.git
cd multimodal- Create and activate a virtual environment:
python -m venv .venv
source .venv/bin/activate # On Windows use: .venv\Scripts\activate- Install the required libraries:
pip install -r requirements.txtTo run the program, follow these steps:
- Start the Python script
python main.py- The webcam will activate, and the program will begin listening for voice commands and tracking hand gestures.
- Use the following voice commands to control the mouse:
- "two": Double-click
- "left": Left-click
- "right": Right-click
- "up": Scroll up
- "down": Scroll down
- Press ESC to exit the program.