π Quick Start
π Installing Requirements
UnityNeuroSpeech requires several programs to be installed.
You can simply run setup.bat
β it will download everything automatically.
Then just import the .unitypackage
into your project.
π‘ What Are These Requirements?
- Ollama β a platform for running large language models (LLMs) locally. You can use models like DeepSeek, Gemma, Qwen, etc. Avoid small models β they might reduce accuracy and context understanding.
- UV β a modern, ultra-fast Python package and environment manager. It replaces traditional tools like
pip
andvenv
. Coqui XTTS uses UV to simplify installation and allows running the TTS command directly, without manual Python setup. - Coqui XTTS β a Text-To-Speech model that can generate speech in any custom voice you want: Chester Bennington, Chris Tucker, Vito Corleone (The Godfather), Cyn (Murder Drones), or any other.
- Whisper β a Speech-To-Text model. You can use lightweight versions like
ggml-tiny.bin
for speed, or heavier ones likeggml-medium.bin
for better accuracy.
ποΈ Voice Files
Donβt forget that you need voice files for AI speech.
Make sure your files meet the following requirements:
- Format:
.wav
- Duration: 5β15 seconds (longer files work, but TTS will load them more slowly)
- Contain only one voice and one language, without background noise
Since UnityNeuroSpeech supports multiple voices for multiple agents simultaneously, files must be named correctly:
<language>_voice<index>.wav
Examples:
- English voice, agent index
0
βen_voice0.wav
- Russian voice, agent index
3
βru_voice3.wav
All voices must be placed in:
Assets/StreamingAssets/UnityNeuroSpeech/Voices/
πΌοΈ Microphone Sprites
Youβll need two sprites for the microphone state (enabled/disabled).
Yes β without them, it wonβt work π€