Instructions to use EpistemeAI/LexiVox with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use EpistemeAI/LexiVox with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-to-speech", model="EpistemeAI/LexiVox")# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("EpistemeAI/LexiVox") model = AutoModelForCausalLM.from_pretrained("EpistemeAI/LexiVox") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- Unsloth Studio new
How to use EpistemeAI/LexiVox with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for EpistemeAI/LexiVox to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for EpistemeAI/LexiVox to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for EpistemeAI/LexiVox to start chatting
Load model with FastModel
pip install unsloth from unsloth import FastModel model, tokenizer = FastModel.from_pretrained( model_name="EpistemeAI/LexiVox", max_seq_length=2048, )
This is a text to speech model, finetuned by Unsloth. Finetuned version of the unsloth/orpheus-3b-0.1-ft
Notes:
Replace model_name = "unsloth/orpheus-3b-0.1-ft" to model_name = "EpistemeAI/LexiVox" in colab
EpistemeAI/LexiVox TTS is a state-of-the-art, Llama-based Speech-LLM designed for high-quality, empathetic text-to-speech generation. This model has been finetuned to deliver human-level speech synthesis, achieving exceptional clarity, expressiveness, and real-time streaming performances.
Model Details
Model Capabilities
- Human-Like Speech: Natural intonation, emotion, and rhythm that is superior to SOTA closed source models
- Zero-Shot Voice Cloning: Clone voices without prior fine-tuning
- Guided Emotion and Intonation: Control speech and emotion characteristics with simple tags
- Low Latency: ~200ms streaming latency for realtime applications, reducible to ~100ms with input streaming
Model Sources
- GitHub Repo: https://github.com/canopyai/Orpheus-TTS
- Blog Post: https://canopylabs.ai/model-releases
- EpisteLabs's Colab Inference Notebook: notebook link
Usage
Check out our Colab (link to Colab) or GitHub (link to GitHub) on how to run easy inference on our finetuned models.
Model Misuse
Do not use our models for impersonation without consent, misinformation or deception (including fake news or fraudulent calls), or any illegal or harmful activity. By using this model, you agree to follow all applicable laws and ethical guidelines. We disclaim responsibility for any use.
Uploaded finetuned model
- Developed by: EpistemeAI
- License: apache-2.0
- Finetuned from model : unsloth/orpheus-3b-0.1-ft
This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
- Downloads last month
- 11
Model tree for EpistemeAI/LexiVox
Base model
meta-llama/Llama-3.2-3B-Instruct