lee101/guided-text-to-speech 📝 → 🖼️
About
Guided Text to Speech Generator

Example Output
Prompt:
"hi whats the weather?"
Output
Performance Metrics
7.72s
Prediction Time
209.40s
Total Time
All Input Parameters
{ "voice": "A female speaker with a slightly low-pitched, quite monotone voice delivers her words at a slightly faster-than-average pace in a confined space with very clear audio.", "prompt": "hi whats the weather?" }
Input Parameters
- voice (required)
- voice description
- prompt (required)
- Voice
Output Schema
Output
Example Execution Logs
Using the model-agnostic default `max_length` (=2580) to control the generation length. We recommend setting `max_new_tokens` to control the maximum length of the generation. Calling `sample` directly is deprecated and will be removed in v4.41. Use `generate` or a custom generation loop instead. --- Logging error --- Traceback (most recent call last): File "/root/.pyenv/versions/3.11.9/lib/python3.11/logging/__init__.py", line 1110, in emit msg = self.format(record) ^^^^^^^^^^^^^^^^^^^ File "/root/.pyenv/versions/3.11.9/lib/python3.11/logging/__init__.py", line 953, in format return fmt.format(record) ^^^^^^^^^^^^^^^^^^ File "/root/.pyenv/versions/3.11.9/lib/python3.11/logging/__init__.py", line 687, in format record.message = record.getMessage() ^^^^^^^^^^^^^^^^^^^ File "/root/.pyenv/versions/3.11.9/lib/python3.11/logging/__init__.py", line 377, in getMessage msg = msg % self.args ~~~~^~~~~~~~~~~ TypeError: not all arguments converted during string formatting Call stack: File "<string>", line 1, in <module> File "/root/.pyenv/versions/3.11.9/lib/python3.11/multiprocessing/spawn.py", line 122, in spawn_main exitcode = _main(fd, parent_sentinel) File "/root/.pyenv/versions/3.11.9/lib/python3.11/multiprocessing/spawn.py", line 135, in _main return self._bootstrap(parent_sentinel) File "/root/.pyenv/versions/3.11.9/lib/python3.11/multiprocessing/process.py", line 314, in _bootstrap self.run() File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/cog/server/worker.py", line 179, in run self._loop() File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/cog/server/worker.py", line 211, in _loop self._predict(ev.payload) File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/cog/server/worker.py", line 221, in _predict result = predict(**payload) File "/src/predict.py", line 15, in predict sample_rate, audio_arr = gen_tts(prompt, voice) File "/src/parlerlib.py", line 91, in gen_tts generation = model.generate( File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context return func(*args, **kwargs) File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/parler_tts/modeling_parler_tts.py", line 2608, in generate outputs = self.sample( File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/transformers/generation/utils.py", line 2584, in sample return self._sample(*args, **kwargs) File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/transformers/generation/utils.py", line 2730, in _sample logger.warning_once( File "/root/.pyenv/versions/3.11.9/lib/python3.11/site-packages/transformers/utils/logging.py", line 329, in warning_once self.warning(*args, **kwargs) Message: '`eos_token_id` is deprecated in this function and will be removed in v4.41, use `stopping_criteria=StoppingCriteriaList([EosTokenCriteria(eos_token_id=eos_token_id)])` instead. Otherwise make sure to set `model.generation_config.eos_token_id`' Arguments: (<class 'FutureWarning'>,)
Version Details
- Version ID
fc0617a394340824a7dd1aa78f76e92c061449abd48e67ee9dbe30a6448c8be2
- Version Created
- May 28, 2024