Spaces:
Runtime error
Runtime error
Upload folder using huggingface_hub
Browse files- .gitattributes +0 -3
- app.py +1 -1
- fish_speech/models/text2semantic/llama.py +1 -1
- tools/api.py +1 -1
- tools/vqgan/inference.py +1 -1
.gitattributes
CHANGED
|
@@ -41,6 +41,3 @@ examples/English.wav filter=lfs diff=lfs merge=lfs -text
|
|
| 41 |
examples/French.wav filter=lfs diff=lfs merge=lfs -text
|
| 42 |
examples/German.wav filter=lfs diff=lfs merge=lfs -text
|
| 43 |
examples/Spanish.wav filter=lfs diff=lfs merge=lfs -text
|
| 44 |
-
022b2161-8f56-4432-a9ae-b4bd514e4821.mp3 filter=lfs diff=lfs merge=lfs -text
|
| 45 |
-
output.wav filter=lfs diff=lfs merge=lfs -text
|
| 46 |
-
ref.wav filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 41 |
examples/French.wav filter=lfs diff=lfs merge=lfs -text
|
| 42 |
examples/German.wav filter=lfs diff=lfs merge=lfs -text
|
| 43 |
examples/Spanish.wav filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
app.py
CHANGED
|
@@ -313,4 +313,4 @@ if __name__ == "__main__":
|
|
| 313 |
inference_fct = get_inference_wrapper(inference_engine)
|
| 314 |
|
| 315 |
app = build_app(inference_fct, args.theme)
|
| 316 |
-
app.queue(api_open=True).launch(show_error=True, show_api=True)
|
|
|
|
| 313 |
inference_fct = get_inference_wrapper(inference_engine)
|
| 314 |
|
| 315 |
app = build_app(inference_fct, args.theme)
|
| 316 |
+
app.queue(api_open=True).launch(show_error=True, show_api=True, server_name="0.0.0.0", server_port=18888)
|
fish_speech/models/text2semantic/llama.py
CHANGED
|
@@ -48,7 +48,7 @@ class BaseModelArgs:
|
|
| 48 |
|
| 49 |
# Codebook configs
|
| 50 |
codebook_size: int = 160
|
| 51 |
-
num_codebooks: int =
|
| 52 |
|
| 53 |
# Gradient checkpointing
|
| 54 |
use_gradient_checkpointing: bool = True
|
|
|
|
| 48 |
|
| 49 |
# Codebook configs
|
| 50 |
codebook_size: int = 160
|
| 51 |
+
num_codebooks: int = 4
|
| 52 |
|
| 53 |
# Gradient checkpointing
|
| 54 |
use_gradient_checkpointing: bool = True
|
tools/api.py
CHANGED
|
@@ -136,7 +136,7 @@ async def other_exception_handler(exc: "Exception"):
|
|
| 136 |
|
| 137 |
|
| 138 |
def load_audio(reference_audio, sr):
|
| 139 |
-
if len(
|
| 140 |
audio_data = reference_audio
|
| 141 |
reference_audio = io.BytesIO(audio_data)
|
| 142 |
|
|
|
|
| 136 |
|
| 137 |
|
| 138 |
def load_audio(reference_audio, sr):
|
| 139 |
+
if len(reference_audio) > 255 or not Path(reference_audio).exists():
|
| 140 |
audio_data = reference_audio
|
| 141 |
reference_audio = io.BytesIO(audio_data)
|
| 142 |
|
tools/vqgan/inference.py
CHANGED
|
@@ -14,7 +14,7 @@ from omegaconf import OmegaConf
|
|
| 14 |
from tools.file import AUDIO_EXTENSIONS
|
| 15 |
|
| 16 |
# register eval resolver
|
| 17 |
-
|
| 18 |
|
| 19 |
|
| 20 |
def load_model(config_name, checkpoint_path, device="cuda"):
|
|
|
|
| 14 |
from tools.file import AUDIO_EXTENSIONS
|
| 15 |
|
| 16 |
# register eval resolver
|
| 17 |
+
OmegaConf.register_new_resolver("eval", eval)
|
| 18 |
|
| 19 |
|
| 20 |
def load_model(config_name, checkpoint_path, device="cuda"):
|