(index-tts) C:\Users\admin\index-tts>uv run webui.py
>> GPT weights restored from: checkpoints\gpt.pth
[2025-09-09 12:32:49,076] [INFO] [real_accelerator.py:254:get_accelerator] Setting ds_accelerator to cuda (auto detect)
test.c
LINK : fatal error LNK1181: 无法打开输入文件“aio.lib”
test.c
LINK : fatal error LNK1181: 无法打开输入文件“cufile.lib”
W0909 12:33:25.637000 4428 .venv\Lib\site-packages\torch\distributed\elastic\multiprocessing\redirects.py:29] NOTE: Redirects are currently not supported in Windows or MacOs.
[2025-09-09 12:33:26,840] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False
GPT2InferenceModel has generative capabilities, as `prepare_inputs_for_generation` is explicitly overwritten. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions.
- If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes
- If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception).
- If you are not the owner of the model architecture class, please contact the model code owner to update it.
[2025-09-09 12:33:26,861] [INFO] [logging.py:107:log_dist] [Rank -1] DeepSpeed info: version=0.17.1+2ce55057, git-hash=2ce55057, git-branch=v0.17.1-branch
[2025-09-09 12:33:26,862] [WARNING] [config_utils.py:70:_process_deprecated_field] Config parameter mp_size is deprecated use tensor_parallel.tp_size instead
[2025-09-09 12:33:26,864] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False
[2025-09-09 12:33:26,864] [INFO] [logging.py:107:log_dist] [Rank -1] quantize_bits = 8 mlp_extra_grouping = False, quantize_groups = 1
[2025-09-09 12:33:26,990] [INFO] [logging.py:107:log_dist] [Rank -1] DeepSpeed-Inference config: {'layer_id': 0, 'hidden_size': 1280, 'intermediate_size': 5120, 'heads': 20, 'num_hidden_layers': -1, 'dtype': torch.float32, 'pre_layer_norm': True, 'norm_type': <NormType.LayerNorm: 1>, 'local_rank': -1, 'stochastic_mode': False, 'epsilon': 1e-05, 'mp_size': 1, 'scale_attention': True, 'triangular_masking': True, 'local_attention': False, 'window_size': 1, 'rotary_dim': -1, 'rotate_half': False, 'rotate_every_two': True, 'return_tuple': True, 'mlp_after_attn': True, 'mlp_act_func_type': <ActivationFuncType.GELU: 1>, 'training_mp_size': 1, 'bigscience_bloom': False, 'max_out_tokens': 1024, 'min_out_tokens': 1, 'scale_attn_by_inverse_layer_idx': False, 'enable_qkv_quantization': False, 'use_mup': False, 'return_single_tuple': False, 'set_empty_params': False, 'transposed_mode': False, 'use_triton': False, 'triton_autotune': False, 'num_kv': -1, 'rope_theta': 10000, 'invert_mask': True}
W0909 12:33:27.396000 4428 .venv\Lib\site-packages\torch\utils\cpp_extension.py:466] Error checking compiler version for cl: [WinError 2] 系统找不到指定的文件。
W0909 12:33:27.397000 4428 .venv\Lib\site-packages\torch\utils\cpp_extension.py:2425] TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
W0909 12:33:27.397000 4428 .venv\Lib\site-packages\torch\utils\cpp_extension.py:2425] If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'] to specific architectures.
信息: 用提供的模式无法找到文件。
>> Failed to load custom CUDA kernel for BigVGAN. Falling back to torch.
Traceback (most recent call last):
File "C:\Users\admin\index-tts\webui.py", line 52, in <module>
tts = IndexTTS2(model_dir=cmd_args.model_dir, cfg_path=os.path.join(cmd_args.model_dir, "config.yaml"),is_fp16=cmd_args.is_fp16)
File "C:\Users\admin\index-tts\indextts\infer_v2.py", line 106, in __init__
self.extract_features = SeamlessM4TFeatureExtractor.from_pretrained("facebook/w2v-bert-2.0")
File "C:\Users\admin\index-tts\.venv\lib\site-packages\transformers\feature_extraction_utils.py", line 384, in from_pretrained
feature_extractor_dict, kwargs = cls.get_feature_extractor_dict(pretrained_model_name_or_path, **kwargs)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\transformers\feature_extraction_utils.py", line 510, in get_feature_extractor_dict
resolved_feature_extractor_file = cached_file(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\transformers\utils\hub.py", line 312, in cached_file
file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\transformers\utils\hub.py", line 557, in cached_files
raise e
File "C:\Users\admin\index-tts\.venv\lib\site-packages\transformers\utils\hub.py", line 470, in cached_files
hf_hub_download(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\file_download.py", line 1010, in hf_hub_download
return _hf_hub_download_to_cache_dir(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\file_download.py", line 1073, in _hf_hub_download_to_cache_dir
(url_to_download, etag, commit_hash, expected_size, xet_file_data, head_call_error) = _get_metadata_or_catch_error(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\file_download.py", line 1546, in _get_metadata_or_catch_error
metadata = get_hf_file_metadata(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\file_download.py", line 1463, in get_hf_file_metadata
r = _request_wrapper(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\file_download.py", line 286, in _request_wrapper
response = _request_wrapper(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\file_download.py", line 309, in _request_wrapper
response = http_backoff(method=method, url=url, **params, retry_on_exceptions=(), retry_on_status_codes=(429,))
File "C:\Users\admin\index-tts\.venv\lib\site-packages\huggingface_hub\utils\_http.py", line 310, in http_backoff
response = session.request(method=method, url=url, **kwargs)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\requests\sessions.py", line 575, in request
prep = self.prepare_request(req)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\requests\sessions.py", line 484, in prepare_request
p.prepare(
File "C:\Users\admin\index-tts\.venv\lib\site-packages\requests\models.py", line 367, in prepare
self.prepare_url(url, params)
File "C:\Users\admin\index-tts\.venv\lib\site-packages\requests\models.py", line 438, in prepare_url
raise MissingSchema(
requests.exceptions.MissingSchema: Invalid URL '""/facebook/w2v-bert-2.0/resolve/main/preprocessor_config.json': No scheme supplied. Perhaps you meant https://""/facebook/w2v-bert-2.0/resolve/main/preprocessor_config.json?
(index-tts) C:\Users\admin\index-tts>python -c "import torch; print(torch.cuda.is_available(), torch.version.cuda)" True 12.8