Application started at 2026-01-12 19:36:34
Set default params
Bind signal...
2026-01-12 19:36:41,024 - INFO - 返回缓存的编解码器 Windows-264: libx264
2026-01-12 19:36:41,026 - INFO - 返回缓存的编解码器 Windows-265: libx265
Load torch...
end
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
self.proxy_str='
http://127.0.0.1:10020',self.uuid=None
进入执行分离阶段 trk.shoud_recogn=True
2026-01-12 19:56:03,552 - INFO - cmd=['ffmpeg', '-hide_banner', '-ignore_unknown', '-threads', '0', '-y', '-i', 'E:/Downloads/10.mp4', '-ac', '1', '-ar', '16000', '-c
......
File "videotrans\task\job.py", line 113, in run
File "videotrans\task\_speech2text.py", line 146, in recogn
File "videotrans\recognition\__init__.py", line 227, in run
File "videotrans\recognition\_base.py", line 80, in run
File "videotrans\recognition\_funasr.py", line 60, in _exec
File "funasr\auto\auto_model.py", line 306, in generate
return self.inference_with_vad(input, input_len=input_len, **cfg)
File "funasr\auto\auto_model.py", line 383, in inference_with_vad
res = self.inference(
File "funasr\auto\auto_model.py", line 345, in inference
res = model.inference(**batch, **kwargs)
File "D:\win-videotrans-v3.67-tiny\_internal\funasr\models\fsmn_vad_streaming\model.py", line 690, in inference
audio_sample = torch.cat((cache["prev_samples"], audio_sample_list[0]))
TypeError: expected Tensor as element 1 in argument 0, but got str
2026-01-12 19:58:58,128 - ERROR - expected Tensor as element 1 in argument 0, but got str
Traceback (most recent call last):
File "videotrans\task\job.py", line 113, in run
File "videotrans\task\_speech2text.py", line 146, in recogn
File "videotrans\recognition\__init__.py", line 227, in run
File "videotrans\recognition\_base.py", line 80, in run
File "videotrans\recognition\_funasr.py", line 60, in _exec
File "funasr\auto\auto_model.py", line 306, in generate
return self.inference_with_vad(input, input_len=input_len, **cfg)
File "funasr\auto\auto_model.py", line 383, in inference_with_vad
res = self.inference(
File "funasr\auto\auto_model.py", line 345, in inference
res = model.inference(**batch, **kwargs)
File "D:\win-videotrans-v3.67-tiny\_internal\funasr\models\fsmn_vad_streaming\model.py", line 690, in inference
audio_sample = torch.cat((cache["prev_samples"], audio_sample_list[0]))
TypeError: expected Tensor as element 1 in argument 0, but got str
ERROR:VideoTrans:expected Tensor as element 1 in argument 0, but got str
Traceback (most recent call last):
File "videotrans\task\job.py", line 113, in run
File "videotrans\task\_speech2text.py", line 146, in recogn
File "videotrans\recognition\__init__.py", line 227, in run
File "videotrans\recognition\_base.py", line 80, in run
File "videotrans\recognition\_funasr.py", line 60, in _exec
File "funasr\auto\auto_model.py", line 306, in generate
return self.inference_with_vad(input, input_len=input_len, **cfg)
File "funasr\auto\auto_model.py", line 383, in inference_with_vad
res = self.inference(
File "funasr\auto\auto_model.py", line 345, in inference
res = model.inference(**batch, **kwargs)
File "D:\win-videotrans-v3.67-tiny\_internal\funasr\models\fsmn_vad_streaming\model.py", line 690, in inference
audio_sample = torch.cat((cache["prev_samples"], audio_sample_list[0]))
TypeError: expected Tensor as element 1 in argument 0, but got str
进入执行语音识别阶段
正在等待线程 SpeechToText 结束...
Successfully killed ffmpeg processes using taskkill