You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
run successful when on CPUExecutionProvider mode.
but run error when on CUDAExecutionProvider mode .
error log :
python app.py
/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/pandas/core/computation/expressions.py:21: UserWarning: Pandas requires version '2.8.4' or newer of 'numexpr' (version '2.8.3' currently installed).
from pandas.core.computation.check import NUMEXPR_INSTALLED
2024-08-22 10:06:06,271 - cv2box - INFO - Use default log level: info, or you can set env 'CV_LOG_LEVEL' to debug/info/warning/error/critical
Running on local URL: http://0.0.0.0:5000
To create a public link, set share=True in launch().
warning: det_size is already set in scrfd model, ignore
pfpld ['CUDAExecutionProvider', 'CPUExecutionProvider']
Model warm up done !
[[gpu] onnx 10 times fps: 42.141950004018966]
[[gpu] onnx 10 times fps: 42.45289932084333]
2024-08-22 10:06:25.617085482 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.617124996 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose
2024-08-22 10:06:25.617650844 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_1 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.617671885 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_1
2024-08-22 10:06:25.618188729 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_2 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.618209622 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_2
2024-08-22 10:06:25.618725877 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_3 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.618746728 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_3
2024-08-22 10:06:25.619162584 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_4 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.619183908 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_4
2024-08-22 10:06:25.619586327 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_5 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.619602145 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_5
xseg ['CUDAExecutionProvider', 'CPUExecutionProvider']
Model warm up done !
[[gpu] onnx 10 times fps: 3.398418237666232]
[[gpu] onnx 10 times fps: 3.8644535658839736]
9O ['CUDAExecutionProvider', 'CPUExecutionProvider']
Model warm up done !
[[gpu] onnx 10 times fps: 4.281553497165233]
[[gpu] onnx 10 times fps: 5.184034725818552]
warning: det_size is already set in scrfd model, ignore
Traceback (most recent call last):
File "/share/ai/ReHiFace-S-main/face_detect/face_align_68.py", line 200, in init
self.ort_session = ort.InferenceSession(onnx_path)
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 360, in init
self._create_inference_session(providers, provider_options, disabled_optimizers)
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 388, in _create_inference_session
raise ValueError(
ValueError: This ORT build has ['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider'] enabled. Since ORT 1.9, you are required to explicitly set the providers parameter when instantiating InferenceSession. For example, onnxruntime.InferenceSession(..., providers=['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider'], ...)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/queueing.py", line 536, in process_events
response = await route_utils.call_process_api(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/route_utils.py", line 288, in call_process_api
output = await app.get_blocks().process_api(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/blocks.py", line 1931, in process_api
result = await self.call_function(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/blocks.py", line 1516, in call_function
prediction = await anyio.to_thread.run_sync( # type: ignore
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync
return await get_async_backend().run_sync_in_worker_thread(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2177, in run_sync_in_worker_thread
return await future
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 859, in run
result = context.run(func, *args)
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/utils.py", line 826, in wrapper
response = f(*args, **kwargs)
File "/share/ai/ReHiFace-S-main/app.py", line 106, in process_image_video
initialize_processor()
File "/share/ai/ReHiFace-S-main/app.py", line 24, in initialize_processor
processor = FaceSwapProcessor(crop_size=opt.input_size)
File "/share/ai/ReHiFace-S-main/app.py", line 32, in init
self.face_alignment = face_alignment_landmark(lm_type=68)
File "/share/ai/ReHiFace-S-main/face_detect/face_align_68.py", line 226, in init
self.fan = pfpld(cpu=False)
File "/share/ai/ReHiFace-S-main/face_detect/face_align_68.py", line 202, in init
raise e("load onnx failed")
TypeError: 'ValueError' object is not callable
The text was updated successfully, but these errors were encountered:
run successful when on CPUExecutionProvider mode.
but run error when on CUDAExecutionProvider mode .
error log :
python app.py
/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/pandas/core/computation/expressions.py:21: UserWarning: Pandas requires version '2.8.4' or newer of 'numexpr' (version '2.8.3' currently installed).
from pandas.core.computation.check import NUMEXPR_INSTALLED
2024-08-22 10:06:06,271 - cv2box - INFO - Use default log level: info, or you can set env 'CV_LOG_LEVEL' to debug/info/warning/error/critical
Running on local URL: http://0.0.0.0:5000
To create a public link, set
share=True
inlaunch()
.warning: det_size is already set in scrfd model, ignore
pfpld ['CUDAExecutionProvider', 'CPUExecutionProvider']
Model warm up done !
[[gpu] onnx 10 times fps: 42.141950004018966]
[[gpu] onnx 10 times fps: 42.45289932084333]
2024-08-22 10:06:25.617085482 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.617124996 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose
2024-08-22 10:06:25.617650844 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_1 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.617671885 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_1
2024-08-22 10:06:25.618188729 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_2 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.618209622 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_2
2024-08-22 10:06:25.618725877 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_3 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.618746728 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_3
2024-08-22 10:06:25.619162584 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_4 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.619183908 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_4
2024-08-22 10:06:25.619586327 [W:onnxruntime:Default, cuda_execution_provider.cc:2310 ConvTransposeNeedFallbackToCPU] Dropping the ConvTranspose node: conv2d_transpose_5 to CPU because it requires asymmetric padding which the CUDA EP currently does not support
2024-08-22 10:06:25.619602145 [W:onnxruntime:Default, cuda_execution_provider.cc:2411 GetCapability] CUDA kernel not supported. Fallback to CPU execution provider for Op type: ConvTranspose node name: conv2d_transpose_5
xseg ['CUDAExecutionProvider', 'CPUExecutionProvider']
Model warm up done !
[[gpu] onnx 10 times fps: 3.398418237666232]
[[gpu] onnx 10 times fps: 3.8644535658839736]
9O ['CUDAExecutionProvider', 'CPUExecutionProvider']
Model warm up done !
[[gpu] onnx 10 times fps: 4.281553497165233]
[[gpu] onnx 10 times fps: 5.184034725818552]
warning: det_size is already set in scrfd model, ignore
Traceback (most recent call last):
File "/share/ai/ReHiFace-S-main/face_detect/face_align_68.py", line 200, in init
self.ort_session = ort.InferenceSession(onnx_path)
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 360, in init
self._create_inference_session(providers, provider_options, disabled_optimizers)
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 388, in _create_inference_session
raise ValueError(
ValueError: This ORT build has ['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider'] enabled. Since ORT 1.9, you are required to explicitly set the providers parameter when instantiating InferenceSession. For example, onnxruntime.InferenceSession(..., providers=['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider'], ...)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/queueing.py", line 536, in process_events
response = await route_utils.call_process_api(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/route_utils.py", line 288, in call_process_api
output = await app.get_blocks().process_api(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/blocks.py", line 1931, in process_api
result = await self.call_function(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/blocks.py", line 1516, in call_function
prediction = await anyio.to_thread.run_sync( # type: ignore
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync
return await get_async_backend().run_sync_in_worker_thread(
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2177, in run_sync_in_worker_thread
return await future
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 859, in run
result = context.run(func, *args)
File "/opt/miniconda3/envs/ReHiFace/lib/python3.10/site-packages/gradio/utils.py", line 826, in wrapper
response = f(*args, **kwargs)
File "/share/ai/ReHiFace-S-main/app.py", line 106, in process_image_video
initialize_processor()
File "/share/ai/ReHiFace-S-main/app.py", line 24, in initialize_processor
processor = FaceSwapProcessor(crop_size=opt.input_size)
File "/share/ai/ReHiFace-S-main/app.py", line 32, in init
self.face_alignment = face_alignment_landmark(lm_type=68)
File "/share/ai/ReHiFace-S-main/face_detect/face_align_68.py", line 226, in init
self.fan = pfpld(cpu=False)
File "/share/ai/ReHiFace-S-main/face_detect/face_align_68.py", line 202, in init
raise e("load onnx failed")
TypeError: 'ValueError' object is not callable
The text was updated successfully, but these errors were encountered: