runtime error

Exit code: 1. Reason: r device for inference. Pipelines loaded with `dtype=torch.float16` cannot run with `cpu` device. It is not recommended to move them to `cpu` as running them will fail. Please make sure to use an accelerator to run the pipeline in inference, due to the lack of support for`float16` operations on this device in PyTorch. Please, remove the `torch_dtype=torch.float16` argument, or use another device for inference. sdxl_lightning_4step_lora.safetensors: 0%| | 0.00/394M [00:00<?, ?B/s] sdxl_lightning_4step_lora.safetensors: 76%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 299M/394M [00:01<00:00, 298MB/s] sdxl_lightning_4step_lora.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 394M/394M [00:01<00:00, 344MB/s] [INFO] Model is loaded! Traceback (most recent call last): File "/home/user/app/app.py", line 110, in <module> models = { File "/home/user/app/app.py", line 111, in <dictcomp> k: StableMultiDiffusionSDXLPipeline(device, hf_key=v, has_i2t=False).cuda() File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1050, in cuda return self._apply(lambda t: t.cuda(device)) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 900, in _apply module._apply(fn) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 900, in _apply module._apply(fn) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 900, in _apply module._apply(fn) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 927, in _apply param_applied = fn(param) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1050, in <lambda> return self._apply(lambda t: t.cuda(device)) File "/usr/local/lib/python3.10/site-packages/torch/cuda/__init__.py", line 319, in _lazy_init torch._C._cuda_init() RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx

Container logs:

Fetching error logs...