KeyError: 'CUDA_VISIBLE_DEVICES'

本文介绍了解决CUDA_VISIBLE_DEVICES环境变量未设置导致的KeyError异常的方法,并提供了如何手动设置该环境变量的具体步骤。

cvd = os.environ['CUDA_VISIBLE_DEVICES'].strip()
  File "/usr/lib/python2.7/UserDict.py", line 23, in __getitem__
    raise KeyError(key)
KeyError: 'CUDA_VISIBLE_DEVICES'

如果没有,自己export一个就行了

export CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6

root@wlcb-dyg:/home/wlcb-dyg# ollama serve Couldn't find '/root/.ollama/id_ed25519'. Generating new private key. Your new public key is: ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAID4ovJFMvZAtq71NIhecQfJmdE0w16e3jZMylgVJ2lq/ 2025/03/13 09:15:50 routes.go:1225: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:2048 OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-03-13T09:15:50.606+08:00 level=INFO source=images.go:432 msg="total blobs: 0" time=2025-03-13T09:15:50.606+08:00 level=INFO source=images.go:439 msg="total unused blobs removed: 0" time=2025-03-13T09:15:50.606+08:00 level=INFO source=routes.go:1292 msg="Listening on 127.0.0.1:11434 (version 0.6.0)" time=2025-03-13T09:15:50.606+08:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-03-13T09:15:51.523+08:00 level=INFO source=types.go:130 msg="inference compute" id=GPU-91af1d5b-5fed-e3b0-93e9-2112bebf3262 library=cuda variant=v12 compute=7.5 driver=12.2 name="Tesla T4" total="14.6 GiB" available="14.5 GiB" time=2025-03-13T09:15:51.523+08:00 level=INFO source=types.go:130 msg="inference compute" id=GPU-6cc43ad6-f63f-84ca-666d-
03-14
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值