执行:ollama run gemma2:27b
Error: llama runner process has terminated: signal: aborted (core dumped) CUDA error: CUBLAS_STATUS_NOT_INITIALIZED
current device: 0, in function cublas_handle at /go/src/github.com/ollama/ollama/llm/llama.cpp/ggml-cuda/common.cuh:826
cublasCreate_v2(&cublas_handles[device])
GGML_ASSERT: /go/src/github.com/ollama/ollama/llm/llama.cpp/ggml-cuda.cu:100: !"CUDA error"
解决方式:
curl -fsSL https://ollama.com/install.sh | sh
Downloading ollama...
######################################################################## 100.0%##O=# #
Installing ollama to /usr/local/bin...
Adding ollama user to render group...
Adding ollama user to video group...
Adding current user to ollama group...
Creating ollama systemd service...
Enabling and starting ollama service...
NVIDIA GPU installed.
(base) root@aplomb-System-Product-Name:~# ollama run gemma2:27b
你是谁?
我是一个大型语言模型,由 Google 训练。