mirror of
https://github.com/ollama/ollama.git
synced 2025-05-10 09:56:28 +02:00
ml/backend/ggml: use default CUDA compression mode (#10314)
This commit is contained in:
parent
dc264be6ff
commit
09bb2e30f6
1 changed files with 1 additions and 1 deletions
|
@ -24,7 +24,7 @@ set(GGML_LLAMAFILE ON)
|
|||
set(GGML_CUDA_PEER_MAX_BATCH_SIZE 128)
|
||||
set(GGML_CUDA_GRAPHS ON)
|
||||
set(GGML_CUDA_FA ON)
|
||||
set(GGML_CUDA_COMPRESSION_MODE none)
|
||||
set(GGML_CUDA_COMPRESSION_MODE default)
|
||||
|
||||
if((CMAKE_OSX_ARCHITECTURES AND NOT CMAKE_OSX_ARCHITECTURES MATCHES "arm64")
|
||||
OR (NOT CMAKE_OSX_ARCHITECTURES AND NOT CMAKE_SYSTEM_PROCESSOR MATCHES "arm|aarch64|ARM64|ARMv[0-9]+"))
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue