ollama/gpu
Daniel Hiltgen f457d63400 Implement linux NUMA detection
If the system has multiple numa nodes, enable numa support in llama.cpp
If we detect numactl in the path, use that, else use the basic "distribute" mode.
2024-08-05 12:56:20 -07:00
..
amd_common.go lint 2024-08-01 17:06:06 -07:00
amd_hip_windows.go lint 2024-08-01 17:06:06 -07:00
amd_linux.go lint 2024-08-01 17:06:06 -07:00
amd_windows.go lint 2024-08-01 17:06:06 -07:00
assets.go lint 2024-08-01 17:06:06 -07:00
cpu_common.go Implement linux NUMA detection 2024-08-05 12:56:20 -07:00
cuda_common.go lint linux 2024-06-04 11:13:30 -07:00
gpu.go lint 2024-08-01 17:06:06 -07:00
gpu_darwin.go lint 2024-08-01 17:06:06 -07:00
gpu_info.h lint 2024-08-01 17:06:06 -07:00
gpu_info_cudart.c Fix bad symbol load detection 2024-06-19 08:39:07 -07:00
gpu_info_cudart.h Refine GPU discovery to bootstrap once 2024-06-14 14:51:40 -07:00
gpu_info_darwin.h gpu: report system free memory instead of 0 (#5521) 2024-07-06 19:35:04 -04:00
gpu_info_darwin.m gpu: report system free memory instead of 0 (#5521) 2024-07-06 19:35:04 -04:00
gpu_info_nvcuda.c Better nvidia GPU discovery logging 2024-07-03 10:50:40 -07:00
gpu_info_nvcuda.h Better nvidia GPU discovery logging 2024-07-03 10:50:40 -07:00
gpu_info_nvml.c Fix bad symbol load detection 2024-06-19 08:39:07 -07:00
gpu_info_nvml.h Reintroduce nvidia nvml library for windows 2024-06-14 14:51:40 -07:00
gpu_info_oneapi.c get real func ptr. 2024-06-19 09:00:51 +08:00
gpu_info_oneapi.h review comments and coverage 2024-06-14 14:55:50 -07:00
gpu_linux.go lint 2024-08-01 17:06:06 -07:00
gpu_oneapi.go support ollama run on Intel GPUs 2024-05-24 11:18:27 +08:00
gpu_test.go lint 2024-06-04 11:13:30 -07:00
gpu_windows.go lint 2024-08-01 17:06:06 -07:00
types.go llm: avoid loading model if system memory is too small (#5637) 2024-07-11 16:42:57 -07:00