mirror of
https://git.adityakumar.xyz/llama.cpp.git
synced 2025-02-22 15:40:02 +00:00
llama : fix unused warning
This commit is contained in:
parent
66841fdb0e
commit
5a5aeb1e91
1 changed files with 2 additions and 0 deletions
|
@ -1053,6 +1053,8 @@ static void llama_model_load_internal(
|
|||
|
||||
fprintf(stderr, "%s: [cublas] total VRAM used: %zu MB\n", __func__, vram_total / 1024 / 1024);
|
||||
}
|
||||
#else
|
||||
(void) n_gpu_layers;
|
||||
#endif
|
||||
|
||||
// loading time will be recalculate after the first eval, so
|
||||
|
|
Loading…
Reference in a new issue