| ▲ | zozbot234 4 hours ago | |||||||
Your link seems to be describing a runtime environment variable, it doesn't need a separate build from source. I'm not sure though (1) why this info is in build.md which should be specific to the building process, rather than some separate documentation; and (2) if this really isn't CUDA-specific, why the canonical GGML variable name isn't GGML_ENABLE_UNIFIED_MEMORY , with the _CUDA_ variant treated as a legacy alias. AIUI, both of these should be addressed with pull requests for llama.cpp and/or the ggml library itself. | ||||||||
| ▲ | rubiquity 3 hours ago | parent [-] | |||||||
You are right that it is an environment variable, and that's how I have it set in my nix config. Thanks for correcting that. Unfortunately llama.cpp is somewhat notorious for having lackluster docs. Most of the CLI tools don't even tell you what they are for. | ||||||||
| ||||||||