Commit History

Author SHA1 Message Date
  Jeffrey Morgan efbf41ed81 llm: dont link cuda with compat libs (#5621) 5 months ago
  Jeffrey Morgan 4e262eb2a8 remove `GGML_CUDA_FORCE_MMQ=on` from build (#5588) 5 months ago
  Daniel Hiltgen 0bacb30007 Workaround broken ROCm p2p copy 5 months ago
  Jeffrey Morgan 4607c70641 llm: add `-DBUILD_SHARED_LIBS=off` to common cpu cmake flags (#5520) 5 months ago
  Jeffrey Morgan 2cc854f8cb llm: fix missing dylibs by restoring old build behavior on Linux and macOS (#5511) 5 months ago
  Jeffrey Morgan 8f8e736b13 update llama.cpp submodule to `d7fd29f` (#5475) 5 months ago
  Daniel Hiltgen b0930626c5 Add back lower level parallel flags 6 months ago
  Jeffrey Morgan 152fc202f5 llm: update llama.cpp commit to `7c26775` (#4896) 6 months ago
  Daniel Hiltgen ab8c929e20 Add ability to skip oneapi generate 6 months ago
  Daniel Hiltgen 646371f56d Merge pull request #3278 from zhewang1-intc/rebase_ollama_main 7 months ago
  Wang,Zhe fd5971be0b support ollama run on Intel GPUs 7 months ago
  Daniel Hiltgen c48c1d7c46 Port cuda/rocm skip build vars to linux 7 months ago
  Roy Yang 5f73c08729 Remove trailing spaces (#3889) 8 months ago
  Daniel Hiltgen cc5a71e0e3 Merge pull request #3709 from remy415/custom-gpu-defs 8 months ago
  Jeremy 440b7190ed Update gen_linux.sh 8 months ago
  Jeremy 52f5370c48 add support for custom gpu build flags for llama.cpp 8 months ago
  Jeremy 7c000ec3ed adds support for OLLAMA_CUSTOM_GPU_DEFS to customize GPU build flags 8 months ago
  Jeremy 8aec92fa6d rearranged conditional logic for static build, dockerfile updated 8 months ago
  Jeremy 70261b9bb6 move static build to its own flag 8 months ago
  Blake Mizerany 1524f323a3 Revert "build.go: introduce a friendlier way to build Ollama (#3548)" (#3564) 8 months ago
  Blake Mizerany fccf3eecaa build.go: introduce a friendlier way to build Ollama (#3548) 8 months ago
  Jeffrey Morgan 63efa075a0 update generate scripts with new `LLAMA_CUDA` variable, set `HIP_PLATFORM` to avoid compiler errors (#3528) 8 months ago
  Daniel Hiltgen 58d95cc9bd Switch back to subprocessing for llama.cpp 9 months ago
  Jeremy dfc6721b20 add support for libcudart.so for CUDA devices (adds Jetson support) 9 months ago
  Daniel Hiltgen d4c10df2b0 Add Radeon gfx940-942 GPU support 9 months ago
  Daniel Hiltgen bc13da2bfe Avoid rocm runner and dependency clash 9 months ago
  Daniel Hiltgen 3dc1bb6a35 Harden for deps file being empty (or short) 9 months ago
  Daniel Hiltgen 6c5ccb11f9 Revamp ROCm support 10 months ago
  Daniel Hiltgen 6d84f07505 Detect AMD GPU info via sysfs and block old cards 10 months ago
  mraiser 4c4c730a0a Merge branch 'ollama:main' into main 11 months ago