MUSA: support ARM64 and enable dp4a .etc (llama/11843) ab96dac Bodhi Bodhi Hu commited on Feb 21, 2025
ggml-cpu: Add CPU backend support for KleidiAI library (llama/11390) 9de6d81 Charles Xu commited on Feb 20, 2025
ggml: aarch64: implement SVE kernels for q3_K_q8_K vector dot (llama/11917) 1a1acd2 Prashant Vithule vithulep ggerganov HF Staff commited on Feb 20, 2025
CUDA: use async data loading for FlashAttention (llama/11894) 5b9980d JohannesGaessler Diego Devesa commited on Feb 17, 2025
vulkan: implement several ops relevant for ggml_opt (llama/11769) 3c2171d Rémy O commited on Feb 17, 2025
vulkan: support multi/vision rope, and noncontiguous rope (llama/11902) 1c7a669 jeffbolznv commited on Feb 16, 2025
metal : fix the crash caused by the lack of residency set support on Intel Macs. (llama/11904) afbd891 Hale Chan commited on Feb 16, 2025
vulkan: initial support for IQ1_S and IQ1_M quantizations (llama/11528) 0d2e888 Rémy O commited on Feb 15, 2025
cuda : add ampere to the list of default architectures (llama/11870) 1d19dec Diego Devesa commited on Feb 14, 2025
ggml: optimize some vec dot functions for LoongArch ASX (llama/11842) e3acbfc Jinyang He commited on Feb 14, 2025
llamafile: use member variable instead of constant for iq4nlt (llama/11780) 0cb2d04 jmorganca commited on Feb 13, 2025
ggml-cpu : add chunking support to mul_mat_id (llama/11666) e59d9a7 Diego Devesa commited on Feb 13, 2025
ggml : x2 speed for WASM by optimizing SIMD (llama/11453) 464a186 Xuan-Son Nguyen camel-cdr commited on Feb 12, 2025
HIP: Remove GCN from list of devices that avoid MMQ (llama/11831) 78aed55 uvos commited on Feb 12, 2025
HIP: Switch to std::vector in rocblas version check (llama/11820) e144c94 uvos commited on Feb 12, 2025
cleanup: fix compile warnings associated with gnu_printf (llama/11811) ef6a968 bandoti commited on Feb 12, 2025
Fix #11802: Compile bug - RegQueryValueExA changed to RegQueryValueEx (llama/11803) 86969ac Sheldon Robinson commited on Feb 11, 2025
CUDA: use arch list for compatibility check (llama/11775) b88e163 JohannesGaessler Diego Devesa commited on Feb 10, 2025
vulkan: Make Vulkan optional at runtime (ggml/11493). (llama/11494) 762f497 Danny Milosavljevic jeffbolznv commited on Feb 10, 2025
vulkan: add environment variable GGML_VK_PREFER_HOST_MEMORY to avoid VRAM allocation (llama/11592) f9fd130 Wagner Bruna commited on Feb 10, 2025
vulkan: account for lookup tables when checking shared memory size (llama/11502) 758970f jeffbolznv commited on Feb 9, 2025
ggml : optimize and build warning fix for LoongArch (llama/11709) b82d241 Jinyang He commited on Feb 7, 2025
SYCL: Adjust support condition for norm operators (llama/11674) 7e1dbe9 qnixsynapse commited on Feb 6, 2025
ggml : fix LoongArch compile error with 128-bit SIMD (llama/11701) f7296aa junchao-zhao commited on Feb 6, 2025
vulkan: use smaller combined allocations to avoid fragmentation (llama/11551) 1b7672d jeffbolznv commited on Feb 6, 2025
metal : avoid breaking build when metal API predates TARGET_OS_VISION (llama/11690) 5bdb244 charles-dyfis-net commited on Feb 6, 2025
metal : adjust support conditions for norm operators (llama/11671) 5eb35ab ggerganov HF Staff commited on Feb 5, 2025
CUDA: support for mat. mul. with ne03 != ne13 (llama/11656) 78e36a2 JohannesGaessler commited on Feb 5, 2025
CUDA: non-contiguous (RMS) norm support (llama/11659) 4c2e171 JohannesGaessler ggerganov HF Staff commited on Feb 4, 2025
whisper : restore big endian support (#2816) 7631e20 unverified Thomas Fitzsimmons Xuan-Son Nguyen commited on Feb 25, 2025