ggml : upgrade init_tensor API to return a ggml_status (llama/11854) d6b6852 William Tambellini slaren commited on Feb 28, 2025
vulkan: use smaller combined allocations to avoid fragmentation (llama/11551) 1b7672d jeffbolznv commited on Feb 6, 2025
CUDA: backwards pass for misc. ops, add tests (llama/11257) 2fbcec1 JohannesGaessler commited on Jan 16, 2025
ggml : remove return from ggml_gallocr_allocate_node (ggml/1048) f9d4408 danbev commited on Dec 14, 2024
ggml : move more prints to the ggml log system (llama/9839) 98d1a6a Diego Devesa commited on Oct 11, 2024
ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (llama/9573) 673df39 slaren commited on Sep 21, 2024
CUDA: fix partial offloading for ne0 % 256 != 0 (llama/8572) afc137c JohannesGaessler commited on Jul 18, 2024
whisper : reorganize source code + improve CMake (#2256) f75c2e3 unverified ggerganov HF Staff commited on Jun 26, 2024