Skip to content

Support BF16 kvcache, rope and attentions for inference of GGUF/GGML models #4387

Support BF16 kvcache, rope and attentions for inference of GGUF/GGML models

Support BF16 kvcache, rope and attentions for inference of GGUF/GGML models #4387

Annotations

5 warnings

Docs

succeeded Dec 30, 2024 in 2m 51s