master-81040f1
81040f10 · llama : do not allocate KV cache for "vocab_only == true" (#682) · Apr 02, 2023