Skip to content
GitLab
Explore
Sign in
Tags
Tags give the ability to mark specific points in history as being important
b1563
3014b541
·
Update docs for yarn_ext_factor <0.0 as unspecified instead of NaN (#4189)
·
Nov 25, 2023
b1561
af19d357
·
server : OAI API compatibility (#4198)
·
Nov 25, 2023
b1560
e9c13ff7
·
llama : set metal log callback correctly (#4204)
·
Nov 24, 2023
b1559
8a052c13
·
ggml-cuda : support stablelm rope (#4156)
·
Nov 24, 2023
b1557
2568a4bf
·
main.swift : fix eos checking (#4197)
·
Nov 24, 2023
b1555
55978ce0
·
Fix incorrect format strings and uninitialized variables. (#4133)
·
Nov 23, 2023
b1554
6b0a7420
·
llama : KV cache view API + better KV cache management (#4170)
·
Nov 23, 2023
b1552
9d5949f0
·
examples : fix typo in parallel example doc comment (#4181)
·
Nov 23, 2023
b1550
8e672efe
·
stablelm : simplify + speedup generation (#4153)
·
Nov 21, 2023
b1547
881800d1
·
main : Add ChatML functionality to main example (#4046)
·
Nov 20, 2023
b1546
f23c0359
·
ci : add flake8 to github actions (python linting) (#4129)
·
Nov 20, 2023
b1545
40a34fe8
·
speculative : fix prompt tokenization in speculative example (#4025)
·
Nov 20, 2023
b1544
dae06c06
·
Revert "finetune : add --n-gpu-layers flag info to --help (#4128)"
·
Nov 19, 2023
b1543
05e8301e
·
finetune : add --n-gpu-layers flag info to --help (#4128)
·
Nov 19, 2023
b1542
936c79b2
·
server : relay error messages (#4131)
·
Nov 19, 2023
b1541
262005ad
·
common : comma should be semicolon (#4137)
·
Nov 19, 2023
b1539
e9370664
·
gguf-py : export chat templates (#4125)
·
Nov 19, 2023
b1538
28a2e6e7
·
tokenize example: Respect normal add BOS token behavior (#4126)
·
Nov 18, 2023
b1536
2923f17f
·
Clean up ggml-cuda.cu warnings when compiling with clang (for ROCM) (#4124)
·
Nov 18, 2023
b1535
bbecf3f4
·
llama : increase max nodes (#4115)
·
Nov 17, 2023
Prev
1
…
44
45
46
47
48
49
50
51
52
…
99
Next