Skip to content

Feature/ocaml coq #8

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 135 commits into
base: master
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
135 commits
Select commit Hold shift + click to select a range
532dd74
Fix some documentation typos/grammar mistakes (#4032)
richardkiss Nov 12, 2023
21fd874
gguf-py: gguf_writer: Use bytearray to build metadata (#4051)
KerfuffleV2 Nov 12, 2023
bb50a79
Add ReLU and SQR CUDA ops to (partially) fix Persimmon offloading (#4…
KerfuffleV2 Nov 13, 2023
4760e7c
sync : ggml (backend v2) (#3912)
ggerganov Nov 13, 2023
c049b37
readme : update hot topics
ggerganov Nov 13, 2023
3d68f36
ggml : sync (im2col, GPU conv, 32-bit arm compat) (#4060)
ggerganov Nov 13, 2023
bd90eca
llava : fix regression for square images in #3613 (#4056)
monatis Nov 13, 2023
b46d12f
convert.py: also look for plain model.safetensors (#4043)
afrideva Nov 14, 2023
36eed0c
stablelm : StableLM support (#3586)
Galunid Nov 14, 2023
6bb4908
Fix MacOS Sonoma model quantization (#4052)
TortoiseHam Nov 14, 2023
1cf2850
ggml-cuda : increase max graph size (#4084)
slaren Nov 15, 2023
a6fc554
llama : restore prefix space in llama tokenizer (#4081)
cebtenzzre Nov 15, 2023
8da4627
gguf : fix potential infinite loops while parsing (#4100)
texmex76 Nov 16, 2023
91f6499
Respect tokenizer.ggml.add_bos_token value when tokenizing (#4040)
KerfuffleV2 Nov 17, 2023
4f447a4
llama : fix data units (#4101)
ggerganov Nov 17, 2023
b83e149
cuda : get_row_rounding F32 (#4095)
AndrewGodfrey Nov 17, 2023
947f64f
finetune : zero the loraB initial vectors (#4082)
AndrewGodfrey Nov 17, 2023
3e916a0
finetune : speed-up ggml_compute_forward_out_prod_f32 via BLAS (#4079)
gwjr Nov 17, 2023
e85bb1a
llama : add functions to get the model's metadata (#4013)
slaren Nov 17, 2023
ba4cf5c
train : move number of gpu layers argument parsing to common/train.cp…
jpodivin Nov 17, 2023
f7d5e97
py : remove superfluous import statements (#4076)
jpodivin Nov 17, 2023
c7cce12
llava : fix compilation warning that fread return value is not used (…
huawei-lin Nov 17, 2023
9e87ef6
common : improve yaml log escaping (#4080)
joennlae Nov 17, 2023
11173c9
py : Falcon HF compatibility (#4104)
cmp-nct Nov 17, 2023
2ab0707
convert : use 'model' value if it exists. This allows karpathy/tinyll…
dmahurin Nov 17, 2023
2fa02b4
examples : add tokenize (#4039)
zakkor Nov 17, 2023
5ad387e
tokenize : fix trailing whitespace
ggerganov Nov 17, 2023
8e93610
build : support ppc64le build for make and CMake (#3963)
bufferoverflow Nov 17, 2023
bbecf3f
llama : increase max nodes (#4115)
slaren Nov 17, 2023
2923f17
Clean up ggml-cuda.cu warnings when compiling with clang (for ROCM) (…
KerfuffleV2 Nov 18, 2023
0b5c3b0
scripts : Remove missed baichuan convert script (#4127)
Galunid Nov 18, 2023
28a2e6e
tokenize example: Respect normal add BOS token behavior (#4126)
KerfuffleV2 Nov 18, 2023
e937066
gguf-py : export chat templates (#4125)
slaren Nov 19, 2023
35985ac
gitignore : tokenize
ggerganov Nov 19, 2023
262005a
common : comma should be semicolon (#4137)
kchro3 Nov 19, 2023
936c79b
server : relay error messages (#4131)
SoftwareRenderer Nov 19, 2023
05e8301
finetune : add --n-gpu-layers flag info to --help (#4128)
csaben Nov 19, 2023
dae06c0
Revert "finetune : add --n-gpu-layers flag info to --help (#4128)"
ggerganov Nov 19, 2023
40a34fe
speculative : fix prompt tokenization in speculative example (#4025)
AutonomicPerfectionist Nov 20, 2023
f23c035
ci : add flake8 to github actions (python linting) (#4129)
Galunid Nov 20, 2023
881800d
main : Add ChatML functionality to main example (#4046)
Sebby37 Nov 20, 2023
dfc7cd4
readme : update ROCm Windows instructions (#4122)
jammm Nov 20, 2023
0b871f1
finetune - update readme to mention llama support only (#4148)
Galunid Nov 20, 2023
8e672ef
stablelm : simplify + speedup generation (#4153)
Galunid Nov 21, 2023
ff8238f
docs : add llama-star arch idea
ggerganov Nov 23, 2023
9d5949f
examples : fix typo in parallel example doc comment (#4181)
danbev Nov 23, 2023
d103d93
readme : update hot topics
ggerganov Nov 23, 2023
6b0a742
llama : KV cache view API + better KV cache management (#4170)
ggerganov Nov 23, 2023
55978ce
Fix incorrect format strings and uninitialized variables. (#4133)
haohui Nov 23, 2023
b35f3d0
readme : use PATH for Windows ROCm (#4195)
jammm Nov 24, 2023
2568a4b
main.swift : fix eos checking (#4197)
eastriverlee Nov 24, 2023
189d684
convert : fix tensors using grad in some models (#4173)
Galunid Nov 24, 2023
8a052c1
ggml-cuda : support stablelm rope (#4156)
slaren Nov 24, 2023
e9c13ff
llama : set metal log callback correctly (#4204)
slaren Nov 24, 2023
af19d35
server : OAI API compatibility (#4198)
ggerganov Nov 25, 2023
04814e7
readme : update hot topics
ggerganov Nov 25, 2023
3014b54
Update docs for yarn_ext_factor <0.0 as unspecified instead of NaN (#…
crasm Nov 25, 2023
f837c3a
llama : grammar `reserve` space in `decode_utf8` (#4210)
MarcusDunn Nov 25, 2023
1ddb52e
scripts : Use mmap in torch load (#4202)
Galunid Nov 25, 2023
22da055
metal : fix yarn (#4220)
jxy Nov 26, 2023
922754a
lookahead : add example for lookahead decoding (#4207)
ggerganov Nov 26, 2023
9656026
readme : update hot topics
ggerganov Nov 26, 2023
3e73d31
lookahead : support `-n -1` infinite generation
ggerganov Nov 26, 2023
f3b2698
ggml : fix -Warray-bounds warning with gcc (#4231)
cebtenzzre Nov 27, 2023
bb03290
examples : iOS example with swift ui (#4159)
bachittle Nov 27, 2023
0dab8cd
readme : add Amica to UI list (#4230)
kasumi-1 Nov 27, 2023
b38a16d
cmake : fix issue with version info not getting baked into LlamaConfi…
bandoti Nov 27, 2023
8406b09
ggml : re-enable BLAS for CPU when src0 != F32 + remove redundant ful…
ggerganov Nov 28, 2023
64e64aa
ggml : restore abort() in GGML_ASSERT (#4242)
cebtenzzre Nov 28, 2023
4fea342
readme : add FreeChat (#4248)
psugihara Nov 29, 2023
1f5cd83
examples : add readme files
ggerganov Nov 29, 2023
e2bd725
py : fix oai proxy (#3972)
rhjdvsgsgks Nov 30, 2023
954e228
llama : fix typical sampling (#4261)
tarcey Nov 30, 2023
f4d973c
convert.py : fix llama/llama2 conversion due to vocab_size=-1 (#4258)
slaren Nov 30, 2023
b18c66c
llama : fix alignment of general.name in print meta (#4254)
danbev Nov 30, 2023
74daaba
readme : fix typo (#4253)
TortillaZHawaii Nov 30, 2023
f7f9e06
cmake : fix the metal file foder path (#4217)
tanliboy Nov 30, 2023
bde629b
batched.swift : update README.md (#4214)
ensan-hcl Nov 30, 2023
3bd2c7c
docker : add finetune option (#4211)
jooray Nov 30, 2023
524907a
readme : fix (#4135)
vodkaslime Nov 30, 2023
8efa0f6
main : pass LOG_TEE callback to llama.cpp log (#4033)
AndrewGodfrey Nov 30, 2023
33c9892
llava : ShareGPT4V compatibility (vision encoder only loading) (#4172)
cmp-nct Nov 30, 2023
15f5d96
build : fix build info generation and cleanup Makefile (#3920)
cebtenzzre Nov 30, 2023
d2809a3
make : fix Apple clang determination bug (#4272)
WillCorticesAI Nov 30, 2023
f43f093
server : add single-client multi-prompt support (#4232)
ziedbha Nov 30, 2023
1d14411
server : add --log-disable to disable logging to file (#4260)
ziedbha Nov 30, 2023
ef47ec1
ggml : add ggml_soft_max_ext (#4256)
ggerganov Dec 1, 2023
8d6d9f0
py : add requirements file for convert-hf-to-gguf.py (#4277)
danbev Dec 1, 2023
880f579
llama : fix integer overflow during quantization (#4284)
ggerganov Dec 1, 2023
37c746d
llama : add Qwen support (#4281)
simonJJJ Dec 1, 2023
03562f3
llama : support attention bias on LLaMA architecture (#4283)
CausalLM Dec 1, 2023
511f52c
build : enable libstdc++ assertions for debug builds (#4275)
cebtenzzre Dec 1, 2023
b220222
swift : fix token_to_piece implementation (#4278)
ensan-hcl Dec 1, 2023
d5a1cbd
llama : support optional tensors (#4283)
ggerganov Dec 1, 2023
5a7d312
llama : avoid using "optional" keyword (#4283)
ggerganov Dec 1, 2023
d7b800b
llama : pad KV cache size (#4280)
ggerganov Dec 3, 2023
6949b50
py : add grammar to oai like api (#4294)
neph1 Dec 3, 2023
33e171d
server : fix OpenAI API `stop` field to be optional (#4299)
Mardak Dec 3, 2023
adf3de4
ggml : fix soft max out-of-bounds access (#4307)
ggerganov Dec 3, 2023
fbbc428
ggml : reuse ggml_get_n_tasks() in ggml_graph_plan() (#4308)
ggerganov Dec 3, 2023
4fa44e8
grammar-parser : fix typo (#4318)
eltociear Dec 4, 2023
5c9f90c
swift : fix prompt tokenization logic (#4321)
ensan-hcl Dec 4, 2023
d208995
swift : fix concatenation method to avoid invalid UTF8 stringfication…
ensan-hcl Dec 4, 2023
23b5e12
simple : update error message for KV cache check (#4324)
danbev Dec 4, 2023
e4b76bb
swift : revert compiler checks for swift package (#4332)
kchro3 Dec 5, 2023
52c8bc3
sampling : custom samplers order (#4285)
MaggotHATE Dec 5, 2023
2b6ff2e
rebased and trimmed down
Nov 21, 2023
5ea96cc
rebased
Dec 5, 2023
2f3ea04
starting boost
Dec 6, 2023
1c86146
working calling python
Dec 6, 2023
7972929
now getting response from python
Dec 6, 2023
7eb27b3
now it is letting the llm control the output
Dec 6, 2023
d6244ff
adding missing files
Dec 6, 2023
ac69c93
linking
Dec 7, 2023
09a48ec
linking, loading, segfaulting
Dec 8, 2023
593985d
update
Dec 8, 2023
9ec7eb1
nodejs
Dec 8, 2023
1193766
for metacall add the cmake
Dec 9, 2023
34cf9d6
not executing
Dec 9, 2023
da1d845
dont forget the code
Dec 9, 2023
1f52231
wip
Dec 9, 2023
1f3a501
working
Dec 9, 2023
da5bbd7
linker error
Dec 9, 2023
d739470
now linking and crashing
Dec 10, 2023
d239bc9
makefile now building and exec crashing
Dec 10, 2023
fa49f64
cmake not working yet
Dec 10, 2023
9592bc5
fixing first bug
Dec 10, 2023
ad7fc54
remove bad cast
Dec 10, 2023
7025832
now working v1
Dec 10, 2023
b2a5e70
now linking locally
Dec 11, 2023
f62bbea
not linking
Dec 11, 2023
f9d918a
now linking
Dec 12, 2023
6d4ca95
now calling parse and failing not crashing
Dec 12, 2023
e30080b
now not stopping
Dec 12, 2023
4676a64
now getting error messages
Dec 12, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions .devops/tools.sh
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,8 @@ elif [[ "$arg1" == '--quantize' || "$arg1" == '-q' ]]; then
./quantize "$@"
elif [[ "$arg1" == '--run' || "$arg1" == '-r' ]]; then
./main "$@"
elif [[ "$arg1" == '--finetune' || "$arg1" == '-f' ]]; then
./finetune "$@"
elif [[ "$arg1" == '--all-in-one' || "$arg1" == '-a' ]]; then
echo "Converting PTH to GGML..."
for i in `ls $1/$2/ggml-model-f16.bin*`; do
Expand All @@ -34,6 +36,8 @@ else
echo " ex: --outtype f16 \"/models/7B/\" "
echo " --quantize (-q): Optimize with quantization process ggml"
echo " ex: \"/models/7B/ggml-model-f16.bin\" \"/models/7B/ggml-model-q4_0.bin\" 2"
echo " --finetune (-f): Run finetune command to create a lora finetune of the model"
echo " See documentation for finetune for command-line parameters"
echo " --all-in-one (-a): Execute --convert & --quantize"
echo " ex: \"/models/\" 7B"
echo " --server (-s): Run a model on the server"
Expand Down
11 changes: 11 additions & 0 deletions .github/workflows/build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -498,6 +498,17 @@ jobs:
path: |
cudart-llama-bin-win-cu${{ matrix.cuda }}-x64.zip

ios-xcode-build:
runs-on: macos-latest

steps:
- name: Checkout code
uses: actions/checkout@v3

- name: Build Xcode project
run: xcodebuild -project examples/llama.swiftui/llama.swiftui.xcodeproj -scheme llama.swiftui -sdk iphoneos CODE_SIGNING_REQUIRED=NO CODE_SIGN_IDENTITY= -destination 'generic/platform=iOS' build


# freeBSD-latest:
# runs-on: macos-12
# steps:
Expand Down
20 changes: 20 additions & 0 deletions .github/workflows/python-lint.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,20 @@
name: flake8 Lint

on: [push, pull_request]

jobs:
flake8-lint:
runs-on: ubuntu-latest
name: Lint
steps:
- name: Check out source repository
uses: actions/checkout@v3
- name: Set up Python environment
uses: actions/setup-python@v4
with:
python-version: "3.11"
- name: flake8 Lint
uses: py-actions/flake8@v2
with:
ignore: "E203,E211,E221,E225,E231,E241,E251,E261,E266,E501,E701,E704"
exclude: "examples/*,examples/*/**,*/**/__init__.py"
58 changes: 58 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -64,6 +64,7 @@ models-mnt
/speculative
/parallel
/train-text-from-scratch
/tokenize
/vdot
/common/build-info.cpp
arm_neon.h
Expand Down Expand Up @@ -98,3 +99,60 @@ tests/test-tokenizer-0-llama
tests/test-tokenizer-0-falcon
tests/test-tokenizer-1-llama
tests/test-tokenizer-1-bpe
/#llama.cpp#
#*
\\#*
\\#
#
*~
.#*
#*
# -*- mode: gitignore; -*-
*~
\#*\#
/.emacs.desktop
/.emacs.desktop.lock
*.elc
auto-save-list
tramp
.\#*

# Org-mode
.org-id-locations
*_archive

# flymake-mode
*_flymake.*

# eshell files
/eshell/history
/eshell/lastdir

# elpa packages
/elpa/

# reftex files
*.rel

# AUCTeX auto folder
/auto/

# cask packages
.cask/
dist/

# Flycheck
flycheck_*.el

# server auth directory
/server/

# projectiles files
.projectile

# directory configuration
.dir-locals.el

# network security
/network-security.data
/data/
Loading