File tree Expand file tree Collapse file tree 19 files changed +58
-58
lines changed
Expand file tree Collapse file tree 19 files changed +58
-58
lines changed Original file line number Diff line number Diff line change @@ -22,13 +22,13 @@ python-docx==1.1.2
2222pyyaml
2323requests
2424rich
25- safetensors == 0.6 .*
25+ safetensors == 0.7 .*
2626scipy
2727sentencepiece
2828tensorboard
2929torchao == 0.14.*
3030transformers == 4.57.*
31- triton-windows == 3.5.1.post21 ; platform_system == "Windows"
31+ triton-windows == 3.5.1.post22 ; platform_system == "Windows"
3232tqdm
3333wandb
3434
@@ -42,10 +42,10 @@ sse-starlette==1.6.5
4242tiktoken
4343
4444# CUDA wheels
45- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
46- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
47- https://github.com/turboderp-org/exllamav3/releases/download/v0.0.17 /exllamav3-0.0.17 +cu128.torch2.7.0-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
48- https://github.com/turboderp-org/exllamav3/releases/download/v0.0.17 /exllamav3-0.0.17 +cu128.torch2.7.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
45+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
46+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
47+ https://github.com/turboderp-org/exllamav3/releases/download/v0.0.18 /exllamav3-0.0.18 +cu128.torch2.7.0-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
48+ https://github.com/turboderp-org/exllamav3/releases/download/v0.0.18 /exllamav3-0.0.18 +cu128.torch2.7.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
4949https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2+cu128.torch2.7.0-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
5050https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2+cu128.torch2.7.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
5151https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2-py3-none-any.whl; platform_system == "Linux" and platform_machine != "x86_64"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
@@ -40,7 +40,7 @@ sse-starlette==1.6.5
4040tiktoken
4141
4242# AMD wheels
43- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+vulkan-py3-none-win_amd64.whl; platform_system == "Windows"
44- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+rocm6.4.4-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
43+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+vulkan-py3-none-win_amd64.whl; platform_system == "Windows"
44+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+rocm6.4.4-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
4545https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2+rocm6.2.4.torch2.6.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
4646https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2-py3-none-any.whl; platform_system != "Darwin" and platform_machine != "x86_64"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
@@ -40,7 +40,7 @@ sse-starlette==1.6.5
4040tiktoken
4141
4242# AMD wheels
43- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+vulkanavx-py3-none-win_amd64.whl; platform_system == "Windows"
44- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+vulkanavx-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
43+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+vulkanavx-py3-none-win_amd64.whl; platform_system == "Windows"
44+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+vulkanavx-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
4545https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2+rocm6.2.4.torch2.6.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
4646https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2-py3-none-any.whl; platform_system != "Darwin" and platform_machine != "x86_64"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
@@ -40,5 +40,5 @@ sse-starlette==1.6.5
4040tiktoken
4141
4242# Mac wheels
43- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0-py3-none-macosx_15_0_x86_64.whl; platform_system == "Darwin" and platform_release >= "24.0.0"
44- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0-py3-none-macosx_14_0_x86_64.whl; platform_system == "Darwin" and platform_release >= "23.0.0" and platform_release < "24.0.0"
43+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0-py3-none-macosx_15_0_x86_64.whl; platform_system == "Darwin" and platform_release >= "24.0.0"
44+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0-py3-none-macosx_14_0_x86_64.whl; platform_system == "Darwin" and platform_release >= "23.0.0" and platform_release < "24.0.0"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
@@ -40,5 +40,5 @@ sse-starlette==1.6.5
4040tiktoken
4141
4242# Mac wheels
43- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0-py3-none-macosx_15_0_arm64.whl; platform_system == "Darwin" and platform_release >= "24.0.0"
44- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0-py3-none-macosx_14_0_arm64.whl; platform_system == "Darwin" and platform_release >= "23.0.0" and platform_release < "24.0.0"
43+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0-py3-none-macosx_15_0_arm64.whl; platform_system == "Darwin" and platform_release >= "24.0.0"
44+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0-py3-none-macosx_14_0_arm64.whl; platform_system == "Darwin" and platform_release >= "23.0.0" and platform_release < "24.0.0"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
@@ -40,5 +40,5 @@ sse-starlette==1.6.5
4040tiktoken
4141
4242# llama.cpp (CPU only, AVX2)
43- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cpuavx2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
44- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cpuavx2-py3-none-win_amd64.whl; platform_system == "Windows"
43+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cpuavx2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
44+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cpuavx2-py3-none-win_amd64.whl; platform_system == "Windows"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
@@ -40,5 +40,5 @@ sse-starlette==1.6.5
4040tiktoken
4141
4242# llama.cpp (CPU only, no AVX2)
43- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cpuavx-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
44- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cpuavx-py3-none-win_amd64.whl; platform_system == "Windows"
43+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cpuavx-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
44+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cpuavx-py3-none-win_amd64.whl; platform_system == "Windows"
Original file line number Diff line number Diff line change @@ -22,13 +22,13 @@ python-docx==1.1.2
2222pyyaml
2323requests
2424rich
25- safetensors==0.6 .*
25+ safetensors==0.7 .*
2626scipy
2727sentencepiece
2828tensorboard
2929torchao==0.14.*
3030transformers==4.57.*
31- triton-windows==3.5.1.post21 ; platform_system == "Windows"
31+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3232tqdm
3333wandb
3434
@@ -42,10 +42,10 @@ sse-starlette==1.6.5
4242tiktoken
4343
4444# CUDA wheels
45- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cu124avx-py3-none-win_amd64.whl; platform_system == "Windows"
46- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cu124avx-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
47- https://github.com/turboderp-org/exllamav3/releases/download/v0.0.17 /exllamav3-0.0.17 +cu128.torch2.7.0-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
48- https://github.com/turboderp-org/exllamav3/releases/download/v0.0.17 /exllamav3-0.0.17 +cu128.torch2.7.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
45+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cu124avx-py3-none-win_amd64.whl; platform_system == "Windows"
46+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cu124avx-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
47+ https://github.com/turboderp-org/exllamav3/releases/download/v0.0.18 /exllamav3-0.0.18 +cu128.torch2.7.0-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
48+ https://github.com/turboderp-org/exllamav3/releases/download/v0.0.18 /exllamav3-0.0.18 +cu128.torch2.7.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
4949https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2+cu128.torch2.7.0-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
5050https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2+cu128.torch2.7.0-cp311-cp311-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.11"
5151https://github.com/turboderp-org/exllamav2/releases/download/v0.3.2/exllamav2-0.3.2-py3-none-any.whl; platform_system == "Linux" and platform_machine != "x86_64"
Original file line number Diff line number Diff line change @@ -20,13 +20,13 @@ python-docx==1.1.2
2020pyyaml
2121requests
2222rich
23- safetensors==0.6 .*
23+ safetensors==0.7 .*
2424scipy
2525sentencepiece
2626tensorboard
2727torchao==0.14.*
2828transformers==4.57.*
29- triton-windows==3.5.1.post21 ; platform_system == "Windows"
29+ triton-windows==3.5.1.post22 ; platform_system == "Windows"
3030tqdm
3131wandb
3232
Original file line number Diff line number Diff line change @@ -23,5 +23,5 @@ sse-starlette==1.6.5
2323tiktoken
2424
2525# CUDA wheels
26- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
27- https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.66 .0/llama_cpp_binaries-0.66 .0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
26+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
27+ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.70 .0/llama_cpp_binaries-0.70 .0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
You can’t perform that action at this time.
0 commit comments