|
[tool.poetry] |
|
name = "llama-cpp-scripts" |
|
version = "0.0.0" |
|
description = "Scripts that ship with llama.cpp" |
|
authors = ["GGML <ggml@ggml.ai>"] |
|
readme = "README.md" |
|
homepage = "https://ggml.ai" |
|
repository = "https://github.com/ggerganov/llama.cpp" |
|
keywords = ["ggml", "gguf", "llama.cpp"] |
|
packages = [{ include = "*.py", from = "." }] |
|
classifiers = [ |
|
"Programming Language :: Python :: 3", |
|
"License :: OSI Approved :: MIT License", |
|
"Operating System :: OS Independent", |
|
] |
|
|
|
[tool.poetry.dependencies] |
|
python = ">=3.9" |
|
numpy = "^1.25.0" |
|
sentencepiece = ">=0.1.98,<=0.2.0" |
|
transformers = ">=4.35.2,<5.0.0" |
|
protobuf = ">=4.21.0,<5.0.0" |
|
gguf = { path = "./gguf-py" } |
|
torch = { version = "^2.2.0", source = "pytorch" } |
|
|
|
[tool.poetry.dev-dependencies] |
|
pytest = "^5.2" |
|
|
|
|
|
|
|
|
|
[[tool.poetry.source]] |
|
name = "pytorch" |
|
url = "https://download.pytorch.org/whl/cpu" |
|
priority = "explicit" |
|
|
|
[build-system] |
|
requires = ["poetry-core>=1.0.0"] |
|
build-backend = "poetry.core.masonry.api" |
|
|
|
[tool.poetry.scripts] |
|
llama-convert-hf-to-gguf = "convert_hf_to_gguf:main" |
|
llama-convert-llama-ggml-to-gguf = "convert_llama_ggml_to_gguf:main" |
|
llama-ggml-vk-generate-shaders = "ggml_vk_generate_shaders:main" |
|
|