-
Notifications
You must be signed in to change notification settings - Fork 0
/
pyproject.toml
58 lines (52 loc) · 1.73 KB
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
[tool.poetry]
name = "ga-prompt-llm"
version = "0.1.0"
description = "An small package that allow you to interact with an LLM using Genetic Algorithms to improve the quality of your prompt or reverse engineering (useful if you want to detect security lacks on your own LLM)"
authors = ["Team 404 <info@dreamlearning.ai>"]
readme = "README.md"
[tool.poetry.dependencies]
python = "^3.11"
torch = [
{version = "^2.1.2+cu118", source = "pytorch-gpu-src", markers = "sys_platform == 'linux'"},
{version = "^2.1.2", source = "PyPI", markers = "sys_platform == 'darwin'"}
]
torchvision = [
{version = "^0.16.2+cu118", source = "pytorch-gpu-src", markers = "sys_platform == 'linux'"},
{version = "^0.16.2", source = "PyPI", markers = "sys_platform == 'darwin'"}
]
torchaudio = [
{version = "^2.1.2+cu118", source = "pytorch-gpu-src", markers = "sys_platform == 'linux'"},
{version = "^2.1.2", source = "PyPI", markers = "sys_platform == 'darwin'"}
]
transformers = "^4.36.2"
datasets = "^2.15.0"
pytest-xprocess = "^0.23.0"
bitsandbytes = "^0.41.3.post2"
scipy = "^1.11.4"
accelerate = "^0.25.0"
sentence-transformers = "^2.2.2"
qqdm = "^0.0.7"
tokenizers = "^0.15.0"
einops = "^0.7.0"
numpy = "^1.26.2"
wonderwords = "^2.2.0"
vadersentiment = "^3.3.2"
starlette = "^0.34.0"
requests = "^2.31.0"
pytest = "^7.4.3"
pyhypercycle-aim = {git = "https://github.com/hypercycle-development/pyhypercycle-aim.git"}
spacy = "^3.7.2"
openai = "^1.6.0"
python-dotenv = "^1.0.0"
nltk = "^3.8.1"
[[tool.poetry.source]]
name = "pytorch-gpu-src"
url = "https://download.pytorch.org/whl/cu118"
priority = "explicit"
[build-system]
requires = ["poetry-core"]
build-backend = "poetry.core.masonry.api"
[tool.pytest.ini_options]
pythonpath = [
"src/ga_prompt_llm"
]