forked from SciPhi-AI/synthesizer
-
Notifications
You must be signed in to change notification settings - Fork 0
/
pyproject.toml
78 lines (69 loc) · 2.32 KB
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
[build-system]
requires = ["poetry-core", "setuptools", "wheel"]
build-backend = "poetry.core.masonry.api"
[tool.poetry]
name = "SciPhi"
version = "0.1.0"
description = "SciPhi: A framework for breaking scaling laws."
authors = ["Owen Colegrove <[email protected]>"]
license = "Apache-2.0"
readme = "README.md"
[tool.poetry.dependencies]
# Required dependencies
numpy = "^1.26.0"
python = ">=3.10,<3.12"
python-dotenv = "^1.0.0"
pandas = "^2.1.0"
# Begin optional dependencies
# anthropic
anthropic = { version = "^0.3.10", optional = true }
# hf
accelerate = { version = "^0.23.0", optional = true }
datasets = { version = "^2.14.5", optional = true }
torch = { version = "^2.0.1", optional = true }
transformers = { version = "^4.33.1", optional = true }
# openai
openai = { version = "0.27.8", optional = true }
matplotlib = {version = "^3.8.0", optional = true}
plotly = {version = "^5.17.0", optional = true}
scipy = {version = "^1.11.2", optional = true}
scikit-learn = {version = "^1.3.1", optional = true}
# vllm
vllm = { version = "0.1.7", optional = true }
# llama-index
llama-index = { version = "^0.8.29.post1", optional = true }
# chroma
chromadb = { version = "^0.4.12", optional = true }
retrying = "^1.3.4"
[tool.poetry.extras]
anthropic_support = ["anthropic"]
hf_support = ["accelerate", "datasets", "torch", "transformers"]
openai_support = ["openai"]
openai_utils_support = ["openai"]
vllm_support = ["accelerate", "torch", "vllm"]
llama_index_support = ["llama-index"]
chroma_support = ["chromadb"]
# omit "accelerate" and "vllm", since it can cause failures w/out cuda
all = ["anthropic", "datasets", "torch", "transformers", "openai", "matplotlib", "plotly", "scipy", "scikit-learn", "llama-index", "chromadb"]
all_with_cuda = ["anthropic", "accelerate" , "datasets", "torch", "transformers", "openai", "matplotlib", "plotly", "scipy", "scikit-learn", "llama-index", "chromadb", "vllm"]
[tool.poetry.group.dev.dependencies]
black = "^23.3.0"
flake8 = "6.1.0"
isort = "5.12.0"
pre-commit = "^3.3.3"
mypy = "^1.5.1"
sourcery = "^1.6.0"
types-requests = "^2.31.0.2"
types-attrs = "^19.1.0"
yapf = "0.40.1"
[tool.black]
line-length = 79
[tool.mypy]
ignore_missing_imports = true
exclude = 'playground/.*'
[[tool.mypy.overrides]]
module = "yaml"
ignore_missing_imports = true
[tool.flake8]
ignore = ["E501", "W503"]
exclude = 'playground/.*'