-
Notifications
You must be signed in to change notification settings - Fork 6.5k
/
Copy pathpyproject.toml
52 lines (43 loc) · 1002 Bytes
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
[build-system]
requires = ["setuptools>=61.0", "wheel"]
build-backend = "setuptools.build_meta"
[project]
name = "llms-from-scratch"
version = "1.0.6"
description = "Implement a ChatGPT-like LLM in PyTorch from scratch, step by step"
readme = "README.md"
requires-python = ">=3.10"
dependencies = [
"torch>=2.3.0",
"jupyterlab>=4.0",
"tiktoken>=0.5.1",
"matplotlib>=3.7.1",
"tensorflow>=2.18.0",
"tqdm>=4.66.1",
"numpy>=1.26,<2.1",
"pandas>=2.2.1",
"pip>=25.0.1",
"pytest>=8.3.5",
]
[tool.uv.sources]
llms-from-scratch = { workspace = true }
[dependency-groups]
dev = [
"build>=1.2.2.post1",
"llms-from-scratch",
"twine>=6.1.0",
]
[tool.ruff]
line-length = 140
[tool.ruff.lint]
exclude = [".venv"]
# Ignored rules (W504 removed)
ignore = [
"C406", "E226", "E402", "E702", "E703",
"E722", "E731", "E741"
]
# `llms_from_scratch` PyPI package
[tool.setuptools]
package-dir = {"" = "pkg"}
[tool.setuptools.packages.find]
where = ["pkg"]