forked from instructlab/instructlab
-
Notifications
You must be signed in to change notification settings - Fork 0
/
requirements.txt
47 lines (46 loc) · 1.46 KB
/
requirements.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
# SPDX-License-Identifier: Apache-2.0
click>=8.1.7,<9.0.0
click-didyoumean>=0.3.0
datasets>=2.18.0
gguf>=0.6.0
GitPython>=3.1.42
# Linux: 4-bit quantization with BitsAndBytes is not ready to use, yet.
# see https://github.com/instructlab/instructlab/issues/579
# bitsandbytes; sys_platform=='linux' and platform_machine=='x86_64'
httpx>=0.25.0
instructlab-eval>=0.0.4
instructlab-quantize>=0.1.0
instructlab-schema>=0.2.0
instructlab-sdg>=0.0.4.1
instructlab-training>=0.0.3
jsonschema>=4.21.1
llama_cpp_python[server]==0.2.79
mlx>=0.5.1,<0.6.0; sys_platform == 'darwin' and platform_machine == 'arm64'
# HabanaLabs / Intel Gaudi env comes with Python 3.10 and slightly older
# versions of some dependencies. Use '3.10' as an indicator.
# Habana installer has NumPy 1.23.5
numpy>=1.23.5,<2.0.0 ; python_version == '3.10'
numpy>=1.26.4,<2.0.0 ; python_version != '3.10'
openai>=1.13.3
peft>=0.9.0
prompt-toolkit>=3.0.38
pydantic>=2.7.4
pydantic_yaml>=1.2.0
PyYAML>=6.0.0
rich>=13.3.1
rouge-score>=0.1.2
sentencepiece>=0.2.0
# "old" version required for vLLM on CUDA to build
tokenizers>=0.11.1
toml>=0.10.2
# Intel Gaudi 1.16.1 has torch 2.2.0a0+git8964477 with oneMKL
torch>=2.3.0a0,<3.0.0 ; python_version == '3.10'
torch>=2.3.0,<3.0.0 ; python_version != '3.10'
tqdm>=4.66.2
# 'optimum' for Intel Gaudi needs transformers <4.39.0,>=4.38.0
transformers>=4.41.2
trl>=0.9.4
wandb>=0.16.4
# the below library should NOT be imported into any python files
# it is for CLI usage ONLY
yamllint>=1.35.1