-
Notifications
You must be signed in to change notification settings - Fork 29
/
Cargo.toml
61 lines (53 loc) · 1.25 KB
/
Cargo.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
[package]
name = "rllama"
version = "0.3.0"
edition = "2021"
authors = ["Mikko Juola"]
description = "Pure Rust implementation of LLaMA-family of models, executable"
documentation = "https://github.com/Noeda/rllama"
homepage = "https://github.com/Noeda/rllama"
repository = "https://github.com/Noeda/rllama"
license = "AGPL-3.0"
keywords = ["llama", "machine-learning"]
categories = ["command-line-utilities"]
[lib]
path = "src/lib.rs"
[[bin]]
name = "rllama"
path = "src/main.rs"
[dependencies]
protobuf = "3.2"
thiserror = "1.0"
half = "2.2"
num-complex = "0.4"
embedded-profiling = "0.3"
rand = "0.8"
approx = "0.5"
rayon = "1.7"
clap = { version = "4.1", features = ["derive"] }
indicatif = "0.17"
colored = "2"
serde = { version = "1", features = ["derive"] }
serde_json = "1"
mimalloc = "0.1"
ocl = { version = "0.19", optional = true }
rocket = { version = "0.4", features = ["sse"], optional = true }
lazy_static = "1.4"
zip = "0.6"
ouroboros = "0.15"
[features]
opencl = ["ocl"]
server = ["rocket"]
# We need protobuf compiler
[build-dependencies]
protobuf-codegen = "3.2"
protobuf-parse = "3.2"
[dev-dependencies]
criterion = "0.4"
[profile.release]
panic = 'abort'
debug = true
[[bench]]
path = "src/benches/benchmark.rs"
name = "benchmark"
harness = false