forked from ruvnet/RuVector
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathCargo.toml
More file actions
109 lines (85 loc) · 2.48 KB
/
Cargo.toml
File metadata and controls
109 lines (85 loc) · 2.48 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
[package]
name = "ruvector-onnx-embeddings"
version = "0.1.0"
edition = "2021"
authors = ["RuVector Team"]
description = "ONNX-based embedding generation for RuVector - Reimagined embedding pipeline in pure Rust"
license = "MIT"
repository = "https://github.com/ruvnet/ruvector"
keywords = ["onnx", "embeddings", "vector-database", "rust", "ml"]
categories = ["science", "algorithms"]
# Make this a standalone package, not part of the workspace
[workspace]
[dependencies]
# ONNX Runtime - Core inference engine
ort = { version = "2.0.0-rc.9", features = ["download-binaries", "half"] }
# Tokenization - HuggingFace tokenizers in Rust
tokenizers = { version = "0.20", default-features = false, features = ["progressbar", "onig"] }
# Tensor operations
ndarray = { version = "0.16", features = ["rayon"] }
half = "2.4"
# Async runtime
tokio = { version = "1.41", features = ["full"] }
# Serialization
serde = { version = "1.0", features = ["derive"] }
serde_json = "1.0"
# Error handling
thiserror = "2.0"
anyhow = "1.0"
# HTTP client for model downloads
reqwest = { version = "0.12", features = ["blocking", "stream"] }
futures-util = "0.3"
# Progress bars and CLI
indicatif = "0.17"
console = "0.15"
# Logging
tracing = "0.1"
tracing-subscriber = { version = "0.3", features = ["env-filter"] }
# File operations
sha2 = "0.10"
hex = "0.4"
tempfile = "3.14"
dirs = "5.0"
# Parallel processing
rayon = "1.10"
# Concurrency
parking_lot = "0.12"
# UUID for vector IDs
uuid = { version = "1.11", features = ["v4"] }
# GPU acceleration (optional)
wgpu = { version = "23.0", optional = true }
bytemuck = { version = "1.14", optional = true, features = ["derive"] }
[dev-dependencies]
criterion = { version = "0.5", features = ["html_reports"] }
approx = "0.5"
[[bench]]
name = "embedding_benchmark"
harness = false
[[bench]]
name = "gpu_benchmark"
harness = false
required-features = ["gpu"]
[[example]]
name = "basic_embedding"
path = "examples/basic.rs"
[[example]]
name = "batch_embedding"
path = "examples/batch.rs"
[[example]]
name = "semantic_search"
path = "examples/semantic_search.rs"
[features]
default = ["download-models"]
download-models = []
cuda = ["ort/cuda"]
tensorrt = ["ort/tensorrt"]
coreml = ["ort/coreml"]
simsimd = [] # Optional SIMD acceleration (not yet implemented)
# GPU acceleration features
gpu = ["dep:wgpu", "dep:bytemuck"]
cuda-wasm = ["gpu"] # CUDA-WASM transpilation (requires gpu)
webgpu = ["gpu"] # WebGPU backend alias
[profile.release]
opt-level = 3
lto = "thin"
codegen-units = 1