Skip to content

Commit 36052c1

Browse files
committed
Update rust, transformers, optimum, onnxruntime, onnx, loguru, pytest
Rust 1.32.1 transformers 4.33.2 optimum 1.13.2 onnxruntime 1.16.0 onnx 1.14.1 loguru 0.7.2 pytest 7.4.2
1 parent ef953ab commit 36052c1

File tree

7 files changed

+108
-106
lines changed

7 files changed

+108
-106
lines changed

Dockerfile

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ ENV LIBRARY_PATH="$CUDA_HOME/lib64/stubs"
8888

8989
## Rust builder ################################################################
9090
# Specific debian version so that compatible glibc version is used
91-
FROM rust:1.72-bullseye as rust-builder
91+
FROM rust:1.72.1-bullseye as rust-builder
9292
ARG PROTOC_VERSION
9393

9494
ENV CARGO_REGISTRIES_CRATES_IO_PROTOCOL=sparse
@@ -255,7 +255,7 @@ COPY proto proto
255255
COPY server server
256256
RUN cd server && make gen-server && pip install ".[accelerate, onnx-gpu]" --no-cache-dir
257257

258-
# Patch codegen model changes into transformers 4.31
258+
# Patch codegen model changes into transformers 4.33.2
259259
RUN cp server/transformers_patch/modeling_codegen.py \
260260
/opt/miniconda/lib/python3.*/site-packages/transformers/models/codegen/modeling_codegen.py
261261

integration_tests/poetry.lock

Lines changed: 4 additions & 4 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

integration_tests/pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ python = "^3.9"
1010
[tool.poetry.group.dev.dependencies]
1111
protobuf = "^4.24.3"
1212
grpcio-tools = "^1.58.0"
13-
pytest = "^7.4.0"
13+
pytest = "^7.4.2"
1414
pytest-asyncio = "^0.21.1"
1515
requests = "^2.31.0"
1616
pyyaml = "^6.0.1"

rust-toolchain.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
[toolchain]
2-
channel = "1.72.0"
2+
channel = "1.72.1"
33
components = ["rustfmt", "clippy"]

server/poetry.lock

Lines changed: 91 additions & 90 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

server/pyproject.toml

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -12,18 +12,18 @@ python = ">=3.9.0,<3.13"
1212
protobuf = "^4.24.3"
1313
grpcio = "^1.58.0"
1414
grpcio-reflection = "^1.58.0"
15-
loguru = "^0.7.0"
15+
loguru = "^0.7.2"
1616
typer = "^0.9.0"
1717
accelerate = { version = "0.23.0", optional = true }
18-
bitsandbytes = { version = "^0.41.0", optional = true }
18+
bitsandbytes = { version = "^0.41.1", optional = true }
1919
scipy = { version = "^1.11.2", optional = true }
2020
safetensors = "^0.3.3"
2121
sentencepiece = "^0.1.99"
22-
transformers = "4.33.1"
23-
optimum = { version = "1.11.0", extras = ["onnxruntime-gpu"], optional = true }
24-
onnxruntime = { version = "1.15.1", optional = true }
25-
onnxruntime-gpu = { version = "1.15.1", optional = true }
26-
onnx = { version = "1.14.0", optional = true }
22+
transformers = "4.33.2"
23+
optimum = { version = "1.13.2", extras = ["onnxruntime-gpu"], optional = true }
24+
onnxruntime = { version = "1.16.0", optional = true }
25+
onnxruntime-gpu = { version = "1.16.0", optional = true }
26+
onnx = { version = "1.14.1", optional = true }
2727
einops = "^0.7.0rc2"
2828

2929
# Explicitly install some transitive dependencies to avoid CVEs
@@ -40,7 +40,7 @@ onnx-gpu = ["optimum", "onnxruntime-gpu", "onnx"]
4040

4141
[tool.poetry.group.dev.dependencies]
4242
grpcio-tools = "^1.58.0"
43-
pytest = "^7.4.0"
43+
pytest = "^7.4.2"
4444

4545
[build-system]
4646
requires = ["poetry-core>=1.0.0"]

server/transformers_patch/modeling_codegen.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -463,6 +463,7 @@ def forward(
463463
if input_ids is not None and inputs_embeds is not None:
464464
raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time")
465465
elif input_ids is not None:
466+
self.warn_if_padding_and_no_attention_mask(input_ids, attention_mask)
466467
input_shape = input_ids.size()
467468
input_ids = input_ids.view(-1, input_shape[-1])
468469
batch_size = input_ids.shape[0]

0 commit comments

Comments
 (0)