Skip to content

Commit 411699b

Browse files
committed
Bump version: 1.1.45 → 1.1.46
1 parent e420a07 commit 411699b

File tree

9 files changed

+21
-21
lines changed

9 files changed

+21
-21
lines changed

.bumpversion.cfg

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
[bumpversion]
2-
current_version = 1.1.45
2+
current_version = 1.1.46
33
commit = True
44
tag = True
55

.pre-commit-config.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,11 +23,11 @@ repos:
2323
- id: end-of-file-fixer
2424
- id: no-commit-to-branch
2525
- repo: https://github.com/psf/black
26-
rev: 26.3.0
26+
rev: 26.3.1
2727
hooks:
2828
- id: black
2929
- repo: https://github.com/astral-sh/ruff-pre-commit
30-
rev: v0.15.5
30+
rev: v0.15.6
3131
hooks:
3232
- id: ruff
3333
types_or: [ python, pyi, jupyter ]

Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,6 @@ ENV HOST=${HOST} \
5454
RUN apt-get update \
5555
&& apt-get install -y ripgrep tree fd-find curl nano libpq-dev \
5656
&& curl -LsSf https://astral.sh/uv/install.sh | sh \
57-
&& uv pip install --system --upgrade --verbose --no-cache --break-system-packages --prerelease=allow vector-mcp[postgres,chromadb,couchbase,qdrant,mongodb,huggingface,agent]>=1.1.45
57+
&& uv pip install --system --upgrade --verbose --no-cache --break-system-packages --prerelease=allow vector-mcp[postgres,chromadb,couchbase,qdrant,mongodb,huggingface,agent]>=1.1.46
5858

5959
CMD ["vector-mcp"]

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
![PyPI - Wheel](https://img.shields.io/pypi/wheel/vector-mcp)
2222
![PyPI - Implementation](https://img.shields.io/pypi/implementation/vector-mcp)
2323

24-
*Version: 1.1.45*
24+
*Version: 1.1.46*
2525

2626
## Overview
2727

compose.yml

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ services:
2929
- "PROVIDER=openai"
3030
- "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
3131
- "LLM_API_KEY=${LLM_API_KEY:-llama}"
32-
- "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
32+
- "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
3333
- "DEBUG=False"
3434
- "ENABLE_WEB_UI=True"
3535
- "ENABLE_OTEL=True"
@@ -94,7 +94,7 @@ services:
9494
- "PROVIDER=openai"
9595
- "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
9696
- "LLM_API_KEY=${LLM_API_KEY:-llama}"
97-
- "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
97+
- "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
9898
- "DEBUG=False"
9999
- "ENABLE_WEB_UI=True"
100100
- "ENABLE_OTEL=True"
@@ -145,7 +145,7 @@ services:
145145
# - "PROVIDER=openai"
146146
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
147147
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
148-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
148+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
149149
# - "DEBUG=False"
150150
# - "ENABLE_WEB_UI=True"
151151
# - "ENABLE_OTEL=True"
@@ -195,7 +195,7 @@ services:
195195
# - "PROVIDER=openai"
196196
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
197197
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
198-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
198+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
199199
# - "DEBUG=False"
200200
# - "ENABLE_WEB_UI=True"
201201
# - "ENABLE_OTEL=True"
@@ -243,7 +243,7 @@ services:
243243
# - "PROVIDER=openai"
244244
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
245245
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
246-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
246+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
247247
# - "DEBUG=False"
248248
# - "ENABLE_WEB_UI=True"
249249
# - "ENABLE_OTEL=True"

mcp.compose.yml

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ services:
7373
# - "PROVIDER=openai"
7474
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
7575
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
76-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
76+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
7777
# - "DEBUG=False"
7878
# - "ENABLE_WEB_UI=True"
7979
# - "ENABLE_OTEL=True"
@@ -186,7 +186,7 @@ services:
186186
- "PROVIDER=openai"
187187
- "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
188188
- "LLM_API_KEY=${LLM_API_KEY:-llama}"
189-
- "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
189+
- "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
190190
- "DEBUG=False"
191191
- "ENABLE_WEB_UI=True"
192192
- "ENABLE_OTEL=True"
@@ -281,7 +281,7 @@ services:
281281
# - "PROVIDER=openai"
282282
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
283283
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
284-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
284+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
285285
# - "DEBUG=False"
286286
# - "ENABLE_WEB_UI=True"
287287
# - "ENABLE_OTEL=True"
@@ -375,7 +375,7 @@ services:
375375
# - "PROVIDER=openai"
376376
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
377377
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
378-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
378+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
379379
# - "DEBUG=False"
380380
# - "ENABLE_WEB_UI=True"
381381
# - "ENABLE_OTEL=True"
@@ -464,7 +464,7 @@ services:
464464
# - "PROVIDER=openai"
465465
# - "LLM_BASE_URL=${LLM_BASE_URL:-http://host.docker.internal:1234/v1}"
466466
# - "LLM_API_KEY=${LLM_API_KEY:-llama}"
467-
# - "MODEL_ID=${MODEL_ID:-qwen/qwen3.5-35b-a3b}"
467+
# - "MODEL_ID=${MODEL_ID:-nvidia/nemotron-3-super}"
468468
# - "DEBUG=False"
469469
# - "ENABLE_WEB_UI=True"
470470
# - "ENABLE_OTEL=True"

pyproject.toml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,15 +4,15 @@ build-backend = "setuptools.build_meta"
44

55
[project]
66
name = "vector-mcp"
7-
version = "1.1.45"
7+
version = "1.1.46"
88
description = "Integrate RAG into AI Agents via MCP Server. Supports multiple Vector database technologies."
99
readme = "README.md"
1010
authors = [{ name = "Audel Rouhi", email = "knucklessg1@gmail.com" }]
1111
license = { text = "MIT" }
1212
classifiers = [ "Development Status :: 5 - Production/Stable", "License :: Public Domain", "Environment :: Console", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3"]
1313
requires-python = ">=3.10"
1414
dependencies = [
15-
"agent-utilities[embeddings-openai,mcp]>=0.2.26",
15+
"agent-utilities[embeddings-openai,mcp]>=0.2.27",
1616
"chromadb>=1.5.1",
1717
"markdownify>=1.2.2",
1818
"beautifulsoup4>=4.14.3",
@@ -50,9 +50,9 @@ huggingface = [
5050
"llama-index-embeddings-huggingface>=0.6.1",
5151
"sentence_transformers>=5.2.2"]
5252
agent = [
53-
"agent-utilities[agent,logfire]>=0.2.26"]
53+
"agent-utilities[agent,logfire]>=0.2.27"]
5454
all = [
55-
"vector-mcp[postgres,chromadb,couchbase,qdrant,mongodb,huggingface,agent]>=1.1.45"
55+
"vector-mcp[postgres,chromadb,couchbase,qdrant,mongodb,huggingface,agent]>=1.1.46"
5656
]
5757

5858
[project.scripts]

vector_mcp/agent_server.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
load_identity,
1313
)
1414

15-
__version__ = "1.1.45"
15+
__version__ = "1.1.46"
1616

1717
logging.basicConfig(
1818
level=logging.INFO,

vector_mcp/mcp_server.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
)
2222
from vector_mcp.retriever.retriever import RAGRetriever
2323

24-
__version__ = "1.1.45"
24+
__version__ = "1.1.46"
2525

2626
logging.basicConfig(
2727
level=logging.DEBUG, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s"

0 commit comments

Comments
 (0)