Skip to content

Commit 9e8c475

Browse files
authored
preliminary embeddinggemma support
1 parent 61d901b commit 9e8c475

File tree

3 files changed

+231
-259
lines changed

3 files changed

+231
-259
lines changed

src/constants.py

Lines changed: 19 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -324,7 +324,7 @@
324324
"timm==1.0.19",
325325
"tokenizers==0.22.0",
326326
"tqdm==4.67.1",
327-
"transformers==4.56.0",
327+
"transformers==4.56.1",
328328
"typing-inspection==0.4.1", # required by pydantic and pydantic-settings
329329
"typing_extensions==4.15.0",
330330
"unstructured-client==0.42.3",
@@ -592,7 +592,7 @@
592592
VECTOR_MODELS = {
593593
'BAAI': [
594594
{
595-
'name': 'bge-small-en-v1.5',# transformers 4.30.0
595+
'name': 'bge-small-en-v1.5',
596596
'dimensions': 384,
597597
'max_sequence': 512,
598598
'size_mb': 134,
@@ -627,22 +627,24 @@
627627
'precision': 'float32',
628628
'license': 'mit',
629629
},
630+
],
631+
# 'Google': [
630632
# {
631-
# 'name': 'bge-code-v1',# transformers 4.49.0
632-
# 'dimensions': 1536,
633-
# 'max_sequence': 4096,
634-
# 'size_mb': 1340,
635-
# 'repo_id': 'BAAI/bge-code-v1',
636-
# 'cache_dir': 'BAAI--bge-code-v1',
633+
# 'name': 'embeddinggemma-300m',
634+
# 'dimensions': 768,
635+
# 'max_sequence': 2048,
636+
# 'size_mb': 1210,
637+
# 'repo_id': 'google/embeddinggemma-300m',
638+
# 'cache_dir': 'google--embeddinggemma-300m',
637639
# 'type': 'vector',
638-
# 'parameters': '1540m',
640+
# 'parameters': '303m',
639641
# 'precision': 'float32',
640-
# 'license': 'apache-2.0',
642+
# 'license': 'gemma - commercial ok',
641643
# },
642-
],
644+
# ],
643645
'IBM': [
644646
{
645-
'name': 'Granite-30m-English',# transformers 4.38.2
647+
'name': 'Granite-30m-English',
646648
'dimensions': 384,
647649
'max_sequence': 512,
648650
'size_mb': 61,
@@ -668,7 +670,7 @@
668670
],
669671
'infly': [
670672
{
671-
'name': 'inf-retriever-v1-1.5b',# transformers 4.48.1
673+
'name': 'inf-retriever-v1-1.5b',
672674
'dimensions': 1536,
673675
'max_sequence': 8192,
674676
'size_mb': 3090,
@@ -680,7 +682,7 @@
680682
'license': 'apache-2.0',
681683
},
682684
{
683-
'name': 'inf-retriever-v1-7b',# transformers 4.44.2
685+
'name': 'inf-retriever-v1-7b',
684686
'dimensions': 3584,
685687
'max_sequence': 8192,
686688
'size_mb': 14130,
@@ -694,7 +696,7 @@
694696
],
695697
'intfloat': [
696698
{
697-
'name': 'e5-small-v2',# 4.29.0.dev0
699+
'name': 'e5-small-v2',
698700
'dimensions': 384,
699701
'max_sequence': 512,
700702
'size_mb': 134,
@@ -732,7 +734,7 @@
732734
],
733735
'Qwen': [
734736
{
735-
'name': 'Qwen3-Embedding-0.6B',# transformers 4.51.3
737+
'name': 'Qwen3-Embedding-0.6B',
736738
'dimensions': 1024,
737739
'max_sequence':8192,
738740
'size_mb': 1190,
@@ -770,7 +772,7 @@
770772
],
771773
'Snowflake': [
772774
{
773-
'name': 'arctic-embed-m-v2.0',# transformers 4.39.3
775+
'name': 'arctic-embed-m-v2.0',
774776
'dimensions': 768,
775777
'max_sequence':8192,
776778
'size_mb': 1220,
@@ -1561,11 +1563,6 @@
15611563
"folder": "cuDF",
15621564
"scraper_class": "PydataThemeScraper"
15631565
},
1564-
# "CuPy": {
1565-
# "URL": "https://docs.cupy.dev/en/stable/",
1566-
# "folder": "cupy",
1567-
# "scraper_class": "PydataThemeScraper"
1568-
# },
15691566
"cycler": {
15701567
"URL": "https://matplotlib.org/cycler/",
15711568
"folder": "cycler"
@@ -1891,11 +1888,6 @@
18911888
"folder": "platformdirs",
18921889
"scraper_class": "FuroThemeScraper"
18931890
},
1894-
# "Playwright": {
1895-
# "URL": "https://playwright.dev/python/",
1896-
# "folder": "playwright",
1897-
# "scraper_class": "DivClassThemeDocMarkdownMarkdownScraper"
1898-
# },
18991891
"pluggy": {
19001892
"URL": "https://pluggy.readthedocs.io/en/stable/",
19011893
"folder": "pluggy",
@@ -3159,7 +3151,6 @@ def _generate_button_style(cls, color_values):
31593151
system_message = "You are a helpful person who clearly and directly answers questions in a succinct fashion based on contexts provided to you. If you cannot find the answer within the contexts simply tell me that the contexts do not provide an answer. However, if the contexts partially address my question I still want you to answer based on what the contexts say and then briefly summarize the parts of my question that the contexts didn't provide an answer."
31603152
rag_string = "Here are the contexts to base your answer on. However, I need to reiterate that I only want you to base your response on these contexts and do not use outside knowledge that you may have been trained with."
31613153

3162-
31633154
r"""
31643155
31653156
====================================================================
@@ -3326,23 +3317,20 @@ def _generate_button_style(cls, color_values):
33263317
* Triton 3.2 works with PyTorch >= 2.6 .
33273318
* Triton 3.1 works with PyTorch >= 2.4 . PyTorch 2.3 and older are not supported.
33283319
3329-
33303320
************
33313321
cuDNN & CUDA
33323322
************
33333323
33343324
# Nvidia promises that all cuDNN 9+ releases are compatible with all CUDA 12.x releases.
33353325
# However, certain version of the torch library are built/tested with certain versions of cuDNN. Doesn't always mean incompatibility.
33363326
3337-
33383327
***********************
33393328
LINUX Flash Attention 2
33403329
***********************
33413330
33423331
# HIGHLY CUDA SPECIFIC
33433332
# Obtained from https://github.com/Dao-AILab/flash-attention/blob/main/.github/workflows/publish.yml
33443333
3345-
33463334
*************************
33473335
WINDOWS Flash Attention 2
33483336
*************************
@@ -3358,7 +3346,6 @@ def _generate_button_style(cls, color_values):
33583346
+--------------+-----------------------------------------+------------+
33593347
# Windows wheels: https://github.com/kingbri1/flash-attention
33603348
3361-
33623349
********
33633350
Xformers
33643351
********
@@ -3382,7 +3369,6 @@ def _generate_button_style(cls, color_values):
33823369
* FA2 support determined by by https://github.com/facebookresearch/xformers/blob/main/xformers/ops/fmha/flash.py
33833370
* CUDA support determined by https://github.com/facebookresearch/xformers/blob/main/.github/actions/setup-build-cuda/action.yml
33843371
3385-
33863372
***************
33873373
**CTRANSLATE2**
33883374
***************

0 commit comments

Comments
 (0)