324324 "timm==1.0.19" ,
325325 "tokenizers==0.22.0" ,
326326 "tqdm==4.67.1" ,
327- "transformers==4.56.0 " ,
327+ "transformers==4.56.1 " ,
328328 "typing-inspection==0.4.1" , # required by pydantic and pydantic-settings
329329 "typing_extensions==4.15.0" ,
330330 "unstructured-client==0.42.3" ,
592592VECTOR_MODELS = {
593593 'BAAI' : [
594594 {
595- 'name' : 'bge-small-en-v1.5' ,# transformers 4.30.0
595+ 'name' : 'bge-small-en-v1.5' ,
596596 'dimensions' : 384 ,
597597 'max_sequence' : 512 ,
598598 'size_mb' : 134 ,
627627 'precision' : 'float32' ,
628628 'license' : 'mit' ,
629629 },
630+ ],
631+ # 'Google': [
630632 # {
631- # 'name': 'bge-code-v1',# transformers 4.49.0
632- # 'dimensions': 1536 ,
633- # 'max_sequence': 4096 ,
634- # 'size_mb': 1340 ,
635- # 'repo_id': 'BAAI/bge-code-v1 ',
636- # 'cache_dir': 'BAAI--bge-code-v1 ',
633+ # 'name': 'embeddinggemma-300m',
634+ # 'dimensions': 768 ,
635+ # 'max_sequence': 2048 ,
636+ # 'size_mb': 1210 ,
637+ # 'repo_id': 'google/embeddinggemma-300m ',
638+ # 'cache_dir': 'google--embeddinggemma-300m ',
637639 # 'type': 'vector',
638- # 'parameters': '1540m ',
640+ # 'parameters': '303m ',
639641 # 'precision': 'float32',
640- # 'license': 'apache-2.0 ',
642+ # 'license': 'gemma - commercial ok ',
641643 # },
642- ],
644+ # ],
643645 'IBM' : [
644646 {
645- 'name' : 'Granite-30m-English' ,# transformers 4.38.2
647+ 'name' : 'Granite-30m-English' ,
646648 'dimensions' : 384 ,
647649 'max_sequence' : 512 ,
648650 'size_mb' : 61 ,
668670 ],
669671 'infly' : [
670672 {
671- 'name' : 'inf-retriever-v1-1.5b' ,# transformers 4.48.1
673+ 'name' : 'inf-retriever-v1-1.5b' ,
672674 'dimensions' : 1536 ,
673675 'max_sequence' : 8192 ,
674676 'size_mb' : 3090 ,
680682 'license' : 'apache-2.0' ,
681683 },
682684 {
683- 'name' : 'inf-retriever-v1-7b' ,# transformers 4.44.2
685+ 'name' : 'inf-retriever-v1-7b' ,
684686 'dimensions' : 3584 ,
685687 'max_sequence' : 8192 ,
686688 'size_mb' : 14130 ,
694696 ],
695697 'intfloat' : [
696698 {
697- 'name' : 'e5-small-v2' ,# 4.29.0.dev0
699+ 'name' : 'e5-small-v2' ,
698700 'dimensions' : 384 ,
699701 'max_sequence' : 512 ,
700702 'size_mb' : 134 ,
732734 ],
733735 'Qwen' : [
734736 {
735- 'name' : 'Qwen3-Embedding-0.6B' ,# transformers 4.51.3
737+ 'name' : 'Qwen3-Embedding-0.6B' ,
736738 'dimensions' : 1024 ,
737739 'max_sequence' :8192 ,
738740 'size_mb' : 1190 ,
770772 ],
771773 'Snowflake' : [
772774 {
773- 'name' : 'arctic-embed-m-v2.0' ,# transformers 4.39.3
775+ 'name' : 'arctic-embed-m-v2.0' ,
774776 'dimensions' : 768 ,
775777 'max_sequence' :8192 ,
776778 'size_mb' : 1220 ,
15611563 "folder" : "cuDF" ,
15621564 "scraper_class" : "PydataThemeScraper"
15631565 },
1564- # "CuPy": {
1565- # "URL": "https://docs.cupy.dev/en/stable/",
1566- # "folder": "cupy",
1567- # "scraper_class": "PydataThemeScraper"
1568- # },
15691566 "cycler" : {
15701567 "URL" : "https://matplotlib.org/cycler/" ,
15711568 "folder" : "cycler"
18911888 "folder" : "platformdirs" ,
18921889 "scraper_class" : "FuroThemeScraper"
18931890 },
1894- # "Playwright": {
1895- # "URL": "https://playwright.dev/python/",
1896- # "folder": "playwright",
1897- # "scraper_class": "DivClassThemeDocMarkdownMarkdownScraper"
1898- # },
18991891 "pluggy" : {
19001892 "URL" : "https://pluggy.readthedocs.io/en/stable/" ,
19011893 "folder" : "pluggy" ,
@@ -3159,7 +3151,6 @@ def _generate_button_style(cls, color_values):
31593151system_message = "You are a helpful person who clearly and directly answers questions in a succinct fashion based on contexts provided to you. If you cannot find the answer within the contexts simply tell me that the contexts do not provide an answer. However, if the contexts partially address my question I still want you to answer based on what the contexts say and then briefly summarize the parts of my question that the contexts didn't provide an answer."
31603152rag_string = "Here are the contexts to base your answer on. However, I need to reiterate that I only want you to base your response on these contexts and do not use outside knowledge that you may have been trained with."
31613153
3162-
31633154r"""
31643155
31653156====================================================================
@@ -3326,23 +3317,20 @@ def _generate_button_style(cls, color_values):
33263317* Triton 3.2 works with PyTorch >= 2.6 .
33273318* Triton 3.1 works with PyTorch >= 2.4 . PyTorch 2.3 and older are not supported.
33283319
3329-
33303320************
33313321cuDNN & CUDA
33323322************
33333323
33343324# Nvidia promises that all cuDNN 9+ releases are compatible with all CUDA 12.x releases.
33353325# However, certain version of the torch library are built/tested with certain versions of cuDNN. Doesn't always mean incompatibility.
33363326
3337-
33383327***********************
33393328LINUX Flash Attention 2
33403329***********************
33413330
33423331# HIGHLY CUDA SPECIFIC
33433332# Obtained from https://github.com/Dao-AILab/flash-attention/blob/main/.github/workflows/publish.yml
33443333
3345-
33463334*************************
33473335WINDOWS Flash Attention 2
33483336*************************
@@ -3358,7 +3346,6 @@ def _generate_button_style(cls, color_values):
33583346+--------------+-----------------------------------------+------------+
33593347# Windows wheels: https://github.com/kingbri1/flash-attention
33603348
3361-
33623349********
33633350Xformers
33643351********
@@ -3382,7 +3369,6 @@ def _generate_button_style(cls, color_values):
33823369* FA2 support determined by by https://github.com/facebookresearch/xformers/blob/main/xformers/ops/fmha/flash.py
33833370* CUDA support determined by https://github.com/facebookresearch/xformers/blob/main/.github/actions/setup-build-cuda/action.yml
33843371
3385-
33863372***************
33873373**CTRANSLATE2**
33883374***************
0 commit comments