Traceback (most recent call last):
File "<frozen runpy>", line 198, in _run_module_as_main
File "<frozen runpy>", line 88, in _run_code
File "__main__.py", line 4, in <module>
main()
File "main.py", line 1162, in main
coder.run()
File "base_coder.py", line 887, in run
self.run_one(user_message, preproc)
File "base_coder.py", line 928, in run_one
message = self.preproc_user_input(user_message)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "base_coder.py", line 917, in preproc_user_input
return self.commands.run(inp)
^^^^^^^^^^^^^^^^^^^^^^
File "commands.py", line 324, in run
return self.do_run(command, rest_inp)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "commands.py", line 296, in do_run
return cmd_method(args)
^^^^^^^^^^^^^^^^
File "commands.py", line 1125, in cmd_help
self.help = Help()
^^^^^^
File "help.py", line 139, in __init__
Settings.embed_model = HuggingFaceEmbedding(model_name="BAAI/bge-small-en-v1.5")
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "base.py", line 160, in __init__
model = SentenceTransformer(
^^^^^^^^^^^^^^^^^^^^
File "SentenceTransformer.py", line 339, in __init__
modules = self._load_auto_model(
^^^^^^^^^^^^^^^^^^^^^^
File "SentenceTransformer.py", line 2112, in _load_auto_model
transformer_model = Transformer(
^^^^^^^^^^^^
File "Transformer.py", line 88, in __init__
config, is_peft_model = self._load_config(model_name_or_path, cache_dir, backend, config_args)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "Transformer.py", line 139, in _load_config
find_adapter_config_file(
File "peft_utils.py", line 83, in find_adapter_config_file
adapter_cached_filename = cached_file(
^^^^^^^^^^^^
File "hub.py", line 276, in cached_file
file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "hub.py", line 507, in cached_files
raise e
File "hub.py", line 419, in cached_files
hf_hub_download(
File "_validators.py", line 89, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "file_download.py", line 1032, in hf_hub_download
return _hf_hub_download_to_cache_dir(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "file_download.py", line 1165, in _hf_hub_download_to_cache_dir
_get_metadata_or_catch_error(
File "file_download.py", line 1699, in _get_metadata_or_catch_error
metadata = get_hf_file_metadata(
^^^^^^^^^^^^^^^^^^^^^
File "_validators.py", line 89, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "file_download.py", line 1622, in get_hf_file_metadata
response = _httpx_follow_relative_redirects(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "file_download.py", line 302, in _httpx_follow_relative_redirects
response = http_backoff(
^^^^^^^^^^^^^
File "_http.py", line 506, in http_backoff
return next(
^^^^^
File "_http.py", line 414, in _http_backoff_base
response = client.request(method=method, url=url, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "_client.py", line 825, in request
return self.send(request, auth=auth, follow_redirects=follow_redirects)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "_client.py", line 901, in send
raise RuntimeError("Cannot send a request, as the client has been closed.")
RuntimeError: Cannot send a request, as the client has been closed.
Aider version: 0.86.2
Python version: 3.12.10
Platform: Windows-11-10.0.26200-SP0
Python implementation: CPython
Virtual environment: No
OS: Windows 11 (64bit)
Git information unavailable
An uncaught exception occurred: