Code for a paper "Multi-Modal Adapter for Vision-Language Models"
- Setup required python version using your preferred method (e.g. pyenv, virtualenv, etc.). For pyenv users:
pyenv install 3.11.6
pyenv local 3.11.6- Install poetry if needed following the instructions at https://python-poetry.org/docs/#installation
- Install dependencies:
poetry install- Set up the pre-commit hooks:
poetry run pre-commit installFirst, create a new conda environment from the environment.yml file:
conda env create -f environment.ymlThen, activate the environment:
conda activate fomoAfter install dependencies with pip using the requirements.txt file:
pip install -r requirements.txtFor some of the datasets, you have to download them from the original source. The datasets are not included in this repository.
Download the dataset from https://www.kaggle.com/datasets/jutrera/stanford-car-dataset-by-classes-folder to data/stanford-cars.