Skip to content

Latest commit

 

History

History
44 lines (34 loc) · 1.67 KB

File metadata and controls

44 lines (34 loc) · 1.67 KB

FARM: Fine-Grained Alignment for Cross-Modal Recipe Retrieval

More information coming soon. Please feel free to reach out to us if you have any questions.

Installation

First create the conda environment from the env.yaml file:

conda env create --name farm --file=env/env.yaml
source activate farm

Following T-Food, we use bootstrap.pytorch.

cd bootstrap.pytorch
pip install -e .

Install CLIP.

pip install git+https://github.com/openai/CLIP.git

Dataset

We use the Recipe1M dataset in this work.

Evaluation

We adopt the evaluation code from T-Food. For the experiments with missing data, we use empty strings for the corresponding recipe components. Pre-trained models available in this link.

Acknowledgements

We would like to express our gratitude to T-Food for their incredible work on the original project. We use their code for training and evaluating the models. The code for the hyperbolic embedding loss has been adopted from UnHyperML.

Citation

If you find this method and/or code useful, please consider citing:

@inproceedings{wahed2024fine,
  title={Fine-Grained Alignment for Cross-Modal Recipe Retrieval},
  author={Wahed, Muntasir and Zhou, Xiaona and Yu, Tianjiao and Lourentzou, Ismini},
  booktitle={Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision},
  pages={5584--5593},
  year={2024}
}