A PyTorch Lightning solution to training CLIP from scratch

train-CLIP

A PyTorch Lightning solution to training CLIP from scratch.

Usage 🚂

This training setup is easily usable right outside the box! Simply provide a training directory or your own dataset and we’ve got the rest covered. To train a model just specify a name from the paper name and tell us your training folder and batch size. All possible models can be seen in the yaml files in models/config

python train.py --model_name RN50 --folder data_dir --batchsize 512

Training with our DataModule 📉

As long as each of the image pairs have the same stem name (i.e. coco_img1.png and coco_img1.txt) all that you need to do is specify the folder on runtime. Any subfolder structure will be ignored, meaning foo/bar/image1.jpg will always find

 

 

 

To finish reading, please visit source site