CLIP
CLIP employs contrastive language-image pre-training to achieve zero-shot prediction, matching performance with labeled data models. By integrating with PyTorch and TorchVision, CLIP facilitates diverse tasks like CIFAR-100 predictions and linear-probe evaluations through its image and text encoding capabilities.