Lists (1)
Sort Name ascending (A-Z)
Stars
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
Emu Series: Generative Multimodal Models from BAAI
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.
Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).
FashionCLIP is a CLIP-like model fine-tuned for the fashion domain.