Stars
Optimized primitives for collective multi-GPU communication
When it comes to optimizers, it's always better to be safe than sorry
FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/
Fast and memory-efficient exact attention
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open Platform for Enterprise AI (OPEA) project.
Open source code for AlphaFold 2.
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction of AlphaFold 2
PyTorch Tutorial for Deep Learning Researchers
Pretrain, finetune and serve LLMs on Intel platforms with Ray
Port of OpenAI's Whisper model in C/C++
Tensors and Dynamic neural networks in Python with strong GPU acceleration
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
General technology for enabling AI capabilities w/ LLMs and MLLMs
A cross platform way to express data transformation, relational algebra, standardized record expression and plans.
A modular acceleration toolkit for big data analytic engines
A blazingly fast JSON serializing & deserializing library
The official home of the Presto distributed SQL query engine for big data