-
Amazon AGI
- New York City
- https://www.linkedin.com/in/shengzha/
- @szha_
Stars
程序员在家做饭方法指南。Programmer's guide about how to cook at home (Simplified Chinese only).
FlashMLA: Efficient Multi-head Latent Attention Kernels
A simple toolchain for moving Remarkable highlights to Readwise
A curated list of projects related to the reMarkable tablet
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
A prize for finding tasks that cause large language models to show inverse scaling
🦜🔗 Build context-aware reasoning applications
PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
Open Academic Research on Improving LLaMA to SOTA LLM
A playbook for systematically maximizing the performance of deep learning models.
This repo includes ChatGPT prompt curation to use ChatGPT and other LLM tools better.
Fast and memory-efficient exact attention
Training and serving large-scale neural networks with auto parallelization.
A list of ICs and IPs for AI, Machine Learning and Deep Learning.
Library for 8-bit optimizers and quantization routines.
Qiskit is an open-source SDK for working with quantum computers at the level of extended quantum circuits, operators, and primitives.
Making large AI models cheaper, faster and more accessible
Full description can be found here: https://discuss.huggingface.co/t/pretrain-gpt-neo-for-open-source-github-copilot-model/7678?u=ncoop57
Training neural networks in TensorFlow 2.0 with 5x less memory
LightSeq: A High Performance Library for Sequence Processing and Generation
This repository contains implementations and illustrative code to accompany DeepMind publications
oneAPI Technical Advisory Board (TAB) Meeting Notes
Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727