Showing 1-6 of 6 projects
An ongoing research project focused on training large language models at scale using transformers.
A Python library for using and fine-tuning over 900 large language models and multimodal models for various AI tasks.
An open-source implementation of large language models with a focus on model parallelism and efficiency.
Comprehensive open-source library of AI research and engineering skills for any AI model.
Official repo for Pai-Megatron-Patch, a large language model and visual language model training framework developed by Alibaba Cloud.
An ongoing research project for training transformer language models at scale, including BERT and GPT-2.
Get weekly updates on trending AI coding tools and projects.