Explore Projects

Discover 4 open source projects

Active filters (1):
Search: multi-modal-learningร—
Clear all

Showing 1-4 of 4 projects

mlfoundations/open_clip

Open source implementation of CLIP, a contrastive learning model for multi-modal tasks like zero-shot classification.

13.5K
Stable
Python
Computer Vision
PyTorch
#computer-vision#contrastive-learning#pretrained-model

OFA-Sys/Chinese-CLIP

Chinese version of CLIP for cross-modal retrieval and representation generation

5.8K
Stable
Jupyter Notebook
Computer Vision
LLM Frameworks
PyTorch
#chinese#clip#computer-vision

lyuchenyang/Macaw-LLM

Macaw-LLM is a multi-modal language modeling framework that integrates image, video, audio, and text data.

1.6K
Archived
Python
LLM Frameworks
Computer Vision
PyTorch
#multi-modal-learning#deep-learning#natural-language-processing

NVlabs/prismer

Prismer: A Vision-Language Model with Multi-Task Experts for image-captioning and vision-language-model applications.

1.3K
Archived
Python
React
#vision-language-model#multi-task-learning#image-captioning

Stay in the loop

Get weekly updates on trending AI coding tools and projects.