- Atlanta, Georgia
-
03:38
- 5h behind - in/trilok-padhi-bb809b147
- @trilokpadhi_
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL
PyTorch Code for the paper "VSE++: Improving Visual-Semantic Embeddings with Hard Negatives"
Official Repo for Open-Reasoner-Zero
Solve Visual Understanding with Reinforced VLMs
📦 Repomix (formerly Repopack) is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) o…
Simple go utility to download HuggingFace Models and Datasets
A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.
[ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs
A holistic way of understanding how Llama and its components run in practice, with code and detailed documentation.
Official implementation of the Law of Vision Representation in MLLMs
This repository collects all relevant resources about interpretability in LLMs
zhijian-liu / lmms-eval
Forked from EvolvingLMMs-Lab/lmms-evalAccelerating the development of large multimodal models (LMMs) with lmms-eval
[ICML'24 Oral] "MagicLens: Self-Supervised Image Retrieval with Open-Ended Instructions"
[NeurIPS 2024] Knowledge Circuits in Pretrained Transformers
Accelerating the development of large multimodal models (LMMs) with one-click evaluation module - lmms-eval.
[CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"
Coding a Multimodal (Vision) Language Model from scratch in PyTorch with full explanation: https://www.youtube.com/watch?v=vAmKB7iPkWw
Utilities intended for use with Llama models.
A library for experimenting with, training and evaluating neural networks, with a focus on adversarial robustness.
A framework for few-shot evaluation of language models.
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024
Implementation of a multimodal diffusion transformer in Pytorch
Slides and recordings of talks hosted by our community
A 4-hour coding workshop to understand how LLMs are implemented and used