Starred repositories
Visualizer for neural network, deep learning and machine learning models
Basic Python interface for MoveIt 2 built on top of ROS 2 actions and services
欢迎来到 LLM-Dojo,这里是一个开源大模型学习场所,使用简洁且易阅读的代码构建模型训练框架(支持各种主流模型如Qwen、Llama、GLM等等)、RLHF框架(DPO/CPO/KTO/PPO)等各种功能。👩🎓👨🎓
Strong and Open Vision Language Assistant for Mobile Devices
[RSS 2024] 3D Diffusion Policy: Generalizable Visuomotor Policy Learning via Simple 3D Representations
Baseline model for "GraspNet-1Billion: A Large-Scale Benchmark for General Object Grasping" (CVPR 2020)
🦜🔗 Build context-aware reasoning applications
This is the third party implementation of the paper Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection.
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models
Octo is a transformer-based robot policy trained on a diverse mix of 800k robot trajectories.
Official repo for "iVideoGPT: Interactive VideoGPTs are Scalable World Models", https://arxiv.org/abs/2405.15223
Curated list of papers and resources focused on 3D Gaussian Splatting, intended to keep pace with the anticipated surge of research in the coming months.
The official implementation of SAGA (Segment Any 3D GAussians)
[ECCV 2022] XMem: Long-Term Video Object Segmentation with an Atkinson-Shiffrin Memory Model
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
CALVIN - A benchmark for Language-Conditioned Policy Learning for Long-Horizon Robot Manipulation Tasks
ALFRED - A Benchmark for Interpreting Grounded Instructions for Everyday Tasks
Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Pre-training Reusable Representations for Robotic Manipulation Using Diverse Human Video Data
A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites
Examples and guides for using the OpenAI API
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
Recent LLM-based CV and related works. Welcome to comment/contribute!
ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial and Multi-Map SLAM