-
Shanghai AI Laboratory
- Shanghai
-
22:30
(UTC +08:00) - yuhangzang.github.io
- https://orcid.org/0000-0003-1110-5062
- @yuhangzang
Highlights
Starred repositories
An open-source RAG-based tool for chatting with your documents.
The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, and more.
A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.
SOTA discrete acoustic codec models with 40 tokens per second for audio language modeling
Qwen2-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
Data annotation toolbox supports image, audio and video data.
The Open-Source Data Annotation Platform
A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。
Agentic components of the Llama Stack APIs
Run PyTorch LLMs locally on servers, desktop and mobile
Anthropic's educational courses
trholding / llama2.c
Forked from karpathy/llama2.cLlama 2 Everywhere (L2E)
Official Repository of MMLONGBENCH-DOC: Benchmarking Long-context Document Understanding with Visualizations
[ECCV 2024] MVSGaussian: Fast Generalizable Gaussian Splatting Reconstruction from Multi-View Stereo
[ArXiv 2024] WildAvatar: Web-scale In-the-wild Video Dataset for 3D Avatar Creation
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
Banishing LLM Hallucinations Requires Rethinking Generalization
An official implementation of ShareGPT4Video: Improving Video Understanding and Generation with Better Captions
Official implementation of Bootstrap3D: Improving 3D Content Creation with Synthetic Data
TRI-ML / linear_open_lm
Forked from mlfoundations/open_lmA repository for research on medium sized language models.
Multimodal language model benchmark, featuring challenging examples