Skip to content

Latest commit

 

History

History
27 lines (19 loc) · 1.94 KB

File metadata and controls

27 lines (19 loc) · 1.94 KB
InternLM

GitHub Org's stars

Welcome to the InternLM (Intern Large Models) organization. Intern-series large models (chinese name: 书生) are developed by Shanghai AI Laboratory and we keep open-sourcing high quality LLMs/MLLMs as well as toolchains for development and application.

Models

  • Intern-S1: a multimodal reasoning model with strong general capabilities and state-of-the-art performance on a wide range of scientific tasks.
  • InternVL: an advanced multimodal large language model (MLLM) series that demonstrates superior general performance.
  • InternLM: a series of multi-lingual foundation models and chat models.
  • InternLM-Math: state-of-the-art bilingual math reasoning LLMs.
  • InternLM-XComposer: a vision-language large model based on InternLM for advanced text-image comprehension and composition.

Toolchain

  • XTuner: a next-generation training engine built for ultra-large MoE models, with superior efficiency.
  • LMDeploy: a toolkit for compressing, deploying and serving LLMs.
  • OpenCompass: a platform for large model evaluation, providing a fair, open, and reproducible benchmark.
  • Lagent: a lightweight framework that allows users to efficiently build LLM-based agents.

Applications

  • HuixiangDou: a domain-specific assistant based on LLMs which can deal with complex techinical questions in group chats.
  • MindSearch: an LLM-based multi-agent framework of web search engine.