skillpack.co
All solutions

Tongyi DeepResearch

active

First fully open-source deep research agent matching closed-source leaders on benchmarks. HLE 32.9 (exceeds OpenAI's 26.6), 30.5B params / 3.3B active (MoE), runs locally. 18.5K stars. Apache 2.0.

Score 82
Tongyi DeepResearch in action

Where it wins

HLE 32.9 — exceeds OpenAI's 26.6 on same benchmark

GAIA 70.9, BrowseComp 43.4, FRAMES 90.6 — competitive across all benchmarks

30.5B params, only 3.3B active per token (MoE) — runs on consumer hardware

18,483 stars in ~6 months — rapid growth

365 HN pts / 153 comments — near-viral reception

Apache 2.0 — full commercial use allowed

Deployed in Alibaba's Gaode Maps — production-validated

Where to be skeptical

No PyPI/npm download stats — adoption beyond GitHub stars unclear

Newer project (~6 months) — less battle-tested than GPT Researcher

Alibaba corporate backing may concern some users

No MCP server available

Editorial verdict

#5 in research. Open-source disruptor — HLE 32.9 exceeds OpenAI's 26.6 on the same benchmark, runs on consumer hardware via MoE. Called 'the DeepSeek moment for AI agents' by VentureBeat. Ranked below GPT Researcher due to less proven adoption data.

Related

Public evidence

Raw GitHub source

GitHub README peek

Constrained peek so you can sanity-check the source material without leaving the site.

<div align="center"> <picture> <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/logo.png" width="100%"> </picture> </div> <hr> <div align="center" style="line-height: 1;"> </div> <p align="center"> <p align="center"> 🤗 <a href="https://huggingface.co/Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" target="_blank">HuggingFace</a> | <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/tongyi.png" width="14px" style="display:inline;"> <a href="https://modelscope.cn/models/iic/Tongyi-DeepResearch-30B-A3B" target="_blank">ModelScope</a> | 💬 <a href="./assets/wechat_new.jpg">WeChat(微信)</a> | 📰 <a href="https://tongyi-agent.github.io/blog/introducing-tongyi-deep-research/">Blog</a> | 📑 <a href="https://arxiv.org/pdf/2510.24701">Paper</a> <p align="center"> <a href="https://trendshift.io/repositories/14895" target="_blank"><img src="https://trendshift.io/api/badge/repositories/14895" alt="Alibaba-NLP%2FDeepResearch | Trendshift" style="width: 250px; height: 55px;" width="250" height="55"/></a>

👏 Welcome to try Tongyi DeepResearch via our <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/tongyi.png" width="14px" style="display:inline;"> Modelscope online demo or 🤗 Huggingface online demo or <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/WebAgent/assets/aliyun.png" width="14px" style="display:inline;"> bailian service!

[!NOTE] This demo is for quick exploration only. Response times may vary or fail intermittently due to model latency and tool QPS limits. For a stable experience we recommend local deployment; for a production-ready service, visit <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/WebAgent/assets/aliyun.png" width="14px" style="display:inline;"> bailian and follow the guided setup.

Introduction

We present <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/tongyi.png" width="14px" style="display:inline;"> Tongyi DeepResearch, an agentic large language model featuring 30.5 billion total parameters, with only 3.3 billion activated per token. Developed by Tongyi Lab, the model is specifically designed for long-horizon, deep information-seeking tasks. Tongyi DeepResearch demonstrates state-of-the-art performance across a range of agentic search benchmarks, including Humanity's Last Exam, BrowseComp, BrowseComp-ZH, WebWalkerQA,xbench-DeepSearch, FRAMES and SimpleQA.

Tongyi DeepResearch builds upon our previous work on the <img src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/tongyi.png" width="14px" style="display:inline;"> WebAgent project.

More details can be found in our 📰 <a href="https://tongyi-agent.github.io/blog/introducing-tongyi-deep-research/">Tech Blog</a>.

<p align="center"> <img width="100%" src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/performance.png"> </p>

Features

  • ⚙️ Fully automated synthetic data generation pipeline: We design a highly scalable data synthesis pipeline, which is fully automatic and empowers agentic pre-training, supervised fine-tuning, and reinforcement learning.
  • 🔄 Large-scale continual pre-training on agentic data: Leveraging diverse, high-quality agentic interaction data to extend model capabilities, maintain freshness, and strengthen reasoning performance.
  • 🔁 End-to-end reinforcement learning: We employ a strictly on-policy RL approach based on a customized Group Relative Policy Optimization framework, with token-level policy gradients, leave-one-out advantage estimation, and selective filtering of negative samples to stabilize training in a non‑stationary environment.
  • 🤖 Agent Inference Paradigm Compatibility: At inference, Tongyi DeepResearch is compatible with two inference paradigms: ReAct, for rigorously evaluating the model's core intrinsic abilities, and an IterResearch-based 'Heavy' mode, which uses a test-time scaling strategy to unlock the model's maximum performance ceiling.

Model Download

You can directly download the model by following the links below.

ModelDownload LinksModel SizeContext Length
Tongyi-DeepResearch-30B-A3B🤗 HuggingFace<br> 🤖 ModelScope30B-A3B128K

News

[2025/09/20]🚀 Tongyi-DeepResearch-30B-A3B is now on OpenRouter! Follow the Quick-start guide.

[2025/09/17]🔥 We have released Tongyi-DeepResearch-30B-A3B.

Deep Research Benchmark Results

<p align="center"> <img width="100%" src="https://raw.githubusercontent.com/Alibaba-NLP/DeepResearch/main/assets/benchmark.png"> </p>

Quick Start

This guide provides instructions for setting up the environment and running inference scripts located in the inference folder.

1. Environment Setup
  • Recommended Python version: 3.10.0 (using other versions may cause dependency issues).
  • It is strongly advised to create an isolated environment using conda or virtualenv.
# Example with Conda
conda create -n react_infer_env python=3.10.0
conda activate react_infer_env
2. Installation

Install the required dependencies:

pip install -r requirements.txt
3. Environment Configuration and Prepare Evaluation Data

Environment Configuration

Configure your API keys and settings by copying the example environment file:

# Copy the example environment file
cp .env.example .env

Edit the .env file and provide your actual API keys and configuration values:

  • SERPER_KEY_ID: Get your key from Serper.dev for web search and Google Scholar
  • JINA_API_KEYS: Get your key from Jina.ai for web page reading
  • API_KEY/API_BASE: OpenAI-compatible API for page summarization from OpenAI
  • DASHSCOPE_API_KEY: Get your key from Dashscope for file parsing
  • SANDBOX_FUSION_ENDPOINT: Python interpreter sandbox endpoints (see SandboxFusion)
  • MODEL_PATH: Path to your model weights
  • DATASET: Name of your evaluation dataset
  • OUTPUT_PATH: Directory for saving results

Note: The .env file is gitignored, so your secrets will not be committed to the repository.

Prepare Evaluation Data

The system supports two input file formats: JSON and JSONL.

Supported File Formats:

Option 1: JSONL Format (recommended)

  • Create your data file with .jsonl extension (e.g., my_questions.jsonl)
  • Each line must be a valid JSON object with question and answer keys:
    {"question": "What is the capital of France?", "answer": "Paris"}
    {"question": "Explain quantum computing", "answer": ""}
    

Option 2: JSON Format

  • Create your data file with .json extension (e.g., my_questions.json)
  • File must contain a JSON array of objects, each with question and answer keys:
    [
      { "question": "What is the capital of France?", "answer": "Paris" },
      { "question": "Explain quantum computing", "answer": "" }
    ]
    

Important Note: The answer field contains the ground truth/reference answer used for evaluation. The system generates its own responses to the questions, and these reference answers are used to automatically judge the quality of the generated responses during benchmark evaluation.

File References for Document Processing:

  • If using the file parser tool, prepend the filename to the question field
  • Place referenced files in eval_data/file_corpus/ directory
  • Example: {"question": "(Uploaded 1 file: ['report.pdf'])\n\nWhat are the key findings?", "answer": "..."}
View on GitHub →