Description
Tongyi DeepResearch is an agentic large language model developed by Tongyi Lab, with 30 billion total parameters activating only 3 billion per token. It's optimized for long-horizon, deep information-seeking tasks and delivers state-of-the-art performance on benchmarks like Humanity's Last Exam, BrowserComp, BrowserComp-ZH, WebWalkerQA, GAIA, xbench-DeepSearch, and FRAMES. This makes it superior for complex agentic search, reasoning, and multi-step problem-solving compared to prior models. The model includes a fully automated synthetic data pipeline for scalable pre-training, fine-tuning, and reinforcement learning. It uses large-scale continual pre-training on diverse agentic data to boost reasoning and stay fresh. It also features end-to-end on-policy RL with a customized Group Relative Policy Optimization, including token-level gradients and negative sample filtering for stable training. The model supports ReAct for core ability checks and an IterResearch-based 'Heavy' mode for max performance through test-time scaling. It's ideal for advanced research agents, tool use, and heavy inference workflows.
API Usage Examples
OpenAI Compatible Endpoint
Use this endpoint with any OpenAI-compatible library. Model: Tongyi DeepResearch 30B A3B (alibaba/tongyi-deepresearch-30b-a3b)
curl https://api.ridvay.com/v1/chat/completions -H "Content-Type: application/json" -H "Authorization: Bearer YOUR_API_KEY" -d '{
"model": "alibaba/tongyi-deepresearch-30b-a3b",
"messages": [
{
"role": "user",
"content": "Explain the capabilities of the Tongyi DeepResearch 30B A3B model"
}
],
"temperature": 0.7,
"max_tokens": 1024
}'Supported Modalities
- Text
API Pricing
- Input: 0.09$ / 1M tokens
- Output: 0.4$ / 1M tokens
Token Limits
- Max Output: 131,072 tokens
- Max Context: 131,072 tokens
Subscription Tiers
- free
- pro
- ultimate