能力 · 2026-05-12
AI Models with Long Context Windows
AI models supporting 200K+ token context windows.
這是什麼?
- Long-context LLMs accept input of 200K tokens or more — enough for entire books, multi-file codebases or hours of transcripts in a single prompt.
- Some models scale to 1M, 2M or even 10M tokens of context.
為什麼重要
- Long context is an alternative (or supplement) to RAG — instead of retrieving chunks, you can paste everything in.
- Beware that effective recall degrades with input length, and per-million-token pricing can make long prompts expensive.
- Some vendors apply tiered pricing above 200K tokens — see each model's detail page for over-200K rates.
397 個模型支援此能力
顯示前 60 項,共 397 項。 用 完整目錄 進一步篩選。
Frequently asked questions
How many AI models support 200K+ 上下文視窗?
397 canonical models in our database currently support 200K+ 上下文視窗. The list is regenerated on every data refresh, so it always reflects the latest model releases from models.dev.
What is the cheapest model with 200K+ 上下文視窗?
Gemini 1.5 Flash-8B from Google is currently the lowest-priced option, at $0.037 per 1M input tokens and $0.150 per 1M output tokens. The full table above is sorted price-ascending.
Which model with 200K+ 上下文視窗 has the largest context window?
Qwen Long (Alibaba (Qwen)) leads on context at 10M tokens. This may matter if you also need long-document understanding alongside 200K+ 上下文視窗.
Which models are available on the most providers?
Production-readiness usually correlates with how many independent providers host the same weights. The top three by provider count are: Kimi K2.5 (45), MiniMax-M2.5 (40), GLM-5 (38).
How is 200K+ 上下文視窗 different from a regular LLM?
Long-context models accept ≥ 200K input tokens — enough for entire books, codebases or hours of transcripts in one prompt. Effective recall and per-token pricing both degrade with input length, so 'big context' is not always the right choice over RAG.
How often is this list updated?
Daily. Our data pipeline pulls models.dev once a day, regenerates the canonical model list, and rebuilds these pages so newly released models appear within 24 hours.
Explore more
Top models with this capability
- Gemini 1.5 Flash-8B$0.04 in / $0.15 out
- Qwen3 235B A22B Instruct 2507$0.10 in / $0.10 out
- Qwen3-235B-A22B-Thinking-2507$0.10 in / $0.10 out
- Qwen3 30B A3B Instruct 2507$0.10 in / $0.10 out
- Qwen3 30B A3B Thinking 2507$0.10 in / $0.10 out
Other capabilities
Best-of lists you might also want
Pricing comparisons
Vendors in this list
最近更新:
Prices in USD per 1M tokens. Unknown means the provider does not publish per-token pricing.
Data is sourced from models.dev and normalized for comparison. Prices and capabilities may change. Always verify critical production decisions with the provider's official documentation.