- Популярные видео
- Авто
- Видео-блоги
- ДТП, аварии
- Для маленьких
- Еда, напитки
- Животные
- Закон и право
- Знаменитости
- Игры
- Искусство
- Комедии
- Красота, мода
- Кулинария, рецепты
- Люди
- Мото
- Музыка
- Мультфильмы
- Наука, технологии
- Новости
- Образование
- Политика
- Праздники
- Приколы
- Природа
- Происшествия
- Путешествия
- Развлечения
- Ржач
- Семья
- Сериалы
- Спорт
- Стиль жизни
- ТВ передачи
- Танцы
- Технологии
- Товары
- Ужасы
- Фильмы
- Шоу-бизнес
- Юмор
This Free AI Rivals Claude — And Runs On A 4090 - Qwen3.6-27B
Alibaba just open-sourced a 27B model that matches Claude Opus on coding benchmarks — and runs on a single gaming GPU. This one changes the math.
On April 22, 2026, Alibaba's Qwen team released Qwen3.6-27B under the Apache 2.0 license. Free weights. Full commercial use. No cloud dependency. The model posted a 77.2% on SWE-bench Verified — within 3.6 points of Claude Opus 4.6 (80.8%) — and tied Claude 4.5 Opus at 59.3% on Terminal-Bench 2.0. On Alibaba's internal QwenWebBench for web agents, it scored 1487. And the real headline: it outperforms Alibaba's own 397B MoE model, Qwen3.5-397B-A17B, on agentic coding. Fifteen times smaller. Better numbers.
The architecture is a hybrid Gated DeltaNet + Gated Attention layout across 64 layers, with three out of every four sublayers using linear attention. Context window: 262,144 tokens native, extensible to 1,010,000. Hidden dim 5120, intermediate FFN dim 17408. Multi Token Prediction enabled at serving time for faster decoding. A new "Thinking Preservation" feature retains reasoning traces across multi-turn agent workflows.
The cost comparison is brutal for closed labs. Qwen3.6-27B: free to self-host, or ~$0.33/million input and ~$1.95/million output on hosted providers. Claude Opus 4.6: $5/$25. That's roughly 15–20x cheaper at API level, or literally free if you own a 4090.
In this episode, Jane Sterling breaks down what shipped, what the benchmarks actually say (with and without the caveats), how the architecture works, why "27B beats 397B" flips the scaling narrative, what the geopolitics of Apache 2.0 Chinese weights look like, and what this means for developers, enterprise buyers, and the closed-lab business model.
⏱ Timestamps
00:00 Scene One — The Drop
03:00 Scene Two — The Numbers
06:00 Scene Three — The Real Story
🔔 Subscribe to Sterling Intelligence for weekly AI coverage that cuts through the hype.
https://www.youtube.com/@SterlingIntelligence
No hype. No filler. Just the signal.
— Jane Sterling, Sterling Intelligence
#Qwen #Qwen36 #Alibaba #OpenSourceAI #AINews #LLM #AgenticAI #ClaudeOpus #SWEBench #TerminalBench #AICoding #LocalLLM #HuggingFace #SterlingIntelligence #JaneSterling #AIRace #AIBenchmarks #ChinaAI #Apache2 #AIWeekly
Видео This Free AI Rivals Claude — And Runs On A 4090 - Qwen3.6-27B канала Sterling Intelligence
On April 22, 2026, Alibaba's Qwen team released Qwen3.6-27B under the Apache 2.0 license. Free weights. Full commercial use. No cloud dependency. The model posted a 77.2% on SWE-bench Verified — within 3.6 points of Claude Opus 4.6 (80.8%) — and tied Claude 4.5 Opus at 59.3% on Terminal-Bench 2.0. On Alibaba's internal QwenWebBench for web agents, it scored 1487. And the real headline: it outperforms Alibaba's own 397B MoE model, Qwen3.5-397B-A17B, on agentic coding. Fifteen times smaller. Better numbers.
The architecture is a hybrid Gated DeltaNet + Gated Attention layout across 64 layers, with three out of every four sublayers using linear attention. Context window: 262,144 tokens native, extensible to 1,010,000. Hidden dim 5120, intermediate FFN dim 17408. Multi Token Prediction enabled at serving time for faster decoding. A new "Thinking Preservation" feature retains reasoning traces across multi-turn agent workflows.
The cost comparison is brutal for closed labs. Qwen3.6-27B: free to self-host, or ~$0.33/million input and ~$1.95/million output on hosted providers. Claude Opus 4.6: $5/$25. That's roughly 15–20x cheaper at API level, or literally free if you own a 4090.
In this episode, Jane Sterling breaks down what shipped, what the benchmarks actually say (with and without the caveats), how the architecture works, why "27B beats 397B" flips the scaling narrative, what the geopolitics of Apache 2.0 Chinese weights look like, and what this means for developers, enterprise buyers, and the closed-lab business model.
⏱ Timestamps
00:00 Scene One — The Drop
03:00 Scene Two — The Numbers
06:00 Scene Three — The Real Story
🔔 Subscribe to Sterling Intelligence for weekly AI coverage that cuts through the hype.
https://www.youtube.com/@SterlingIntelligence
No hype. No filler. Just the signal.
— Jane Sterling, Sterling Intelligence
#Qwen #Qwen36 #Alibaba #OpenSourceAI #AINews #LLM #AgenticAI #ClaudeOpus #SWEBench #TerminalBench #AICoding #LocalLLM #HuggingFace #SterlingIntelligence #JaneSterling #AIRace #AIBenchmarks #ChinaAI #Apache2 #AIWeekly
Видео This Free AI Rivals Claude — And Runs On A 4090 - Qwen3.6-27B канала Sterling Intelligence
AI news artificial intelligence AI updates AI weekly AI tools AI agents generative AI LLM news AI model releases AI for business AI industry news AI startups AI trends 2026 enterprise AI practical AI OpenAI Anthropic Google AI Claude AI machine learning future of AI AI analysis tech news AI breakthroughs sterling intelligence Jane Sterling
Комментарии отсутствуют
Информация о видео
26 апреля 2026 г. 19:00:20
00:10:11
Другие видео канала




















