<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>AI-Deployment on Pavel Nasovich's Blog</title><link>https://forcewake.me/tags/ai-deployment/</link><description>Recent content in AI-Deployment on Pavel Nasovich's Blog</description><generator>Hugo -- 0.157.0</generator><language>en-us</language><copyright>Copyright 2026</copyright><lastBuildDate>Thu, 07 Aug 2025 23:39:36 +0200</lastBuildDate><atom:link href="https://forcewake.me/tags/ai-deployment/index.xml" rel="self" type="application/rss+xml"/><item><title>Qwen3-30B-A3B Deep Dive: How 128 Experts Achieve Frontier Performance at 10% Active Parameters</title><link>https://forcewake.me/qwen3-30b-a3b-deep-dive-how-128-experts-achieve-frontier-performance-at-10-active-parameters/</link><pubDate>Thu, 07 Aug 2025 00:00:00 +0000</pubDate><guid>https://forcewake.me/qwen3-30b-a3b-deep-dive-how-128-experts-achieve-frontier-performance-at-10-active-parameters/</guid><description>An architectural analysis of Qwen3-30B-A3B reveals how 128-expert MoE design achieves 91.0 ArenaHard performance with only 3.3B active parameters. This deep technical exploration examines the model&amp;#39;s hybrid thinking framework, sophisticated routing mechanisms, and strategic architectural decisions that enable frontier capabilities at 10x parameter efficiency. We dissect the training methodology, quantization resilience, and deployment flexibility that positions this model as a paradigm shift in efficient AI—demonstrating how intelligent design trumps raw scale in the evolving landscape of large language models.</description></item></channel></rss>