<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Parameter-Efficiency on Pavel Nasovich's Blog</title><link>https://forcewake.me/tags/parameter-efficiency/</link><description>Recent content in Parameter-Efficiency on Pavel Nasovich's Blog</description><generator>Hugo -- 0.157.0</generator><language>en-us</language><copyright>Copyright 2026</copyright><lastBuildDate>Sat, 09 Aug 2025 00:31:26 +0200</lastBuildDate><atom:link href="https://forcewake.me/tags/parameter-efficiency/index.xml" rel="self" type="application/rss+xml"/><item><title>II-Search-4B: A Love Letter to Small Models (Or How I Learned to Stop Worrying and Embrace 4B Parameters)</title><link>https://forcewake.me/ii-search-4b-a-love-letter-to-small-models-or-how-i-learned-to-stop-worrying-and-embrace-4b-parameters/</link><pubDate>Fri, 08 Aug 2025 00:00:00 +0000</pubDate><guid>https://forcewake.me/ii-search-4b-a-love-letter-to-small-models-or-how-i-learned-to-stop-worrying-and-embrace-4b-parameters/</guid><description>A technical analysis of II-Search-4B reveals how this 4-billion parameter model achieves 91.8% SimpleQA accuracy through specialized web search capabilities, outperforming base models by 256% on multi-hop reasoning tasks. We dissect the practical deployment strategies from 8-GPU tensor parallelism to single-GPU quantization and Apple Silicon optimization, demonstrating how focused specialization enables frontier search performance at 1/50th the parameter count of comparable general-purpose models—proving that intelligent constraints and targeted training trump raw scale in the evolving landscape of AI agents.</description></item><item><title>Qwen3-30B-A3B Deep Dive: How 128 Experts Achieve Frontier Performance at 10% Active Parameters</title><link>https://forcewake.me/qwen3-30b-a3b-deep-dive-how-128-experts-achieve-frontier-performance-at-10-active-parameters/</link><pubDate>Thu, 07 Aug 2025 00:00:00 +0000</pubDate><guid>https://forcewake.me/qwen3-30b-a3b-deep-dive-how-128-experts-achieve-frontier-performance-at-10-active-parameters/</guid><description>An architectural analysis of Qwen3-30B-A3B reveals how 128-expert MoE design achieves 91.0 ArenaHard performance with only 3.3B active parameters. This deep technical exploration examines the model&amp;#39;s hybrid thinking framework, sophisticated routing mechanisms, and strategic architectural decisions that enable frontier capabilities at 10x parameter efficiency. We dissect the training methodology, quantization resilience, and deployment flexibility that positions this model as a paradigm shift in efficient AI—demonstrating how intelligent design trumps raw scale in the evolving landscape of large language models.</description></item></channel></rss>