Running Local LLMs on Microsoft Surface Pro 11: NPU Acceleration with DeepSeek Models

Introduction The computing industry is witnessing a paradigm shift with the integration of dedicated AI accelerators in consumer devices. Microsoft’s Copilot+ PCs, including the Surface Pro 11, represent a strategic investment in on-device AI processing capabilities. The ability to run sophisticated Large Language Models (LLMs) locally, without constant cloud connectivity, offers compelling advantages in terms of privacy, latency, and offline functionality. This report investigates the practical aspects of developing and deploying local LLMs on the Surface Pro 11 SD X Elite with 32GB RAM, focusing specifically on leveraging the Neural Processing Unit (NPU) acceleration through ONNX runtime and the implementation of the DeepSeek R1 7B and 14B distilled models. By examining the developer experience, performance characteristics, and comparing with Apple’s M4 silicon, we aim to provide a comprehensive understanding of the current state and future potential of on-device AI processing. ...

May 12, 2025 · 18 min · 3751 words · Pavel Nasovich

Beyond the Prompt: Securing Your LLM's Connection to the World

Large Language Models (LLMs) are revolutionizing how we interact with technology. But their true potential often unlocks when they break free from their digital sandbox and interact with the real world – fetching live data, triggering actions via APIs, or using specialized software tools. Enter the Model-Context Protocol (MCP) and similar frameworks, designed to be the universal adapter, the “USB-C port,” connecting these powerful models to the vast ecosystem of external tools and data sources. ...

April 14, 2025 · 10 min · 1993 words · Pavel Nasovich

Mastering Enterprise AI: A Deep Dive into Azure AI Gateway

Generative AI is revolutionizing business, offering incredible potential. Yet, for many enterprises, adopting powerful models like those in Azure OpenAI feels like navigating the Wild West. How do you unleash innovation without facing runaway costs, complex security threats, inconsistent usage, and the immense challenge of governing AI responsibly at scale? The answer lies in establishing robust, centralized control. Enter the Azure AI Gateway. It’s absolutely critical to understand this: Azure AI Gateway is not a standalone product. Instead, it refers to a powerful set of capabilities integrated directly into Azure API Management (APIM). Microsoft leverages the mature, battle-tested foundation of APIM to provide a centralized control plane, purpose-built for the unique demands of managing Generative AI workloads within the enterprise. Forget deploying separate gateway software; Azure builds this intelligence into the platform you may already use for API management. ...

April 4, 2025 · 11 min · 2265 words · Pavel Nasovich