Corti
Francesco Corti
Docker
6 giugno 2025

DevOps & Cloud

Is Local AI a Thing? Examining the Viability of On-Device LLMs vs. Cloud-Based AI Services

In the rapidly evolving AI landscape, organizations and developers often encounter certain challenges when utilizing cloud-based LLM services like ChatGPT, Claude, and Gemini. While these platforms offer powerful capabilities, they can present scaling cost considerations and raise important questions around data security and privacy governance. Many businesses require complete control over where their sensitive information is processed and stored as part of their compliance frameworks. Additionally, SaaS solutions are not one-size-fits-all, with emerging use cases where specialized, lightweight models can deliver satisfactory results for specific domains. Businesses increasingly seek models optimized for particular tasks, requiring flexibility that generic cloud services don’t always efficiently provide. The combination of high costs, security governance requirements, and the need for targeted solutions has created momentum toward exploring alternative approaches.
This talk will showcase the potential of local LLM usage during development as a viable alternative to cloud-based services. We'll share performance insights from locally-run models and explore how some models can perform well with limited resources in real-world conditions. The presentation will highlight the growing ecosystem of specialized models designed for specific domains or programming languages, demonstrating how these purpose-built systems can effectively address targeted applications. Through practical examples, we'll illustrate a development lifecycle that leverages local LLMs, giving developers a clear path to building, testing, and deploying AI-powered applications with full control over their data and infrastructure.