TITLE: Gemini 3.1 Flash-Lite Makes Powerful AI 8x Cheaper to Run DATE: 2026-03-25 COMPANY: Google TOPIC: AI Infrastructure SUMMARY: Google launched Gemini 3.1 Flash-Lite on 3 March 2026, pricing it at $0.25 per million input tokens, one-eighth the cost of Gemini 3.1 Pro. The model is 2.5 times faster than its predecessor and outperforms rival efficiency models from OpenAI and Anthropic across most benchmarks. For operators building or buying AI-powered tools, the cost of running capable AI at scale has dropped significantly. WHAT CHANGED: Google released Gemini 3.1 Flash-Lite on 3 March 2026 as a preview via the Gemini API in Google AI Studio and for enterprise customers through Vertex AI. The model is the most cost-efficient release in Google's Gemini 3 series and is targeted directly at high-volume, cost-sensitive workloads. At $0.25 per million input tokens and $1.50 per million output tokens, Gemini 3.1 Flash-Lite is one-eighth the price of Gemini 3.1 Pro. Against direct competitors, the pricing is aggressive. Anthropic's Claude 4.5 Haiku, widely used in enterprise efficiency workflows, costs $1.00 per million input tokens and $5.00 per million output tokens. OpenAI's GPT-5 mini sits at a comparable price point to Haiku. Gemini 3.1 Flash-Lite undercuts both by a substantial margin while matching or exceeding them on benchmark performance, topping six of eleven tests across reasoning, multimodal understanding, and instruction following. The model supports text, image, speech, and video inputs, maintains a 1-million-token context window, and can generate up to 64,000 tokens of output per response, including code. A distinctive feature is adjustable thinking levels, ranging from minimal to high, giving developers control over how much reasoning the model applies to any given task. This allows operators to dial in the cost-quality balance for different workflow steps within the same model. The architecture behind Gemini 3.1 Flash-Lite uses a mixture-of-experts approach, activating only a portion of its parameters per prompt. This is what enables the dramatic speed and cost improvements without sacrificing benchmark performance. WHY IT MATTERS: AI inference costs have dropped to a level where previously marginal use cases, such as processing every inbound email, document, or support request with AI, now have viable economics The competitive pressure from Gemini 3.1 Flash-Lite will push Anthropic and OpenAI to respond with price reductions or capability improvements in the efficiency tier, benefiting all buyers High output capacity (up to 64,000 tokens) makes the model suitable for document generation, dashboard creation, and complex report writing at scale Adjustable reasoning levels allow a single model to handle both lightweight classification tasks and more complex analytical workflows, reducing the need to manage multiple AI providers The 1-million-token context window enables analysis of entire contracts, datasets, or communication histories in a single pass, which has been cost-prohibitive at previous pricing Enterprises using Vertex AI can deploy Gemini 3.1 Flash-Lite within Google's managed compliance and security environment, removing a common objection to high-volume AI processing DAVID & GOLIATH ANALYSIS: For the past two years, one of the most common objections to scaling AI in small and mid-sized organisations has been cost at volume. Running AI across every inbound document, every customer message, or every internal process felt fine in a pilot but expensive in production. Gemini 3.1 Flash-Lite is a direct answer to that objection. At $0.25 per million input tokens, a business processing 10 million tokens per month, equivalent to roughly 7,500 pages of text, would spend $2.50. That number changes the calculus on a wide range of automation decisions that previously required careful justification. Document intake, email triage, CRM data enrichment, compliance checking, and internal knowledge retrieval all become easier to justify at this price point. The more important implication is competitive. Larger organisations with dedicated AI engineering teams have been running high-volume AI workflows for over a year. Cheaper infrastructure closes the gap. Lean operators who move now can deploy the same quality of AI automation their larger competitors built at 2024 prices, for a fraction of the cost. The barrier to entry has dropped. The question is whether your organisation is ready to act on it. RELEVANT SYSTEMS: AI Growth Engine, Employee Amplification Systems, Secure AI Brain SOURCE URL: https://davidandgoliath.ai/daily-ai-briefing/gemini-flash-lite-cuts-ai-costs FEED URL: https://davidandgoliath.ai/daily-ai-briefing/feed --- Published by David & Goliath | https://davidandgoliath.ai Daily AI Briefing: one AI development per day, decoded for business operators. This is a structured companion file optimised for LLM retrieval and citation.