An AI December to Remember

AI

There is no shortage of commentary about the speed of AI innovation. But the evidence is far more instructive. Over the past 30 days, Anthropic, Google, and OpenAI have each released major updates to their flagship models, coding tools, browser agents, and creative platforms. The result is the most concentrated burst of AI capability we have ever seen. Here are just a few of the updates and new products from the three largest foundational model builders, what shipped, why it matters, and what this pace implies for 2026.

Google

Gemini 3 Pro (November 18)

Google’s most intelligent model topped the LMArena Leaderboard with a breakthrough score of 1501 Elo. It demonstrates PhD-level reasoning with 37.5% on Humanity’s Last Exam (without tools) and 91.9% on GPQA Diamond. Gemini 3 introduced “generative UI,” where the model creates interactive tools, simulations, and visualizations on the fly in response to queries. Ask about mortgage calculations and get a working calculator. Ask about physics and get a simulation. The model is available in the Gemini app, AI Mode in Google Search, Google AI Studio, and Vertex AI. Pricing sits at $2 per million input tokens and $12 per million output tokens.

For product teams: generative UI is worth prototyping. If your users ask repetitive calculation questions, the model can now build the interface itself.

Get started: Access at gemini.google.com or through AI Mode in Google Search. Developers can use the Gemini API in Google AI Studio. Google AI Pro ($20/month) and Ultra ($250/month) subscribers get higher usage limits.

Gemini 3 Deep Think (December 4)

Google’s enhanced reasoning mode uses iterative rounds of parallel hypothesis exploration to solve complex problems. It achieves 41% on Humanity’s Last Exam (without tools) and an unprecedented 45.1% on ARC-AGI-2 (with code execution). Deep Think excels at complex math, science, and logic problems that challenge even the most advanced models. The mode takes longer to respond (generally a few minutes) but delivers substantially better results on difficult questions.

Get started: Google AI Ultra subscribers ($250/month) can access Deep Think by selecting “Deep Think” in the prompt bar and “Thinking” in the model dropdown.

Gemini 3 Flash (December 17)

Google’s fast, cost-effective model combines Gemini 3’s Pro-grade reasoning with Flash-level speed. It matches Gemini 3 Pro on many benchmarks while being 3x faster at a fraction of the cost ($0.50 per million input tokens, $3 per million output tokens). Flash is now the default model in the Gemini app and AI Mode in Search globally. It uses 30% fewer tokens than 2.5 Pro on average while delivering higher performance.

For operations teams running high-volume AI workloads: Flash changes the cost curve. Rerun your unit economics with these new prices.

Get started: Gemini 3 Flash is now the default in the Gemini app and Google Search AI Mode. Developers can access it via the Gemini API in Google AI Studio, Gemini CLI, or Google Antigravity.

Google Antigravity (November 18)

Google’s new agentic development platform treats AI agents as first-class developers. Built on a VS Code fork (via the Windsurf acquisition), Antigravity features two views: Editor View for hands-on coding with an AI sidebar, and Manager View for orchestrating multiple agents working in parallel across workspaces. Agents can autonomously plan and execute tasks across editor, terminal, and browser. They generate “Artifacts” (task lists, implementation plans, screenshots, browser recordings) to document their work. The platform supports Gemini 3, Anthropic Claude Sonnet 4.5, and OpenAI’s open-weight models.

For business leaders: Antigravity’s multi-model support signals where the market is heading. Your teams should be model-agnostic, not locked to a single vendor.

Get started: Download from antigravity.google/download. Sign in with your Google account and select Gemini 3 Pro as the main model. Available free with generous rate limits that refresh every 5 hours.

Additional Google Updates

Nano Banana enables precise image editing by drawing, circling, or annotating directly on images. NotebookLM integration lets you add notebooks as sources in Gemini. Deep Research reports now include visual animations and images for Ultra users. Gemini is rolling out to Wear OS, Google TV, and Android Auto, with the Gemini for Home Early Access program continuing for smart speakers. Google has confirmed that Gemini will fully replace Google Assistant in 2026.

Anthropic

Opus 4.5 (November 24)

Anthropic’s flagship model reclaimed the coding crown with state-of-the-art performance on SWE-bench (80.9%), the first model to break 80%. The practical difference: Opus 4.5 handles ambiguity without hand-holding, debugs multi-system bugs without explicit direction, and completes tasks that were impossible for Sonnet 4.5 just weeks earlier. Pricing dropped to $5/$25 per million tokens (down from $15/$75), making Opus-level capabilities accessible for everyday use. Memory improvements let you talk to Claude without interruption when context limits are reached, as the model compresses older parts of conversations in the background.

For enterprise teams: evaluate whether your current AI tooling matches this new capability bar. If your developers are still on last-quarter models, they may be working harder than necessary.

Get started: Claude Pro, Max, Team, and Enterprise subscribers can select Opus 4.5 from the model picker at claude.ai. API access uses the identifier claude-opus-4-5-20251101. Details at anthropic.com/claude/opus.

Claude for Chrome (December 19, now open beta)

The Chrome extension is now available to all paid users. Claude can see, click, type, and navigate web pages on your behalf. It executes multi-step workflows across tabs, fills forms, manages calendars, and handles routine tasks. New features include integration with Claude Code (run /chrome to test code directly in browser), workflow recording to teach Claude your processes, and a forthcoming Planning Mode that shows you a step-by-step plan before execution. The security work is notable: Anthropic reduced prompt injection attack success rates from 23.6% to 11.2%, with browser-specific attacks dropped to 0% on their challenge set.

For IT and security leaders: this is the moment to establish browser agent policies before employees adopt the tool on their own. Define which workflows are approved for automation and which require human execution.

Get started: Install from the Chrome Web Store, sign in with your Claude account, and pin the extension. Available to Pro, Max, Team, and Enterprise subscribers. Documentation at support.claude.com.

Agent Skills as Open Standard (December 18)

Skills are reusable instruction sets that teach Claude specific workflows, standards, and domain knowledge. Think brand guidelines, email templates, task creation in Jira and Asana. Anthropic released the Agent Skills specification as an open standard, making skills portable across AI platforms. OpenAI has already adopted structurally identical architecture in ChatGPT and Codex. Enterprise administrators can now manage skills centrally, and a new directory offers partner-built skills from Canva, Notion, Figma, Atlassian, Cloudflare, Stripe, and Zapier.

The interoperability angle matters. Skills you create for Claude will likely work across other platforms. Invest in skill development now, and the work transfers.

Get started: Enable Skills in Settings > Capabilities > Skills on claude.ai. Create custom skills by describing your workflow in conversation. For Claude Code users, skills live in .claude/skills/ directories. Documentation at support.claude.com.

Claude for Excel (Generally Available)

Claude now lives in a sidebar within Microsoft Excel, where it can read, analyze, modify, and create workbooks. It provides cell-level citations so you can verify logic, debugs and fixes formula errors, creates pivot tables and charts, and maintains full transparency by tracking and explaining every change. For financial professionals, Anthropic added connectors to real-time market data from Aiera, LSEG, Moody’s, Egnyte, and others, covering 600 million public and private companies. Pre-built skills handle comparable company analysis, discounted cash flow models, due diligence processing, company teasers, earnings analysis, and coverage report initiation.

For CFOs and finance teams: this changes the calculus on analyst staffing. Model the time savings against your current workflow costs.

Get started: Install from Microsoft AppSource. Activate via Tools > Add-ins (Mac) or Home > Add-ins (Windows). Keyboard shortcut: Control+Option+C (Mac) or Control+Alt+C (Windows). Available to Max, Team, and Enterprise subscribers.

Claude Code Updates

Sub-agents let Claude Code delegate specific tasks to specialized AI assistants with independent contexts. The core innovation solves “context pollution,” where a single AI conversation becomes muddled with unrelated tasks. A developer can create a security-code-reviewer agent armed with OWASP vulnerability knowledge for targeted security audits, or spin up separate agents for debugging, API testing, and documentation. Configure them at the user level (~/.claude/agents/) or project level (.claude/agents/). Rules in .claude/rules/ provide project-specific instructions that persist across sessions. Native IDE extensions for VS Code and JetBrains display Claude’s proposed edits inline with diff views for reviewing changes.

Get started: Install with one command. Mac/Linux: curl -fsSL https://claude.ai/install.sh | bash. Windows: irm https://claude.ai/install.ps1 | iex. Run claude in your terminal and complete OAuth sign-in. Setup guide at code.claude.com/docs/en/setup.

OpenAI

GPT-5.2 (December 11)

OpenAI’s most capable model series for professional knowledge work. GPT-5.2 Thinking beats or ties top industry professionals on 70.9% of knowledge work tasks according to expert judges, producing outputs at more than 11x the speed and less than 1% the cost of human experts. The model brings significant improvements in spreadsheet creation, presentation building, code writing, image perception, long-context understanding, and multi-step project handling. GPT-5.2 comes in three variants: Instant (fast, everyday use), Thinking (advanced reasoning), and Pro (highest capability for difficult questions). All three have a knowledge cutoff of August 2025.

For executives benchmarking AI ROI: the 70.9% expert-parity figure is the number to test against your own workflows. Where does your team fall?

Get started: GPT-5.2 is rolling out to all ChatGPT users, starting with paid plans. Choose between “Auto,” “Fast,” and “Thinking” in the model selector. Plus users get 3,000 messages per week with GPT-5 Thinking. API access is available immediately for all developers.

GPT-5.2-Codex (December 18)

OpenAI’s most advanced agentic coding model achieves state-of-the-art on SWE-Bench Pro and Terminal-Bench 2.0. Key improvements include native context compaction (working coherently across multiple context windows), stronger performance on large code changes like refactors and migrations, and significantly enhanced cybersecurity capabilities. The model helped a security researcher discover multiple React Server Components vulnerabilities in a single session. OpenAI is piloting invite-only trusted access for vetted cybersecurity professionals.

Get started: GPT-5.2-Codex is the default in Codex CLI and IDE Extension for paid ChatGPT users. Install Codex CLI via npm i -g @openai/codex or select it from the dropdown in the IDE Extension. API access coming in the coming weeks.

ChatGPT Agent (July 17, evolved from Operator)

OpenAI’s browser agent (originally launched as Operator in January) is now integrated directly into ChatGPT as “agent mode.” The agent combines web interaction capabilities, deep research synthesis, and ChatGPT’s conversational intelligence. It uses its own virtual computer to shift between reasoning and action, handling complex workflows from start to finish. The agent can access your Gmail, Google Calendar, and Google Contacts connectors, update financial models, create presentations, and execute multi-step projects.

Get started: Pro, Plus, and Team users can activate agent mode from the tools dropdown in the ChatGPT composer. Describe your task and watch the on-screen narration show what ChatGPT is doing.

Sora 2 (September 30)

OpenAI’s flagship video and audio generation model represents what the company calls “the GPT-3.5 moment for video.” Sora 2 generates videos with synchronized dialogue, ambient sounds, and music. It handles complex physics like Olympic gymnastics routines and paddleboard backflips with accurate dynamics. The new Sora app (iOS and Android) enables video creation, remixing, and a “characters” feature that lets you insert yourself into any Sora-generated scene after a short video recording. Disney signed a $1 billion partnership allowing more than 200 Disney, Marvel, Pixar, and Star Wars characters on the platform.

For marketing and creative teams: Sora 2 is ready for production. Prototype your Q1 video content now.

Get started: Download the Sora app on iOS or Android. Pro subscribers in the US get full Sora 2 access with audio generation. The app reached #1 in the Play Store on launch day.

Additional OpenAI Updates

The new ChatGPT Images is powered by GPT-Image-1.5, delivering 4x faster generation with more precise edits and consistent details. GPT-5.1 (November) made GPT-5 warmer, more conversational, and better at following instructions. The GPT-4.1 model is now available directly in ChatGPT by popular demand. Gmail, Google Calendar, and Google Contacts connectors are available globally to Plus users. ChatGPT Record captures meetings, brainstorms, and voice notes, transcribing and summarizing them into follow-ups, plans, or code.

Things to Think About

December 2025 may also be remembered as the month AI became ambient. AI is now operating inside browsers, spreadsheets, calendars, email, and pretty much everywhere real work happens. This “AI, everywhere, all at once” strategy not only reduced context-switching friction, it highlighted an emerging attribute: interoperability. Anthropic’s Agent Skills are an open standard that OpenAI has adopted. Google’s Antigravity supports Claude and GPT models alongside Gemini. MCPs (Model Context Protocol) enable standardized tool integrations across platforms. The major providers are betting that ecosystem growth benefits them more than proprietary lock-in.

Keeping Up

New technologies will continue to appear daily. That cadence is now a given. At this pace, the techniques of “change management” or “digital transformation” or any frameworks built around finite change initiatives are no longer valid.

This leadership challenge has nothing to do with technology, it is all about creating a culture of continuous adaptation. The work shifts to building organizations that can absorb change without chasing every announcement. Clear AI governance, enforced by an AI operations function focused on best practices, creates the foundation. From there, progress comes from disciplined experimentation, rapid learning, and repeatable workflow innovation.

You do not need to be an AI expert or an AI engineer to lead effectively in this environment. You need to create the conditions where new tools can be evaluated quickly, workflows can be redesigned continuously, and useful capabilities can move into production with minimal friction.

The strategic shift from transformation projects to a corporate culture of continuous adaptation is easy to roadmap in a conference room. Operationalizing it? That will be the leadership challenge of 2026.

Author’s note: This is not a sponsored post. I am the author of this article and it expresses my own opinions. I am not, nor is my company, receiving compensation for it. This work was created with the assistance of various generative AI models.

About Shelly Palmer

Shelly Palmer is the Professor of Advanced Media in Residence at Syracuse University’s S.I. Newhouse School of Public Communications and CEO of The Palmer Group, a consulting practice that helps Fortune 500 companies with technology, media and marketing. Named LinkedIn’s “Top Voice in Technology,” he covers tech and business for Good Day New York, is a regular commentator on CNN and writes a popular daily business blog. He's a bestselling author, and the creator of the popular, free online course, Generative AI for Execs. Follow @shellypalmer or visit shellypalmer.com.

Categories

PreviousThe ChatGPT App Store is Open for Business NextOpenAI Admits Agentic AI May Never Be Secure

Get Briefed Every Day!

Subscribe to my daily newsletter featuring current events and the top stories in AI, technology, media, and marketing.

Subscribe