Hello, dear readers. Happy belated Thanksgiving and Black Friday!
This yr I felt like I used to be living in a everlasting Deva Day. Every week, some lab releases a latest model, a latest agent platform, or a latest “this changes everything” demo. It’s overwhelming. But this is also the first yr that I feel like AI is finally diversifying — not only one or two pioneering cloud models, but the entire ecosystem: open and closed, giant and small, Western and Chinese, cloud and on-premises.
Here’s what I’m really grateful for about the 2025 AI Thanksgiving release in 2025 – releases that feel like they are going to make a difference over the next 12-24 months, not only during this week’s hype cycle.
1. OpenAI maintained good shipping: GPT-5, GPT-5.1, Atlas, Sora 2 and open scales
As the company that undeniably ushered in the era of “generative AI” with its viral ChatGPT product in late 2022, OpenAI likely had one of the most difficult tasks for any AI company in 2025: continuing its growth trajectory at the same time as well-funded competitors like Google with its Gemini models and other startups like Anthropic introduced their very own highly competitive offers.
Fortunately, OpenAI rose to the challenge, and quite a bit. Its important act was GPT-5, unveiled in August as one other pioneering reasoning model, followed in November by GPT-5.1 with latest Instant and Thinking variants that dynamically adjust the amount of “thinking time” spent on a task.
In practice, the GPT-5 launch has been bumpy – VentureBeat documented early math and coding errors and better-than-expected community response, concluding that “OpenAI’s GPT-5 rollout has not been smooth,” but this was quickly corrected based on user feedback, and as an on a regular basis user of the model, I’m personally pleased and impressed with it.
At the same time, corporations that really use these models are seeing solid profits. Global ZenDeskfor example, it claims that agents using GPT-5 now resolve greater than half of customer tickets, with some customers seeing resolution rates of 80-90%. It’s a quiet story: these models don’t all the time impress the chattering classes at X, but they’re starting to move real KPIs.
On the tooling side, OpenAI has finally given developers a serious AI engineer with GPT-5.1-Codex-Max, a latest coding model that may execute long, agent-based workflows and is already OpenAI’s default Codex environment. VentureBeat detailed this in “OpenAI debuts GPT-5.1-Codex-Max encoding model and has already completed the 24-hour task internally.”
Then there’s ChatGPT Atlas, a full browser with ChatGPT built right into Chrome itself – sidebar summaries, on-page evaluation, and search tightly integrated into regular browsing. This is the clearest sign that “assistant” and “browser” are on a collision course.
On the media side, Sora 2 transformed the original Sora video demo into a full video and audio model with higher physics, synchronized audio and dialogue, and more control over the style and structure of shots, in addition to a dedicated Sora app with a full-fledged social networking component, allowing any user create your personal TV network in your pocket.
Finally – and perhaps most symbolically – OpenAI released gpt-oss-120B and gpt-oss-20B, open-weight MoE inference models under an Apache 2.0-style license. Regardless of what you think of their quality (and early adopters of open source have been vocal in their complaints), this is the first time since GPT-2 that OpenAI has given serious attention to the public space.
2. China’s open source wave hits the mainstream
If 2023-24 was about Llama and Mistral, 2025 belongs to China’s openweight ecosystem.
This was demonstrated by a study conducted by MIT and Hugging Face China currently barely leads the United States in global downloads under the open modelthanks in large part to DeepSeek and Alibaba’s Qwen family.
Overview of the most vital events:
-
DeepSeek-R1 launched in January as an open-source inference model competing with OpenAI’s o1, with MIT-licensed weights and a family of smaller models. VentureBeat followed the story from launch, through its impact on cybersecurity, to performance-optimized R1 variants.
-
Kimi K2’s considering from Moonshot, an open source “thinking” model that reasons step by step with tools, very much in the type of o1/R1, and is positioned as the best open reasoning model in the world to date.
-
Z.ai delivered GLM-4.5 and GLM-4.5-Air as “agent” models, open source databases, and hybrid reasoning variants on GitHub.
-
Baidu ERNIE 4.5 Family emerged as a fully open, multimodal MoE package under Apache 2.0, featuring a dense 0.3B model and visual “Thinking” variants focusing on graphs, STEM, and tool use.
-
Alibaba Qwen3 line – which incorporates Qwen3-Coder, large inference models, and the Qwen3-VL series released in summer and fall 2025 – continues to set a high bar for open scales in coding, translation, and multimodal reasoning, which led me to consider this past summer as “
Summer Qwen.”
VentureBeat is tracking these developments, including Chinese math and reasoning models like Light-R1-32B and Weibo’s tiny VibeThinker-1.5B, which are outperforming DeepSeek’s baselines on modest training budgets.
If you care about open ecosystems or on-premise options, this is the yr when the Chinese open-weight scene stopped being a curiosity and became a serious alternative.
3. Small and local models grow up
Another thing I’m grateful for: we finally get it Good small models, not only toys.
Liquid AI spent 2025 developing variants of the Liquid Foundation Models (LFM2) and LFM2-VL vision language, designed from day one for low-latency and device-aware deployments – edge devices, robots and constrained servers, not only giant clusters. Newer LFM2-VL-3B focuses on embedded robotics and industrial autonomy, with demonstrations scheduled for ROSCon.
When it comes to big tech, Google’s Gemma 3 line is strong proof that “small” can still have possibilities. Gemma 3 covers parameters from 270M to 27B, all with open scales and multimodal support on larger variants.
A standout is the Gemma 3 270M, a compact model built specifically for tuning and structured text tasks – corresponding to custom formatters, routers, and watchdogs – discussed each on the Google Developer Blog and in community discussions in local LLM circles.
These models may never gain traction in X, but they’re exactly what you wish for privacy-sensitive workloads, offline workflows, thin-client devices, and “agent swarms” where you do not need every tool call going to a giant borderline LLM.
4. Meta + Midjourney: aesthetics as a service
One of the weirder twists this yr: Meta teamed up with Midjourney as an alternative of just trying to beat him.
In August, Meta announced a deal to license Midjourney’s “aesthetic technology” – an image and video generation stack – and to integrate it with future Meta models and products, from Facebook and Instagram feeds to Meta AI features.
VentureBeat covered the partnership in the article “Meta partners with Midjourney and will license its technology for future models and products,” raising an obvious query: Does this decelerate or change Midjourney’s API roadmap? I’m still waiting for a response, but unfortunately it states that API release plans have yet to materialize, which suggests that it has.
But for creators and brands, the immediate takeaway is easy: Midjourney-level visuals are starting to appear across mainstream social tools, relatively than being locked away in the Discord bot. This could normalize higher-quality AI art for a much wider audience and force rivals like OpenAI, Google and Black Forest Labs to continually raise the bar.
5. Google Gemini 3 and Nano Banana Pro
Google tried to respond to GPT-5 with Gemini 3, touted as the strongest model ever, with higher reasoning, coding and understanding of multimodality, in addition to a latest Deep Think mode for slow and difficult problems.
VentureBeat’s coverage of “Google unveils Gemini 3 as it emerges as a leader in math, science, multimodal and agent-based AI” framed it as a direct shot at pioneering benchmarks and agent-based workflows.
However, the surprise hit is Nano Banana Pro (Gemini 3 Pro Image), Google’s latest flagship image generator. It specializes in infographics, diagrams, multi-topic scenes and multilingual text that is actually readable in 2K and 4K resolutions.
In the world of enterprise AI — where charts, product diagrams, and images that “visually explain the system” matter greater than fantasy dragons — that is a big deal.
6. Wild cards I concentrate to
A few more releases I’m grateful for, even if they do not fit neatly into one bucket:
-
Flux.2 by Black Forest Labs image models that launched just earlier this week with the goal of difficult each Nano Banana Pro and Midjourney in terms of quality and control. VentureBeat delved into the details in the article “Black Forest Labs Launches Flux.2 AI Image Models to Challenge Nano Banana Pro and Midjourney.”
-
Claude Opus by Anthropic 4.5a latest flagship product aimed at cheaper, more efficient coding and long-term task execution, described in the article “Anthropic’s Claude Opus 4.5 is here: Cheaper AI, infinite chats and coding skills that beat humans.”
-
The constant heartbeat of open math/inference models – from Light-R1 to VibeThinker and others – that show you do not need $100 million price of coaching to move the needle.
Final thought (for now)
If 2024 was the yr of “one big cloud model,” then 2025 is the yr the map exploded: multiple frontiers at the top, China leading the way in open models, small and efficient systems maturing rapidly, and creative ecosystems like Midjourney being pulled into the big tech stacks.
I’m grateful not only for any single model, but that we now have it options — closed and open, local and hosted, reasoning first and media first. For journalists, designers and corporations, this diversity is the real story of 2025.
Merry Christmas and all the best to you and your family members!
