No big frontier-moving news from any of the major generative AI companies this week, but lots of interesting little tidbits. So, Iām taking a slightly different approach to covering this weekās news and focusing on each of Google, OpenAI, and Anthropic and all the news theyāve dropped.
In ethics news thereās been some good articles on the environment with a new tool from Hugging Face that estimates the energy usage of your chats as well as a study on future data centre costs that estimates that they are becoming 1.34x more energy efficient every year.
I also highly recommend the two posts that Simon Willison has posted this week in Long Reads - thereās one on how good OpenAIās o3 is at guessing where an image is taken from and also one on how o3 is now very reliable at AI-assisted search. Well worth a read!
Google
Thereās a really interesting rivalry developing between Googleās 2.5 Pro and OpenAIās o3 models, the two models right on the frontier of capability right now. 2.5 Pro tops the Chatbot Arena leaderboard, but o3 has taken the lead in the FictionLive benchmark that tests working with large amounts of text that Google has been dominating. AI Explained has a great video comparing the two models thatās worth checking out.
Google also released some interesting new information as part of Alphabetās Q1 2025 earnings call. AI overviews now reach more than 1.5 billion people every month (but I still think there are some hallucination issues to address).
Lastly, Google stated in a recent court filing that Gemini now has 350m active users. They also showed a slide estimating that ChatGPT has 600m weekly active users. It was only two weeks that Sam Altman announced that they had 500m weekly active users in a Ted Talk. As a reminder, ChatGPT only surpassed 300m weekly active users in December, and 400m weekly active users back in February. It certainly feels like consumer adoption of frontier generative AI models is accelerating this year!
Source
OpenAI
There were lots of little news drop from OpenAI this week as well. The biggest announcement is that theyāve now made their new viral image generation model available in the API. This is allowing other applications to use their model and for developers to be able to generate large volume of images programmatically. This might lead to a big increase in AI slop on the internet, but at least its better slop than we had a few weeks ago!
OpenAI also announced a ālightweightā version of their Deep Research tool for free users. If you havenāt had a chance to use a Deep Research tool from OpenAI, Google, or Anthropic yet I really encourage you to give this a go. Itās amazing how quickly it can research anything for you online and Iām sure weāll find lots of use cases for this going forwards.
In some other smaller news, OpenAI said theyā be interested in buying Googleās Chrome browser if the company was broken up. This completely plays into OpenAIās current strategy with Sam Altman increasingly talking about themselves as a product company and not an AI research company. Itās not secret that Sam Altman wants OpenAI to become the largest consumer digital platform and heās clearly gunning to take Googleās crown.
This might come sooner than people realise, as there was also a report this week that ChatGPT search, that directly competes with Google, is growing quickly in the EU. Theyāre currently serving 41.3m average monthly users with their search product, up from 11.2m in October, so itās grown by 3.5x in 6 months.
Lastly, Sam Altman admitted that people being polite to ChatGPT was costing them tens of millions of dollars, but he thinks itās money well spent!
Source
Anthropic
We also had some interesting comments and news come out of Anthropic this week. We had a piece published by Dario Amodei, their CEO, about the urgent need to understand how models work, and the company warned that AI employees are probably only about a year away.
Jason Clinton, Anthropicās CISO, believes that we will see AI agents in the next year that have their own memories, their own roles in companies, and have a level of autonomy that far exceeds what AI agents have today. Heās focused on figuring out what this might mean from a corporate security perspective - do these AI Agents need their own individual corporate accounts and passwords, what kind of network access should they be allowed, and who should be responsible for managing its actions?
Anthropic also announce that theyāre launching a new program to study AI āmodel welfareā. This is a research program to determine whether the āwelfareā of an AI model deserves moral consideration and whether the models can suffer from distress. There are still disagreements over how much we should anthropomorphise AI models, but regardless of peopleās beliefs I think looking at these kind of things will only help us whatās going on inside these models and can only be a good thing.
Source
Adobe releases new Firefly image generation models
These new image generation models from Adobe aim to compete with OpenAIās new image generation model. Image Model 4 (not the most creatively named!) is positioned as an everyday model while the āUltraā variant is when you need more detail and realism in your images.
Adobeās 4th generation models improve on the 3rd generation in terms of quality, speed, and control over things like structure, style, camera angles, and zoom.
Firefly has become a really interesting platform for Adobe now as its positioned as a ācommercially safeā creative AI solution that hosts image and video generation capabilities using its own models and other third party models (like OpenAIās image generating models). It will soon be integrated into Adobeās wider Creative Cloud which will be a big feature as Adobeās platform is the most widely used across the creative industries.
Character.AI unveils AvatarFX, an AI video model to create lifelike chatbots
I havenāt written much about Character.AI since August last year when their CEO left to return to Google mostly because there hasnāt been much news. So itās interesting to see this new release from them that will allow users to animate their characters in a variety of styles and voices.
There are plenty of issues with this though, as the platform could be used to create deepfakes with just a photo. To try and combat this, Character.ai will apply watermarks to the videos to make it clear that the footage isnāt real and they will block the creation of videos of minors as well as try to filter images of real people to make them less recognisable.
Character.ai has been a controversial platform since its launch a few years ago, and is currently facing lots of safety issues surrounding its technology where users have formed emotional attachments to their characters. Itās an interesting area to keep an eye on as it might help us better understand our relationship with technology as it becomes more human-like.
AI Ethics News
This tool estimates how much electricity your chatbot messages consume
Within six years, building the leading AI data center may cost $200B
Collective licence to ensure UK authors get paid for works used to train AI
Public comments to White House on AI policy touch on copyright, tariffs
OpenAIās GPT-4.1 may be less aligned than the companyās previous AI models
An AI-generated radio host in Australia went unnoticed for months
OpenAI seeks to make its upcoming āopenā AI model best-in-class
Crowdsourced AI benchmarks have serious flaws, some experts say
Long Reads
Simon Willison - Watching o3 guess a photoās location is surreal, dystopian and wildly entertaining
Dario Amodei - The Urgency of Interpretability
Simon Willison - AI assisted search-based research actually works now
Stratechery - Apple and the Ghosts of Companies Past
TED 2025 - Sam Altman Talks the Future of AI, Safety and Power
Velvet Shark - Why do AI Company Logos Look Like Buttholes?
āThe future is already here, itās just not evenly distributed.ā
William Gibson








