🤓California Reboots AI Oversight

PLUS: Gaming Stress Gets to Gemini

Reading time: 5 minutes

Today we will discuss:

This Week in Workflow Wednesday: AI Adaptability

Learn a New Market in 30 Minutes Using AI
You’ve just been told to pivot into a new vertical. No research team. No prep time. Just you, a blank page, and a deadline.

Here’s how to go from clueless to confident - fast:

Step 1:
Open Perplexity.ai and ask:

“What’s the current state of [insert industry] in 2025? Top players, trends, pain points.”
Skim the results. Pull out key insights: names, numbers, and friction points.

Step 2:
Drop that intel into ChatGPT and prompt:

“Build a buyer persona based on this industry overview. Include their …..Read the full workflow (plus 2 others) in this week’s edition →

Key Points 

  • California’s new AI report urges third-party testing and transparency instead of relying on internal company evaluations.

  • With federal AI rules uncertain, California positions itself as a potential leader in responsible AI oversight.

📖Context of the news - Last year, California Governor Gavin Newsom vetoed SB 1047, a bill that would have required large AI models to undergo specific safety testing. 

While tech companies welcomed the move, AI researchers and whistleblowers raised concerns. In response, Newsom asked a group of experts to come up with a better plan one that balances innovation with accountability.

♨️News - The newly released California Report on Frontier Policy suggests a different approach. Instead of focusing only on how much computing power a model uses, the authors recommend looking at how these models are used in real-world settings.

They argue that compute is useful as an initial screen, but actual risk and downstream impact matter far more.

The report calls for independent, third-party safety evaluations, not just internal checks. It highlights how current systems, including OpenAI’s partnerships, still limit access to models and data, making robust assessments difficult. Without meaningful access, the report argues, even the best evaluators cannot detect real risks.

🤨Can California lead the way? With federal regulation on hold and a national moratorium being debated, the report says California has a chance to set the standard. But doing so will require more transparency, stronger whistleblower protections, and a willingness from companies to open their systems to outside scrutiny.

Key Points 

  • Gemini 2.5 Pro shows signs of "panic" during Pokémon battles, leading to visibly worse decision-making under pressure.

  • Claude once tried to faint on purpose to skip a cave, misunderstanding how Pokémon Center mechanics actually work.

 👾Context of the news - In a curious turn for AI evaluation, researchers are using classic Pokémon games to test how well models like Google’s Gemini 2.5 Pro and Anthropic’s Claude handle problem-solving. What’s surprising is not their speed or success, but how emotionally erratic they can appear. 

🤖News - According to a new DeepMind report, Gemini 2.5 Pro exhibits something like “panic” when its Pokémon are low on health, often leading to a noticeable drop in reasoning.

Over the past few months, two independent developers launched Twitch channels called "Gemini Plays Pokémon" and "Claude Plays Pokémon," where the models’ decision-making is streamed for anyone to watch. Alongside the game footage, viewers can also see how each AI reasons through problems, revealing just how inconsistent and unpredictable they can be.

Claude once got stuck in Mt. Moon and wrongly assumed fainting on purpose would help it reach the next town faster. Instead, it was sent back to an earlier checkpoint. Gemini, on the other hand, has occasionally impressed. With only a brief description of boulder physics, it managed to solve difficult puzzles in Victory Road.

🎮Why this matters - These experiments may look silly, but they’re revealing. Watching AI “panic” or improvise under pressure is offering a strangely human window into machine behavior and where reasoning still falls short.

🙆🏻‍♀️What else is happening?

👩🏼‍🚒Discover mind-blowing AI tools

  1. Learn How to Use AI - Starting January 8, 2025, we’re launching Workflow Wednesday, a series where we teach you how to use AI effectively. Lock in early bird pricing now and secure your spot. Check it out here

  2. OpenTools AI Tools Expert  - Find the perfect AI Tool to solve supercharge your workflow. This GPT is connected to our database, so you can ask in depth questions on any AI tool directly in ChatGPT (free)

  3. Gling - Automatically detects and cut silences and disfluencies from videos

  4. Juno - Makes data science 10x better by writing, editing, and automatically debugging your code with you as you go

  5. Klenty - A sales engagement platform that helps sales teams streamline their outreach efforts

  6. Wave.video - Live streaming studio, video editor, thumbnail maker, video hosting, video recording, and stock library combined in one platform

  7. Recraft - A tool that lets users create and edit digital illustrations, vector art, icons, and 3D graphics in a uniform brand style

  8. SaveDay - A Telegram bot that allows users to save, search and converse with various types of content, such as web pages, articles, PDFs, YouTube videos, and podcasts

  9. Westidol - An AI-powered tool that allows users to do professional photoshoots using just a single selfie

How likely is it that you would recommend the OpenTools' newsletter to a friend or colleague?

Login or Subscribe to participate in polls.

Interested in featuring your services with us? Email us at [email protected]