• Ainsider
  • Posts
  • Ainsider AI Newsletter – Edition 54

Ainsider AI Newsletter – Edition 54

The most important AI upgrades & updates from last week | New AI Tools and Models

Table of Contents

Your weekly digest of the most groundbreaking updates, upgrades, and events in the AI space.

Google Unveils Agent2Agent (A2A) Protocol for AI Interoperability

Google kicked off the week with the launch of the Agent2Agent (A2A) protocol, a groundbreaking open standard designed to enable AI agents from different vendors to communicate and collaborate seamlessly. Supported by over 50 tech partners, including Atlassian, Salesforce, and Deloitte, A2A aims to boost productivity, reduce costs, and foster innovation by allowing agents to work together across platforms and frameworks. This protocol could usher in a new era of collaborative AI systems, solving complex problems more efficiently.

Google introduced Firebase Studio
It’s a cloud-based AI development environment, as part of announcements at Cloud Next 2025, detailed at Firebase Blog and VentureBeat. Launched in April 2025, Firebase Studio is an agentic cloud-based development environment that helps developers build and ship production-quality full-stack AI apps, including APIs, backends, frontends, mobile, and more. Key features include:

  • Rapid prototyping, building, and shipping of full-stack AI-infused apps directly from the browser.

  • Unification of Project IDX with specialized AI agents and assistance from Gemini in Firebase, providing a collaborative workspace accessible from anywhere.

  • Ability to import existing projects or start new ones with templates supporting various languages and frameworks.

  • Additional coding and preview features like the App Prototyping agent.

  • Availability in preview with no cost for 3 workspaces, with more for Google Developer Program members (10 for members, 30 for Premium members).

  • Aimed at both developers and non-develops, enabling creation, launch, iteration, and monitoring of apps, APIs, backends, and frontends directly from browsers.

Firebase Studio is experiencing high demand, as noted in VentureBeat, and is part of Google’s broader strategy to enhance developer productivity with AI, as discussed in Medium and The Register.

Google’s Deep Research Now Powered by Gemini 2.5 Pro

Google upgraded its Deep Research tool with Gemini 2.5 Pro, enhancing its capabilities for data analysis and insight generation. Available to Gemini Advanced users, the tool now excels at:

  • Synthesizing complex information.

  • Generating detailed, insightful reports.

  • Advanced analytical reasoning.
    This update makes Deep Research a powerful ally for researchers and professionals tackling intricate topics across web, Android, and iOS platforms.

Canva’s Visual Suite 2.0

Canva unveiled Visual Suite 2.0 at Canva Create 2025 on April 10, 2025, introducing a suite of AI-powered tools to make design more accessible and efficient for everyone. This update blends creativity with productivity, helping both individuals and teams create designs seamlessly.

Key Features:

  • Unified Design Platform: Create presentations, videos, whiteboards, and websites in one format, reducing the need to switch tools (Presentations, Videos, Whiteboards, Websites).

  • Canva Sheets: A visual spreadsheet with AI features like Magic Insights and Data Connectors for Google Analytics, HubSpot, and more, addressing data anxiety for 70% of professionals (Canva Sheets).

  • Magic Charts: Turn complex data into interactive visuals like scrollable reports and infographics.

  • Magic Studio at Scale: Scale content creation with dynamic data, used 16 billion times, ideal for campaigns and outreach (Magic Studio).

  • Canva AI: A voice-enabled design companion for brainstorming and editing via prompts (Canva AI).

  • Canva Code: Create interactive designs like calculators without coding, enhancing functionality for landing pages and presentations (Canva Code).

  • Advanced Photo Editing: Edit images within designs with AI tools like Background Generator and object removal.

Meta Releases Llama 4 Models

Meta dropped a major update with the release of Llama 4 Scout and Llama 4 Maverick, their most advanced multimodal AI models yet. These open-source models are setting new benchmarks:

  • Llama 4 Scout: A 17B-parameter model with a 10M token context window, 16 experts, and 109B total parameters.

  • Llama 4 Maverick: Also 17B parameters but with 128 experts, 400B total parameters, and a 1M context window, excelling in image grounding and matching GPT-4o’s performance at half the parameters.

  • Llama 4 Behemoth: A 2-trillion-parameter model still in training, already surpassing GPT-4.5 and Claude Sonnet 3.7 in STEM benchmarks, with 288B active parameters and 16 experts.

These open-source models are natively multimodal, supporting text, images, and other media, and multilingual in 12 languages (Arabic, English, French, German, Hindi, Indonesian, Italian, Portuguese, Spanish, Tagalog, Thai, Vietnamese). Trained on publicly available data, licensed data, and Meta-proprietary data (e.g., Instagram, Facebook posts, Meta AI interactions) with a data cutoff in August 2024, Llama 4 is available on Llama.com and Hugging Face. Meta AI, with 700 million monthly active users, has been updated to use Llama 4 in 40 countries, potentially expanding with a standalone app in Q2 2025.

Microsoft Copilot: Personalized AI Companion and Security Enhancements

On April 4, 2025, Microsoft marked its 50th anniversary with significant updates to Copilot, transforming it into a personalized AI companion. This update, as per Microsoft Blog, tailors Copilot to individual user needs, values, and expectations, enhancing relevance for personal and enterprise use. Additionally, Microsoft introduced new AI security features through Microsoft Security Copilot, as reported in Help Net Security, including agents for phishing triage, alert triage for data loss prevention and insider risk management, conditional access optimization, vulnerability remediation, and threat intelligence briefing. These agents continuously pull information from various disciplines, providing both manual and automated recommendations, addressing growing concerns about AI-related security risks.

Further, Microsoft’s Copilot Studio saw updates, including:

  • Autonomous Agents: Respond to critical events in real-time, such as budget alerts or low-supply notifications, as detailed in Microsoft Copilot Blog.

  • Deep Reasoning: Enhanced capabilities for complex problem-solving, improving task efficiency.

  • Model Context Protocol (MCP): A new protocol in public preview, streamlining the connection of AI apps, APIs, and data sources into agents, enabling seamless integration of knowledge servers and external tools, with SDK support for custom MCP servers and access to pre-built connectors in the marketplace. This is detailed at Microsoft Copilot Blog.

These updates enhance Copilot’s role as a versatile AI assistant, particularly for enterprise environments.

Rumors of Quasar: A New GPT at OpenRouter

On April 4, 2025, OpenRouter surprised the AI community by releasing Quasar Alpha, a "stealth" model with a 1 million token context window, primarily optimized for coding but also capable of general tasks. Detailed at Quasar Alpha and 16x Prompt, it’s described as a pre-release from one of OpenRouter’s partner labs, available for free to gather community feedback. Early benchmarks show it scoring 55% on the aider polyglot coding benchmark, competitive with o3-mini-medium, DeepSeek V3, and Claude 3.5 Sonnet. Speculation, as per Hacker News and Reddit, suggests it might be from OpenAI, given its performance, but its origins remain unconfirmed, adding intrigue to its release.

Model Context Protocol (MCP) Upgrades

As part of Microsoft’s Copilot Studio updates, the Model Context Protocol (MCP) entered public preview, as mentioned in Microsoft Copilot Blog. MCP streamlines the integration of AI apps, APIs, and data sources into agents, enabling seamless connections with knowledge servers and external tools. It includes SDK support for custom MCP servers and access to pre-built connectors in the marketplace, enhancing interoperability and efficiency in AI ecosystems.

Amazon Introduces Nova Sonic: A Speech-to-Speech AI Breakthrough

Amazon unveiled Nova Sonic, a new foundation model that unifies speech understanding and generation for more humanlike voice interactions. Unlike traditional voice systems that rely on multiple models, Nova Sonic handles tone, inflection, and pacing in a single model, enabling real-time, emotionally intelligent conversations. An example demo showcased a travel assistant adjusting its tone to reassure a customer worried about costs, highlighting its potential for customer service applications.

ChatGPT Gets a Memory Upgrade for Better Conversations

ChatGPT received a significant update with improved memory retention, allowing it to maintain context over longer conversations. This upgrade results in more coherent and relevant responses, enhancing user experience in extended interactions. While specifics on the technical improvements weren’t detailed, this move keeps ChatGPT competitive in the conversational AI space.

NVIDIA and Stanford Pioneer One-Minute AI Cartoons with Test-Time Training

NVIDIA and Stanford University collaborated on a new AI technique called Test-Time Training (TTT), enabling the generation of one-minute animated cartoons with improved temporal and spatial coherence. Demonstrated with Tom and Jerry-style videos, this method uses TTT layers in a pre-trained Diffusion Transformer to produce smooth, single-pass videos without post-processing. This innovation could revolutionize AI-driven video generation for entertainment and beyond.

Subscribe to keep reading

This content is free, but you must be subscribed to Ainsider to continue reading.

Already a subscriber?Sign In.Not now