Archives AI News

Building and Optimizing Intelligent Machine Learning Pipelines with TPOT for Complete Automation and Performance Enhancement

23f1

We begin this tutorial to demonstrate how to harness TPOT to automate and optimize machine learning pipelines practically. By working directly in Google Colab, we ensure the setup is lightweight, reproducible, and accessible. We walk through loading data, defining a custom scorer, tailoring the search space with advanced models like XGBoost, and setting up a […] The post Building and Optimizing Intelligent Machine Learning Pipelines with TPOT for Complete Automation and Performance Enhancement appeared first on MarkTechPost.

Set up custom domain names for Amazon Bedrock AgentCore Runtime agents

Screenshot 2025 08 20 at 11.41.18 AM 1024x483 1

In this post, we show you how to create custom domain names for your Amazon Bedrock AgentCore Runtime agent endpoints using CloudFront as a reverse proxy. This solution provides several key benefits: simplified integration for development teams, custom domains that align with your organization, cleaner infrastructure abstraction, and straightforward maintenance when endpoints need updates.

Detect Amazon Bedrock misconfigurations with Datadog Cloud Security

ML 18870 datadog cloudcraft 1

We’re excited to announce new security capabilities in Datadog Cloud Security that can help you detect and remediate Amazon Bedrock misconfigurations before they become security incidents. This integration helps organizations embed robust security controls and secure their use of the powerful capabilities of Amazon Bedrock by offering three critical advantages: holistic AI security by integrating AI security into your broader cloud security strategy, real-time risk detection through identifying potential AI-related security issues as they emerge, and simplified compliance to help meet evolving AI regulations with pre-built detections.

Google Cloud’s open ecosystem for Apache Iceberg

AI is transforming data into a strategic asset, driving demand for flexible, integrated, and real-time data architectures. But yesterday’s data tools can’t handle AI’s demand for massive volumes of real-time and multi-modal data. Data lakes, for instance, offer flexibility for raw data but lack enforcement and consistency. Meanwhile, traditional data marts, warehouses, and lake architectures often result in silos, and require costly ETL to bridge analytical, unstructured, and operational data. The shift to open lakehouses relies on open table formats like Apache Iceberg, which has emerged as the de facto open-source table format for data lakes. Today, alongside our partners Confluent, Databricks, dbt, Fivetran, Informatica and Snowflake, we’re excited to reiterate our commitment to this open standard. Whether you’re integrating best-of-breed services from diverse providers or navigating a complex data landscape because of a merger and acquisition, adopting an open table format like Iceberg can help you dismantle your traditional data silos. aside_block <ListValue: [StructValue([('title', '$300 in free credit to try Google Cloud data analytics'), ('body', <wagtail.rich_text.RichText object at 0x3ee9c51bae20>), ('btn_text', ''), ('href', ''), ('image', None)])]> United in our Iceberg support At its core, Iceberg provides a metadata layer that enables efficient query planning and data management. This crucial layer, encompassing table schema, partitioning, and data file locations, powers advanced features like time travel and data pruning, which allows data teams to swiftly pinpoint relevant data, streamline performance, and accelerate insights. The data management industry is coalescing around the open Apache Iceberg standard. At Google Cloud, we recently delivered innovations which leverage Google’s Cloud Storage (GCS) to provide an enterprise-grade experience for managing and interoperating with Iceberg data, including BigLake tables for Apache Iceberg and BigLake Metastore with a new REST Catalog API.  Databricks recently announced Iceberg support with their Unity Catalog, allowing users to read and write managed Iceberg tables across a variety of catalogs. Similarly, Snowflake supports interoperable storage with Apache Iceberg tables, allowing organizations to access Iceberg data within Snowflake, minimizing the latency associated with ingesting or copying data. "This open, standard interface allows any Iceberg-compatible engine — including BigQuery, Apache Spark, Databricks, and Snowflake — to operate on the same, single copy of Iceberg data. This powerful architecture even bridges the gap between analytical and operational workloads. By supporting Iceberg and other open table formats in Unity Catalog, we’re unifying data and governance across the enterprise to truly democratize data and AI. No matter what table format our customers choose, we ensure it’s accessible, optimized, and governed for Business and Technical users." - Ryan Blue, Original creator of Apache Iceberg, Member of Technical Staff, Databricks "Customers shouldn’t have to choose between open formats and best-in-class performance or business continuity. Snowflake’s native support for open source standards unifies data while preserving flexibility and choice, paving the way to build and securely scale high-performing lakehouses without silos or operational overhead.” - Rithesh Makkena, Senior Regional Vice President of Partner Solutions Engineering, Snowflake At Google Cloud, we’re committed to an open Data Cloud that lets data teams build modern, data-driven applications wherever their workloads are, while using open source, open standards and open formats like Apache Iceberg.  We partner closely with an extensive ecosystem of partners including Confluent, dbt Labs, Fivetran, and Informatica on Apache Iceberg initiatives. "Apache Iceberg has emerged as a critical enabler for the open data ecosystem, providing the flexibility, interoperability, and consistency that modern, real-time data architectures demand. At Confluent, we're dedicated to helping customers leverage this power. Our Tableflow innovation, by representing Apache Kafka topics as open Iceberg tables, exemplifies how this open format eliminates complex ETL and ensures data is always fresh, accurate, and instantly actionable for critical real-time analytics and AI." - Shaun Clowes, Chief Product Officer, Confluent "dbt was born out of an open source project to help people transform data. Open data ecosystems are at the core of what we do. Supporting Iceberg in dbt ensures that our customers will have standards and choices for how they use their transformed data in their AI and data workflows." - Ryan Segar, Chief Product Office, dbt Labs "Open table formats like Apache Iceberg make it possible to reduce data copies by decoupling your data from the compute engines used to access it. Fivetran’s Managed Data Lake service ensures data is delivered to cloud storage as transactionally consistent tables in a way that preserves the structure from the source. Fivetran’s Managed Data Lake seamlessly integrates with Google Cloud Storage and BigLake metastore, providing a single governance layer within customers' Google projects and making Iceberg tables just as easy to query as native BigQuery tables." - Dan Lynn, Vice President of Product Management, Databases, Fivetran "Our collaboration with Google on the Iceberg format is ushering in a new era of open, interoperable data architecture. Together, we're enabling organizations to unify their data effortlessly, accelerate insights and innovate without limits by eliminating silos and unlocking the full power of the modern data ecosystem." - Rik Tamm-Daniels, GVP, Technology Alliances The power of shared data By adopting Iceberg, customers can share data across different query engines and platforms, leveraging shared datasets for a multitude of workloads and improving interoperability. Organizations can now share data from Snowflake to BigQuery, unlocking powerful BigQuery ML capabilities such as text generation or machine translation, and simplifying ML model development and deployment. Likewise, data teams can share data with BigQuery from Databricks to achieve cost efficiencies, leverage built-in ML, or implement agentic workflows. Customers like Global Payments embraced Iceberg for more flexibility across their diverse data tools. BigQuery and Snowflake serve millions of merchants, and allow the business to analyze transaction data and unlock deep customer insights. Likewise, Unilever has transformed its data management approach with Iceberg, which allows it to manage large datasets more efficiently, particularly in a data lakehouse architecture. Using a combination of Google Cloud and Databricks, Unilever stores and analyzes large amounts of complex data, allowing them and their suppliers to take action wherever and whenever needed.  Whether you create your Iceberg tables in BigQuery, Databricks, or Snowflake, you can leverage the resulting data from any platform and have your tables stay continuously up-to-date. This interoperability will help you operate with greater efficiency and security, drastically reducing the time you spend moving or duplicating datasets, and eliminating the need for complex pipelines to utilize your preferred tools, platforms, and processing systems. Get started today with BigQuery and BigLake for your AI-ready lakehouse. You can learn how to build an open data lakehouse with BigQuery and Snowflake by watching a tutorial, then diving into the Quickstart Guide. Learn how to connect and build an open data lakehouse with BigQuery and Databricks.

From clicks to clusters: Expanding Confidential Computing with Intel TDX

image1 6YVpf1C

Privacy-protecting Confidential Computing has come a long way since we introduced Confidential Virtual Machines (VMs) five years ago. The technology, which can protect data while in use, strengthens a security gap beyond data encryption at rest and in transit.  Since then, customers have used Confidential Computing to protect patient medical data, comply with privacy guidance of GDPR and Schrems II for U.S.-Europe data transfers, and run high-performance computing (HPC) workloads securely.  By isolating workloads in hardware-based Trusted Execution Environments (TEEs), Confidential Computing empowers customers to process their most sensitive information in the public cloud with assurance.  As part of the advancements we’ve made with Confidential Computing, we added even more security capabilities with the introduction of Confidential VMs with Intel Trust Domain Extensions (TDX) last year. Intel TDX creates an isolated trust domain (TD) in a VM, uses hardware extensions for managing and encrypting memory to protect cloud workloads, and offers hardware-based remote attestation for verification.  Today, we are excited to highlight our greatly expanded, and generally available, Intel TDX-based offerings, which includes Confidential GKE Nodes, Confidential Space, Confidential GPU, and more regions and zones where customers can use Confidential Computing. Click to create a Confidential VM Google Cloud Console now offers Google Compute Engine (GCE) customers a new interface for Intel TDX — no code changes required. To get started, follow these steps: Start at the GCE Create an instance page Go to the Security tab and under Confidential VM service, click Enable Then select Intel TDX from the dropdown menu and click Confirm.  It’s that simple to create a Confidential VM. Create a new Confidential VM with Intel TDX in the Google Cloud console. Get Confidential Computing in more regions and zones Confidential VMs with Intel TDX were first available with support for three regions (and nine zones.) To accommodate growing demand, we’ve expanded support for Intel TDX on the C3 machine series to 10 regions (and 21 zones,) and we are planning more for the future. The full list is available here. As regional availability and scalability are critical, your account team is available to help you plan early to ensure your capacity needs are met.  Confidential GKE Nodes with Intel TDX, now generally available Confidential GKE Nodes are built on top of Confidential VM and deliver hardware-based protections to your Google Kubernetes Engine (GKE) clusters and node pools to ensure that your containerized workloads remain encrypted in memory. Today, Confidential GKE Nodes are generally available with Intel TDX on GKE Standard and GKE Autopilot.  Confidential GKE Nodes with Intel TDX on the C3 machine series can be created on GKE Standard via CLI, API, UI, and Terraform. The confidential setting can be set at the cluster level or the node pool level with no code changes. You can learn more here. Confidential GKE Nodes with Intel TDX on the C3 machine series can also be created on GKE Autopilot. It can be enabled through the use of custom compute classes. In GKE, a compute class is a profile that consists of a set of node attributes that GKE uses to provision the nodes that run your workloads during autoscaling events. Check out our documentation to get started. aside_block <ListValue: [StructValue([('title', '$300 in free credit to try Google Cloud security products'), ('body', <wagtail.rich_text.RichText object at 0x3ee9a9f83040>), ('btn_text', ''), ('href', ''), ('image', None)])]> Confidential Space with Intel TDX, now generally available Also built on Confidential VM, our Confidential Space offering is a robust solution for many common issues including addressing insider threats, enabling joint machine-learning training and private gen AI inference, and fostering multi-party collaboration on sensitive data. Here are just a few examples of what our customers have built with Confidential Space: Confidential matching enabled customers to securely connect their first-party data for Google Ads measurement and audience solutions. Symphony demonstrated with its Confidential Cloud how SaaS companies can guarantee isolation of customer data from privileged insiders in the highly regulated financial industry. Duality delivered privacy-preserving federated learning solutions for a broad range of use cases in healthcare, financial services, and the public sector. Flare spearheaded innovation in verifiable AI on blockchain. Previously, Confidential Space was only available with AMD-based technology and hardware (on the N2D, C2D, C3D, and C4D machine series), but now it is also available with Intel-based technology and hardware. This is ideal for those wanting attestation guarantees with a hardware root of trust and for those focused on Intel’s C3 machine series.  Additionally, Confidential Space with Intel TDX is measured into runtime measurement registers (RTMR) and the measurements are verified by Google Cloud Attestation. Note that for Confidential VMs with Intel TDX, RTMRs are now populated as well. Confidential Space benefits are highlighted in the NCC Group’s latest independent security evaluation.  Confidential VM and Confidential GKE Nodes with NVIDIA H100 GPUs, now generally available If you’re looking for performance and security while protecting data in use, Confidential VM and Confidential GKE Nodes with NVIDIA H100 GPUs on the accelerator-optimized A3 machine series are now generally available. These offerings deliver Google Cloud’s first Confidential GPUs, focus on ease of use to meet the demand for secure computing, and extend security to data-intensive, AI and ML workloads by having Intel TDX enabled on the CPU and NVIDIA Confidential Computing enabled on the GPU. You now have the ability to secure your data performantly during inference and training across models. Confidential VM with NVIDIA H100 GPUs is available with the a3-highgpu-1g machine type and in three zones: europe-west4-c, us-central1-a, and us-east5-a. No code changes are needed for most AI and ML workloads. For pricing details, see here. Confidential GKE Nodes with NVIDIA H100 GPUs are generally available on both GKE Standard and GKE Autopilot (through custom compute class). To get started, click here.  And, we also have Confidential Space with NVIDIA H100 GPUs in preview. Intel has a free tier for independent attestation Intel’s attestation verifier service, Intel Tiber Trust Authority, now has a free tier. Google Cloud Confidential VMs and Confidential Space are both integrated with Intel Tiber Trust Authority as a third party attestation service, but now Intel Tiber Trust Authority is making secure attestation more accessible for all by offering a free tier (with optional paid support).  When Confidential VM and Confidential Space customers use Intel Tiber Trust Authority, they can gain stronger separation of duties security guarantees. Click here to learn more.  What our customers say "Thanks to the joint efforts of Super Protocol, Google Cloud, and NVIDIA, the world now gains a new layer of possibility — unlocking Confidential AI without cloud borders. With A3 Confidential VMs built on NVIDIA H100 GPUs now integrated into Super’s decentralized infrastructure and marketplace, companies can securely run, monetize, and collaborate on sensitive AI and data — across any environment. This enables seamless collaboration between Google Cloud customers and partners in other clouds — with no need for shared trust, manual agreements, or compromise. For the broader market, A3 instances at scale accelerate global access, while Super ensures confidentiality, verifiability, and self-sovereignty — fully automated and requiring no expertise in confidential computing. We are excited to open this next chapter of Confidential AI, built to work wherever you and your partners are," said Nukri Basharuli, founder and CEO, Super Protocol. “We’re proud to have partnered with Google Cloud to validate their Confidential Computing-enabled GPU solution — a major step forward in securing sensitive data for AI and machine learning workloads, without compromising on performance or scalability. Confidential Computing allows organizations to process sensitive workloads in the cloud while protecting sensitive data and models from both the cloud provider and the organization's insiders and internal threats. However, for gen AI and agentic AI use cases, protecting the CPU alone isn’t enough — both CPU and GPU must also run in confidential mode with mutual trust. With Google Cloud’s new offering, Anjuna can now launch Confidential Containers that leverage Intel TDX and NVIDIA H100 GPUs in confidential mode. This ensures that data, configurations, secrets, and code remain protected end-to-end from any untrusted entity, bringing state-of-the-art security for sensitive data.” said Steve Van Lare, CTO, Anjuna Security. “With data processing worldwide growing up to three times faster than ever before and doubling every six months, the future of cloud computing must be built on trust. In collaboration with Google, Modelyo leverages Confidential VMs on the A3 machine series with NVIDIA H100 GPUs, transforming Confidential Computing into a seamless, intuitive, and fully integrated cloud experience. This enables us to deliver end-to-end managed solutions across interconnected environments, empowering organizations to innovate confidently knowing their data remains effortlessly protected at every stage.” said Benny Meir, CEO, Modelyo. How to get started with Confidential Computing To add that extra layer of protection and privacy to your sensitive workloads, check out our documentation for Confidential VMs and Confidential GKE Nodes today.

Simplify complex eventing at Scale with Eventarc Advanced

1 XV0Yogc.max 1000x1000 1

Modern application development requires organizations to invest not only in scale but also in simplification and central governance. This means more than message routing; it requires a simple, unified messaging platform that can intelligently filter, transform, and govern the flow of information in real-time, taming complexity all in one place. Today, we are excited to announce the general availability of Eventarc Advanced, a unified, serverless eventing platform that goes beyond simple routing by combining real-time filtering, transformation, management, and delivery in one place — for a complex, multi-source event-driven architecture. Evolving Eventarc to handle complexity Eventarc Advanced is an evolution of Eventarc Standard and offers out-of-the-box integration patterns to simplify your eventing needs. With Eventarc Advanced, organizations can Integrate existing services using Publish API and leverage Google Cloud events to build sophisticated event-driven applications. Centrally manage, secure, and observe the flow of messages across services with support for per-message fine-grained access control. Intelligently route messages to appropriate destinations based on flexible message criteria. Transform and convert events in real-time, with support for multiple payload formats and built-in capability to transform event attributes. Publish to Google Cloud services using HTTP binding. With Eventarc Advanced, you can build sophisticated eventing systems. In contrast, Eventarc Standard is best for simple one-to-one eventing needs involving Google Cloud events (comparison). Eventarc Advanced’s key technical features include: Publish API to ingest custom and third-party messages using CloudEvents format (details). Message bus that acts as the central nervous system of your event-driven architecture, providing centralized observability, security and management. Message bus is based on Envoy and uses the policy engine of Cloud Load Balancers and Cloud Service Mesh. Your existing systems can publish messages to a central message bus that can be intelligently routed to appropriate consumers based on flexible criteria. The message bus simplifies event management and reduces operational overhead.  You can gain insights into your message flows with centralized monitoring, logging, and tracing capabilities. Logs are captured in Cloud Logging, providing detailed information about event processing and errors. Out-of-the-box event mediation capabilities to adapt messages on the fly without modifying your source or destination services, and to handle different events through support for multiple payload formats (Avro, JSON, Protobuf) and built-in capability to transform event attributes. Eventarc Advanced incorporates error-handling by offering reliable event delivery and graceful recovery from transient failures. Empowering developers and operators We designed Eventarc Advanced to cater to the needs of both developers and operators: “Simplicity” for developers: Focus on building your core application features, not on complex event routing logic. Eventarc Advanced provides a unified API and a consistent experience, letting you build decoupled, reliable, and scalable services including real-time transformations. “Centralized governance” for platform operators: Simplify the setup and management of your eventing infrastructure. Centralized governance across projects / teams, plus monitoring and logging make it easier to identify and resolve issues, reducing operational overhead.  How Eventarc Advanced works Imagine an order processing system where orders are created, payments are processed, and items are shipped. Each action is an "event," and in a complex system, managing this flow can be challenging. This is where Eventarc Advanced comes in. It provides a centralized way to manage, observe, and route all your application's events. Let's explore how it works. Set up your message busAt the heart of Eventarc Advanced is a message bus that acts as the central nervous system for your event-driven application. Every event, regardless of its origin, is sent to the message bus to be analyzed and routed. This central hub is where you can define security policies, controlling exactly who can send events and what kind are allowed. In our example, you would create a message bus to receive all order-related events. Whether an order is newly created, its payment is confirmed, or its status changes to "shipped," the events land here. Connect your event sourcesNext, connect your sources that generate order events. Event sources are the services and applications that generate events and feed them into your message bus. Eventarc Advanced makes this easy, supporting a wide range of sources, including: Google API events External apps or custom systems via Publish API In our example, the event source could be a custom service using the Publish API. Every time a new order is saved or an existing one is updated, it automatically sends an event to your message bus. Configure pipelines and destinationsThis is another area where Eventarc Advanced shines. With events flowing into your message bus, you can configure pipelines to intelligently route them to the correct destinations, allowing you to filter, transform, and direct events with precision. In the above example, New order notification: You can set up a filter that looks for events with status: "new". This pipeline routes these events to a notification service that sends an order confirmation email to the customer. Fraud detection: For high-value orders (e.g., amount > $1000), you can apply a transformation and route it to a specialized fraud detection service for analysis. Unlocking new possibilities Eventarc Advanced opens up new possibilities for your applications and workflows: Large-scale application integration: Connect numerous services and agents, enabling them to communicate asynchronously and reliably, even across different event formats and schemas. Event streaming for AI and analytics: Handle the influx of data from IoT devices and AI workloads by filtering and transforming them before feeding them into your analytics pipelines. Hybrid and multi-cloud deployments: Extend your event-driven architectures beyond Google Cloud, integrating with on-premises systems and other cloud providers.  What's next As today’s applications become increasingly agentic, distributed and data-driven, the need for efficient and secure event orchestration is more critical than ever. With upcoming native support for Service Extensions to insert custom code into the data path and services like Model Armor, Eventarc Advanced’s message bus provides security and networking controls for agent communications.  Eventarc Advanced is available today. To learn more about Eventarc Advanced, see the documentation. To learn more about event-driven architectures, visit our Architecture Center based on Google Cloud best practices. Get ready to take your event-driven architectures to the next level!

Framework actually did it: I upgraded a laptop’s entire GPU in just three minutes

framework 16 upgrade

On Tuesday, I told you how the modular computer company Framework was finally fulfilling its promise of the "holy grail for gamers" - a laptop with modular, swappable discrete graphics cards so easy to swap, practically anyone can do it at home. The first futureproof gaming laptop, perhaps? Today, I can confirm the system actually […]

What’s new with Google Cloud

52 Whats new

Want to know the latest from Google Cloud? Find it here in one handy location. Check back regularly for our newest updates, announcements, resources, events, learning opportunities, and more. Tip: Not sure where to find what you’re looking for on the Google Cloud blog? Start here: Google Cloud blog 101: Full list of topics, links, and resources. Aug 25 - 29 Applied AI for Modern Manufacturers: New original growth series, hosted by Jake Hall, The Manufacturing Millennial, that dives into leading trends, best practices, and what companies are doing right now with AI in manufacturing. Hear from industry thought leaders - Rick Bullotta, Jonathan Wise, Walker Reynolds and Berardino Baratta - and Google Cloud experts - Praveen Rao, Eric Lam, Dave Nguyen Ph.D., Geoffrey Hirschheim, and Jim Anderson. Watch Modules 1 and  2 now, where we delve into the AI Innovation and trends and AI Costs and ROI in the Era of Digital Manufacturing. Next module kicks off Tuesday, Sep 2. Join now Firestore with MongoDB compatibility is now generally available (GA): Developers can now build cost-effective, scalable, and highly reliable apps on Firestore's serverless database using a familiar MongoDB-compatible API. With the general availability of Firestore with MongoDB compatibility, the 600,000 active developers within the Firestore community can now use existing MongoDB application code, drivers, and tools, as well as the open-source MongoDB ecosystem, with Firestore's serverless service. Firestore offers benefits like multi-region replication, virtually unlimited scalability, up to 99.999% SLA, single-digit millisecond read performance, integrated Google Cloud governance, and pay-as-you-go pricing. Register now for the webinar on September 9th for a deep dive into Firestore with MongoDB compatibility. Aug 18 - 22 Earth Engine in BigQuery is now Generally Available, bringing advanced geospatial analytics directly to your BigQuery workflows. Unlock insights with satellite data! Aug 11 - Aug 15 New HPC VM and Slurm-gcp Images: A new HPC VM Image (under the project cloud-hpc-image-public) is now available, featuring a Rocky Linux 8-based image, IntelMPI v2021.16, and RDMA drivers. In partnership with SchedMD, new Slurm images (Slurm 25.05) have also been released. These are based on the latest HPC VM Image and are available for Ubuntu 22.04/24.04 Accelerator Images (ARM/AMD64) and Debian 12. These releases allow for the deployment of Slurm-ready clusters on GCP, providing the advantages of an HPC-optimized and performance-tested foundation. Read more. Scaling our Gemini Embedding model in Vertex AI. Following increased popularity from its General Availability launch in May, we've recently increased quota and input size limits for customers of Vertex AI's most powerful text embedding model, gemini-embedding-001. Customers can now send up to 250 input texts per request (generating 250 embeddings) instead of only a single piece of text, bringing improved throughput and decreased round-trip network latency to large-scale embedding applications. We've increased quota limits for this model by 10x for most users, allowing hassle-free scaling of embedding applications to millions of tokens per minute and beyond.Get started with Gemini Embeddings today! Aug 4 - Aug 8 GKE Node Memory Swap in private preview: You can now configure swap space on your GKE Standard nodes to provide a crucial buffer against Out-of-Memory (OOM) errors for memory-intensive applications, especially during unexpected usage spikes. Enabling swap can improve workload resilience, reduce pod evictions due to memory pressure, and enhance overall application stability and cost-effectiveness. This feature is currently available in a private preview. Contact your Google Cloud account team for more information and to request access. If you'd like to see more configurations, please contact your account team or make a feature request on our issue tracker! Unlock Peak Performance: GKE Topology Manager is Now Generally Available: For customers running performance-sensitive workloads like AI/ML and HPC, GKE Topology Manager is now GA and ready to optimize your performance through NUMA alignment. By ensuring CPU, memory, and GPU resources are allocated on the same NUMA node, the Topology Manager minimizes cross-socket latency and maximizes throughput for your most demanding applications. Configure your alignment policies via the NodeConfig API to achieve significant performance gains. Achieve these performance gains by configuring your alignment policies via the NodeConfig API. If you'd like to see more expansion of Topology manager, please contact your account team or make a feature request on our issue tracker! Fine-Tune at Scale: A Massive GKE NodeConfig Expansion for All Workloads: GKE has massively expanded node customization capabilities, adding nearly 130 new Sysctl and Kubelet configurations. This gives you finer-grained control for any workload needing node customization, performance requirements, or application-specific tuning. By replacing complex DaemonSets with native controls, you can benefit from enhanced security, high flexibility, faster node startup times, and less operational management. Check out our public documentation to learn how to consume these new NodeConfig options. If you'd like to see more configurations, please contact your account team or make a feature request on our issue tracker! New capability for managing licenses in Compute Engine: We are announcing a new capability in Compute Engine which allows users to easily change the OS licenses on their VMs. Users can now append, remove, or replace OS licenses, enabling seamless transitions between license types—such as converting Red Hat Enterprise Linux from pay-as-you-go (PAYG) to bring-your-own subscription (BYOS), or upgrading from Ubuntu to Ubuntu Pro—without needing to redeploy instances. This feature empowers customers to meet their evolving licensing with speed and flexibility. To learn more, read about managing licenses on Compute Engine. GKE Turns 10 Hackathon: Calling all developers! Google Kubernetes Engine (GKE) is turning 10, and we're celebrating with a hackathon! Join us to build powerful AI agents that interact with microservice applications using Google Kubernetes Engine and Google AI models. Compete for over $50,000 in prizes and demonstrate the power of building agentic AI on GKE. Submissions are open from Aug 18, 2025 to Sept, 22 2025 Learn more and register: gketurns10.devpost.com Jul 28 - Aug 1 Now GA: C4 VMs with Local SSD, bare metal, and larger shapes, on Intel Xeon 6: C4's expanded shapes are now GA! This expansion introduces C4 shapes with Google’s next-gen Titanium Local SSD, C4 bare metal instances, and new extra-large shapes, all powered by the latest Intel Xeon 6 processors, Granite Rapids. We’re excited to be the first leading hyperscaler to bring Xeon 6 to customers, delivering performance gains of up to 30% for general compute and up to 60% for ML recommendation workloads, and up to 35% lower access latency on Titanium Local SSD shapes. Learn more here! Jul 14 - 18 DMS SQL Server to PostgreSQL migrations are now generally available! Accelerate your SQL Server modernization to Cloud SQL for PostgreSQL or AlloyDB for PostgreSQL with: Automatic database schema and code conversion  Gemini augmented code conversion  Gemini assisted PostgreSQL training and code improvements Low-downtime, CDC based data movement Learn more and start your migration journey today! Jul 7 - 11 Level up your AI Agent game with "The Agent Factory," a new video podcast for developers! We're going beyond the buzz to explore practical design, build, deploy, & management strategies for production-ready AI agents using Google Cloud. Expect code snippets, architecture deep dives, and integrations with open-source frameworks. Subscribe now! Jun 23 - 27 Announcing partnership between Maxim AI and Google Cloud's Vertex AI to evaluate agentic applications — Maxim AI offers a comprehensive platform to help teams build, evaluate, and observe their AI agents with greater speed and confidence, covering the entire AI lifecycle from prompt engineering to production monitoring. This new partnership deeply integrates Vertex AI's Gen AI evaluation service directly within the Maxim AI environment, allowing users to leverage Gemini to power assistant responses and evaluate them using Vertex AI's comprehensive suite of evaluators. This provides access to metrics such as helpfulness, relevance, safety, and trajectory. The setup allows users to simulate, evaluate, and trace complex multi-turn interactions on Maxim, helping teams bring reliable AI products to market faster through a seamless developer experience. To learn more, check out this blog from Maxim AI Run non-request workloads at scale with Cloud Run Worker Pools, now in Public Preview — Looking for the ease-of-use and scalability of serverless, without being limited to HTTP request-driven workloads? Cloud Run Worker Pools provide the same elasticity and high-quality developer experience as Cloud Run Services, but are designed for non-request workloads. Worker Pools are ideal for pull-based use cases like processing messages from Pub/Sub or Kafka, and other backend processing.  Check out the public documentation to learn more about how to choose between Services, Jobs, and Worker Pools. Then give Worker Pools a try by deploying a sample Worker Pool. Building a Multi-Agent Research Assistant for Financial Analysis with Schroders & Google Cloud — Financial analysts spend hours grappling with ever-increasing volumes of market and company data to extract key signals, combine diverse data sources, and produce company research. To maximise its edge as an active manager, Schroders wants to enable its analysts to shift from data collection to the higher-value strategic thinking that is critical for business scalability and client investment performance.  To achieve this, Schroders and Google Cloud collaborated to build a multi-agent research assistant prototype using Vertex AI Agent Builder. Find out more here. Jun 16 - 20 Simplify Your Multi-Cloud Strategy with Cloud Location Finder, now in Public Preview: As cloud environments expand beyond traditional architectures to include multiple clouds, managing your infrastructure effectively becomes more complex. Imagine effortlessly accessing consistent and up-to-date location information across different cloud providers, so your multi-cloud applications are designed and optimized with performance, security, and regulatory compliance in mind. Today, we are making this a reality with Cloud Location Finder, a new Google Cloud service which provides up-to-date location data across Google Cloud, Amazon Web Services (AWS), Azure, and Oracle Cloud Infrastructure (OCI). Now, you can strategically deploy workloads across different cloud providers with confidence and control. Cloud Location Finder is accessible via REST APIs and gcloud CLI, explore the Cloud Location Finder documentation and blog to learn more. SOTA Gemini Text Embedding is Now Generally Available in Vertex AI: We recently launched a new Gemini Embedding text model (gemini-embedding-001) through the Vertex AI GenAI API. This groundbreaking model, leveraging Gemini's core language understanding, sets a new benchmark for text embeddings. It's the first unified model to excel across English, multilingual text, and code, outperforming previous models (text-embedding-005, text-multilingual-embedding-002) and achieving top ranking on the MTEB Multilingual leaderboard (100+ tasks). Our internal benchmarks demonstrate substantial performance improvements across various industry verticals, including retail, news, finance, healthcare, legal, and code. Detailed results are available in our technical report. Backup vaults now support disk backups and multi-regions: We’ve added exciting new features to Google Cloud Backup and Disaster Recovery service! You can now secure your Persistent Disk and Hyperdisk backups in backup vaults, protecting them from cyber attacks and accidental data loss. In addition, backup vaults can now be created in multi-region storage locations, maximizing your data resilience and supporting compliance with business continuity requirements. Check out the blog to learn more! DeepSeek R1, a powerful 671B parameters model, is now available as a fully managed API on Vertex AI in Preview, making advanced AI capabilities more accessible to developers. This Model as a Service (MaaS) offering eliminates the need for extensive GPU resources and infrastructure management, allowing developers to focus on building applications. DeepSeek R1 on Vertex AI provides a simple, scalable API with features like transparent "chain-of-thought" reasoning and enterprise-ready security. It's currently available at no additional cost during the preview, and can be accessed via UI, REST API, or the OpenAI Python API Client Library. Learn more. Jun 9 - 13 Serverless Spark Now GA in BigQuery: Unified Analytics, Accelerated: Google Cloud Serverless for Apache Spark is now generally available in BigQuery, offering a unified developer experience in BigQuery Studio. Run Spark and SQL side-by-side on the same data, powered by the Lightning Engine for up to 3.6x faster performance and enhanced with Gemini productivity. Simplify your data pipelines and accelerate insights with this deeply integrated, zero-ops solution. Cloud Pub/Sub introduced Pub/Sub Single Message Transforms (SMTs) to make it easy to perform simple data transformations right within Pub/Sub: An overarching goal of Pub/Sub is to simplify streaming architectures. We already greatly simplified data movement with Import Topics and Export Subscriptions, which removed the need to use additional services for ingesting raw streaming data through Pub/Sub into destinations like BigQuery. Pub/Sub Single Message Transforms (SMTs), designed to be a suite of features making it easy to validate, filter, enrich, and alter individual messages as they move in real time. The first SMT is available now: JavaScript User-Defined Functions (UDFs), which allows you to perform simple, lightweight modifications to message attributes and/or the data directly within Pub/Sub via snippets of JavaScript code. JavaScript UDFs as the first Single Message Transform is generally available starting today for all users. You'll find the new "Add Transform" option in the Google Cloud console when you create a topic or subscription in your Google Cloud project. You can also use gcloud CLI to start using JavaScript Single Message Transforms today. This analysis evaluates the efficiency of fine-tuning a Llama 3-8B model on Vertex AI using both a single A100 GPU and a distributed four-A100 setup with Axolotl. While both methods achieved similar model convergence, the results underscore the power of distributed training. The process, which took 1 day and 20 hours on a single device, was completed in just 11 hours in the distributed environment—a dramatic acceleration. This speed was achieved with consistently high GPU utilization (94%), though at the cost of higher system and GPU memory overhead. For a detailed breakdown of the methodology, resource utilization metrics, and performance curves, you can review the complete work here. May 26 - 30 Cloud Run GPUs are now GA: NVIDIA GPU support for Cloud Run is now generally available, offering a powerful runtime for a variety of use cases that’s also remarkably cost-efficient. Developers can now get on-demand access to GPUs with our serverless runtime, Cloud Run. Follow the footsteps of customers like MidJourney, vivo, and Wayfair. Read blog. Datastream now supports MongoDB as a source! Seamlessly ingest data from MongoDB (Replica Sets, Sharded Clusters, self-hosted, AtlasDB) into BigQuery/Cloud Storage. Enjoy scalable, fully-managed data streaming with backfill and CDC, enabling real-time insights and data-driven decisions. Link May 19 - May 23 Beyond cuts and fades: Understanding narrative flow with Gemini for accurate scene transition detection — Google Cloud's Gemini models are revolutionizing video understanding by accurately detecting narrative scene transitions, moving beyond simple cuts and fades. This breakthrough technology understands the holistic context of videos by analyzing visual, audio, and textual elements simultaneously. Media companies can now convert passive video assets into structured data, enabling intelligent content discovery, strategic ad placement, and personalized viewing experiences. The result? Up to 38% increased viewer engagement and 27% reduced abandonment rates.  Read more on the medium blog.  Learn more and access the code repository: View Code Repo Announced at I/O: Deploy AI apps to Cloud Run from AI Studio and MCP — We are making AI deployments easier and more accessible by introducing new ways to deploy your apps to Cloud Run. You can deploy applications developed in AI Studio with a click of a button to Cloud Run, including Gemma 3.  Model Context Protocol(MCP) is becoming a popular open protocol standardizing how AI agents interact with other tools. Now with Cloud Run MCP server, you can deploy apps from compatible AI agents like from Claude or VS Code Copilot. Read blog to learn more. May 12 - May 16 Google for Startups Accelerator: AI For Energy now accepting applications!Applications are now open for startups headquartered in Europe and Israel, working on solutions for utilities, grid operators and energy developers; solutions for residential and commercial end-use customers focused on demand flexibility and solutions for industrial customers. This equity-free program offers 10 weeks of intensive mentorship and technical project support to startups integrating AI into their core energy services or products. Selected startups will collaborate with a cohort of peer founders and engage with leaders across Google and the energy sector. The curriculum will provide founders with access to AI tools and include workshops on tech and infrastructure, UX and product, growth, sales, leadership and more. Learn more and apply before June 30th, 2025.  Extending Google Cloud Workstations containers to run any GUI based programAre you having difficulty customizing Google Cloud Workstations to run a GUI program outside of the supported configurations of IDE’s? If so, you’re not alone. In this article we discuss how to use the base Workstations Docker image and build it to run a terminal and Google Chrome. Google Cloud Marketplace simplifies deals and improves economics. Announcing three initiatives that build upon Google Cloud Marketplace as a growth engine for customers and partners: Improving partner deal economics to help partners retain more earnings by moving to a variable revenue share model Simplifying commit drawdown for purchases through channel partners Unlocking new workloads with the Marketplace Customer Credit Program incentiveLearn more 2025 Google Cloud DORA Awards are now open for submission!Has your team achieved remarkable success through DORA principles? It's time to shine. We're thrilled to announce the launch of the 2025 Google Cloud DORA Awards, celebrating outstanding achievements in technology delivery and operational performance. Submit your story today! May 5 - May 9 AI assisted development with MCP Toolbox for DatabasesWe are excited to announce new updates to MCP Toolbox for Databases. Developers can now use Toolbox from their preferred IDE, such as Cursor, Windsurf, Claude Desktop, more and leverage our new pre-built tools such as execute_sql and list_tables for AI-assisted development with Cloud SQL for PostgreSQL, AlloyDB and self-managed PostgreSQL. Get Started with MCP Toolbox for Databases Apr 28 - May 2 Itching to build AI agents? Join the Agent Development Kit Hackathon with Google Cloud! Use ADK to build multi-agent systems to solve challenges around complex processes, customer engagement, content creation, and more. Compete for over $50,000 in prizes and demonstrate the power of multi-agent systems with ADK and Google Cloud. Submissions are open from May 12, 2025 to June 23, 2025. Learn more and register here. Apr 21 - Apr 25 Iceland’s Magic: Reliving Solo Adventure through GeminiEmbark on a journey through Iceland's stunning landscapes, as experienced on Gauti's Icelandic solo trip. From majestic waterfalls to the enchanting Northern Lights, Gautami then takes these cherished memories a step further, using Google's multi-modal AI, specifically Veo2, to bring static photos to life. Discover how technology can enhance and dynamically relive travel experiences, turning precious moments into immersive short videos. This innovative approach showcases the power of AI in preserving and enriching our memories from Gauti's unforgettable Icelandic travels. Read more. Introducing ETLC - A Context-First Approach to Data Processing in the Generative AI Era: As organizations adopt generative AI, data pipelines often lack the dynamic context needed. This paper introduces ETLC (Extract, Transform, Load, Contextualize), adding semantic, relational, operational, environmental, and behavioral context. ETLC enables Dynamic Context Engines for context-aware RAG, AI co-pilots, and agentic systems. It works with standards like the Model Context Protocol (MCP) for effective context delivery, ensuring business-specific AI outputs. Read the full paper. Apr 14 - Apr 18 What’s new in Database CenterWith general availability, Database Center now provides enhanced performance and health monitoring for all Google Cloud databases, including Cloud SQL, AlloyDB, Spanner, Bigtable, Memorystore, and Firestore. It delivers richer metrics and actionable recommendations, helps you to optimize database performance and reliability, and customize your experience. Database Center also leverages Gemini to deliver assistive performance troubleshooting experience. Finally, you can track the weekly progress of your database inventory and health issues.  Get started with Database Center today Access Database Center in Google Cloud console Review the documentation to learn more Apr 7 - Apr 11 This week, at Google Cloud Next, we announced an expansion of Bigtable's SQL capabilities and introduced continuous materialized views. Bigtable SQL and continuous materialized views empower users to build fully-managed, real-time application backends using familiar SQL syntax, including specialized features that preserve Bigtable's flexible schema — a vital aspect of real-time applications. Read more in this blog. DORA Report Goes Global: Now Available in 9 Languages!Unlock the power of DevOps insights with the DORA report, now available in 9 languages, including Chinese, French, Japanese, Korean, Portuguese, and Spanish. Global teams can now optimize their practices, benchmark performance, and gain localized insights to accelerate software delivery. The report highlights the significant impact of AI on software development, explores platform engineering’s promises and challenges, and emphasizes user-centricity and stable priorities for organizational success. Download the DORA Report Now New Google Cloud State of AI Infrastructure Report ReleasedIs your infrastructure ready for AI? The 2025 State of AI Infrastructure Report is here, packed with insights from 500+ global tech leaders. Discover the strategies and challenges shaping the future of AI and learn how to build a robust, secure, and cost-effective AI-ready cloud. Download the report and enhance your AI investments today. Download the 2025 AI infrastructure report now Google Cloud and Oracle Accelerate Enterprise Modernization with New Regions, Expanded CapabilitiesAnnouncing major Oracle Database@Google Cloud enhancements! We're launching the flexible Oracle Base Database Service and powerful new Exadata X11M machines. We're rapidly expanding to 20 global locations, adding new Partner Cross-Cloud Interconnect options, and introducing Cross-Region Disaster Recovery for Autonomous Database. Benefit from enhanced Google Cloud Monitoring, integrated Backup & DR, plus expanded support for enterprise applications like SAP. Customers can run critical Oracle workloads with more power, resilience, and seamless Google Cloud integration. Get started right away from your Google Cloud Console or learn more here. Mar 17 - Mar 21 Cloud CISO Perspectives: 5 tips for secure AI success - To coincide with new AI Protection capabilities in Security Command Center, we’re offering 5 tips to set up your organization for secure AI success. Our 4-6-3 rule for strengthening security ties to business: The desire to quickly transform a business can push leaders to neglect security and resilience, but prioritizing security can unlock value. Follow these 4 principles, 6 steps, and 3 metrics to use a security-first mindset to drive business results. The new Data Protection Tab in Compute Engine ensures your resources are protected: Not only have we co-located your backup options, but we also have introduced smart default data protection for any Compute Engine instance created via Cloud Console. Here’s how it works. DORA report - Impact of Generative AI in Software DevelopmentThis report builds on and extends DORA's research into AI. We review the current landscape of AI adoption, look into its impact on developers and organizations, and outline a framework and practical guidance for successful integration, measurement, and continuous improvement. Download the report! Mar 10 - Mar 14 Protecting your APIs from OWASP’s top 10 security threats: We compare OWASP’s top 10 API security threats list to the security capabilities of Apigee. Here’s how we hold up. Project Shield makes it easier to sign up, set up, automate DDoS protection: It’s now easier than ever for vulnerable organizations to apply to Project Shield, set up protection, and automate their defenses. Here’s how. How Google Does It: Red teaming at Google scale - The best red teams are creative sparring partners for defenders, probing for weaknesses. Here’s how we do red teaming at Google scale. AI Hypercomputer is a fully integrated supercomputing architecture for AI workloads – and it’s easier to use than you think. Check out this blog, where we break down four common use cases, including reference architectures and tutorials, representing just a few of the many ways you can use AI Hypercomputer today.  Transform Business Operations with Gemini-Powered SMS-iT CRM on Google Cloud: SMS-iT CRM on Google Cloud unifies SMS, MMS, email, voice, and 22+ social channels into one Smart Inbox. Enjoy real-time voice interactions, AI chatbots, immersive video conferencing, AI tutors, AI operator, and unlimited AI agents for lead management. Benefit from revenue-driven automation, intelligent appointment scheduling with secure payments, dynamic marketing tools, robust analytics, and an integrated ERP suite that streamlines operations from project management to commerce. This comprehensive solution is designed to eliminate inefficiencies and drive exponential growth for your business. Experience the Future Today. Join us for a new webinar, Smarter CX, Bigger Impact: Transforming Customer Experiences with Google AI, where we'll explore how Google AI can help you deliver exceptional customer experiences and drive business growth. You'll learn how to: Transform Customer Experiences:  With conversational AI agents that provide personalized customer engagements. Improve Employee Productivity & Experience: With AI that monitors customers sentiment in real-time, and assists customer service representatives to raise customer satisfaction scores. Deliver Value Faster: With  30+ data connectors and 70+ action connectors to the most commonly used CRMs and information systems.Register here Mar 3 - Mar 7 Hej Sverige! Google Cloud launches new region in Sweden - More than just another region, it represents a significant investment in Sweden's future and Google’s ongoing commitment to empowering businesses and individuals with the power of the cloud. This new region, our 42nd globally and 13th in Europe, opens doors to opportunities for innovation, sustainability, and growth — within Sweden and across the globe. We're excited about the potential it holds for your digital transformations and AI aspirations. [March 11th webinar] Building infrastructure for the Generative AI era: insights from the 2025 State of AI Infra report: Staying at the forefront of AI requires an infrastructure built for AI. Generative AI is revolutionizing industries, but it demands a new approach to infrastructure. In this webinar, we'll unveil insights from Google Cloud's latest research report and equip tech leaders with a practical roadmap for building and managing gen AI workloads, including: the top gen AI use cases driving the greatest return on investment, current infrastructure approaches and preferences for Generative AI workloads, the impact of performance benchmarks, scalability, and security on cloud provider selection. Register today. Cloud CISO Perspectives: Why PQC is the next Y2K, and what you can do about it: Much like Y2K 25 years ago, post-quantum cryptography may seem like the future’s problem — but it will soon be ours if IT doesn’t move faster, explains Google Cloud’s Christiane Peters. Here's how business leaders can get going on PQC prep. How Google Does It: Using threat intelligence to uncover and track cybercrime — How does Google use threat intelligence to uncover and track cybercrime? Google Threat Intelligence Group’s Kimberly Goody takes you behind the scenes. 5 key cybersecurity strategies for manufacturing executives — Here are five key governance strategies that can help manufacturing executives build a robust cybersecurity posture and better mitigate the evolving risks they face. Datastream now offers Salesforce source in Preview. Instantly connect, capture changes, and deliver data to BigQuery, Cloud Storage, etc. Power real-time insights with flexible authentication and robust backfill/CDC. Unlock Salesforce data for Google Cloud analytics, reporting, and generative AI. Read the documentation to learn more. Find out how much you can save with Spanner - According to a recent Forrester Total Economic Impact™ study, by migrating to Spanner from a traditional database, a $1 billion per year B2C organization could get a 132% return on investment (ROI) with a 9-month payback period, and realize $7.74M in total benefits over the three years. To see how, check out the blog or download the report.  GenAI Observability for Developers series: The Google Cloud DevRel team hosted a four-part webinar series, "Gen AI Observability for Developers," demonstrating observability best practices in four programming languages. Participants learned to instrument a sample application deployed on Cloud Run for auditing Vertex AI usage, writing structured logs, tracking performance metrics, and utilizing OpenTelemetry for tracing. The series covered Go, Java, NodeJS, and Python, using common logging and web frameworks. Missed it? Recordings and hands-on codelabs are available to guide you at: Gen AI O11y for Go Developers Gen AI O11y for Java Developers Gen AI O11y for NodeJS Developers Gen AI O11y for Python DevelopersStay tuned for future events at cloudonair.withgoogle.com. Feb 24 - Feb 28 Rethinking 5G: Ericsson and Google Cloud are collaborating to redefine 5G mobile core networks with a focus on autonomous operations. By leveraging AI and cloud infrastructure, we aim to enhance efficiency, security, and innovation in the telecommunications industry. This partnership addresses the increasing demands of 5G and connected devices, paving the way for a more dynamic and intelligent network future, and setting the stage for next-generation technologies like 6G. Learn more here. Adopt a principles-centered well-architected framework to design, build, deploy, and manage Google Cloud workloads that are secure, resilient, efficient, cost-efficient, and high-performing. Also get industry and technology-focused well-architected framework guidance, like for AI and ML workloads. Feb 17 - Feb 21 Easier Default Backup Configuration for Compute Engine Instances - The Create a Compute Instance page in the Google Cloud console now includes enhanced data protection options to streamline backup and replication configurations. By default, an option to back up data is pre-selected, ensuring recoverability in case of unforeseen events. Learn more here. Feb 10 - Feb 14 [Webinar] Generative AI for Software Delivery: Strategies for IT Leaders: Generative AI is transforming the way organizations build and deploy software. Join Google Cloud experts on February 26th to learn how organizations can leverage AI to streamline their software delivery, including: the role of gen AI in software development, how to use gen AI for migration and modernization, best practices for integrating gen AI into your existing workflows, and real-world applications of gen AI in software modernization and migration through live demos. Register here. Feb 3 - Feb 7 SQL is great but not perfect. We’d like to invite you to reimagine how you write SQL with Google’s newest invention: pipe syntax (public available to all BigQuery and Cloud Logging users). This new extension to GoogleSQL brings a modern, streamlined approach to data analysis. Now you can write simpler, shorter and more flexible queries for faster insights. Check out this video to learn more.  Jan 13 - Jan 17 C4A virtual machines with Titanium SSD—the first Axion-based, general-purpose instance with Titanium SSD, are now generally available. C4A virtual machines with Titanium SSDs are custom designed by Google for cloud workloads that require real-time data processing, with low-latency and high-throughput storage performance. Titanium SSDs enhance storage security and performance while offloading local storage processing to free up CPU resources. Learn more here. Jan 6 - Jan 10 A look back on a year of Earth Engine advancements: 2024 was a landmark year for Google Earth Engine, marked by significant advancements in platform management, cloud integration, and core functionality and increased interoperability between Google Cloud tools and services. Here’s a round up of 2024’s top Earth Engine launches. Get early access to our new Solar API data and features: We're excited to announce that we are working on 2 significant expansions to the Solar API from Google Maps Platform and are looking for trusted testers to help us bring them to market. These include improved and expanded buildings coverage and greater insights for existing solar installations with Detected Arrays. Learn more. Google for Startups Accelerator: Women Founders applications are now open for women-led startups headquartered in Europe and Israel. Discover why this program could be the perfect fit for your startup and apply before January 24th, 2025. Best of N: Generating High-Quality Grounded Answers with Multiple Drafts - We are excited to announce that Check Grounding API has released a new helpfulness score feature. Building on top of our existing groundedness score, we now enable users to implement Best of N to improve RAG response quality without requiring extensive model retraining. Learn more about Best of N and how it can help you here.

Introducing auto scaling on Amazon SageMaker HyperPod

ml 19629 1

In this post, we announce that Amazon SageMaker HyperPod now supports managed node automatic scaling with Karpenter, enabling efficient scaling of SageMaker HyperPod clusters to meet inference and training demands. We dive into the benefits of Karpenter and provide details on enabling and configuring Karpenter in SageMaker HyperPod EKS clusters.