Gemini’s New Beta Delivers Proactive Responses Using Your Photos, Emails, and More

Google has launched a beta feature that allows Gemini to connect directly with user apps like Gmail and Photos for more personalized responses. The update aims to make Gemini more proactive while keeping privacy controls in user hands.

By Samantha Reed Edited by Maria Konash Published: Updated:
Gemini’s New Beta Delivers Proactive Responses Using Your Photos, Emails, and More
Google debuts Gemini Personal Intelligence for deeper, app-connected AI personalization. Photo: Solen Feyissa / Unsplash

Google has introduced a new personalization feature for its Gemini AI assistant that allows users to connect their Google apps with a single tap. The feature, called Personal Intelligence, is rolling out in beta to eligible Google AI Pro and AI Ultra subscribers in the United States and is designed to make Gemini more context-aware, proactive, and useful in everyday tasks.

When enabled, Personal Intelligence allows Gemini to securely access data from apps such as Gmail, Google Photos, YouTube, and Search. Users choose whether to turn the feature on and which apps to connect, with the option to disable access at any time. Google said the setup is designed to be simple while giving users granular control over how their information is used.

The feature is intended to help Gemini reason across multiple sources and retrieve specific details from personal content. Gemini can combine information from text, images, and video to answer questions or make suggestions tailored to an individual’s history and preferences. For example, Gemini can locate details from emails, identify information embedded in photos, and reference past activity to provide more relevant recommendations.

Privacy Controls and Safeguards

Google emphasized that privacy was a central design principle for Personal Intelligence. App connections are turned off by default, and Gemini only accesses linked data in response to user requests. The company said personal data stored in Gmail or Photos is not directly used to train its AI models. Instead, Gemini references that information to generate responses, while model training relies on filtered or obfuscated prompts and outputs.

Gemini also aims to provide transparency by explaining which sources it used to generate an answer. Users can ask follow-up questions to verify responses, regenerate replies without personalization, or use temporary chats that do not draw on connected apps. Guardrails are in place to limit proactive assumptions about sensitive topics, including health-related data, unless explicitly requested.

Google acknowledged that errors and over-personalization may still occur during the beta phase. The company encouraged users to provide feedback when responses feel inaccurate or make incorrect assumptions. Addressing nuance, such as changes in relationships or evolving interests, remains an active area of development.

Availability and Broader AI Strategy

Personal Intelligence works across web, Android, and iOS, and is compatible with all models available in the Gemini model picker. Google said it plans to expand availability to additional countries and eventually to the free tier. The feature is also expected to come to AI Mode in Search. At launch, it is limited to personal Google accounts and does not support Workspace enterprise or education users.

The update comes as Google’s Gemini models gain wider industry adoption, including a recent multi-year agreement under which Apple will use Gemini models and Google’s cloud infrastructure to power future AI features and upcoming Siri upgrades.

Netflix Buys Ben Affleck’s AI Startup for Up to $600 Million

Netflix has acquired InterPositive, an AI startup focused on post-production editing tools co-founded by Ben Affleck. The deal could reach $600 million as streaming platforms expand AI use in filmmaking.

By Samantha Reed Edited by Maria Konash Published:
Netflix Buys Ben Affleck’s AI Startup for Up to $600 Million
Netflix acquires Ben Affleck’s AI startup for up to $600M to develop post-production editing tools for filmmakers. Image: Jakob Owens / Unsplash

Netflix has acquired InterPositive, an artificial intelligence company focused on post-production tools for filmmakers. The startup was co-founded by actor and filmmaker Ben Affleck and develops software designed to assist editors and production teams in refining footage during the editing process.

According to Bloomberg, the transaction could be worth up to $600 million, potentially making it one of the largest acquisitions in Netflix’s history. Netflix has not publicly confirmed the full financial terms of the deal.

Sources familiar with the agreement told Bloomberg that the upfront cash payment may be lower, with additional payouts tied to performance targets. If those targets are met, the total value of the acquisition could approach the reported figure.

The largest acquisition previously completed by Netflix was the purchase of the Roald Dahl Story Company for about $700 million.

AI Tools Designed for Film Post-Production

InterPositive develops artificial intelligence tools aimed at helping filmmakers work more efficiently during the editing phase of production. The company’s software focuses on improving workflows rather than generating entirely new content.

Its technology can assist with tasks such as identifying continuity issues, enhancing scenes, and streamlining the process of reviewing and organizing large volumes of footage. These capabilities are intended to reduce time spent on manual editing work while maintaining creative control for filmmakers.

The company’s tools also avoid using copyrighted footage without permission or generating new scenes from scratch, which has been a point of concern for many professionals in the entertainment industry as generative AI becomes more widely adopted.

Netflix has not yet announced how InterPositive’s technology will be integrated into its production pipeline. However, the acquisition aligns with the company’s ongoing efforts to incorporate artificial intelligence into film and television production workflows.

Streaming Platforms Accelerate AI Adoption

Netflix has already experimented with AI-assisted visual effects in its original content. One example includes the use of generative AI technology to create a building-collapse sequence in the Argentine series The Eternaut.

Other entertainment companies are also expanding AI initiatives. Amazon has been developing internal AI teams focused on film and television production, while Disney recently signed an agreement with OpenAI to explore the use of artificial intelligence in media workflows.

The growing role of AI in filmmaking has also raised concerns among industry professionals. Workers in the film and television sectors have warned that AI tools could affect employment in editing, visual effects, and other creative roles.

There are also ongoing debates about how AI models should compensate creators when training on copyrighted material. Industry unions and advocacy groups have called for clearer guidelines to ensure that creative professionals are fairly credited and compensated as AI technologies become more integrated into production pipelines.

Palantir and Nvidia Launch Sovereign AI Data Center Architecture

Palantir and Nvidia unveiled a sovereign AI OS reference architecture designed to deliver turnkey AI data centers. The platform integrates Nvidia Blackwell systems with Palantir’s enterprise AI software stack.

By Olivia Grant Edited by Maria Konash Published:
Palantir and Nvidia Launch Sovereign AI Data Center Architecture
Palantir and Nvidia unveil a sovereign AI OS architecture combining Blackwell GPUs with Palantir platforms. Image: GuerrillaBuzz / Unsplash

Palantir Technologies has announced a new sovereign AI operating system reference architecture developed in partnership with Nvidia. The system is designed to provide organizations with a turnkey AI data center environment that integrates hardware infrastructure, data management, and AI application deployment.

The platform, called the Palantir AI OS Reference Architecture (AIOS-RA), combines Nvidia’s AI hardware and software stack with Palantir’s enterprise data and AI platforms. The companies said the architecture delivers a production-ready system capable of running advanced AI workloads while maintaining strict data sovereignty and operational control.

The reference architecture is built on Nvidia’s Enterprise Reference Architectures and is optimized to run Palantir’s full software ecosystem. This includes the company’s Artificial Intelligence Platform (AIP), Foundry data platform, Apollo deployment system, Rubix security framework, and AIP Hub.

Full-Stack AI Data Center Design

The system integrates Nvidia’s Blackwell Ultra AI infrastructure, including servers equipped with eight Blackwell Ultra GPUs and Spectrum-X Ethernet networking designed for AI training and inference workloads.

Alongside the hardware layer, the architecture includes a hardened Kubernetes-based compute environment for running Foundry services such as data cataloging, pipeline development, and analytics workloads.

Palantir’s Rubix platform provides a zero-trust security model for Kubernetes infrastructure, while the Apollo system automates deployment and lifecycle management across distributed environments.

The architecture also incorporates Nvidia’s AI Enterprise software stack, CUDA-X libraries, Nemotron open AI models, and Magnum IO data acceleration tools to improve training and inference performance.

Supporting Sovereign and Edge AI Deployments

The companies said the architecture is particularly suited for organizations operating in sensitive environments where data control and regulatory compliance are critical. These include government agencies, defense organizations, and industries with strict data residency requirements.

By enabling on-premise and edge deployments, the sovereign AI architecture allows organizations to run AI workloads locally without relying on public cloud infrastructure. This approach can also reduce latency for mission-critical applications and enable AI deployments in geographically distributed environments.

“From our first deployment with the United States government and in every deployment since, our software has had to meet the moment in the most complex and sensitive environments where customers must maintain control,” said Palantir Chief Architect Akshay Krishnaswamy.

Nvidia’s enterprise AI platforms team said the collaboration reflects growing demand for integrated AI infrastructure solutions.

“AI is redefining the infrastructure stack — demanding, latency-sensitive and data-sovereign environments require a full-stack architecture,” said Nvidia Vice President Justin Boitano.

The companies said the joint architecture is designed to help organizations convert large volumes of operational data into actionable intelligence while maintaining control over infrastructure, models, and applications.

AI & Machine Learning, Cloud & Infrastructure, Enterprise Tech, News

Musk Confirms Tesla–xAI Collaboration on Digital Optimus AI Agent Project

Elon Musk revealed Digital Optimus, a joint Tesla–xAI AI agent project powered by Grok and Tesla hardware. The announcement deepens links between the companies amid ongoing shareholder litigation.

By Daniel Mercer Edited by Maria Konash Published:
Musk Confirms Tesla–xAI Collaboration on Digital Optimus AI Agent Project
Elon Musk unveils Digital Optimus, a Tesla–xAI AI agent powered by Grok and Tesla hardware. Image: Soly Moses / Pexels

Elon Musk has announced a new artificial intelligence project called Digital Optimus, a collaboration between Tesla and xAI designed to build a computer-controlling AI agent. The system will combine xAI’s Grok large language model with Tesla hardware and software infrastructure.

Musk described the system as a dual-layer AI architecture in which Grok performs high-level reasoning while a Tesla-built component processes real-time computer interactions. The system is designed to analyze screen activity and user inputs, including keyboard and mouse actions, from the previous several seconds.

The approach mirrors psychologist Daniel Kahneman’s “dual-process” theory of cognition. According to Musk, Tesla’s component will handle fast, instinctive responses while Grok functions as the strategic reasoning layer.

In a post on X, Musk said the project will run on Tesla’s AI4 chip alongside Nvidia-powered cloud infrastructure used by xAI. He also introduced the internal nickname “Macrohard,” a reference aimed at Microsoft, and claimed the architecture could eventually emulate complex organizational workflows.

The initiative forms part of Tesla’s previously disclosed $2 billion investment agreement with xAI, expanding the relationship between the electric vehicle maker and Musk’s artificial intelligence startup.

Reversal From Earlier Tesla Strategy

The announcement represents a shift from Musk’s earlier statements about the relationship between the two companies. In September 2024, Musk said Tesla had “no need to license anything from xAI,” arguing that Tesla’s real-world AI systems differed significantly from large language models.

At the time, Musk maintained that Tesla’s AI development was focused on autonomous driving and robotics systems rather than generative AI models. The statement also came shortly after Tesla shareholders filed a lawsuit alleging that Musk had breached fiduciary duties by creating xAI outside Tesla while redirecting resources and talent.

Digital Optimus now explicitly integrates xAI’s Grok model with Tesla hardware, signaling deeper technical collaboration between the companies. The system positions Grok as the central reasoning engine while Tesla supplies the specialized hardware and interface layer.

The partnership also follows Tesla’s $2 billion investment in xAI earlier this year during the startup’s Series E funding round, which valued the company at roughly $230 billion.

Legal and Strategic Context

The growing relationship between Tesla and xAI continues to draw scrutiny from investors and courts. The ongoing lawsuit filed in Delaware Chancery Court alleges that Musk diverted Tesla resources, including engineers and computing hardware, to build xAI as a separate company under his control.

Plaintiffs in the case argue that technologies developed within xAI could have been built inside Tesla, particularly given Musk’s long-standing positioning of Tesla as a leader in artificial intelligence and robotics. The lawsuit seeks remedies that could potentially transfer Musk’s xAI ownership stake to Tesla.

Further developments have intensified the connection between Musk’s companies. Earlier this year, SpaceX completed an all-stock acquisition of xAI that valued the combined entity at roughly $1.25 trillion, with plans for a future public offering. Tesla’s investment in xAI therefore represents an indirect stake in the merged SpaceX-xAI structure.

The Digital Optimus project also aligns with Musk’s broader vision for Tesla’s Optimus humanoid robot. xAI executives previously told investors that their long-term goal was to develop advanced AI systems capable of powering robotics platforms such as Tesla’s Optimus.

Taken together, the project signals a tighter technological link between Tesla’s hardware platforms and xAI’s AI models, as Musk increasingly positions the companies’ technologies to work together in building advanced AI-driven systems.

Google Launches Gemini Embedding 2 Multimodal Model

Google has introduced Gemini Embedding 2, a multimodal embedding model that processes text, images, video, audio, and documents in a unified vector space. The model is available in preview through the Gemini API and Vertex AI.

By Daniel Mercer Edited by Maria Konash Published:
Google Launches Gemini Embedding 2 Multimodal Model
Google launches Gemini Embedding 2, a multimodal model for text, images, video, audio, and documents. Image: Google

Google has released Gemini Embedding 2, its first fully multimodal embedding model built on the Gemini architecture. The model is now available in public preview through the Gemini API and Vertex AI platform.

The new system expands beyond traditional text embeddings by mapping multiple data types into a single vector space. Gemini Embedding 2 can process text, images, video, audio, and documents, enabling applications that require semantic understanding across different media formats.

Embeddings are numerical representations of content that allow AI systems to understand relationships between pieces of information. They are widely used in tasks such as semantic search, recommendation systems, clustering, and retrieval-augmented generation.

By integrating multiple modalities, the model allows developers to simplify pipelines that previously required separate systems for different types of data. According to Google, the unified embedding space can capture relationships between media types, improving the accuracy of multimodal search and analysis.

Multimodal Capabilities and Input Support

Gemini Embedding 2 supports a wide range of inputs across several formats. The model can process up to 8,192 tokens of text, six images per request in PNG or JPEG format, and up to 120 seconds of video in MP4 or MOV format.

The model can also ingest audio directly without requiring transcription and can embed PDF documents of up to six pages. Developers can combine multiple input types within a single request, such as pairing images with text prompts to generate richer semantic representations.

This multimodal capability allows AI systems to analyze complex datasets where information appears in different formats. For example, developers could use a photo or video clip as a search query to retrieve related media assets or documents.

Flexible Embedding Dimensions and Performance Gains

Like earlier Google embedding models, Gemini Embedding 2 uses Matryoshka Representation Learning, a technique that compresses information across nested vector dimensions. This allows developers to adjust embedding sizes depending on performance and storage requirements.

The model’s default vector size is 3,072 dimensions, though developers can scale outputs to smaller dimensions such as 1,536 or 768 for more efficient storage and processing.

Google said the model demonstrates improved performance across text, image, video, and speech tasks compared with previous embedding models. The company noted that the model’s multimodal capabilities could enable new applications in areas such as media search, sentiment analysis, and data clustering.

Early partners are already using the technology to build advanced search tools. Paramount Skydance reported that the model improved its ability to locate video assets based on text queries, allowing the system to identify visual expressions and retrieve related media more accurately.

Gemini Embedding 2 is now accessible through Google’s developer tools and can integrate with AI frameworks including LangChain, LlamaIndex, and vector databases such as Weaviate, Qdrant, and ChromaDB. Google said the model is designed to serve as a foundation for next-generation multimodal AI systems that work with diverse forms of data.

AI & Machine Learning, News

Meta Unveils New MTIA Chips for AI Data Centers

Meta introduced four new in-house MTIA chips designed for AI training and inference as the company accelerates data center expansion. The chips aim to improve performance and reduce reliance on external hardware suppliers.

By Olivia Grant Edited by Maria Konash Published:
Meta Unveils New MTIA Chips for AI Data Centers
Meta unveils MTIA AI chips for data centers, targeting AI training and inference while reducing reliance on GPUs. Photo: Brian Kostiuk / Unsplash

Meta has unveiled four new custom chips designed for artificial intelligence workloads as part of its expanding data center infrastructure. The processors are part of the company’s Meta Training and Inference Accelerator (MTIA) family, a line of chips built to handle specific AI tasks within Meta’s platforms.

The announcement marks the latest step in Meta’s push to reduce reliance on external chip vendors by designing its own silicon optimized for internal workloads. According to Meta Vice President of Engineering Yee Jiun Song, custom chips allow the company to improve price-performance efficiency across its data centers while diversifying its hardware supply chain.

“This also provides us with more diversity in terms of silicon supply and insulates us from price changes to some extent,” Song said.

Meta first revealed the MTIA architecture in 2023 and released a second-generation version in 2024. The new lineup significantly expands the platform as the company scales its AI infrastructure to support recommendation systems and generative AI features across its products.

Four New Chips Target Different AI Workloads

The first of the new chips, MTIA 300, has already been deployed in Meta data centers. It is designed to train smaller AI models used for core platform tasks such as ranking and recommendation algorithms. These systems determine which posts, ads, and videos appear in users’ feeds across services including Facebook and Instagram.

Three additional chips are currently in development. The MTIA 400 is nearing deployment after completing testing, while the MTIA 450 and MTIA 500 are scheduled to become operational by 2027.

Unlike the MTIA 300, the upcoming chips will focus on generative AI inference workloads such as generating images and videos from text prompts. However, Meta said the processors will not be used to train large language models.

Song noted that Meta plans to release new chips roughly every six months as the company rapidly expands computing capacity. Each chip generation is expected to remain in service for more than five years.

Data Center Expansion and AI Infrastructure

The new processors will support Meta’s massive data center expansion. The company is currently building a large facility in Louisiana and additional centers in Ohio and Indiana. Reports also indicate that Meta is exploring leasing space at a major AI data center site in Texas.

The custom chips are manufactured by Taiwan Semiconductor Manufacturing Company, though Meta did not confirm whether production will occur at the company’s new fabrication facilities in Arizona.

Meta’s in-house silicon strategy follows a broader industry trend among major technology companies developing application-specific integrated circuits, or ASICs, for AI workloads. These chips are typically smaller and more energy-efficient than general-purpose GPUs but are optimized for narrower tasks.

Google pioneered this approach with its Tensor Processing Units in 2015, and Amazon followed with custom chips for its cloud services in 2018. Unlike those companies, Meta uses its MTIA chips exclusively for internal operations rather than offering them through a public cloud platform.

Despite building its own silicon, Meta continues to rely heavily on external hardware suppliers. The company recently signed agreements to deploy millions of Nvidia GPUs and up to six gigawatts of AMD GPUs across its data centers.

Song acknowledged that securing high-bandwidth memory remains a potential constraint as AI infrastructure spending increases across the technology industry. However, he said Meta has diversified its supply chain and believes it has secured the resources required for its planned deployments.

AI & Machine Learning, Cloud & Infrastructure, News