OpenAI Expands Codex Into Full Software Development Assistant

OpenAI has upgraded Codex with computer control, memory, and workflow integrations. The update pushes Codex beyond coding into a full development lifecycle assistant.

By Daniel Mercer Edited by AIstify Team Published: Updated:

OpenAI has released a major update to Codex, significantly expanding its capabilities beyond code generation into a broader software development assistant. The update targets the more than 3 million developers using Codex weekly and reflects a growing push toward AI systems that can manage end-to-end workflows rather than isolated tasks.

The new version introduces “computer use,” allowing Codex to interact directly with a user’s device by seeing screens, clicking, and typing via its own cursor. Multiple AI agents can run in parallel on macOS without interfering with user activity. This enables tasks such as testing applications, iterating on front-end designs, and working with tools that lack APIs.

Codex now also includes an in-app browser, enabling developers to annotate web pages and guide the AI in real time. This feature is aimed at improving workflows in frontend and game development, with plans to expand toward broader browser automation. In parallel, Codex gains image generation capabilities through integration with OpenAI’s image model, allowing developers to create and refine visual assets such as mockups and UI concepts directly within the development process.

Deeper Integration Across Developer Tools

The update introduces more than 90 new plugins, expanding Codex’s ability to connect with commonly used tools and services. These include integrations with platforms such as Jira, GitLab, CircleCI, and Microsoft Office, among others. The plugins combine app integrations and external servers to give Codex more context and execution capabilities across workflows.

Within the Codex app, new features support key development tasks such as reviewing pull requests, addressing code review comments, and managing multiple terminal sessions. Developers can also connect to remote development environments via SSH and access files with rich previews for documents, spreadsheets, and presentations. A new summary pane helps track agent actions, sources, and outputs.

These additions are designed to reduce context switching, allowing developers to move between writing code, reviewing outputs, and collaborating with AI in a single environment.

Automation, Memory, and Long-Running Tasks

Codex is also gaining stronger automation capabilities. Users can now reuse conversation threads to preserve context across sessions and schedule tasks for the future. The system can “wake up” and continue work on long-running processes, spanning days or weeks.

A preview of memory features allows Codex to retain user preferences, corrections, and previously gathered information. This enables more personalized and efficient task execution over time, reducing the need for repeated instructions.

Additionally, Codex can proactively suggest next steps based on project context. For example, it can identify unresolved comments in documents, pull updates from tools like Slack or Notion, and generate a prioritized task list to help users resume work quickly.

Availability and Direction

The update is rolling out to Codex desktop users signed in with ChatGPT, with some features such as memory and personalization expanding later to enterprise and regional users. Computer control capabilities are initially limited to macOS.

The release highlights a broader shift in AI development tools. Codex is evolving from a coding assistant into a system that can coordinate tasks, manage workflows, and assist with decision-making across the software lifecycle. This positions it closer to an autonomous development partner rather than a reactive tool, reflecting how developers are increasingly using AI not just to write code, but to manage complex projects end to end.

AI & Machine Learning, News

Anthropic CEO Set for White House Meeting Over Mythos AI

Anthropic CEO Dario Amodei is expected to meet White House officials as tensions ease over its Mythos AI model. The talks signal potential renewed government collaboration.

By Maria Konash Published:
Dario Amodei meets White House as U.S. weighs Mythos AI for cybersecurity despite Pentagon tensions. Image: David Everett Strickler / Unsplash

Anthropic CEO Dario Amodei is scheduled to meet Susie Wiles at the White House on Friday, according to a report by Axios, signaling a possible breakthrough in the company’s dispute with the U.S. Department of Defense. The meeting comes as the Trump administration reassesses the strategic value of Anthropic’s latest AI model, Claude Mythos Preview.

The reported discussions follow a period of tension between Anthropic and the Pentagon, which had previously cut business ties with the company after a contract disagreement. Despite that setback, U.S. officials are now said to be recognizing the model’s advanced capabilities, particularly in cybersecurity contexts where it can simulate or test defense systems against sophisticated threats.

Mythos was introduced earlier this month as part of Anthropic’s “Project Glasswing,” a controlled deployment initiative that allows select organizations to access the model for defensive cybersecurity applications. The system has drawn attention for its ability to model high-level cyberattack scenarios, raising both interest and concern within government circles.

Government Interest in Advanced AI Capabilities

According to the report, the Trump administration is considering broader use of the technology across federal agencies. A separate report by Bloomberg indicated that a version of the Mythos model could be made available to major government departments, suggesting a shift toward closer collaboration despite earlier disputes.

Sources cited by Axios argue that limiting access to such advanced AI systems could undermine U.S. competitiveness, particularly against geopolitical rivals like China. The argument reflects a growing view within policy circles that frontier AI capabilities are becoming strategically important assets, especially in cybersecurity and defense.

Anthropic has not publicly commented on the reported meeting, and Reuters noted it could not independently verify the details. However, the company has previously confirmed ongoing discussions with the administration. Co-founder Jack Clark said earlier this week that conversations with government officials were continuing even after the Pentagon ended its formal relationship with the company.

From Dispute to Potential Partnership

The planned meeting suggests a potential reset in relations between Anthropic and U.S. defense stakeholders. While earlier disagreements led to a breakdown in cooperation, the renewed interest in Mythos highlights how rapidly evolving AI capabilities are reshaping government priorities.

The situation also underscores a broader tension in AI governance: balancing national security interests with concerns about misuse. Models like Mythos, designed to simulate advanced cyber capabilities, can serve both defensive and potentially offensive purposes, making controlled access and oversight critical.

If discussions lead to formal agreements, Anthropic could re-emerge as a key partner in U.S. government AI initiatives, particularly in cybersecurity. The outcome may also influence how other AI developers engage with federal agencies, as governments increasingly seek access to cutting-edge systems while navigating safety and policy constraints.

OpenAI Launches GPT-Rosalind for Biology and Drug Discovery

OpenAI has introduced GPT-Rosalind, a specialized AI model for life sciences research. The system aims to accelerate drug discovery and biological analysis workflows.

By Laura Bennett Edited by Maria Konash Published:
OpenAI unveils GPT-Rosalind, a life sciences model accelerating drug discovery and genomics. Image: OpenAI

OpenAI has introduced GPT-Rosalind, a new domain-specific AI model designed to support research in biology, drug discovery, and translational medicine. The model is being released as a research preview through a controlled access program, reflecting both its advanced capabilities and the sensitivity of its potential applications.

GPT-Rosalind is built to address one of the most complex challenges in life sciences: the fragmented and time-intensive workflows that underpin early-stage discovery. Developing a new drug can take 10 to 15 years, with early research decisions having compounding effects on downstream outcomes. The model is designed to help scientists navigate large volumes of literature, datasets, and experimental variables more efficiently, while also generating and testing new hypotheses.

The system is available through ChatGPT, Codex, and the API for qualified enterprise users. OpenAI is also launching a Life Sciences research plugin for Codex, enabling integration with more than 50 scientific databases and tools. Early collaborators include major pharmaceutical and research organizations such as Amgen, Moderna, Allen Institute, and Thermo Fisher Scientific.

Built for Complex Scientific Workflows

Unlike general-purpose AI models, GPT-Rosalind is optimized for reasoning across specialized domains including chemistry, genomics, protein engineering, and disease biology. It is designed to assist with multi-step research tasks such as literature review, experimental planning, sequence analysis, and data interpretation.

OpenAI reports that the model shows improved performance on benchmarks related to biochemical reasoning, including protein structure analysis, phylogenetics, and experimental design. It also demonstrates stronger ability to use external tools and databases within complex workflows, a critical requirement for real-world scientific research.

In industry evaluations, GPT-Rosalind achieved leading results on bioinformatics benchmarks such as BixBench and outperformed earlier models on several tasks in LABBench2, including molecular cloning design. In collaboration with Dyno Therapeutics, the model also ranked above most human experts on certain RNA prediction tasks.

Controlled Access and Research Integration

Given the potential risks associated with advanced biological research tools, OpenAI is deploying GPT-Rosalind through a “trusted access” model. Organizations must meet criteria related to legitimate scientific use, governance, and security controls before gaining access. The rollout initially focuses on enterprise users in the United States.

The accompanying Life Sciences plugin provides an orchestration layer for scientific workflows, connecting researchers to public datasets, literature sources, and domain-specific tools. This allows the model to move beyond static responses and actively support research processes such as protein structure lookup, sequence search, and dataset discovery.

OpenAI said the system was developed with enhanced security measures and is intended for use in controlled research environments. During the preview phase, usage will not consume standard API credits, though safeguards are in place to prevent misuse.

The release marks the first step in a broader effort to build AI systems tailored to scientific discovery. OpenAI says future iterations will expand the model’s capabilities for long-horizon, tool-intensive workflows, with ongoing collaborations across academia, biotech, and national laboratories aimed at advancing areas such as protein and catalyst design.

AI & Machine Learning, News, Research & Innovation

NTU Singapore Develops AI-Powered Biochip for Rapid Disease Detection

Researchers at NTU Singapore have developed an AI-powered biochip that detects disease-linked microRNAs in minutes. The system could enable faster, more precise diagnostics.

By Laura Bennett Edited by Maria Konash Published:
NTU Singapore unveils AI biochip detecting microRNA in 20 minutes for faster, high-accuracy diagnostics. Image: Nanyang Technological University

A research team from Nanyang Technological University has developed a new AI-powered biochip capable of rapidly detecting microRNAs, tiny genetic markers linked to diseases including cancer and cardiovascular conditions. The system, described in the journal Advanced Materials, combines nanophotonic sensing with automated image analysis to significantly reduce diagnostic time.

The platform can analyze a small blood sample and detect multiple microRNA biomarkers in about 20 minutes, compared to several hours required by traditional methods such as PCR (polymerase chain reaction). Researchers say the system achieves high sensitivity and accuracy, detecting extremely low concentrations of microRNAs, even down to a few molecules.

The work was led by Associate Professor Chen Yu-Cheng, who said the goal is to create a scalable diagnostic platform capable of screening multiple disease markers quickly and accurately. Initial testing focused on microRNAs linked to non-small cell lung cancer, demonstrating the system’s ability to identify multiple targets simultaneously without complex sample preparation.

How the Technology Works

At the core of the system is a nanophotonic chip embedded with nanocavities, microscopic structures that enhance fluorescent signals when microRNAs bind to specific probes. These cavities amplify weak signals, making it possible to detect even single molecules.

The chip is paired with an AI imaging system that captures and analyzes thousands of signals in a single snapshot. Using a deep learning model based on Mask R-CNN, the system automatically identifies and classifies microRNA signals, removing the need for manual counting and reducing the risk of human error.

Unlike conventional approaches, which often require amplification or labeled probes, the NTU platform directly measures microRNAs in liquid samples. The researchers report accuracy levels exceeding 99 percent across test scenarios, including experiments using both cancer cell extracts and synthetic samples.

Toward Faster, Scalable Diagnostics

The team has also built a compact prototype that includes a camera and a mobile application for real-time analysis. This setup could support point-of-care testing, where results are generated quickly without the need for specialized laboratory infrastructure.

Researchers believe the platform could eventually be adapted for large-scale screening, potentially analyzing hundreds or thousands of biomarkers from blood, saliva, or urine samples. This could open the door to earlier disease detection, better monitoring of treatment response, and more personalized healthcare.

Independent experts note that microRNAs have long been considered promising biomarkers but have been difficult to measure reliably due to their small size and similarity. A system that can accurately detect multiple microRNAs could improve clinical decision-making, particularly in oncology and chronic disease management.

The project is supported by Singapore’s research funding programs, and the team has filed a technology disclosure through NTU’s commercialization arm. Future work will focus on clinical validation and scaling the platform for broader use in healthcare and pharmaceutical research.

AI & Machine Learning, News, Research & Innovation

Anthropic Introduces Identity Verification for Claude Users

Anthropic is rolling out identity verification for Claude users to strengthen safety and compliance. The move introduces ID checks for certain features and use cases.

By Daniel Mercer Edited by Maria Konash Published:
Anthropic adds ID verification to Claude via Persona to curb abuse and meet compliance. Image: Farhat Altaf / Unsplash

Anthropic has begun rolling out identity verification requirements for users of its Claude platform, signaling a stronger push toward safety, compliance, and misuse prevention as AI systems become more powerful. The new process will apply selectively, with users prompted to verify their identity when accessing certain features or during routine integrity checks.

The verification system is powered by Persona, a third-party provider specializing in digital identity checks. Users are required to submit a government-issued photo ID and, in some cases, complete a live selfie capture using a phone or webcam. The process typically takes a few minutes and is designed to confirm identity without collecting unnecessary data.

Anthropic says the verification rollout is tied to broader efforts to enforce its usage policies and comply with legal obligations, particularly as advanced AI capabilities raise concerns about misuse. The company emphasized that verification data is used solely for identity confirmation and not for training AI models or other secondary purposes.

Accepted identification includes passports, driver’s licenses, and national ID cards, provided they are physical, valid, and clearly legible. The system explicitly rejects digital IDs, photocopies, or non-government credentials such as student cards or employee badges. Failed verification attempts can result from poor image quality, expired documents, or technical issues, though users are allowed multiple retries.

From a data handling perspective, Anthropic positions itself as the data controller, while Persona processes the information on its behalf. Importantly, identity documents and selfies are stored on Persona’s systems rather than Anthropic’s infrastructure. The company says all data is encrypted in transit and at rest, and Persona is contractually restricted from using the data beyond verification and fraud prevention purposes.

Anthropic also clarified that identity data will not be shared with third parties for marketing or advertising. Access is limited to verification and compliance workflows, with exceptions only in cases where legal obligations require disclosure.

Accounts may still face suspension or bans after verification if they violate platform rules, including repeated misuse, operating from unsupported regions, or breaching terms of service. Users who believe enforcement actions are incorrect can submit appeals for review.

Why This Matters

Identity verification marks a shift toward stricter governance in AI platforms. As models gain more advanced capabilities, companies face increasing pressure to prevent harmful use cases, particularly in areas like cybersecurity, fraud, and misinformation.

For businesses and developers, this introduces an additional compliance step that may affect onboarding and user experience. However, it could also improve trust in AI systems by reducing anonymous misuse and enforcing accountability.

For users, the tradeoff is clear: access to more powerful features may require sharing sensitive identity information, even if safeguards are in place.

Context

Anthropic’s move aligns with a broader industry trend toward tighter controls on AI access. Competitors like OpenAI have also explored verification, tiered access, and usage restrictions for advanced AI tools.

The rollout comes alongside Anthropic’s increasing focus on safety frameworks, including recent efforts to limit high-risk capabilities and introduce safeguards in newer models. As regulators worldwide examine AI risks more closely, identity verification may become a standard requirement across leading platforms.

AI & Machine Learning, News

Claude Opus 4.7 Launches With Stronger Coding, Vision Capabilities

Anthropic has released Claude Opus 4.7 with improved coding, vision, and reliability features. The update also introduces new safety controls for cybersecurity use cases.

By Daniel Mercer Edited by Maria Konash Published:
Anthropic unveils Claude Opus 4.7 with stronger coding, vision, and safety for enterprise AI. Image: Anthropic

Anthropic has announced the general availability of its latest AI model, Claude Opus 4.7, positioning it as a direct upgrade over Opus 4.6 with significant gains in advanced software engineering and multimodal capabilities. The release comes as the company continues to iterate toward more powerful systems, while cautiously testing safety mechanisms ahead of broader deployment of its more advanced Claude Mythos Preview.

Anthropic says Opus 4.7 performs better on complex, long-running coding tasks, allowing users to delegate work that previously required close oversight. The model shows improved instruction-following, with a more literal interpretation of prompts, which may require developers to adjust existing workflows. It also introduces stronger self-verification behavior, meaning it attempts to validate its outputs before returning results.

A key upgrade is in multimodal performance. Opus 4.7 can process images up to 2,576 pixels on the long edge, more than triple the resolution of earlier Claude models. This enables use cases such as analyzing dense screenshots, extracting data from diagrams, and supporting pixel-precise design workflows. Internally, Anthropic reports improved performance in domains such as finance, legal reasoning, and document analysis, including stronger results on third-party benchmarks measuring economically valuable knowledge work.

The model is now available across Anthropic’s ecosystem, including its API and integrations with platforms such as Amazon Bedrock, Google Vertex AI, and Microsoft Foundry. Pricing remains unchanged from Opus 4.6 at $5 per million input tokens and $25 per million output tokens.

Anthropic is also introducing new controls for developers, including an “xhigh” effort level that balances reasoning depth and latency, as well as task budgeting tools to manage token usage in longer workflows. Additional features in its coding environment include an automated code review tool and expanded autonomous execution modes.

On safety, Opus 4.7 is the first model released under Anthropic’s new cybersecurity framework introduced with Project Glasswing. The system includes safeguards that detect and block high-risk cyber-related queries. For vetted professionals, the company has launched a Cyber Verification Program to allow legitimate security research and testing.

Why This Matters

The release reflects a broader shift in enterprise AI toward reliability and autonomy. Improvements in coding and long-task execution make models like Opus 4.7 more viable for real-world development workflows, reducing the need for constant human supervision.

Enhanced vision capabilities also expand AI’s role in design, analytics, and operations, where interpreting complex visuals is critical. At the same time, the introduction of cybersecurity safeguards highlights growing concerns about misuse as models become more capable.

For businesses, the combination of higher performance and unchanged pricing could accelerate adoption, particularly in software development, finance, and knowledge work automation.

Context

Anthropic has been steadily iterating on its Claude model family, competing with offerings from companies like OpenAI and Google. The company’s strategy emphasizes safety alongside capability, often limiting access to its most advanced systems while testing controls on intermediate models.

The mention of Claude Mythos Preview suggests Anthropic is preparing for a next generation of more powerful AI systems, but is proceeding cautiously due to potential risks, particularly in cybersecurity.

The addition of finer-grained control over compute effort and token usage also reflects an industry-wide trend toward giving developers more control over cost-performance tradeoffs, as AI systems are increasingly deployed in production environments.

Exit mobile version