Announcements

REMINDER: Sign up for the Foundry Newsletter to receive a summary of new products, features, and improvements across the platform directly to your inbox. For more information on how to subscribe, see the Foundry Newsletter and Product Feedback channels announcement.

Share your thoughts about these announcements in our Developer Community Forum ↗.


Claude Opus 4.6 available from Anthropic, AWS Bedrock, and Google Vertex

Date published: 2026-02-12

Claude Opus 4.6 is now available from Anthropic, AWS Bedrock, and Google Vertex on non-georestricted enrollments. For US and EU non-georestricted enrollments, the model is available from AWS Bedrock and Google Vertex.

Model overview

Anthropic’s latest flagship model, Claude Opus 4.6, sets a new standard for advanced LLMs across coding, agentic workflows, and knowledge work. Opus 4.6 builds on its predecessor with stronger coding skills, deeper planning, longer agent autonomy, and improved code review and debugging. It operates more reliably in large codebases and can sustain complex, multi-step tasks with minimal intervention. For more information, review Anthropic's model documentation ↗.

  • Context window: 200,000 tokens
  • Modalities: Text, image
  • Capabilities: Extended thinking, function calling

Getting started

To use these models:

Your feedback matters

We want to hear about your experiences using language models in the Palantir platform and welcome your feedback. Share your thoughts with Palantir Support channels or on our Developer Community ↗ using the  language-model-service tag ↗.


Model Studio, our no-code model training tool, is now generally available

Date published: 2026-02-12

Model Studio, a new workspace that allows users to train and deploy machine learning models, will be generally available and ready for use in all environments the week of February 9. This follows a successful public beta period that started in October 2025.

Model Studio transforms the complex task of building production-grade models into a streamlined no-code process that makes advanced machine learning more accessible. Whether you are a data scientist looking to accelerate your workflow, or a business user eager to unlock insights from your data, Model Studio provides essential tools and a user-friendly interface that simplifies the journey from data to model.

The Model Studio application home page displaying recent training runs and run details.

The Model Studio application home page, displaying recent training runs and run details.

What is Model Studio?

Model Studio is a no-code model development tool that allows you to train models in tasks such as forecasting, classification, and regression. With Model Studio, you can maximize model performance for your use cases by training models with custom data while retaining customization and control over the training process with optional parameter configuration.

Building useful, production-ready models traditionally requires deep technical expertise and significant time investment, but Model Studio changes that by providing the following features:

  • A streamlined point-and-click interface for configuring model training jobs; no coding required.
  • Built-in production-grade model trainers tailored for common use cases such as time series forecasting, regression, and classification.
  • Smart defaults and guided workflows that empower you to get started quickly, even if you are new to machine learning.
  • In-depth experiment tracking with integrated performance metrics that allow you to monitor and refine your models with confidence.
  • Full data lineage and secure access controls built on top of the Palantir platform, ensuring transparency and security at every step.

Who should use Model Studio?

Model Studio is perfect for technical and non-technical users alike. Business users who want to leverage machine learning without coding and data scientists who want to accelerate prototyping and model deployment can both benefit from Model Studio's tools and simplified process. Additionally, organizations can benefit from Model Studio by lowering the barrier to AI adoption and empowering more teams to build and use models.

Getting started

To get started with model training, open the Model Studio application and follow these steps:

  • Select the best model trainer for your use case (time series forecasting, classification, or regression).
  • Choose your input datasets.
  • Configure your model using intuitive options, or stick with the recommended defaults.

After configuring your model, you can launch a training run and review model performance in real time with clear metrics and experiment tracking.

What's next on the development roadmap?

As Model Studio continues to evolve, we are committed to enhancing the user experience. To do so, we will introduce features such as:

  • Enhanced experiment logging for deeper training performance insights, including AI-readable content.
  • An expanded set of supported modeling tasks.
  • Full marketplace support for models built in Model Studio.
  • Direct support for time series inputs on top of datasets.

Learn more about Model Studio.

Tell us what you think

As we continue to develop Model Studio, we want to hear about your experiences and welcome your feedback. Share your thoughts through Palantir Support channels or our developer community ↗.


Create unscoped Developer Console applications

Date published: 2026-02-10

Developer Console applications can now be unscoped, giving you full access to Developer Console features that were previously unavailable with standalone OAuth clients, including:

  • OSDK usage
  • Documentation (OSDK, Platform APIs, development)
  • Marketplace integration
  • Website hosting
  • Metrics

Previously, the only unscoped option was building standalone OAuth clients, and using them meant sacrificing these features entirely. As a result of this improvement, we have deprecated standalone OAuth clients.

How to make your application unscoped

All Developer Console applications are created scoped by default. To make an application unscoped, follow these steps:

  • Navigate to the OAuth & scopes tab in your Developer Console application's sidebar
  • Under the Application scope section, select Unscoped.
  • Confirm your action.
  • Select Switch to unscoped.

Application scope section in Developer Console.

Application scope section in Developer Console.

You can change your application’s scope between scoped and unscoped at any time.

Review the documentation in Developer Console.

Client restrictions

Client enablement project access restrictions are not compatible with Developer Console applications, whether scoped or unscoped. Instead, leave project access and marking restrictions in Control Panel > Third party applications as unrestricted and manage client restrictions directly through Developer Console.

Avoid adding any client restrictions within Control Panel. Set Project access and Marking restrictions to unrestricted.

Avoid adding any client restrictions within Control Panel. Set Project access and Marking restrictions to unrestricted.

Add project and API restrictions to your client within Developer Console  Platform SDK.

Add project and API restrictions to your client within Developer Console > Platform SDK.

Support for marking restrictions in Developer Console is coming soon.

Your feedback matters

We want to hear about your experiences using Developer Console and welcome your feedback. Share your thoughts through Palantir Support channels or on our Developer Community ↗.


Presentation mode in Workflow Lineage

Date published: 2026-02-10

Presentation mode is now available in Workflow Lineage for all enrollments. Use the presentation mode for Workflow Lineage to create and organize visual frames of your workflow graph, making it easier to present your work. To get started, select the project screen icon in the bottom of the graph.

The edit presentation frames entry point in Workflow Lineage.

The edit presentation frames entry point in Workflow Lineage.

Once you edit the presentation mode, you can capture frames. You can do this by saving snapshots of your graph’s current state including node arrangement, layout, colors, and zoom level.

An example of a presentation frame in Workflow Lineage.

An example of a presentation frame in Workflow Lineage.

Manage existing frames in the bottom window. You can easily rename, reorder, or delete frames as needed.

Screenshot of where to delete a  presentation frame.

Screenshot of where to delete a presentation frame.

You can also hide frames you do not want to delete but do not want to show in your presentations.

Example of hiding presentation frames.

Example of hiding presentation frames.

Use the , and . hotkeys to move back and forth through your presentation.

You must save your graph before using presentation mode. We recommend adding text nodes to each frame to guide your audience step by step with custom descriptions.

Try it out and make your workflow presentations more dynamic and engaging. Learn more about presentation mode.


Claude Sonnet 4.5, Claude Haiku 4.5 now available in Japan region

Date published: 2026-02-05

Claude Sonnet 4.5 and Claude Haiku 4.5 models are now available from AWS Bedrock on Japan georestricted enrollments.

Model overviews

Claude Sonnet 4.5 ↗ is Anthropic's latest medium weight model with strong performance in coding, math, reasoning, and tool calling, all at a reasonable cost and speed.

  • Context window: 200k tokens
  • Knowledge cutoff: January 2025
  • Modalities: Text, image
  • Capabilities: Tool calling, vision, coding

Claude Haiku 4.5 ↗ is Anthropic's most powerful small model, ideal for real-time, lightweight tasks where speed, cost, and performance are critical.

  • Context window: 200k tokens
  • Knowledge cutoff: February 2025
  • Modalities: Text, image
  • Capabilities: Tool calling, vision, coding

Getting started

To use these models:

Your feedback matters

We want to hear about your experiences using language models in the Palantir platform and welcome your feedback. Share your thoughts in Palantir Support channels or on our Developer Community ↗ using the language-model-service tag.


Deploy containers with compute modules

Date published: 2026-02-05

Compute modules are now generally available in Foundry. With compute modules, you can run containers that scale dynamically based on load, bringing your existing code, in any language, into Foundry without rewriting it.

What you can build

Compute modules enable several key workflows in Foundry:

Custom functions and APIs: Create functions that can be called from Workshop, Slate, Ontology SDK applications, and other Foundry environments. Host custom or open-source models from platforms like Hugging Face and query them directly from your applications.

Data pipelines: Connect to external data sources and ingest data into Foundry streams, datasets, or media sets in real time. Use your own transformation logic to process data before writing it to outputs.

Legacy code integration: Bring business-critical code written in any language into Foundry without translation. Use this code to back pipelines, Workshop modules, AIP Logic functions, or custom Ontology SDK applications.

An example of a compute module overview in Foundry with information about the job status functions and container metadeta.

An example of a compute module overview in Foundry, with information about the job status, functions, and container metadeta.

Why it matters

Compute modules solve the challenge of integrating existing code into Foundry. Instead of rewriting your logic in a Foundry-supported language, containerize it and deploy it directly. The platform handles scaling, authentication, and connections to other Foundry resources automatically.

Key features include:

  • Dynamic horizontal scaling based on current and predicted load
  • Zero-downtime updates when deploying new container versions
  • Native connections to Foundry datasets, Ontology resources, and APIs
  • External connections using REST, WebSockets, SSE, or other protocols
  • Marketplace compatibility for sharing modules across organizations

Get started

Review the compute modules documentation to build your first function or pipeline.


Deploy document extraction workflows with AIP Document Intelligence

Date published: 2026-02-03

AIP Document Intelligence will be generally available on February 4, 2026 and is enabled by default for all AIP enrollments. AIP Document Intelligence is a low-code application for configuring and deploying document extraction workflows. Users can upload sample documents, experiment with different extraction strategies, and evaluate results based on quality, speed, and cost—all before deploying at scale. AIP Document Intelligence then generates Python transforms that can process entire document collections using the selected strategy, converting PDFs and images into structured Markdown with preserved tables and formatting.

Learn more about AIP Document Intelligence.

Result of Layout-aware OCR  Vision LLM extraction with metrics on cost speed and token usage.

Result of Layout-aware OCR + Vision LLM extraction with metrics on cost, speed, and token usage.

Compare extraction strategies

AIP Document Intelligence provides multiple extraction approaches, from traditional OCR to vision-language models. You can test each method on your specific documents and view side-by-side comparisons of extraction quality, processing time, and compute costs. This experimentation phase helps teams select the right approach for their use case without writing custom code.

Comparison of Vision LLM Extraction vs. Layout-aware OCR  Vision LLM Extraction shows drastic improvement in complex table extraction quality.

Comparison of Vision LLM Extraction vs. Layout-aware OCR + Vision LLM Extraction shows drastic improvement in complex table extraction quality.

Deploy extraction pipelines in one click

Once a strategy is configured, AIP Document Intelligence generates production-ready Python transforms that process documents at scale. The latest deployment uses lightweight transforms rather than Spark, significantly improving processing speed. Workflows that previously took days extracting data from document collections can now complete the same work in hours. Refer to the documentation for more detailed instruction on how to deploy and customize your Python transforms.

Choose a validated extraction strategy and deploy to a Python transforms to batch process documents.

Choose a validated extraction strategy and deploy to a Python transforms to batch process documents.

Maintain quality across diverse document types

Enterprise documents vary widely in structure, formatting, and content density. AIP Document Intelligence handles this diversity through configurable extraction strategies that can adapt to multi-column layouts, embedded tables, and mixed-language content. Users working with maintenance manuals, regulatory filings, invoices have successfully extracted structured data while preserving critical formatting and relationships.

When to use AIP Document Intelligence

AIP Document Intelligence is designed for workflows where document content needs to be extracted and structured for downstream AI applications. This includes:

  • Populating vector databases for retrieval-augmented generation (RAG) systems
  • Extracting tabular data from reports, invoices, or forms for analysis
  • Converting legacy documentation into searchable, structured formats
  • Preparing training data for domain-specific language models

For workflows that require extracting specific entities (like part numbers, dates, or named entities) rather than full document content, upcoming entity extraction capabilities will provide more targeted functionality.

What's next on the development roadmap?

  • Entity extraction from documents: The team is developing capabilities to extract structured entities, such as equipment identifiers, monetary values, dates, and custom domain concepts, directly from documents. This will enable direct population of Ontology objects from unstructured sources.
  • Broader use of AIP Document Intelligence: Allow extraction configurations to be called directly from AIP Logic and Ontology functions to expand Document Intelligence beyond Python transforms to broader workflow automation scenarios.

Your feedback matters

We want to hear about your experiences using AIP Document Intelligence and welcome your feedback. Share your thoughts with Palantir Support channels or on our Developer Community ↗ using the aip-document-intelligence tag ↗.


GPT-5.2 Codex now available in AIP

Date published: 2026-02-03

GPT-5.2 Codex is now available directly from OpenAI for non-georestricted enrollments.

Model overviews

GPT-5.2 Codex ↗ is a coding optimized version of the GPT-5.2 model from OpenAI, with improvements in agentic coding capabilities, context compaction, and stronger performance on large code changes like refactors and migrations.

  • Context window: 400,000 tokens
  • Knowledge cutoff: August 2025
  • Modalities: Text, image
  • Capabilities: Responses API, structured outputs, function calling

Getting started

To use these models:

Your feedback matters

We want to hear about your experiences using language models in the Palantir platform and welcome your feedback. Share your thoughts with Palantir Support channels or on our Developer Community ↗ using the language-model-service tag.


Multi-ontology support in Workflow Lineage

Date published: 2026-02-03

Workflow Lineage just became a lot more robust - you can now visualize resources across multiple ontologies in one unified graph. Instantly identify cross-ontology relationships, spot external resources at a glance, and switch between ontologies without leaving your workflow view.

The Workflow Lineage graph now displays resources across multiple ontologies with visual indicators highlighting nodes from outside the selected ontology.

The Workflow Lineage graph now displays resources across multiple ontologies, with visual indicators highlighting nodes from outside the selected ontology.

What's new?

  • Unified visualization: The Workflow Lineage graph now displays all resource nodes across different ontologies in a single view.
  • Cross-ontology awareness: Object, interface and action nodes from other ontologies appear grayed out with a warning icon, so you can instantly identify their origin.
  • Smart warnings: When multiple ontologies are present, a warning icon appears next to the ontology icon at the upper right side of the graph, keeping you informed at a glance.
  • Ontology switching: View all ontologies present in your graph and easily switch between them using the ontology icon.

Easily switch between ontologies using the ontology blue cube icon to view and navigate all ontologies present in your graph.

Easily switch between ontologies using the ontology (blue cube) icon to view and navigate all ontologies present in your graph.

For action-type nodes from outside the selected ontology, functionality is limited. For example, bulk updates are only possible for function-backed actions within your currently selected ontology.

Share your thoughts

We welcome your feedback about Workflow Lineage in our Palantir Support channels, and on our Developer Community ↗ using the workflow-lineage tag ↗.