intelligence.join(analytics, on=["smarts"], how="full outer")

Palantir Foundry provides a vertically integrated operating system for Generative AI that contrasts with the distributed primitive approach of cloud hyperscalers (AWS/Azure/GCP). While hyperscalers offer raw compute and isolated model APIs, Foundry abstracts the integration layer via the Ontology. This architecture allows engineering teams to focus on function development, logic orchestration, and evaluation pipelines rather than infrastructure plumbing (e.g., vector database synchronization, context window management, and stateful tool execution).
The barrier to entry for Generative AI is low. Any developer can grab an API key from OpenAI or deploy Llama 3 on Hugging Face. The real gap is between a quick demo and a production-grade enterprise application that is reliable, secure, and maintainable.
For Architects and CTOs, the decision usually falls into two camps:
Primitive Approach: Requires manual orchestration of Azure OpenAI, Pinecone (Vector DB), LangChain (Application Logic), and FastAPI (Tool Execution). Engineers are responsible for maintaining the “glue code” that synchronizes state between these disparate services.
Platform Approach: Unifies the Data Layer, Logic Layer, and Application Layer. The Ontology serves as both the semantic graph and the vector store, on top of which we can create operational applications and LLM orchestrations.

The primitive approach maximizes flexibility and is attractive to teams that want fine-grained control over every component. The trade-off is a significant “Integration Tax” in Time-to-Market and ongoing maintenance. Let’s see why, in many enterprise scenarios, Foundry ends up delivering faster and with less operational overhead.
The hardest part of GenAI isn’t the model; it’s the Business Context. To make an LLM smart, you need to feed it clean, structured data about your business—your specific customers, assets, and orders
In a typical Azure-based stack (e.g., Azure OpenAI + Databricks + Azure AI Search):
This works, but you now manage:
Foundry’s Ontology is a semantic layer that maps data to business objects (e.g., Incident, Asset, Plant) and their relationships. That live, governed data graph is directly consumable by AI workflows:
In practice, this often reduces the time spent on building and maintaining bespoke RAG infrastructure - from weeks of custom plumbing to days of Ontology modeling before you can build the first useful AI workflow.
With a DIY Azure/AWS stack, you typically write Python/TypeScript code to:
This approach is powerful and highly customizable, but it tends to produce “glue code” that is hard to debug and expensive to maintain over time.
Foundry’s AIP Logic:
Cloud providers are introducing similar orchestration capabilities (e.g., Azure AI Studio, Prompt Flow, Bedrock Agents), but these still need to be integrated with your data platform, security model, and frontend framework. In Foundry, those elements live in one environment.
The industry is shifting from basic “Chat + RAG” interfaces to agentic workflows-systems where AI uses tools to perform multi-step tasks.
In a custom Azure stack, building an agent typically requires you to:
Palantir AIP Agent Studio simplifies this pattern:
Tools as Ontology Actions
Bind existing Ontology Actions (e.g., Create_WorkOrder, Reschedule_Shipment) as tools. Tool definitions and wiring are handled by the platform, so you don’t manually maintain schemas and function-calling conventions. Additionally you can create custom tools using Typescript and just plug it in the agent studio.
Application context awareness
Agents running inside a Workshop app can read application variables directly (e.g., the currentIncident the user is viewing).
From assistant to automation
The same agent can act:
DIY stacks can implement similar capabilities, but they usually require multiple services and custom code to achieve what Agent Studio offers as a unified environment.
Success in GenAI depends on speed of improvement, not just speed to first release. Models will hallucinate, business rules will shift, and requirements will evolve.
In a AWS/Azure stack, evaluation is often ad-hoc:
Foundry’s AIP Evals builds evaluation into the platform:
Azure and AWS allow you to assemble similar evaluation pipelines, but doing so generally requires combining multiple tools and building custom evaluation frameworks. In Foundry, this capability is available as part of the core development lifecycle.
Cloud IAM models (Azure AD, AWS IAM) are comprehensive but fragmented when applied across a multi-service GenAI architecture. You typically manage:
Any misalignment between these layers can lead to data leakage or slow security reviews.
Foundry’s model:
For organizations with stringent compliance and audit requirements, this integrated approach often shortens security review cycles compared to a fully custom stack.
Typical tech stack:
Say you have the following Ontology objects already created and powering up existing operational workflows. Now you want to add a Agentic workflow that automates the creation of work order based on the incident’s description.
Ontology have 4 objects - Incident, Asset, Work Order, Maintenance Event
Incidents are linked to Assets and Work Orders and Assets ar linked to Maintenance events
There’s also an action associated with Work Order to create a new work order.

In Agent Studio, we can create agent that leverages the ontology objects and links bind the actions as tools and configure the agent via prompts and settings rather than raw JSON schemas.
We can select one of the LLM model available on foundry and give it below system prompt.
You are an Incident Triage Specialist. When an incident is selected, investigate its description and summarize the incident in 1-2 lines. Use the Query objects with SQL tool to find past Maintenance Events for the linked Asset. If a similar issue occurred before, suggest the same fix and output a resolution note in 1-2 lines.
Finally, use the Create Work Order tool to draft a resolution.
If Incident is already closed don't suggest any resolution or create work order.

Setup the tools for LLM to use

Set up AIP Eval, to test the agent on historical incidents,
Use keyword checker to check if LLM is identifying correct keywords and Use LLM-as-a-judge to score quality of generated summary and resolution text.

Now using Workshop, you can build an app that displays Incident list and detail view bound directly to Ontology objects. Whenever a row is clicked on table it open up a overlay, where the agent panel uses the selected Incident as context and summarizes and proposes a work order draft action and ask user for approval for better governance.
The user can additionally give feedback about the reply using the thumbs up/down button.

All of this is done within a single platform, with fewer separate services to integrate or writing any complex to code. This reduces end-to-end delivery time to a few weeks or less, with a more straightforward security and governance story.
Cloud primitives from Azure, AWS, and GCP are powerful and mature. They are often the right choice when teams require deep infrastructure control, want to experiment with highly customized architectures, or already have strong internal platforms and libraries. However, for many enterprises trying to operationalize GenAI on top of complex and sensitive data, the Integration Tax of assembling and maintaining a DIY stack is substantial.
Palantir Foundry tends to accelerate GenAI programs by: Providing a shared Ontology for live, structured business context, offering AIP Logic and Agent Studio to reduce glue code and orchestration complexity, embedding evaluation into the development lifecycle and enforcing security and governance consistently across data, models, and applications.
In practice, this often translates into faster Time-to-Market and faster Time-to-Value, with teams spending more effort on solving business problems and less on stitching together infrastructure.