[ Open Compute Format for Ai ]

Reuse, Optimize, and Govern your AI Compute

Define once, run anywhere—Xorq lets teams catalog, share, and serve ML in pure Python with declarative, portable, and governed expressions.

Simplify ML data processing with open source Xorq

30+ Source integrations
Powered by open source
Reuse
ML Training
Define and run portable, multi-step ML training pipelines with deferred caching & observability.
Reuse
Inference
Serve and reuse models anywhere with low-latency Arrow Flight and consistent scoring logic.
Optimize
Compute Offloading
Push heavy logic to cheaper backends like DuckDB or Postgres with into_backend.
Optimize
Hash-based Caching
Avoid redundant computation by caching expressions locally or to the cloud.
Govern
Lineage and Observability
Track column-level lineage and collect OTel metrics for every expression automatically.
[ meet xorq ]

The Missing Layer in AI Infrastructure

Modern storage formats like Iceberg solve data reproducibility.
But compute—the transformations, features, models—remains brittle and siloed.

Xorq is the compute catalog: a unified layer to declare, reuse, and observe every expression of compute—across engines, teams, and environments.

Legacy

With Xorq

Thomas McGeehan, 66 Degrees
"Xorq is a new compute framework with quietly radical ideas. Deferred pipelines. Cross-engine execution. Portable Python UDFS. Arrow-native caching."
Wes McKinney, Posit PBC
"I'm excited about Xorq! Ibis and Apache DataFusion brought together to orchestrate multi-engine AI compute, all powered by Apache Arrow"
Julien Hurault, Boring Data
"Xorq was the only Python library that provided the openness, composability, and simplicity we needed to build the simplest, MCP-ready semantic layer on Earth."
Daniel Ashy, Yendo
"The Xorq framework greatly simplified our ML pipelines, improving performance 10x and reducing the compute and storage resources required to run them."
Reusability

Next-level Reusability

You've never experienced reuse like this.

Reusable Xorq "expressions" integrate seamlessly with Python. Easy to share and discover with compile-time validation and clear lineage.

Governance

Your AI Engineering Catalog

Innovate fast and confidently. Xorq automatically catalogs your AI compute artifacts as you go to facilitate reuse, troubleshooting, and quality.

Optimization

Fast Iteration. Lower Compute Costs.

By the time you have it working locally, it's already optimized for production, with caching, millisecond data exchange, and other high-performance features.

Collaboration

Put an End to ML Silos

One platform to support your entire AI engineering organization. Securely share and discover reusable artifacts across individuals, teams, and partners.

[ The Compute Catalog ]

Accelerate AI Innovation

A compute catalog is a powerful asset. Share and discover reusable expressions, combine them into new composite expressions, and observe and troubleshoot their behavior.

Just build, run, and serve.

Reusability, optimization, and governance are automatic.
Build

Declare and compile multi-engine Python

Xorq catalogs every expression: versioned, composable, lineage-tracked artifacts represented as YAML format.

Run

Build Once, Run Anywhere.

Run Xorq expressions as portable UDXFs with cross-engine optimizations, caching & observability.

Serve

High-performance Catalog Servers

Create portable Catalog Servers, with millisecond data transfer via Apache Arrow Flight.

Use cases

Plug xorq into your stack

From experimentation to production, xorq fits into the workflows your team already uses—without forcing new tools or rewrites.

View all
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Install xorq

Get started with Xorq

Start yourself or request your free build.

Install Xorq

Spin up your first Xorq engine in minutes—locally or in the cloud.

Install with pip
pip install xorq
Or use nix instead
nix run github:xorq-labs/xorq
# of stars
# of followers

Request a demo

Not sure where to start? We’ll build your first Xorq UDXF—free. Tailored just to your stack, your use case, and your goals.

Start with a template

Start with a pre-built pipeline tailored to real-world ML tasks—modify it, run it, and make it yours in minutes.

browse all templates
30+ ML integrations
the pipeline

Tutorials, Insights & Updates

Ideas and insights from a team building the most portable pipeline runtime.

FAQs

Find answers to common questions about Xorq below.

What is xorq?

Xorq is an open-source compute format for AI; a unified layer to declare, catalog, reuse, and observe every expression of compute—across engines, teams, and environments.

Why was Xorq created?

Xorq was created by a team of data scientists on a mission to help others accelerate AI innovation by simplifying and standardizing the declaration, reuse, portability, and governance of ML data processing.

Is Xorq easy to use?

Xorq is very easy to adopt. The open source library enhances Python with a declarative pandas-style syntax for defining AI data processing. It abstracts away implementation and data engineering details that normally complicate AI data processing and slow down production deployment.

Where can I learn more?

You can explore our documentation for detailed guides and tutorials. Additionally, our blog features insights and updates on xorq's capabilities. Join our community for discussions and shared learning experiences.

Still have questions?

We're here to help you!

Simpler ML,
Faster AI innovation.

Try Xorq today, or request a walkthrough.

Copied to clipboard!