Q

2.0

Products Data Management Athyna · Cloud Data Studio
Athyna — Cloud Data Studio AI Data Team Plain English Semantic Layer · SemantIQ

Data prep and transformation, in plain English — with your AI Data Team.

Ask in plain English. Build the workflow as you talk. No-Code. No Hassle.

Describe what you need. Your AI Data Analyst and AI Data Engineer turn plain English into a trusted, repeatable workflow — grounded in the Semantic Layer powered by SemantIQ. Your data never moves. Save the output as a Virtual Live Dataset or share it as a DaaS API — a reusable asset, not a dead screenshot.

Built for analysts, stewards, business users, and data scientists who've had enough of waiting on data engineering. Athyna turns a data request into a data product — before your coffee gets cold.

Athyna
Conversational data prep · live
Building
Athyna conversational workflow building animation Animation demonstrating a user describing data preparation tasks in plain English — dedup customers, encrypt SSN, impute null age values with median, and group by age demographics — and Athyna constructing a workflow of four connected nodes in real time. USER PROMPT ➔ "Dedup customers" "Encrypt SSN" "Impute null age values with median age" "Group by age demographics" BUILDING WORKFLOW ↓ DEDUP_CUST Dedup Customers MASK_SSN Encrypt SSN IMPUTE_AGE Impute Age (median) GRP_BY_DEMO Group by Demographics Workflow ready Save as Virtual Live Dataset · or share as DaaS API
Athyna in production
20×
Faster Prep
10×
Lower TCO
0
Lines of Code
Zero
Data Copy
<500ms
Median Transform
Why Athyna Exists

Data prep is 60–80% of the work.

Analysts spend most of their day wrangling files, chasing nulls, joining sheets, and writing one-off SQL that nobody reuses. The work is slow, repetitive, and it stays trapped in notebooks and Slack DMs. By the time the answer is ready, the question has changed.

Athyna compresses that entire cycle. Pair with the AI Data Analyst or AI Data Engineer, describe the prep in English or drag-and-drop a visual workflow, and Athyna compiles it, runs it on an in-memory SQL engine, and saves the result as a governed, reusable data product. The Semantic Layer — powered by SemantIQ — keeps every transform grounded in the same business definitions your BI, governance, and ML pipelines already use. The work compounds instead of disappearing.

Data prep isn't a chore. It's a product in the making.

The old way
Every ad-hoc analysis dies in someone's notebook.
01
Business users wait weeks for a data engineer to wrangle three files into one clean table.
02
One-off Python scripts sprawl across laptops. No lineage. No reuse. No governance.
03
The same cleanse-and-join gets redone by four different analysts — with four different answers.
The AI Data Team in Athyna

Two AI agents. One studio.

Athyna is where two members of the xAQUA AI Data Team do their work — the AI Data Analyst and the AI Data Engineer. Both speak plain English. Both work against the same Semantic Layer. Both leave behind a versioned, reusable workflow.

Athyna with the AI Data Analyst and AI Data Engineer working alongside human users
📊
AI Data Analyst
For analysts, stewards, and business users
The AI Data Analyst profiles columns, suggests transformations, writes SQL from your description, and flags quality issues before they hit your dashboard. Ask in English; it shows you the result and the lineage. Athyna is one of three modules where this agent works — alongside Infera (text-to-SQL) and Reeve (data product catalog).
Profile & explore NL → SQL Quality flags Visual workflows
⚙️
AI Data Engineer
For data engineers and power users
The AI Data Engineer optimizes SQL, masks PII, builds joins across federated sources, and promotes Athyna workflows into production pipelines via Composer. Hand it a workflow; it hands you a versioned, governed DAG with quality gates and observability wired in.
SQL optimization PII masking Federated joins Workflow → pipeline
Eight Capabilities · One Studio

Powered by the AI Data Team and the Semantic Layer.

Connect anywhere. Build a workflow. Transform with AI. Validate. Publish as a live data product. All from one codeless canvas — with your data staying where it is.

🔌
Universal Connectivity
Connect Snowflake, Databricks, Oracle, Redshift, S3, Salesforce, files — anything. Schema and metadata inferred automatically.
💬
Natural-Language Prep
Describe the transform. The AI Data Analyst converts your intent into SQL and runs it on an in-memory engine.
🔀
Visual Workflow Builder
Chain tasks — filter, pivot, join, mask, impute — into reusable, versioned workflows. Validate with sample data before you run on production.
🧠
Semantic Layer
Every transform grounded in SemantIQ — so "customer," "revenue," "in-force" mean the same thing here as in BI, governance, and ML.
In-Memory SQL Engine
Transforms run on an embedded columnar engine. Millions of rows in sub-second.
🔒
Zero Copy Architecture
Your data stays where it is. Athyna queries in place — no egress, no shadow lake, no compliance drama.
📦
Virtual Live Datasets
Save the workflow, not just the output. Downstream stays live as sources update.
🔗
Data-as-a-Service API
Publish clean datasets as live API endpoints. Partners and apps consume them fresh.
One Canvas · Every Task

Connect anywhere. Build any workflow.

From Excel to Snowflake — with twelve built-in tasks a few clicks away. Profile, schema, and visualization update live as you build.

Athyna interface — universal connectivity, task library, and live profile + schema + visualization tabs

Connect any source — Snowflake, Databricks, Oracle, Redshift, MySQL, Salesforce, S3, Excel — and chain together a workflow from a full library of tasks. Profile, Data, Schema, and Visualization tabs update at every step.

Twelve Tasks · Zero Code

Every transform you'd script by hand.

A point-and-click toolkit covering the work that fills 80% of an analyst's day. Each task is a versioned, reusable step in your workflow.

🎚️
Filter
🔁
Pivot
📊
Group by Summary
🧬
Merge Columns
✂️
Split Columns
↕️
Sort Columns
🏷️
Rename Columns
🛡️
Mask Columns
📈
Outlier Detection
🧮
Impute Columns
🧪
Clone & Transform
⚙️
SQL Query
What Athyna Does

A studio built for the 80% of data work.

💬
Ask, Don't Script
Describe the transform the way you'd describe it to a colleague. Athyna compiles your intent into SQL and runs it — 20× faster than writing the script yourself.
  • Natural-language to SQL with semantic grounding
  • Context-aware suggestions on every column
  • Edit, rerun, and version every prompt
🔀
Visual Workflows
For those who prefer a canvas — chain tasks visually into reusable, versioned workflows. Validate with sample data and preview intermediate results before running on production.
  • Twelve built-in tasks: filter, pivot, group, merge, split, mask, impute, outlier detection, and more
  • Step-by-step preview at each stage
  • Workflow validation, version control, and sharing
Instant Execution
Athyna runs on an in-memory columnar engine — DuckDB-class speed. Profile, pivot, and join millions of rows in milliseconds. No warehouse warm-up. No waiting.
  • Sub-second transforms at warehouse scale
  • No extract — query in place when possible
  • Interactive feedback, every step
🔍
See Your Data at Every Step
Profile, schema, and distributions update automatically as you transform. No guessing. No surprises at the end.
  • Column-level profile and pattern view
  • Distribution, null, and cardinality charts
  • Side-by-side before/after comparison
📦
Publish as a Data Product
Save your prep as a Virtual Live Dataset. Reeve publishes it to the catalog. Your analysis becomes something the whole org can find, reuse, and trust.
  • One-click publish to the data product catalog
  • Live refresh on source change
  • Consume via SQL, BI, or DaaS API
🛡️
Inherit Governance, Always
Every transform runs inside the semantic layer. Masking, classification, access, and lineage follow the data — you don't have to wire it up.
  • PII masking applied automatically
  • End-to-end column lineage captured
  • Role-based access preserved downstream
How It Works

Connect. Explore. Transform. Observe. Decide.

Five steps from raw data to actionable insight — with governance built in and your data never leaving your boundary.

1
🔌
Connect
Link any source — warehouse, file, API, object store. Schema and metadata inferred automatically.
2
🔍
Explore
Dive deep. Ask questions. Generate insights. Get a feel for what's there before you touch it.
3
Transform
Refine with AI suggestions, SQL-like commands, or natural language. Athyna writes the SQL — you review.
4
👁️
Observe
Profile, schema, and behavioral patterns update live. Spot drift, nulls, and skew before you commit.
5
🎯
Decide
Save as a Virtual Live Dataset. Publish to the catalog. Share via DaaS API. The org reuses it.
Prompt · Compile · Run

From a sentence to a clean dataset.

You describe the prep the way a colleague would describe it — "Mask SSN," "Create Customer Age Group," "Calculate the corrected age & income" — and the AI Data Analyst grounds the request against the Semantic Layer, generates SQL, runs it in memory, and shows you the result with full lineage and profile.

Every prompt becomes a task in your workflow. Chain them. Branch them. Reuse them. The whole sequence becomes a Virtual Live Dataset — so the next analyst doesn't have to start from scratch.

  • Works on any connected source — warehouse, file, SaaS API
  • Every prompt is versioned, auditable, and reproducible
  • Save the workflow as a Virtual Live Dataset — reuse across teams
  • Masking and classification applied automatically
Athyna · member_dataset.prep
# Described by analyst · compiled by Athyna
ask "merge 3 member files, dedupe on member_id,
     coalesce SSN from first non-null,
     flag rows where DOB > today"

── compiled SQL ──
WITH unified AS (
  SELECT * FROM members_q1
  UNION ALL SELECT * FROM members_q2
  UNION ALL SELECT * FROM members_q3
),
deduped AS (
  SELECT member_id,
    COALESCE(ssn_a, ssn_b, ssn_c) AS ssn,
    dob, ROW_NUMBER()
    OVER (PARTITION BY member_id) AS rn
  FROM unified
)
SELECT *, (dob > CURRENT_DATE) AS invalid_dob
FROM deduped WHERE rn = 1;

── execution ──
  scanned   8,213,044 rows
  duration  340ms         ✓ PASS
  output    7,913,812 clean rows
  flagged   14 invalid DOBs

STATUS: READY · save as Virtual Live Dataset?
Athyna natural language workflow — workflow nodes, SQL editor, and member conversation panel

Workflow tasks — Determine Transaction Outliers · Mask SSN · Create Customer Age Group · Transaction by Demography — chained into a reusable workflow. Ask in English; the AI Data Analyst writes the SQL; Cezu narrates what happened.

From Workflow to Reusable Asset

Two ways your prep earns its keep.

A clean dataset shouldn't die in a notebook. Athyna gives every workflow two lives — discoverable in the catalog and consumable as an API.

📦
Virtual Live Dataset
Save the workflow · publish to the catalog
The clean, curated workflow gets saved as a Live Data Product. Anyone can search, access, query, and analyze it from the xAQUA catalog. The workflow runs at consumption time — so the data is always fresh.
  • Searchable in Reeve, the data product catalog
  • Refreshes live as sources update — no stale snapshots
  • Lineage, masking, and classification carry through
  • Reuse across analysts, BI tools, and pipelines
🔗
Data-as-a-Service API
Share live data with apps and partners
Publish any Virtual Live Dataset as a governed REST endpoint. Internal apps, downstream services, and external partners consume fresh, masked, policy-aware data — without ever touching raw records.
  • One-click endpoint generation
  • Token-based access, row- and column-level policy
  • Automatic PII masking by classification
  • Live refresh — partners always read current data
Use Cases

Where Athyna moves the needle.

💼
Analyst Self-Service
Finance · Ops · Marketing analytics
Analysts merge messy source files, reconcile definitions across business units, and build clean datasets on demand — without filing an engineering ticket.
✓ Days-to-weeks compressed into minutes
🧠
ML Training Data Prep
Data Scientists · ML Engineers
Profile, sample, split, and engineer features visually. Publish the training set as a Virtual Live Dataset so every retrain reads from the same clean source.
✓ Reproducible training data, version-controlled
🏛️
Legacy Data Migration
Modernization · Re-platforming
Reshape, cleanse, and map data from legacy systems into the new model — interactively. What used to take 18 months of custom scripts finishes in weeks.
✓ Migrations that ship, not stall
🔗
Partner Data Sharing
DaaS · API delivery
Build a clean, masked, governed dataset and publish it as a live API. Partners consume fresh data without ever touching raw records.
✓ Monetizable, compliant data products
Why Athyna

Not another data prep tool.

Athyna is a module of a unified platform — not a standalone prep tool bolted onto someone else's catalog.

AI Data Team in your studio
The AI Data Analyst and AI Data Engineer work alongside you — not as bolt-on assistants, but as members of the same team. Different cockpits, same governed engine, same compounding asset.
Grounded in the Semantic Layer
Every prompt is grounded in SemantIQ — the same semantic layer your BI, governance, and ML pipelines use. "Customer," "revenue," "in-force" — one definition, everywhere.
Zero copy · your data stays put
Athyna queries in place. No shadow lake. No egress. Performance is fast because the engine is embedded — not because we copied your data somewhere convenient.
Workflows, not notebooks
Every transform is a versioned task in a reusable workflow. Validate with sample data. Share across teams. Your work compounds — it doesn't die in someone's laptop.
In-memory speed · no cluster to warm up
Athyna runs on an embedded columnar SQL engine. Profile and transform millions of rows in sub-second — no Spark ticket, no warehouse queue.
English or drag-and-drop · same engine
Business users speak. Power users drag. Engineers edit SQL. Every path lands on the same execution engine — and the same governed output.
Validation built in
Test workflows with sample data before you run them on production. Quality gates, lineage, and profile checks happen before publish — not after the dashboard breaks.
Governance inherited, not bolted on
PII masking, lineage, classification, and access are applied automatically. You don't wire it up — the platform does.
Deploys where your data lives
Private VPC, air-gapped, on-prem, or cloud. No data leaves your boundary. Transforms happen next to the data.
Athyna + Composer

Athyna explores. Composer productionizes.

Athyna is the interactive studio. Composer — xAQUA's pipeline builder — picks up where you stop. Promote any Athyna workflow to a scheduled, governed, CI/CD'd Airflow DAG with a single click.

Ask the AI Data Engineer in plain English: "Promote yesterday's member prep into a daily pipeline" — it wraps the workflow, adds quality gates, versions it in Git, and deploys to production.

Meet the AI Data Engineer →
Ask the AI Data Engineer AI Data Engineer
👤
Promote yesterday's member prep into a daily pipeline.
⚙️
Converted Athyna workflow to Airflow DAG:

· members_daily_prep.dag — 5 operators
· Schedule: 0 2 * * * · SLA 10min
· Quality gate: dedupe > 95%, null < 2%
· Git: main/pipelines/members_daily
· Deployed to K8s · observability wired

Running tonight. First output ready by 02:10.

Ready to stop waiting on prep?

See Athyna profile a live dataset, run a natural-language transform, and publish the result as a governed data product — in under fifteen minutes.

Or email us directly at sales@xaqua.io