SQL with superpowers for analytics
The Trilogy language is an experiment in better SQL for analytics - a streamlined SQL that replaces tables/joins with a lightweight semantic binding layer and provides easy reuse and composability. It compiles to SQL - making it easy to debug or integrate into existing workflows - and can be run against any supported SQL backend.
pytrilogy is the reference implementation, written in Python.
- Speed - write faster, with concise, powerful syntax
- Efficiency - write less SQL, and reuse what you do
- Fearless refactoring - change models without breaking queries
- Testability - built-in testing patterns with query fixtures
- Easy to use - for humans and LLMs alike
Trilogy is especially powerful for data consumption, providing a rich metadata layer that makes creating, interpreting, and visualizing queries easy and expressive.
We recommend starting with the studio to explore Trilogy. For integration, pytrilogy can be run locally to parse and execute trilogy model [.preql] files using the trilogy CLI tool, or can be run in python by importing the trilogy package.
Tip
Try it now: Open-source studio | Interactive demo | Documentation
Install
pip install pytrilogySave in hello.preql
const prime <- unnest([2, 3, 5, 7, 11, 13, 17, 19, 23, 29]);
def cube_plus_one(x) -> (x * x * x + 1);
WHERE
prime_cubed_plus_one % 7 = 0
SELECT
prime,
@cube_plus_one(prime) as prime_cubed_plus_one
ORDER BY
prime asc
LIMIT 10;Run it in DuckDB
trilogy run hello.preql duckdbFor humans and AI. Enjoy flexible, one-shot query generation without any DB access or security risks.
(full code in the python API section.)
query = text_to_query(
executor.environment,
"number of flights by month in 2005",
Provider.OPENAI,
"gpt-5-chat-latest",
api_key,
)
# get a ready to run query
print(query)
# typical output
'''where local.dep_time.year = 2020
select
local.dep_time.month,
count(local.id2) as number_of_flights
order by
local.dep_time.month asc;'''Versus SQL, Trilogy aims to:
Keep:
- Correctness
- Accessibility
Improve:
- Simplicity
- Refactoring/maintainability
- Reusability/composability
- Expressivness
Maintain:
- Acceptable performance
| Backend | Status | Notes |
|---|---|---|
| BigQuery | Core | Full support |
| DuckDB | Core | Full support |
| Snowflake | Core | Full support |
| SQL Server | Experimental | Limited testing |
| Presto | Experimental | Limited testing |
Save the following code in a file named hello.preql
# semantic model is abstract from data
type word string; # types can be used to provide expressive metadata tags that propagate through dataflow
key sentence_id int;
property sentence_id.word_one string::word; # comments after a definition
property sentence_id.word_two string::word; # are syntactic sugar for adding
property sentence_id.word_three string::word; # a description to it
# comments in other places are just comments
# define our datasource to bind the model to data
# for most work, you can import something already defined
# testing using query fixtures is a common pattern
datasource word_one(
sentence: sentence_id,
word:word_one
)
grain(sentence_id)
query '''
select 1 as sentence, 'Hello' as word
union all
select 2, 'Bonjour'
''';
datasource word_two(
sentence: sentence_id,
word:word_two
)
grain(sentence_id)
query '''
select 1 as sentence, 'World' as word
union all
select 2 as sentence, 'World'
''';
datasource word_three(
sentence: sentence_id,
word:word_three
)
grain(sentence_id)
query '''
select 1 as sentence, '!' as word
union all
select 2 as sentence, '!'
''';
def concat_with_space(x,y) -> x || ' ' || y;
# an actual select statement
# joins are automatically resolved between the 3 sources
with sentences as
select sentence_id, @concat_with_space(word_one, word_two) || word_three as text;
WHERE
sentences.sentence_id in (1,2)
SELECT
sentences.text
;Run it:
trilogy run hello.preql duckdbTrilogy can be run directly in python through the core SDK. Trilogy code can be defined and parsed inline or parsed out of files.
A BigQuery example, similar to the BigQuery quickstart:
from trilogy import Dialects, Environment
environment = Environment()
environment.parse('''
key name string;
key gender string;
key state string;
key year int;
key yearly_name_count int; int;
datasource usa_names(
name:name,
number:yearly_name_count,
year:year,
gender:gender,
state:state
)
address `bigquery-public-data.usa_names.usa_1910_2013`;
''')
executor = Dialects.BIGQUERY.default_executor(environment=environment)
results = executor.execute_text('''
WHERE
name = 'Elvis'
SELECT
name,
sum(yearly_name_count) -> name_count
ORDER BY
name_count desc
LIMIT 10;
''')
# multiple queries can result from one text batch
for row in results:
# get results for first query
answers = row.fetchall()
for x in answers:
print(x)Connect to your favorite provider and generate queries with confidence and high accuracy.
from trilogy import Environment, Dialects
from trilogy.ai import Provider, text_to_query
import os
executor = Dialects.DUCK_DB.default_executor(
environment=Environment(working_path=Path(__file__).parent)
)
api_key = os.environ.get(OPENAI_API_KEY)
if not api_key:
raise ValueError("OPENAI_API_KEY required for gpt generation")
# load a model
executor.parse_file("flight.preql")
# create tables in the DB if needed
executor.execute_file("setup.sql")
# generate a query
query = text_to_query(
executor.environment,
"number of flights by month in 2005",
Provider.OPENAI,
"gpt-5-chat-latest",
api_key,
)
# print the generated trilogy query
print(query)
# run it
results = executor.execute_text(query)[-1].fetchall()
assert len(results) == 12
for row in results:
# all monthly flights are between 5000 and 7000
assert row[1] > 5000 and row[1] < 7000, rowTrilogy can be run through a CLI tool, also named 'trilogy'.
Basic syntax:
trilogy run <cmd or path to trilogy file> <dialect>With backend options:
trilogy run "key x int; datasource test_source(i:x) grain(x) address test; select x;" duckdb --path <path/to/database>Format code:
trilogy fmt <path to trilogy file>BigQuery:
- Uses applicationdefault authentication (TODO: support arbitrary credential paths)
- In Python, you can pass a custom client
DuckDB:
--path- Optional database file path
Postgres:
--host- Database host--port- Database port--username- Username--password- Password--database- Database name
Snowflake:
--account- Snowflake account--username- Username--password- Password
The CLI can pick up default configuration from a config file in the toml format. Detection will be recursive form parent directories of the current working directory, including the current working directory.
This can be used to set
- default engine and arguments
- parallelism for execute for the CLI
- any startup commands to run whenever creating an executor.
# Trilogy Configuration File
# Learn more at: https://github.com/trilogy-data/pytrilogy
[engine]
# Default dialect for execution
dialect = "duck_db"
# Parallelism level for directory execution
# parallelism = 2
# Startup scripts to run before execution
[setup]
# startup_trilogy = []
sql = ['setup/setup_dev.sql']- Interactive demo
- Public model repository - Great place for modeling examples
- Full documentation
Are stable and should be sufficient for executing code from Trilogy as text.
from pytrilogy import Executor, DialectAre also stable, and should be used for cases which programatically generate Trilogy statements without text inputs or need to process/transform parsed code in more complicated ways.
from pytrilogy.authoring import Concept, Function, ...Are likely to be unstable. Open an issue if you need to take dependencies on other modules outside those two paths.
Trilogy is straightforward to run as a server/MCP server; the former to generate SQL on demand and integrate into other tools, and MCP for full interactive query loops.
This makes it easy to integrate Trilogy into existing tools or workflows.
You can see examples of both use cases in the trilogy-studio codebase here and install and run an MCP server directly with that codebase.
If you're interested in a more fleshed out standalone server or MCP server, please open an issue and we'll prioritize it!
Not exhaustive - see documentation for more details.
import [path] as [alias];Types:
string | int | float | bool | date | datetime | time | numeric(scale, precision) | timestamp | interval | array<[type]> | map<[type], [type]> | struct<name:[type], name:[type]>
Key:
key [name] [type];Property:
property [key].[name] [type];
property x.y int;
# or multi-key
property <[key],[key]>.[name] [type];
property <x,y>.z int;Transformation:
auto [name] <- [expression];
auto x <- y + 1;datasource <name>(
<column_and_concept_with_same_name>,
# or a mapping from column to concept
<column>:<concept>,
<column>:<concept>,
)
grain(<concept>, <concept>)
address <table>;
datasource orders(
order_id,
order_date,
total_rev: point_of_sale_rev,
customomer_id: customer.id
)
grain orders
address orders;
Basic SELECT:
WHERE
<concept> = <value>
SELECT
<concept>,
<concept>+1 -> <alias>,
...
HAVING
<alias> = <value2>
ORDER BY
<concept> asc|desc
;CTEs/Rowsets:
with <alias> as
WHERE
<concept> = <value>
select
<concept>,
<concept>+1 -> <alias>,
...
select <alias>.<concept>;Persist to table:
persist <alias> as <table_name> from
<select>;Export to file:
COPY INTO <TARGET_TYPE> '<target_path>' FROM SELECT
<concept>, ...
ORDER BY
<concept>, ...
;Show generated SQL:
show <select>;Validate Model
validate all
validate concepts abc,def...
validate datasources abc,def...Clone repository and install requirements.txt and requirements-test.txt.
Please open an issue first to discuss what you would like to change, and then create a PR against that issue.
Trilogy combines two aspects: a semantic layer and a query language. Examples of both are linked below:
Semantic layers - tools for defining a metadata layer above SQL/warehouse to enable higher level abstractions:
Better SQL has been a popular space. We believe Trilogy takes a different approach than the following, but all are worth checking out. Please open PRs/comment for anything missed!
