Kysely for PostgreSQL
Kysely is a type-safe TypeScript SQL query builder. This skill provides patterns for writing effective queries, managing migrations, and avoiding common pitfalls.
When to Use This Skill
Use this skill when:
- Working in a Node.js/TypeScript project with Kysely installed
- Writing database queries for PostgreSQL
- Creating or modifying database migrations
- Debugging type inference issues in Kysely queries
Reference Files
For detailed examples, see these topic-focused reference files:
- select-where.ts - Basic SELECT patterns, WHERE clauses, AND/OR conditions
- joins.ts - Simple joins, callback joins, subquery joins, cross joins
- aggregations.ts - COUNT, SUM, AVG, GROUP BY, HAVING
- orderby-pagination.ts - ORDER BY, NULLS handling, DISTINCT, pagination
- ctes.ts - Common Table Expressions, multiple CTEs, recursive CTEs
- json-arrays.ts - JSONB handling, array columns, jsonBuildObject, jsonAgg
- relations.ts - jsonArrayFrom, jsonObjectFrom for nested data
- mutations.ts - INSERT, UPDATE, DELETE, UPSERT, INSERT FROM SELECT
- expressions.ts - CASE, $if, subqueries, eb.val/lit/not, standalone expressionBuilder
Core Principles
- Prefer Kysely methods over raw SQL: Almost everything you can do in SQL, you can do in Kysely without `sql``
- Use the ExpressionBuilder (eb): The
ebparameter in callbacks is the foundation of type-safe query building - Let TypeScript guide you: If it compiles, it's likely correct SQL
ExpressionBuilder (eb) - The Foundation
The eb parameter in select/where callbacks provides all expression methods:
.select((eb) => [
eb.ref("column").as("alias"), // Column reference
eb.fn<string>("upper", [eb.ref("email")]), // Function call (typed!)
eb.fn.count("id").as("count"), // Aggregate function
eb.fn.sum("amount").as("total"), // SUM
eb.fn.avg("rating").as("avgRating"), // AVG
eb.fn.coalesce("nullable_col", eb.val(0)), // COALESCE
eb.case().when("status", "=", "active") // CASE expression
.then("Active").else("Inactive").end(),
eb("quantity", "*", eb.ref("unit_price")), // Binary expression
eb.exists(subquery), // EXISTS
eb.not(expression), // NOT / negation
eb.cast(eb.val(" "), "text"), // Cast value to type
eb.and([...]), // AND conditions
eb.or([...]), // OR conditions
])
eb.val() vs eb.lit()
// eb.val() - Creates a parameterized value ($1, $2, etc.) - PREFERRED for user input
// Note: eb.val() alone may fail with "could not determine data type of parameter"
// Use eb.cast(eb.val(...), "text") for string values in function arguments
eb.val("user input") // Becomes: $1 with parameter "user input"
eb.cast(eb.val("safe"), "text") // Becomes: $1::text - always works
// eb.lit() - Creates a literal value in SQL
// ONLY accepts: numbers, booleans, null - NOT strings (throws "unsafe immediate value")
eb.lit(1) // Becomes: 1 (directly in SQL)
eb.lit(true) // Becomes: true
eb.lit(null) // Becomes: NULL
// For string literals, use sql`` template instead
sql`'active'` // Becomes: 'active' (directly in SQL)
sql<string>`'label'` // Typed string literal
Standalone ExpressionBuilder
For reusable helpers outside query callbacks:
import { expressionBuilder } from "kysely";
import type { DB } from "./db.d.ts";
// Create standalone expression builder
const eb = expressionBuilder<DB, "user">();
// Use in helper functions
function isActiveUser() {
return eb.and([
eb("is_active", "=", true),
eb("role", "!=", "banned"),
]);
}
Conditional Expressions with Arrays
Build dynamic filters by collecting expressions:
.where((eb) => {
const filters: Expression<SqlBool>[] = [];
if (firstName) filters.push(eb("first_name", "=", firstName));
if (lastName) filters.push(eb("last_name", "=", lastName));
if (minAge) filters.push(eb("age", ">=", minAge));
// Combine all filters with AND (empty array = no filter)
return eb.and(filters);
})
String Concatenation
Use the || operator with sql template for clean string concatenation:
// RECOMMENDED - Clean and type-safe with eb.ref()
.select((eb) => [
sql<string>`${eb.ref("first_name")} || ' ' || ${eb.ref("last_name")}`.as("full_name"),
])
// Output: "first_name" || ' ' || "last_name"
// ALTERNATIVE - Pure eb() chaining (parameterized literals)
.select((eb) => [
eb(eb("first_name", "||", " "), "||", eb.ref("last_name")).as("full_name"),
])
// Output: "first_name" || $1 || "last_name"
// VERBOSE - concat() function (avoid unless you need NULL handling)
.select((eb) => [
eb.fn<string>("concat", [
eb.ref("first_name"),
eb.cast(eb.val(" "), "text"),
eb.ref("last_name"),
]).as("full_name"),
])
Note: concat() treats NULL as empty string, while || propagates NULL. Use concat() only when you need that NULL behavior.
Query Patterns
Basic SELECT
// Select all columns
const users = await db.selectFrom("user").selectAll().execute();
// Select specific columns with aliases
const users = await db
.selectFrom("user")
.select(["id", "email", "first_name as firstName"])
.execute();
// Single row (returns T | undefined)
const user = await db.selectFrom("user").selectAll()
.where("id", "=", userId).executeTakeFirst();
// Single row that must exist (throws if not found)
const user = await db.selectFrom("user").selectAll()
.where("id", "=", userId).executeTakeFirstOrThrow();
WHERE Clauses
// Equality, comparison, IN, LIKE
.where("status", "=", "active")
.where("price", ">", 100)
.where("role", "in", ["admin", "manager"])
.where("name", "like", "%search%")
.where("deleted_at", "is", null)
// Multiple conditions (chained = AND)
.where("is_active", "=", true)
.where("role", "=", "admin")
// OR conditions
.where((eb) => eb.or([
eb("role", "=", "admin"),
eb("role", "=", "manager"),
]))
// Complex AND/OR
.where((eb) => eb.and([
eb("is_active", "=", true),
eb.or([
eb("price", "<", 50),
eb("stock", ">", 100),
]),
]))
JOINs
// Inner join
.innerJoin("order", "order.user_id", "user.id")
// Left join
.leftJoin("category", "category.id", "product.category_id")
// Self-join with alias
.selectFrom("category as c")
.leftJoin("category as parent", "parent.id", "c.parent_id")
// Multiple joins
.innerJoin("order", "order.id", "order_item.order_id")
.innerJoin("product", "product.id", "order_item.product_id")
.innerJoin("user", "user.id", "order.user_id")
Complex JOINs (Callback Format)
Use the callback format when you need:
- Multiple join conditions (composite keys)
- Mixed column-to-column and column-to-literal comparisons
- OR conditions within joins
- Subquery joins (derived tables)
Join Builder Methods:
onRef(col1, op, col2)- Column-to-column comparisonon(col, op, value)- Column-to-literal comparisonon((eb) => ...)- Complex expressions with OR logic
// Multi-condition join (composite key + filter)
.leftJoin("invoice as i", (join) =>
join
.onRef("sp.service_provider_id", "=", "i.service_provider_id")
.onRef("sp.year", "=", "i.year")
.onRef("sp.month", "=", "i.month")
.on("i.status", "!=", "invalidated")
)
// Join with OR conditions
.leftJoin("order as o", (join) =>
join
.onRef("o.user_id", "=", "u.id")
.on((eb) =>
eb.or([
eb("o.status", "=", "completed"),
eb("o.status", "=", "shipped"),
])
)
)
// Subquery join (derived table) - two callbacks
.leftJoin(
(eb) =>
eb
.selectFrom("order")
.select((eb) => [
"user_id",
eb.fn.count("id").as("order_count"),
eb.fn.max("created_at").as("last_order_at"),
])
.groupBy("user_id")
.as("order_stats"), // MUST have alias!
(join) => join.onRef("order_stats.user_id", "=", "u.id")
)
// Cross join (always-true condition) - for joining aggregated CTEs
.leftJoin("summary_cte", (join) =>
join.on(sql`true`, "=", sql`true`)
)
Aggregations
.select((eb) => [
"status",
eb.fn.count("id").as("count"),
eb.fn.sum("total_amount").as("totalAmount"),
eb.fn.avg("total_amount").as("avgAmount"),
])
.groupBy("status")
.having((eb) => eb.fn.count("id"), ">", 5)
ORDER BY
// Simple ordering
.orderBy("created_at", "desc")
.orderBy("name", "asc")
// NULLS FIRST / NULLS LAST - use order builder callback
.orderBy("category_id", (ob) => ob.asc().nullsLast())
.orderBy("priority", (ob) => ob.desc().nullsFirst())
// Multiple columns - chain orderBy calls (array syntax is deprecated)
.orderBy("category_id", "asc")
.orderBy("price", "desc")
.orderBy("name", "asc")
CTEs (Common Table Expressions)
Use CTEs for complex queries with multiple aggregation levels:
const result = await db
.with("order_totals", (db) =>
db.selectFrom("order")
.innerJoin("user", "user.id", "order.user_id")
.select((eb) => [
"user.id as userId",
"user.email",
eb.fn.sum("order.total_amount").as("totalSpent"),
eb.fn.count("order.id").as("orderCount"),
])
.groupBy(["user.id", "user.email"])
)
.selectFrom("order_totals")
.selectAll()
.orderBy("totalSpent", "desc")
.execute();
JSON Aggregation (PostgreSQL)
import { jsonBuildObject } from "kysely/helpers/postgres";
// Note: jsonAgg is accessed via eb.fn.jsonAgg(), not imported
.with("tasks", (db) =>
db.selectFrom("task")
.leftJoin("user", "user.id", "task.assignee_id")
.select((eb) => [
"task.job_id",
eb.fn.jsonAgg(
jsonBuildObject({
id: eb.ref("task.id"),
status: eb.ref("task.status"),
assignee: jsonBuildObject({
id: eb.ref("user.id"),
name: eb.fn<string>("concat", [
eb.ref("user.first_name"),
eb.cast(eb.val(" "), "text"),
eb.ref("user.last_name"),
]),
}),
})
)
.filterWhere("task.id", "is not", null) // Filter nulls from left join
.as("tasks"),
])
.groupBy("task.job_id")
)
JSON, JSONB, and Array Handling
JSONB Columns
NO JSON.stringify or JSON.parse needed! The pg driver handles JSONB automatically:
// INSERT - pass objects directly
await db
.insertInto("user")
.values({
email: "test@example.com",
metadata: { preferences: { theme: "dark" }, count: 42 },
})
.execute();
// UPDATE - pass objects directly
await db
.updateTable("user")
.set({
metadata: { preferences: { theme: "light" } },
})
.where("id", "=", userId)
.execute();
// READ - returns parsed object, not string
const user = await db
.selectFrom("user")
.select(["id", "metadata"])
.executeTakeFirst();
console.log(user.metadata.preferences.theme); // "dark" - already an object!
Array Columns (text[], int[], etc.)
NO JSON.stringify needed for array columns! The pg driver handles arrays natively:
// INSERT with array - pass array directly
await db
.insertInto("product")
.values({
name: "Product",
tags: ["phone", "electronics", "premium"], // Direct array!
})
.execute();
// READ - returns as native JavaScript array
const product = await db
.selectFrom("product")
.select(["name", "tags"])
.executeTakeFirst();
console.log(product.tags); // ["phone", "electronics", "premium"]
// UPDATE array
await db
.updateTable("product")
.set({ tags: ["updated", "tags"] })
.where("id", "=", productId)
.execute();
Querying Arrays
// Array contains all values (@>) - operator works natively!
.where("tags", "@>", sql`ARRAY['phone', 'premium']::text[]`)
// Arrays overlap (&&) - operator works natively!
.where("tags", "&&", sql`ARRAY['premium', 'basic']::text[]`)
// Array contains value (ANY) - type-safe with eb.fn
.where((eb) => eb(sql`${searchTerm}`, "=", eb.fn("any", [eb.ref("tags")])))
// eb.ref("tags") validates column exists - eb.ref("invalid") would be a TS error
Querying JSONB
// Key exists (?) - operator works natively!
.where("metadata", "?", "theme")
// Any key exists (?|) - operator works natively!
.where("metadata", "?|", sql`array['theme', 'language']`)
// All keys exist (?&) - operator works natively!
.where("metadata", "?&", sql`array['theme', 'notifications']`)
// JSONB contains (@>) - operator works natively!
.where("metadata", "@>", sql`'{"notifications": true}'::jsonb`)
// Extract field as text (->> as operator) - type-safe!
.where((eb) => eb(eb("metadata", "->>", "theme"), "=", "dark"))
// eb("metadata", ...) validates column - eb("invalid", ...) would be TS error
// Extract nested path (#>> still needs sql``)
.where(sql`metadata#>>'{preferences,theme}'`, "=", "dark")
// In SELECT - type-safe with eb()
.select((eb) => [
eb("metadata", "->", "preferences").as("prefs"), // Returns JSONB
eb("metadata", "->>", "theme").as("theme"), // Returns text
])
// Nested paths still need sql``
.select(sql`metadata#>'{preferences,theme}'`.as("t")) // Nested as JSONB
.select(sql<string>`metadata#>>'{a,b}'`.as("t")) // Nested as text
JSONPath (PostgreSQL 12+)
// JSONPath match (@@) - works as native operator!
.where("metadata", "@@", sql`'$.preferences.theme == "dark"'`)
// JSONPath exists (@?) - NOT in Kysely's allowlist, use function instead
// Use jsonb_path_exists() for type-safe column validation
.where((eb) =>
eb.fn("jsonb_path_exists", [eb.ref("metadata"), sql`'$.preferences.theme'`])
)
// eb.ref("metadata") validates column - eb.ref("invalid") would be TS error
// Extract with JSONPath - type-safe with eb.fn
.select((eb) => [
"id",
eb.fn("jsonb_path_query_first", [eb.ref("metadata"), sql`'$.preferences.theme'`]).as("theme"),
])
// JSONPath with variables
const searchValue = "dark";
.where((eb) =>
eb.fn("jsonb_path_exists", [
eb.ref("metadata"),
sql`'$.preferences.theme ? (@ == $val)'`,
sql`jsonb_build_object('val', ${searchValue}::text)`,
])
)
Conditional Queries ($if)
Use $if() for runtime-conditional query modifications:
const result = await db
.selectFrom("user")
.selectAll()
.$if(!includeInactive, (qb) => qb.where("is_active", "=", true))
.$if(includeMetadata, (qb) => qb.select("metadata"))
.$if(!!searchTerm, (qb) => qb.where("name", "like", `%${searchTerm}%`))
.$if(!!roleFilter, (qb) => qb.where("role", "in", roleFilter!))
.execute();
Type behavior: Columns added via $if become optional in the result type since inclusion isn't guaranteed at compile time.
Relations (jsonArrayFrom / jsonObjectFrom)
Kysely is NOT an ORM - it uses PostgreSQL's JSON functions for nested data:
import { jsonArrayFrom, jsonObjectFrom } from "kysely/helpers/postgres";
// One-to-many: User with their orders
const users = await db
.selectFrom("user")
.select((eb) => [
"user.id",
"user.email",
jsonArrayFrom(
eb
.selectFrom("order")
.select(["order.id", "order.status", "order.total_amount"])
.whereRef("order.user_id", "=", "user.id")
.orderBy("order.created_at", "desc")
).as("orders"),
])
.execute();
// Many-to-one: Product with its category
const products = await db
.selectFrom("product")
.select((eb) => [
"product.id",
"product.name",
jsonObjectFrom(
eb
.selectFrom("category")
.select(["category.id", "category.name"])
.whereRef("category.id", "=", "product.category_id")
).as("category"),
])
.execute();
Critical: Use explicit .select() instead of .selectAll() with nested json helpers
When using jsonObjectFrom containing a nested jsonArrayFrom (or vice versa), using selectAll("table") breaks TypeScript's type inference. The result type becomes unknown or loses the nested structure, requiring $castTo to fix.
// WRONG - selectAll() breaks type inference for nested json helpers
const invoice = await db
.selectFrom("invoices")
.selectAll("invoices")
.select((eb) => [
jsonObjectFrom(
eb
.selectFrom("payment_plans")
.selectAll() // ❌ This breaks type inference!
.select((eb2) => [
jsonArrayFrom(
eb2.selectFrom("installments").selectAll()
.whereRef("installments.plan_id", "=", "payment_plans.id")
).as("installments"),
])
.whereRef("payment_plans.invoice_id", "=", "invoices.id")
).as("payment_plan"), // Type is unknown or broken
])
.executeTakeFirst();
// RIGHT - explicit select() preserves type inference
const invoice = await db
.selectFrom("invoices")
.selectAll("invoices")
.select((eb) => [
jsonObjectFrom(
eb
.selectFrom("payment_plans")
.select([ // ✅ Explicit columns!
"payment_plans.id",
"payment_plans.invoice_id",
"payment_plans.notes",
"payment_plans.created_at",
])
.select((eb2) => [
jsonArrayFrom(
eb2.selectFrom("installments").selectAll()
.whereRef("installments.plan_id", "=", "payment_plans.id")
).as("installments"),
])
.whereRef("payment_plans.invoice_id", "=", "invoices.id")
).as("payment_plan"), // Type is properly inferred!
])
.executeTakeFirst();
Why this happens: Kysely's type inference for nested json helpers relies on tracking the selected columns through the query chain. selectAll() returns all columns dynamically, which confuses TypeScript when combined with additional .select() calls that add nested json helpers. Using explicit column names gives TypeScript the static information it needs.
Rule of thumb: When combining jsonObjectFrom/jsonArrayFrom with nested json helpers, always use explicit .select([...columns]) instead of .selectAll() on the subquery containing the nested helper.
Reusable Helpers
Create composable, type-safe helper functions using Expression<T>:
import { Expression, sql } from "kysely";
// Helper that takes and returns Expression<string>
function lower(expr: Expression<string>) {
return sql<string>`lower(${expr})`;
}
// Use in queries
.where(({ eb, ref }) => eb(lower(ref("email")), "=", email.toLowerCase()))
Splitting Query Building and Execution
Build queries without executing, useful for dynamic query construction:
// Build query (doesn't execute)
let query = db
.selectFrom("user")
.select(["id", "email"]);
// Add conditions dynamically
if (role) {
query = query.where("role", "=", role);
}
if (isActive !== undefined) {
query = query.where("is_active", "=", isActive);
}
// Execute when ready
const results = await query.execute();
// Or compile to SQL without executing
const compiled = query.compile();
console.log(compiled.sql); // The SQL string
console.log(compiled.parameters); // Bound parameters
Subqueries
// Subquery in WHERE
.where("id", "in",
db.selectFrom("order").select("user_id").where("status", "=", "completed")
)
// EXISTS subquery
.where((eb) =>
eb.exists(
db.selectFrom("review")
.select(sql`1`.as("one"))
.whereRef("review.product_id", "=", eb.ref("product.id"))
)
)
INSERT Operations
// Single insert with returning
const user = await db
.insertInto("user")
.values({ email: "test@example.com", first_name: "Test", last_name: "User" })
.returning(["id", "email"])
.executeTakeFirst();
// Multiple rows
await db
.insertInto("user")
.values([
{ email: "a@example.com", first_name: "A", last_name: "User" },
{ email: "b@example.com", first_name: "B", last_name: "User" },
])
.execute();
// Upsert (ON CONFLICT) - type-safe with expression builder
await db
.insertInto("product")
.values({ sku: "ABC123", name: "Product", stock_quantity: 10 })
.onConflict((oc) =>
oc.column("sku").doUpdateSet((eb) => ({
stock_quantity: eb("product.stock_quantity", "+", eb.ref("excluded.stock_quantity")),
}))
)
.execute();
// eb("product.invalid_column", ...) would be a TypeScript error!
// Insert from SELECT
await db
.insertInto("archive")
.columns(["user_id", "data", "archived_at"])
.expression(
db.selectFrom("user")
.select(["id", "metadata", sql`now()`.as("archived_at")])
.where("is_active", "=", false)
)
.execute();
UPDATE Operations
// Simple update
await db
.updateTable("user")
.set({ is_active: false })
.where("id", "=", userId)
.execute();
// Update with expression
await db
.updateTable("product")
.set((eb) => ({
stock_quantity: eb("stock_quantity", "+", 10),
}))
.where("sku", "=", "ABC123")
.returning(["id", "stock_quantity"])
.executeTakeFirst();
Migrations
Configuration (kysely.config.ts)
import { PostgresDialect } from "kysely";
import { defineConfig } from "kysely-ctl";
import pg from "pg";
export default defineConfig({
dialect: new PostgresDialect({
pool: new pg.Pool({
connectionString: process.env.DATABASE_URL,
}),
}),
migrations: {
migrationFolder: "src/db/migrations",
},
seeds: {
seedFolder: "src/db/seeds",
},
});
Migration Commands
npx kysely migrate:make migration-name # Create migration
npx kysely migrate:latest # Run all pending migrations
npx kysely migrate:down # Rollback last migration
npx kysely seed make seed-name # Create seed
npx kysely seed run # Run all seeds
Migration File Structure
import type { Kysely } from "kysely";
import { sql } from "kysely";
// Always use Kysely<any> - migrations should be frozen in time
export async function up(db: Kysely<any>): Promise<void> {
await db.schema
.createTable("user")
.addColumn("id", "bigint", (col) => col.primaryKey().generatedAlwaysAsIdentity())
.addColumn("email", "text", (col) => col.notNull().unique())
.addColumn("created_at", "timestamptz", (col) => col.notNull().defaultTo(sql`now()`))
.execute();
// IMPORTANT: Always index foreign key columns!
await db.schema.createIndex("idx_order_user_id").on("order").column("user_id").execute();
}
export async function down(db: Kysely<any>): Promise<void> {
await db.schema.dropTable("user").execute();
}
Recommended Column Types
// Primary keys: Use identity columns (SQL standard, prevents accidental ID conflicts)
.addColumn("id", "bigint", (col) => col.primaryKey().generatedAlwaysAsIdentity())
// NOT serial/bigserial - those allow manual ID inserts that can cause conflicts
// Timestamps: Always use timestamptz (stores UTC, converts to client timezone)
.addColumn("created_at", "timestamptz", (col) => col.notNull().defaultTo(sql`now()`))
// NOT timestamp - loses timezone information
// Money: Use numeric with precision (exact decimal, no floating point errors)
.addColumn("price", "numeric(10, 2)", (col) => col.notNull())
// NOT float/real/double precision - those have rounding errors
// Strings: Use text (no length limit, same performance as varchar)
.addColumn("name", "text", (col) => col.notNull())
// varchar(n) only if you need a hard length constraint
// JSON: Use jsonb (binary, indexable, faster queries)
.addColumn("metadata", "jsonb")
// NOT json - stored as text, no indexing, slower
// Foreign keys: Create indexes manually (PostgreSQL doesn't auto-index FKs)
await db.schema.createIndex("idx_order_user_id").on("order").column("user_id").execute();
Data Type Gotchas
// CORRECT - Space after comma in numeric types
.addColumn("price", "numeric(10, 2)")
// WRONG - Will fail with "invalid column data type"
.addColumn("price", "numeric(10,2)")
// For complex types, use sql template
.addColumn("price", sql`numeric(10, 2)`)
Type Generation
Use kysely-codegen to generate types from your database:
npx kysely-codegen --url "postgresql://..." --out-file src/db/db.d.ts
Generated types use:
Generated<T>for auto-increment columns (optional on insert)ColumnType<Select, Insert, Update>for different operation typesTimestampfor timestamptz columns
Common Pitfalls to Avoid
1. Don't Resort to `sql`` When Kysely Has a Method
// WRONG
.select(sql`count(*)`.as("count"))
// RIGHT
.select((eb) => eb.fn.countAll().as("count"))
2. Don't Forget .execute()
Queries are lazy - they won't run without calling an execute method:
// This does nothing!
db.selectFrom("user").selectAll();
// This runs the query
await db.selectFrom("user").selectAll().execute();
3. Use whereRef for Column-to-Column Comparisons
// WRONG - Compares to string literal "other.column"
.where("table.column", "=", "other.column")
// RIGHT - Compares to actual column value
.whereRef("table.column", "=", "other.column")
4. Type Your Function Returns
// Better type inference
eb.fn<string>("concat", [...])
eb.fn<number>("length", [...])
5. PostgreSQL Does NOT Auto-Index Foreign Keys
Always create indexes on foreign key columns:
await db.schema.createIndex("idx_order_user_id").on("order").column("user_id").execute();
6. Always Type sql Template Literals
When using sql template literals, the inferred type is unknown since Kysely can't know what the SQL expression resolves to. Always provide an explicit type:
// WRONG - Returns unknown type
eb.fn.coalesce("some_json_col", sql`'{}'::jsonb`)
// RIGHT - Explicit type annotation
eb.fn.coalesce("some_json_col", sql<Record<string, unknown>>`'{}'::jsonb`)
// For complex types (e.g., JSON column from a CTE), use typeof with eb.ref
// This ensures the fallback type matches the column type exactly
eb.fn
.coalesce(
eb.ref("jobs_agg.jobs"),
sql<typeof eb.ref<"jobs_agg.jobs">>`'[]'::json`
)
.as("jobs")
Key rule: Every sql template literal should have a type parameter: sql<TYPE>. This ensures proper type inference throughout your query chain.
7. DATE Columns Cause Timezone Issues
By default, the pg driver converts DATE columns to JavaScript Date objects. This causes timezone problems:
Database: 2025-01-01 (just a date, no time)
JS Date: 2025-01-01T00:00:00.000Z (interpreted as UTC midnight)
User in NYC sees: Dec 31, 2024 (5 hours behind UTC)
Solution: Parse DATE as string and let the frontend handle formatting
Step 1: Configure pg to return DATE as string:
import pg from "pg";
// Tell pg to return DATE columns as strings instead of Date objects
const DATE_OID = 1082;
pg.types.setTypeParser(DATE_OID, (val: string) => val);
Step 2: Update kysely-codegen to generate matching types:
npx kysely-codegen \
--url="$DATABASE_URL" \
--out-file=server/db/db.d.ts \
--dialect=postgres \
--date-parser=string
Now DATE columns return strings like "2025-01-01" and the frontend can parse/format respecting the user's timezone.
Note: This applies to DATE columns only. TIMESTAMPTZ columns already handle timezones correctly by storing UTC and converting on read.
PostgreSQL Helpers Summary
All helpers from kysely/helpers/postgres:
import {
jsonArrayFrom, // One-to-many relations (subquery → array)
jsonObjectFrom, // Many-to-one relations (subquery → object | null)
jsonBuildObject, // Build JSON object from expressions
mergeAction, // Get action performed in MERGE query (PostgreSQL 15+)
} from "kysely/helpers/postgres";
Note: jsonAgg is NOT imported - use eb.fn.jsonAgg() instead.
mergeAction (PostgreSQL 15+)
For MERGE queries, get which action was performed:
import { mergeAction } from "kysely/helpers/postgres";
const result = await db
.mergeInto("person")
.using("person_updates", "person.id", "person_updates.id")
.whenMatched()
.thenUpdateSet({ name: eb.ref("person_updates.name") })
.whenNotMatched()
.thenInsertValues({ id: eb.ref("person_updates.id"), name: eb.ref("person_updates.name") })
.returning([mergeAction().as("action"), "id"])
.execute();
// result[0].action is 'INSERT' | 'UPDATE' | 'DELETE'
Extending Kysely
Custom Helper Functions
Most extensions use the sql template tag with RawBuilder<T>:
import { sql, RawBuilder } from "kysely";
// Create a typed helper function
function json<T>(value: T): RawBuilder<T> {
return sql`CAST(${JSON.stringify(value)} AS JSONB)`;
}
// Use in queries
.select((eb) => [
json({ name: "value" }).as("data"),
])
Custom Expression Classes
For reusable expressions, implement the Expression<T> interface:
import { Expression, OperationNode, sql } from "kysely";
class JsonValue<T> implements Expression<T> {
readonly #value: T;
constructor(value: T) {
this.#value = value;
}
get expressionType(): T | undefined {
return undefined;
}
toOperationNode(): OperationNode {
return sql`CAST(${JSON.stringify(this.#value)} AS JSONB)`.toOperationNode();
}
}
Note: Module augmentation and inheritance-based extension are not recommended.
Handling "Excessively Deep Types" Error
The Problem
Complex queries with many CTEs can overwhelm TypeScript's type instantiation limits:
Type instantiation is excessively deep and possibly infinite
This commonly occurs with 12+ with clauses, as Kysely's nested helper types accumulate.
The Solution: $assertType
Use $assertType to simplify the type chain at intermediate points:
const result = await db
.with("cte1", (qb) =>
qb.selectFrom("user")
.select(["id", "email"])
.$assertType<{ id: number; email: string }>() // Simplify type here
)
.with("cte2", (qb) =>
qb.selectFrom("cte1")
.select("email")
.$assertType<{ email: string }>()
)
// ... more CTEs
.selectFrom("cteN")
.selectAll()
.execute();
Key points:
- The asserted type must structurally match the actual type (full type safety preserved)
- Apply to several intermediate
withclauses in large queries - TypeScript cannot automatically simplify these types - explicit assertion is required