Analytics Engineer (dbt)dbt Performance2 promptsIntermediate → Advanced2 single promptsFree to use

dbt Performance AI Prompts

2 Analytics Engineer (dbt) prompts in dbt Performance. Copy ready-to-use templates and run them in your AI workflow. Covers intermediate → advanced levels and 2 single prompts.

AI prompts in dbt Performance

2 prompts
AdvancedSingle prompt
01

dbt Project Scalability

Scale this dbt project to support a growing team and larger data volumes. Current project size: {{model_count}} models, {{team_size}} engineers Pain points: {{pain_points}} (slo...

Prompt text
Scale this dbt project to support a growing team and larger data volumes. Current project size: {{model_count}} models, {{team_size}} engineers Pain points: {{pain_points}} (slow CI, difficult navigation, merge conflicts, inconsistent standards) Warehouse: {{warehouse}} 1. Multi-project architecture (dbt mesh): Split the monorepo into multiple smaller dbt projects connected via cross-project refs: - Platform project: shared staging and dimension models consumed by all - Domain projects: finance, marketing, product — each with their own team and release cycle - Cross-project ref: {{ ref('platform', 'dim_customers') }} Benefits: independent deployments, clear ownership, faster CI (smaller projects) 2. Model governance with groups and contracts: Groups (assign ownership): groups: - name: finance owner: name: Finance Analytics Team email: finance-analytics@company.com Model contract (enforce public model schema): models: - name: fct_revenue group: finance access: public # can be referenced from other projects config: contract: enforced: true # CI fails if schema drifts columns: - name: revenue_usd data_type: numeric 3. Slim CI for large projects: - Use dbt state artifacts to run only modified models and their dependents - Target: CI time < 10 minutes regardless of project size - Store production manifest.json in S3; download in CI as the comparison state 4. Model access levels: - private: only accessible within the same group/project - protected: accessible from the same project - public: can be referenced cross-project Enforce: downstream teams can only depend on public models 5. Style guide enforcement: - sqlfluff: SQL linter with dbt dialect support sqlfluff lint models/ --dialect snowflake - pre-commit hooks: run sqlfmt, sqlfluff, and yamllint before every commit - Standardized model config template in .dbt/config.yml Return: dbt mesh architecture design, contract enforcement setup, slim CI configuration, access level policy, and style guide tooling.
IntermediateSingle prompt
02

dbt Query Performance Optimization

Optimize slow dbt models for this warehouse. Slow model: {{model_name}} Current runtime: {{runtime}} seconds Warehouse: {{warehouse}} Model type: {{model_type}} (incremental, fu...

Prompt text
Optimize slow dbt models for this warehouse. Slow model: {{model_name}} Current runtime: {{runtime}} seconds Warehouse: {{warehouse}} Model type: {{model_type}} (incremental, full table, view) 1. Diagnose the bottleneck: - Run: dbt build --select {{model_name}} and check the query profile in the warehouse console - Identify: full table scans, missing clustering/partitioning, large cross-joins, excessive CTEs 2. Partitioning and clustering: BigQuery: config( partition_by={"field": "order_date", "data_type": "date"}, cluster_by=["customer_id", "order_status"] ) Snowflake: config( cluster_by=['TO_DATE(order_date)', 'order_status'] ) Redshift: config( sort=['order_date'], dist='customer_id' ) 3. Incremental optimization: - Ensure the WHERE clause in the incremental filter uses the partition column - Wrong: WHERE id > (SELECT MAX(id) FROM {{this}}) — full table scan on the source - Right: WHERE updated_at >= (SELECT MAX(updated_at) FROM {{this}}) — if updated_at is the partition key 4. CTE vs temp table trade-off: - Many nested CTEs can confuse the optimizer on some warehouses - Snowflake: CTEs are generally fine - BigQuery: deeply nested CTEs with repeated references can be slow — consider intermediate tables - Redshift: complex CTEs may benefit from being broken into separate models 5. Reduce data early: - Push filters as early as possible in the CTE chain - Do not JOIN before filtering: filter first, then join - Avoid SELECT * in intermediate CTEs — project only needed columns 6. Warehouse-specific tuning: Snowflake: configure warehouse size per model: config(snowflake_warehouse='LARGE_WH') BigQuery: enable BI Engine for sub-second queries on frequently used tables Redshift: ANALYZE after large loads; VACUUM for reclaiming deleted rows space Return: diagnosis approach, partitioning / clustering config for the warehouse, incremental filter optimization, and CTE vs table strategy.

Recommended dbt Performance workflow

1

dbt Project Scalability

Start with a focused prompt in dbt Performance so you establish the first reliable signal before doing broader work.

Jump to this prompt
2

dbt Query Performance Optimization

Review the output and identify what needs follow-up, cleanup, explanation, or deeper analysis.

Jump to this prompt

Frequently asked questions

What is dbt performance in analytics engineer (dbt) work?+

dbt Performance is a practical workflow area inside the Analytics Engineer (dbt) prompt library. It groups prompts that solve closely related tasks instead of leaving users to search through one flat list.

Which prompt should I start with?+

Start with the most general prompt in the list, then move toward the more specific or advanced prompts once you have initial output.

What is the difference between a prompt and a chain?+

A single prompt gives you one instruction and one output. A chain is a multi-step sequence designed to build on earlier results and produce a more complete workflow.

Can I use these prompts outside MLJAR Studio?+

Yes. They work in other AI tools too. MLJAR Studio is still the best fit when you want local execution, visible code, and notebook-based reproducibility.

Where should I go next after this category?+

Good next stops are dbt Advanced Patterns, dbt Modeling, dbt Documentation depending on what the current output reveals.

Explore other AI prompt roles

💼
Business Analyst
50 prompts
Browse Business Analyst prompts
🧩
Citizen Data Scientist
24 prompts
Browse Citizen Data Scientist prompts
☁️
Cloud Data Engineer
20 prompts
Browse Cloud Data Engineer prompts
🛡️
Compliance & Privacy Analyst
12 prompts
Browse Compliance & Privacy Analyst prompts
📊
Data Analyst
72 prompts
Browse Data Analyst prompts
🏗️
Data Engineer
35 prompts
Browse Data Engineer prompts
🧠
Data Scientist
50 prompts
Browse Data Scientist prompts
📈
Data Visualization Specialist
23 prompts
Browse Data Visualization Specialist prompts
🗃️
Database Engineer
18 prompts
Browse Database Engineer prompts
🔧
DataOps Engineer
16 prompts
Browse DataOps Engineer prompts
🛒
Ecommerce Analyst
20 prompts
Browse Ecommerce Analyst prompts
💹
Financial Analyst
22 prompts
Browse Financial Analyst prompts
🩺
Healthcare Data Analyst
25 prompts
Browse Healthcare Data Analyst prompts
🤖
LLM Engineer
20 prompts
Browse LLM Engineer prompts
📣
Marketing Analyst
30 prompts
Browse Marketing Analyst prompts
🤖
ML Engineer
42 prompts
Browse ML Engineer prompts
⚙️
MLOps
35 prompts
Browse MLOps prompts
🧭
Product Analyst
16 prompts
Browse Product Analyst prompts
🧪
Prompt Engineer
18 prompts
Browse Prompt Engineer prompts
🧪
Prompts Engineer
18 prompts
Browse Prompts Engineer prompts
📉
Quantitative Analyst
27 prompts
Browse Quantitative Analyst prompts
🔬
Research Scientist
32 prompts
Browse Research Scientist prompts
🧮
SQL Developer
16 prompts
Browse SQL Developer prompts
📐
Statistician
17 prompts
Browse Statistician prompts