Senior Data Engineer / Analytics Engineer (India-Based)

Filled
December 18, 2025

Job Description

Mercor is partnering with a cutting-edge AI research lab to hire a **Senior Data/Analytics Engineer** with expertise across **DBT** and **Snowflake’s Cortex CLI**. In this role, you will build and scale Snowflake-native data and ML pipelines, leveraging Cortex’s emerging AI/ML capabilities while maintaining production-grade DBT transformations. You will work closely with data engineering, analytics, and ML teams to prototype, operationalise, and optimise AI-driven workflows—defining best practices for Snowflake-native feature engineering and model lifecycle management.
This is a high-impact role within a modern, fully cloud-native data stack.
• * *

## **Responsibilities**

– Design, build, and maintain **DBT models, macros, and tests** following modular data modeling and semantic best practices.

– Integrate **DBT workflows with Snowflake Cortex CLI**, enabling:

– Feature engineering pipelines

– Model training & inference tasks

– Automated pipeline orchestration

– Monitoring and evaluation of Cortex-driven ML models

– Establish best practices for **DBT–Cortex architecture and usage patterns**.

– Collaborate with data scientists and ML engineers to **produce Cortex workloads** in Snowflake.

– Build and optimise **CI/CD pipelines** for dbt (GitHub Actions, GitLab, Azure DevOps).

– Tune Snowflake compute and queries for **performance and cost efficiency**.

– Troubleshoot issues across DBT arti-facts, Snowflake objects, lineage, and data quality.

– Provide guidance on **DBT project governance, structure, documentation, and testing frameworks**.
• * *

## **Required Qualifications**

– **3+ years** experience with **DBT Core or DBT Cloud**, including macros, packages, testing, and deployments.

– Strong expertise with **Snowflake** (warehouses, tasks, streams, materialised views, performance tuning).

– Hands-on experience with **Snowflake Cortex CLI**, or strong ability to learn it quickly.

– Strong SQL skills; working familiarity with **Python** for scripting and DBT automation.

– Experience integrating DBT with orchestration tools (Airflow, Dagster, Prefect, etc.).

– Solid understanding of **modern data engineering, ELT patterns, and version-controlled analytics development**.
• * *

## **Nice-to-Have Skills**

– Prior experience operationalising **ML workflows inside Snowflake**.

– Familiarity with **Snow-park**, Python UDFs/UDTFs.

– Experience building **semantic layers** using DBT metrics.

– Knowledge of **MLOps / DataOps** best practices.

– Exposure to **LLM workflows, vector search, and unstructured data pipelines**.
• * *

## **Why Join**

– You will be an **hourly contractor through Mercor**, working **20–40 hours per week** with flexibility.

– Direct opportunity to build **next-generation Snowflake AI/ML systems** with Cortex.

– High-impact ownership of DBT and Snowflake architecture across production pipelines.

– Work alongside top-tier ML engineers, data scientists, and research teams.

– Fully remote, high-autonomy environment focused on innovation, velocity, and engineering excellence.