Remote Senior Data Engineer / Analytics Engineer (India-Based) - AI Trainer ($35-$70 per hour)
hace 3 días
Toledo
Mercor is partnering with a cutting-edge AI research lab to hire a Senior Data/Analytics Engineer with expertise across DBT and Snowflake’s Cortex CLI. In this role, you will build and scale Snowflake-native data and ML pipelines, leveraging Cortex’s emerging AI/ML capabilities while maintaining production-grade DBT transformations. You will work closely with data engineering, analytics, and ML teams to prototype, operationalise, and optimise AI-driven workflows—defining best practices for Snowflake-native feature engineering and model lifecycle management.This is a high-impact role within a modern, fully cloud-native data stack. * * * ## Responsibilities - Design, build, and maintain DBT models, macros, and tests following modular data modeling and semantic best practices. - Integrate DBT workflows with Snowflake Cortex CLI, enabling: - Feature engineering pipelines - Model training & inference tasks - Automated pipeline orchestration - Monitoring and evaluation of Cortex-driven ML models - Establish best practices for DBT–Cortex architecture and usage patterns. - Collaborate with data scientists and ML engineers to produce Cortex workloads in Snowflake. - Build and optimise CI/CD pipelines for dbt (GitHub Actions, GitLab, Azure DevOps). - Tune Snowflake compute and queries for performance and cost efficiency. - Troubleshoot issues across DBT arti-facts, Snowflake objects, lineage, and data quality. - Provide guidance on DBT project governance, structure, documentation, and testing frameworks. * * * ## Required Qualifications - 3+ years experience with DBT Core or DBT Cloud, including macros, packages, testing, and deployments. - Strong expertise with Snowflake (warehouses, tasks, streams, materialised views, performance tuning). - Hands-on experience with Snowflake Cortex CLI, or strong ability to learn it quickly. - Strong SQL skills; working familiarity with Python for scripting and DBT automation. - Experience integrating DBT with orchestration tools (Airflow, Dagster, Prefect, etc.). - Solid understanding of modern data engineering, ELT patterns, and version-controlled analytics development. * * * ## Nice-to-Have Skills - Prior experience operationalising ML workflows inside Snowflake. - Familiarity with Snow-park, Python UDFs/UDTFs. - Experience building semantic layers using DBT metrics. - Knowledge of MLOps / DataOps best practices. - Exposure to LLM workflows, vector search, and unstructured data pipelines. * * * ## Why Join - You will be an hourly contractor through Mercor, working 20–40 hours per week with flexibility. - Direct opportunity to build next-generation Snowflake AI/ML systems with Cortex. - High-impact ownership of DBT and Snowflake architecture across production pipelines. - Work alongside top-tier ML engineers, data scientists, and research teams. - Fully remote, high-autonomy environment focused on innovation, velocity, and engineering excellence.