Data Engineer
Anblicks
Job Description
Job Title: Data Engineer – Snowflake & dbt Location: Hyderabad Experience: 4–8 years Role Overview We are looking for a highly skilled Data Engineer with strong hands-on experience in Snowflake and dbt to design, build, and optimize modern cloud data platforms. The ideal candidate will work closely with Data Architects, Analytics Engineers, and business stakeholders to deliver scalable, high‑performance data solutions following modern ELT and analytics engineering best practices. Key Responsibilities Data Platform & Pipeline Development Design, develop, and maintain end-to-end ELT pipelines on Snowflake Build scalable and reliable batch data pipelines using SQL and Python Implement and manage Bronze → Silver → Gold data layer architectures Ensure high data quality, reliability, and performance across pipelines dbt & Analytics Engineering Develop and maintain dbt models , tests, snapshots, and documentation Implement dimensional and analytical data models for reporting and BI Apply dbt best practices: modular models, testing, freshness checks, and exposures Optimize transformation logic for cost and performance in Snowflake Performance & Optimization Tune SQL queries and Snowflake warehouse usage for performance and efficiency Manage clustering, partitioning, and resource monitoring in Snowflake Identify and resolve data quality and pipeline performance issues Collaboration & Delivery Work closely with Data Architects, BI teams, and business stakeholders Participate in requirement gathering, design discussions, and code reviews Contribute to Agile/Scrum ceremonies and sprint deliveries Maintain clear documentation for data models, pipelines, and processes Required Technical Skills Strong hands-on experience with Snowflake Warehouses, schemas, roles, performance tuning, cost optimization Strong experience with dbt (Core or Cloud) Models, macros, tests, snapshots, CI/CD integration Advanced SQL skills (analytical queries, optimization techniques) Working experience with Python for data processing or orchestration Experience with data modeling: Dimensional modeling (Star/Snowflake schema) Analytical and reporting-oriented models Experience with Git-based version control and CI/CD pipelines Good to Have Experience with ingestion tools such as Fivetran, Informatica, Matillion, Kafka, or Airflow Exposure to cloud platforms ( AWS / Azure / GCP ) Experience supporting BI tools like Power BI, Looker, Tableau, Sigma Prior experience with data migrations from on‑prem or legacy warehouses Familiarity with data governance, data quality frameworks, and monitoring Soft Skills & Competencies Strong problem-solving and analytical skills Ability to work independently and in team-based delivery models Clear communication with technical and non-technical stakeholders Ownership-driven mindset with attention to detail Comfortable working in fast‑paced, client-facing environments Education Bachelor’s degree in Computer Science, Engineering, Data Science, or a related field