Data Engineer - Analyst
Canopus Infosystems - A CMMI Level 3 Company
Bangalore, India Full Time Engineering Jobs India
Job Description
Job Title: Data Engineer - Analyst
Experience: 2.5 to 5 Years
Location: PAN India (Remote/On-site as applicable)
Role Overview:
We are seeking skilled Data Engineer - Analyst to design, develop, and manage robust data pipelines and analytics-ready datasets. In this role, you will enable data-driven decision-making by supporting BI reporting, product analytics, and business insights. You will collaborate with multiple teams, work with diverse data sources, and ensure high data quality across the entire data lifecycle.
Key Responsibilities:
- Develop and manage scalable ETL/ELT pipelines (both batch and incremental) using SQL and Python
- Ingest and integrate data from multiple sources including databases, APIs, SaaS platforms, event streams, and flat files
- Design and implement analytics-ready data models (such as star schema and data marts) for reporting and analysis
- Build and optimize data transformations within cloud-based warehouses/lakehouses (Snowflake, BigQuery, Redshift, Synapse, Databricks)
- Collaborate with stakeholders to define KPIs, metrics, and reporting requirements
- Develop, maintain, and enhance dashboards and reports using BI tools like Power BI, Tableau, Looker, or Sigma
- Ensure data reliability by implementing validation checks, monitoring systems, alerting mechanisms, and proper documentation
- Optimize data performance and cost efficiency through techniques like incremental loading, partitioning, query tuning, and efficient file formats
Required Skills:
- Strong proficiency in SQL (including CTEs, window functions, joins, aggregations, and performance tuning)
- Strong programming skills in Python for data processing and automation
- Hands-on experience with at least one cloud platform: AWS, Azure, or GCP
- Practical experience with modern data warehouses/lakehouses such as Snowflake, BigQuery, Redshift, Synapse, or Databricks
- Solid understanding of ETL/ELT concepts, including incremental processing, retries, idempotency, and basic CDC (Change Data Capture)
- Good understanding of data modeling principles for analytics and BI use cases
- Experience building user-friendly reports and dashboards using BI tools (Power BI, Tableau, Looker, or Sigma)
Good to have:
- Experience with orchestration tools like Airflow, dbt, Dagster, Prefect, Azure Data Factory (ADF), or AWS Glue
- Familiarity with streaming/event-driven data platforms such as Kafka, Kinesis, or Pub/Sub
- Knowledge of monitoring and logging tools like CloudWatch, Azure Monitor, GCP Monitoring, or Datadog
- Exposure to CI/CD practices and Git-based version control for managing data pipelines
Posted April 19, 2026