ETL Developer
Experience – 4+
Location – Bengaluru
Position Summary
The ETL Developer is responsible for designing, developing, and maintaining scalable data pipelines and transformation logic supporting enterprise analytics platforms. The role focuses on advanced SQL-based data processing, Python-driven transformations, and integration with AWS-based data platforms such as Redshift, Databricks, and Athena.
The incumbent ensures reliable, high-quality data delivery for downstream semantic models and reporting solutions while adhering to enterprise data governance and performance standards. The role also provides limited support for Power BI datasets and reporting layers.
Key Responsibilities
- Strategic Leadership & Planning
- Support design and standardization of ETL frameworks and data transformation patterns.
- Contribute to data quality, governance, and consistency across data pipelines.
- Identify opportunities to optimize data processing and reduce latency in data delivery.
Project Oversight & System Ownership
- Develop and maintain complex ETL pipelines using advanced SQL and Python.
- Design and optimize data transformations, joins, aggregations, and incremental loading strategies.
- Build and manage data pipelines leveraging AWS data platforms (Redshift, Databricks).
- Ensure data pipelines consume and produce governed, high-quality datasets.
- Support integration of transformed data into semantic layers and reporting systems.
- Maintain documentation of ETL logic, data flows, and transformation rules.
Operations & Compliance
- Monitor ETL jobs, data pipelines, and workflow executions to ensure reliability and performance.
- Troubleshoot and resolve data pipeline failures, performance bottlenecks, and data inconsistencies.
- Perform root cause analysis for data-related issues impacting reporting and analytics.
- Ensure compliance with enterprise data governance and IT policies.
- Maintain operational runbooks and support documentation.
Team Leadership
- Promote best practices in SQL development, Python coding, and ETL design.
- Share knowledge on data transformation optimization and pipeline performance tuning.
Cross-functional Collaboration
- Collaborate with Data Engineers, BI Developers, and Architects to ensure alignment across data layers.
- Work with BA/BI developers to support data availability and usability for reporting.
- Coordinate with platform teams to resolve upstream data or infrastructure issues.
- Communicate data pipeline changes and impacts to stakeholders.
Qualifications
Education & Experience
- Bachelor’s degree in Computer Science, Information Systems, Engineering, or related field.
- 4–6 years of experience in ETL development, data engineering, or related roles.
- Hands-on experience with enterprise data platforms (AWS preferred).
Skills & Knowledge
Core (Must-have):
- Advanced SQL (complex queries, performance tuning, joins, aggregations)
- Python (data transformation, scripting, working with large datasets).
- Strong understanding of ETL/ELT design patterns and data pipeline architecture.
Intermediate:
- Experience with Databricks / PySpark or similar distributed processing frameworks.
- Understanding of Redshift and Athena data processing.
Basic / Exposure:
- Power BI dataset understanding and basic troubleshooting.
- Awareness of semantic models and reporting layer dependencies.
Additional:
- Data quality validation and troubleshooting skills.
- Strong analytical and problem-solving capabilities.
- Solid documentation and communication skills.
- Experience with Agile development practices.
ETL and AWS and “data engineering” and Python and SQL and (Databricks or PySpark)
If you are interested, Please share to resume to careeropportunities@vaisesika.in