PE

Data Engineer (Pentaho)

People tech
Hyderabad8-21 LPA Posted 18 Feb 2026
FULL TIME
Amazon S3
Amazon Rds
Amazon Redshift
AWS Glue

Job Description

Data Engineering & ETL

  • Design, develop, and maintain scalable data pipelines and ETL workflows using Pentaho and AWS
  • Build and optimize data lake and data warehouse architectures
  • Migrate existing data infrastructure to modern cloud-based platforms
  • Develop reverse ETL and customer segmentation pipelines

Cloud & Infrastructure

  • Work extensively with AWS services such as S3, RDS, Redshift, Athena, Glue, and MWAA
  • Manage and migrate large-scale databases (multi-terabyte range)
  • Build data APIs and data delivery services for operational and analytical applications

Data Orchestration & Automation

  • Develop and manage workflows using Apache Airflow
  • Automate manual data processes and optimize data delivery
  • Improve infrastructure for scalability and performance

Development & Collaboration

  • Write high-quality, scalable code in Python and SQL
  • Work in Agile/SCRUM development environments
  • Collaborate with cross-functional and offshore teams
  • Perform testing, debugging, and performance tuning

Required Skills & Qualifications:

  • Strong hands-on experience with Pentaho (ETL development)
  • Deep knowledge of AWS data services (S3, RDS, Redshift, Athena, Glue, MWAA)
  • Experience with Apache Airflow for orchestration
  • Strong SQL and SQL analytics expertise
  • Proficiency in Python
  • Experience with Postgres and MySQL
  • Experience building and optimizing large-scale data pipelines
  • Strong understanding of data warehousing concepts
  • Experience with Git or similar code repositories
  • Strong analytical and problem-solving skills

Join WhatsApp Channel