PEPeople tech
Data Engineer (Pentaho)
Hyderabad ₹8-21 LPA Posted 18 Feb 2026
FULL TIME
Amazon S3
Amazon Rds
Amazon Redshift
AWS Glue
Job Description
Data Engineering & ETL
- Design, develop, and maintain scalable data pipelines and ETL workflows using Pentaho and AWS
- Build and optimize data lake and data warehouse architectures
- Migrate existing data infrastructure to modern cloud-based platforms
- Develop reverse ETL and customer segmentation pipelines
Cloud & Infrastructure
- Work extensively with AWS services such as S3, RDS, Redshift, Athena, Glue, and MWAA
- Manage and migrate large-scale databases (multi-terabyte range)
- Build data APIs and data delivery services for operational and analytical applications
Data Orchestration & Automation
- Develop and manage workflows using Apache Airflow
- Automate manual data processes and optimize data delivery
- Improve infrastructure for scalability and performance
Development & Collaboration
- Write high-quality, scalable code in Python and SQL
- Work in Agile/SCRUM development environments
- Collaborate with cross-functional and offshore teams
- Perform testing, debugging, and performance tuning
Required Skills & Qualifications:
- Strong hands-on experience with Pentaho (ETL development)
- Deep knowledge of AWS data services (S3, RDS, Redshift, Athena, Glue, MWAA)
- Experience with Apache Airflow for orchestration
- Strong SQL and SQL analytics expertise
- Proficiency in Python
- Experience with Postgres and MySQL
- Experience building and optimizing large-scale data pipelines
- Strong understanding of data warehousing concepts
- Experience with Git or similar code repositories
- Strong analytical and problem-solving skills