FUFusion Plus Solutions
python,pyspark
Hyderabad ₹5-8 LPA Posted 18 Jul 2025
FULL TIME
Pyspark
Sql
Airflow
Aws
Databricks
Job Description
- Develop and optimize data processing jobs using PySpark for complex transformations and aggregations
- Design and implement robust data pipelines on AWS, ensuring scalability and performance
- Utilize AWS services like EC2 and S3 for data processing and storage solutions
- Manage SQL database schema design, query optimization, and performance tuning
- Design and maintain scalable data warehouses using best practices in data modeling and ETL
- Leverage modern data platforms for collaborative data science across varied sources
- Maintain Databricks clusters and Spark jobs for optimal performance and data quality
- Implement security, backup procedures, and disaster recovery plans using AWS best practices
- Manage source code and automate deployment via GitHub with CI/CD pipelines
- Troubleshoot and optimize PySpark scripts, Databricks notebooks, SQL queries, and Airflow DAGs
- Stay current with advancements in cloud data technologies and recommend useful tools
- Use Apache Airflow for orchestrating and automating data workflows reliably
- Collaborate with data scientists and analysts to build models and pipelines for analytics and ML projects