FU

python,pyspark

Fusion Plus Solutions
Hyderabad5-8 LPA Posted 18 Jul 2025
FULL TIME
Pyspark
Sql
Airflow
Aws
Databricks

Job Description

  • Develop and optimize data processing jobs using PySpark for complex transformations and aggregations
  • Design and implement robust data pipelines on AWS, ensuring scalability and performance
  • Utilize AWS services like EC2 and S3 for data processing and storage solutions
  • Manage SQL database schema design, query optimization, and performance tuning
  • Design and maintain scalable data warehouses using best practices in data modeling and ETL
  • Leverage modern data platforms for collaborative data science across varied sources
  • Maintain Databricks clusters and Spark jobs for optimal performance and data quality
  • Implement security, backup procedures, and disaster recovery plans using AWS best practices
  • Manage source code and automate deployment via GitHub with CI/CD pipelines
  • Troubleshoot and optimize PySpark scripts, Databricks notebooks, SQL queries, and Airflow DAGs
  • Stay current with advancements in cloud data technologies and recommend useful tools
  • Use Apache Airflow for orchestrating and automating data workflows reliably
  • Collaborate with data scientists and analysts to build models and pipelines for analytics and ML projects
Join WhatsApp Channel