AE

Data Architect

Aeries Technology
Bangalore10-15 LPA Posted 17 Jun 2025
FULL TIME
Database Management Systems
Power Bi
Data Architecture
Data Warehousing
Saas
+3 more

Job Description

  • We are looking for a Senior Data Engineer with an architect s mindset and thought leadership approach on the Data & Analytics team. This role goes beyond development its about end-to-end thinking, mentoring and collaborating with a global team.
  • As a key technical leader, you will align data strategies with business needs, and drive best practices in data modeling, governance, and performance optimization. you will play a pivotal role in streamlining tooling for Democratized Development within Snowflake, DBT related data platforms and build high impact data models.
  • You will drive performance optimization, cost efficiency, data governance, and model architecture, ensuring that our data infrastructure is scalable, secure, and high-performing.
  • This role requires deep expertise in cloud-based data engineering, a strong problem-solving mindset, and the ability to collaborate with business and analytics teams . You will work on optimizing compute resources, improving data quality, and enforcing governance standards , support code reviews, and mentor team members, fostering a high-performing data engineering culture.

Key Responsibilities

  • Data Acquisition & Pipeline Development
  • Develop and maintain scalable data pipelines for efficient data ingestion, transformation, and integration.
  • Work with Fivetran, Python, and other ETL/ELT tools to automate and optimize data acquisition from various sources.
  • Ensure reliable data movement from SaaS platforms (eg, Salesforce, Gong, Google Analytics) and operational databases into Snowflake.
  • Monitor and enhance pipeline performance, identifying areas for optimization and fault tolerance.
  • Evaluate and recommend new technologies for data ingestion, transformation, and orchestration.
  • Democratized Development & Tooling
  • Enable and streamline self-service data development for analysts and data practitioners.
  • Drive best practices in DBT and Snowflake for modular, reusable, and governed data modeling.
  • Design and maintain CI/CD pipelines to support version control, testing, and deployment in a modern data stack.
  • Performance Optimization & Cost Efficiency
  • Optimize and tune Snowflake queries and workloads for performance and cost efficiency.
  • Implement warehouse resource scaling strategies to reduce compute costs while maintaining SLAs.
  • Monitor and analyze query performance, storage consumption, and data usage patterns to identify optimization opportunities.
  • Data Quality, Governance & Security
  • Establish data quality monitoring frameworks, integrating automated validation and anomaly detection.
  • Enforce data governance policies, including access controls, lineage tracking, and compliance standards.
  • Work closely with security teams to implement data protection strategies in Snowflake.
  • Data Model Development & Architecture Review
  • Design and develop scalable, well-structured data models in Snowflake.
  • Perform data model reviews to ensure consistency, efficiency, and alignment with business needs.
  • Collaborate with Analytics & BI teams to define metrics layers and transformation logic in DBT.
  • Collaboration & Agile Execution
  • Collaborate with BI teams to ensure data models meet reporting requirements in Power BI.
  • Partner with cross-functional teams including Sales, Marketing, Customer Support, Finance, and Product to deliver trusted, high-quality data solutions.
  • Work within an Agile framework, delivering iterative improvements to data infrastructure.
  • Stay ahead of industry trends in modern data engineering, Snowflake, and DBT to drive innovation.

Qualifications

  • Bachelor s or Master s degree in Computer Science, Information Technology, or a related field.
  • 5+ years of experience in Data Engineering with strong expertise in Snowflake.
  • 2+ years of hands-on experience in DBT for data modeling and transformation.
  • 2+ years of experience in Python, particularly for data pipelines and automation.
  • Strong expertise in SQL optimization and performance tuning.
  • Deep understanding of ETL/ELT architectures, data warehousing, and cloud data management best practices.
  • Experience implementing cost monitoring and optimization techniques in Snowflake.
  • Strong problem-solving skills and ability to troubleshoot complex data issues.
  • Excellent communication skills and ability to work in collaborative, cross-functional teams.
  • Experience with Agile methodologies and iterative data development processes is a plus.