Job Description
Responsibilities
- Design, develop, and maintain high-throughput ETL/ELT pipelines using AWS Glue (PySpark), orchestrated via Apache Airflow or AWS Step Functions.
- Own and optimize large-scale Amazon Redshift clusters and manage high concurrency workloads for a very large user base:
- Lead and contribute to migration projects from Snowflake or traditional RDBMS to Redshift, ensuring minimal downtime and robust validation.
- Integrate and normalize data from heterogeneous sources, including REST APIs, AWS Aurora (MySQL/Postgres), streaming inputs, and flat files.
- Implement intelligent caching strategies and leverage EC2 and serverless compute (Lambda, Glue) for custom transformations and processing at scale.
- Write advanced SQL for analytics, data reconciliation, and validation, demonstrating strong SQL development and tuning experience.
- Implement comprehensive monitoring, alerting, and logging for all data pipelines to ensure reliability, availability, and cost optimization.
- Collaborate directly with product managers, analysts, and client-facing teams to gather requirements and deliver insights-ready datasets.
- Champion data governance, security, and lineage, ensuring data is auditable and well-documented across all environments.
- 2-4 years of core data engineering experience, especially focused on Amazon Redshift hands-on performance tuning and large-scale management capacity.
- Demonstrated experience handling multi-terabyte Redshift clusters, concurrent query loads, and managing complex workload segmentation and queue priorities.
- Strong experience with AWS Glue (PySpark) for large-scale ETL jobs.
- Solid understanding and implementation experience of workflow orchestration using Apache Airflow or AWS Step Functions.
- Strong proficiency in Python, advanced SQL, and data modeling concepts.
- Familiarity with CI/CD pipelines, Git, DevOps processes, and infrastructure-as-code concepts.
- Experience with Amazon Athena, Lake Formation, or S3-based data lakes.
- Hands-on participation in Snowflake, BigQuery, or Teradata migration projects.
Looking to get Placed? Try our Placement Guarantee Plan
- AWS Certifications such as
- AWS Certified Data Analytics - Specialty.
- AWS Certified Solutions Architect - Associate/Professional.
- Exposure to real-time streaming architectures or Lambda architectures.
- Excellent communication skills are a must; must be able to confidently engage with both technical and non-technical stakeholders, including clients.
- Strong problem-solving mindset and a keen attention to performance, scalability, and reliability.
- Demonstrated ability to work independently, lead tasks, and take ownership of large-scale systems.
- Comfortable working in a fast-paced, dynamic, and client-facing environment.
Skills
Data AnalyticsPythonData GovernanceData ModelingEtlImplementationMysqlSnowflakeTeradataData EngineerAnalyticsSqlIf an employer asks you to pay any kind of fee, please notify us immediately. Jobaaj does not charge any fee from the applicants and we do not allow other companies also to do so.
Important dates & deadlines?
Application Deadline
03 Apr 26, 05:38 PM IST
Similar Jobs
View All



