Please click on the Apply to verify the status of jobs posted more than 15 days ago, as they may have expired. Similar Jobs
Job Description
JOB DESCRIPTION
Senior Data Engineer
Experience 6+ Years
Location - Navi Mumbai (onsite)
Job Summary:
We are seeking a highly skilled Data Engineer with deep expertise in Apache Kafka integration with Databricks, structured streaming, and large-scale data pipeline design using the Medallion Architecture. The ideal candidate will demonstrate strong hands-on experience in building and optimizing real-time and batch pipelines, and will be expected to solve real coding problems during the interview.
:
Design, develop, and maintain real-time and batch data pipelines in Databricks.
Integrate Apache Kafka with Databricks using Structured Streaming.
Implement robust data ingestion frameworks using Databricks Autoloader.
Build and maintain Medallion Architecture pipelines across Bronze, Silver, and Gold layers.
Implement checkpointing, output modes, and appropriate processing modes in structured streaming jobs.Design and implement Change Data Capture (CDC) workflows and Slowly Changing Dimensions (SCD) Type 1 and Type 2 logic.
Develop reusable components for merge/upsert operations and window function-based transformations.
Handle large volumes of data efficiently through proper partitioning, caching, and cluster tuning techniques.
Collaborate with cross-functional teams to ensure data availability, reliability, and consistency.
Must Have:
Apache Kafka: Integration, topic management, schema registry (Avro/JSON).
Databricks & Spark Structured Streaming:
*Processing Modes: Append, Update, Complete
*Output Modes: Memory, Console, File, Kafka, Delta
*Checkpointing and fault tolerance
Databricks Autoloader: Schema inference, schema evolution, incremental loads.
Medallion Architecture implementation expertise.
Performance Optimization:
*Data partitioning strategies
Looking to get Placed? Try our Placement Guarantee Plan
*Caching and persistence
*Adaptive query execution and cluster configuration tuning
SQL & Spark SQL: Proficiency in writing efficient queries and transformations.
Data Governance: Schema enforcement, data quality checks, and monitoring.
Good to Have:
Strong coding skills in Python and PySpark.
Experience working in CI/CD environments for data pipelines.
Exposure to cloud platforms (AWS/Azure/GCP).
Understanding of Delta Lake, time travel, and data versioning.
Familiarity with orchestration tools like Airflow or Azure Data Factory.
Immediate Joiners Preferred
Please share your CVs at [HIDDEN TEXT]
Skills
PythonData GovernanceImplementationData EngineerSqlIf an employer asks you to pay any kind of fee, please notify us immediately. Jobaaj does not charge any fee from the applicants and we do not allow other companies also to do so.
About Company
Important dates & deadlines?
Application Deadline
31 May 26, 06:46 PM IST
Similar Jobs
View All

