Staff Engineer - Data

Department Icon Data Science Analytics & Machine Learning
149+ Applicants
Posted: 1 month ago
0-1 years
Pune, Maharashtra
work from office

Posted: 1 month ago
|
Applicants: 150+
Job Description
Similar Jobs
Please verify your account first! Send OTP

Please click on the Apply to verify the status of jobs posted more than 15 days ago, as they may have expired. Similar Jobs

Job Description

About Our Group
The Global Wafer Systems (GWS) group is responsible for the design, development, and support of essential business solutions across factory locations in Asia, Europe, and the United States. The team specializes in the advancement and maintenance of factory control systems, utilizing Artificial Intelligence in fields including image processing, automated data monitoring, recommendation, and automated decision-making.
About The Role - You Will
  • Lead and architect data platform strategy, AI-ready data foundations, and engineering best practices at scale.
  • Lead the architecture and evolution of enterprise-scale data platforms supporting smart factory operations, including operational, parametric, sensor, and image data.
  • Own end-to-end data architecture for AI and GenAI use cases, including LLM/RAG pipelines, feature stores, vector-ready datasets, metadata frameworks, and data lineage.
  • Define and enforce data engineering standards for data modeling, ETL/ELT patterns, data quality, observability, governance, and lifecycle management.
  • Design scalable and resilient data pipelines across batch and streaming workloads using Hadoop, distributed SQL engines, and cloud/on?prem hybrid architectures.
  • Partner with AI/ML, application, and product teams to translate business problems into robust data solutions and reusable data assets.
  • Mentor and technically guide junior and senior data engineers, providing architectural direction, code reviews, and design feedback.
  • Lead complex data problem investigations, driving root-cause analysis, and systemic fixes rather than local workarounds.
  • Influence roadmap and technical strategy across teams and sites, aligning data capabilities with factory, analytics, and AI transformation goals.
  • Drive GenAI readiness at scale, ensuring datasets are discoverable, well-documented, semantically consistent, and safe for AI/agent consumption.
About You
  • Strong passion for data platform excellence, AI-ready data, and GenAI enablement.
  • Ability to think architecturally while remaining deeply technical.
  • Comfort influencing without authority across global, cross-functional teams.
  • A proactive mindset toward data quality, reliability, and long-term maintainability.
  • Willingness to operate in a global environment with flexible working hours when required.
Your Experience Includes
  • Advanced expertise in SQL, including performance tuning and optimization in distributed data platforms.
  • Deep experience designing and building large-scale ETL/ELT pipelines and data architectures.
  • Strong proficiency in Python for data engineering, automation, and framework development.
  • Expert-level data modeling skills, including dimensional, wide-table, and AI/feature-oriented models.
  • Hands-on experience with Oracle PL/SQL, including complex procedures and performance optimization.
  • Proven experience preparing data for AI/GenAI systems, including:
  • Looking to get Placed? Try our Placement Guarantee Plan

    Vector-friendly transformations
  • Metadata extraction and enrichment
  • Structured and unstructured data processing
  • Data validation for AI consumption
  • Experience leading technical design discussions and making architectural trade-offs.
  • Experience with GenAI frameworks and patterns, including RAG architectures, embeddings, inference optimization, and prompt/data observability.
  • Strong hands-on experience with distributed query engines (e.g., Trino) and big data ecosystems.
  • Experience with workflow orchestration tools such as Airflow or NiFi at scale.
  • Experience designing streaming ingestion frameworks using Spark Streaming, Kafka, or equivalent.
  • Background in data governance, data quality frameworks, and observability tooling.
Location:
Our site in Pune is dynamic, both in our cutting-edge, innovative work, as well as our vibrant on-site food, and athletic and personal development opportunities for our employees. You can enjoy breakfast, lunch, or dinner from one of four cafeterias in the park. Take a break from your workday and participate in one of our many walkathons or compete against your colleagues in carrom, chess and table tennis. Learn about a technical topic outside your area of expertise at one of our monthly Technical Speaker Series, or attend one of the frequent on-site cultural festivals, celebrations, and community volunteer opportunities.
Location: Pune, India
Travel: None

Skills

Artificial IntelligenceData ValidationBig DataBusiness SolutionsPythonData ArchitectureData GovernanceData ModelingEtlData ProcessingAi/mlAnalyticsAiMlSql

If an employer asks you to pay any kind of fee, please notify us immediately. Jobaaj does not charge any fee from the applicants and we do not allow other companies also to do so.

Important dates & deadlines?

Application Deadline

29 May 26, 01:55 PM IST

Similar Jobs

View All
Loading...
Bag Logo
Jobaaj
Don't Miss out any Updates

Subscribe now for the latest job alerts
and never miss an update

Job Alert
Google hiring for Specific Roles Apply Now!
1 min ago
New Opportunity
Amazon is hiring freshers Apply Now!
5 min ago
Featured Jobs
Microsoft opening 50+ positions Apply Now!
10 min ago

Staff Engineer - Data

Share with