Please click on the Apply to verify the status of jobs posted more than 15 days ago, as they may have expired. Similar Jobs
Job Description
- Strong proficiency in Python programming.
- Hands-on experience with PySpark and Apache Spark.
- Knowledge of Big Data technologies (Hadoop, Hive, Kafka, etc.).
- Experience with SQL and relational/non-relational databases.
- Familiarity with distributed computing and parallel processing.
- Understanding data engineering best practices.
- Experience with REST APIs, JSON/XML, and data serialization.
- Exposure to cloud computing environments.
- 5+ years of experience in Python and PySpark development.
- Experience with data warehousing and data lakes.
- Knowledge of machine learning libraries (e.g., MLlib) is a plus.
- Strong problem-solving and debugging skills.
- Excellent communication and collaboration abilities.
- Develop and maintain scalable data pipelines using Python and PySpark.
- Design and implement ETL (Extract, Transform, Load) processes.
- Optimize and troubleshoot existing PySpark applications for performance.
- Collaborate with cross-functional teams to understand data requirements.
- Write clean, efficient, and well-documented code.
- Conduct code reviews and participate in design discussions.
- Ensure data integrity and quality across the data lifecycle.
- Integrate with cloud platforms like AWS, Azure, or GCP.
Looking to get Placed? Try our Placement Guarantee Plan
Skills
Big DataPythonData IntegrityData WarehousingEtlMachine LearningSqlIf an employer asks you to pay any kind of fee, please notify us immediately. Jobaaj does not charge any fee from the applicants and we do not allow other companies also to do so.
About Company
Important dates & deadlines?
Application Deadline
30 May 26, 04:23 PM IST
Similar Jobs
View All

