About the job
Key Qualifications
Proven expertise in Python programming.
Practical experience with PySpark and Apache Spark.
Familiarity with Big Data technologies such as Hadoop, Hive, and Kafka.
Strong background in SQL and both relational and non-relational databases.
Understanding of distributed computing and parallel processing.
Knowledge of data engineering best practices.
Experience with REST APIs, JSON/XML, and data serialization techniques.
Exposure to cloud computing platforms.
Additionally, we prefer candidates with:
Over 5 years of experience in Python and PySpark development.
Experience managing data warehousing and data lakes.
Knowledge of machine learning libraries, such as MLlib, is a plus.
Strong analytical and debugging skills.
Excellent communication and teamwork abilities.

