Minimum 2 years of professional experience as a programmer working with large datasets.Experience in at least 1 project involving the processing of large datasets.Experience in at least 1 project programming with Python.Experience in at least 1 project within an on-premise computing environment.Proven experience programming with Apache Spark.Proven experience programming with Python.Proven experience programming with Apache Airflow.Proven experience programming with SQL.Familiarity with Hadoop concepts.Proven experience in programming ELT/ETL processes.Understanding of CI/CD principles and practices.Proficiency in using a version control system (Git).Strong self-organization and problem-solving skills.
About the job
Key Responsibilities:
Design, develop, and implement efficient ELT/ETL processes to handle extensive datasets.
Construct and enhance data processing workflows utilizing Apache Spark.
Employ Python for data manipulation, transformation, and analysis tasks.
Create and oversee data pipelines with Apache Airflow.
Write and refine SQL queries for data extraction, transformation, and loading.
Collaborate with data scientists, analysts, and fellow engineers to comprehend data requirements and deliver impactful solutions.
Operate within an on-premise computing environment for data processing and storage.
Guarantee data quality, integrity, and performance throughout the data lifecycle.
Engage in the implementation and maintenance of CI/CD pipelines for data processes.
Utilize Git for version control and collaborative development.
Diagnose and rectify issues related to data pipelines and infrastructure.
Contribute to the documentation of data processes and systems.
About Inetum
Inetum Polska is a prominent member of the global Inetum Group, dedicated to facilitating the digital transformation of businesses and public institutions. With a presence in cities such as Warsaw, Poznan, Katowice, Lublin, Rzeszow, and Lodz, the company offers a diverse array of IT services. Inetum Polska is committed to employee development, fully financing training, certifications, and attendance at technology conferences. The company also engages in local social initiatives, including charitable projects and the promotion of an active lifestyle. It takes pride in creating a diverse and inclusive work environment, ensuring equal opportunities for all.Globally, Inetum operates across 19 countries and employs over 28,000 professionals, focusing on four key areas:Consulting (Inetum Consulting): Providing strategic advisory services that assist organizations in defining and implementing innovative solutions.Infrastructure and Application Management: Delivering robust infrastructure services to support business operations.Digital Services: Offering a suite of digital solutions tailored to enhance business efficiency.Data and Analytics: Empowering organizations with data-driven insights to facilitate informed decision-making.
This job posting is no longer active and is not accepting applications.
Minimum 2 years of professional experience as a programmer working with large datasets.Experience in at least 1 project involving the processing of large datasets.Experience in at least 1 project programming with Python.Experience in at least 1 project within an on-premise computing environment.Proven experience programming with Apache Spark.Proven experience programming with Python.Proven experience programming with Apache Airflow.Proven experience programming with SQL.Familiarity with Hadoop concepts.Proven experience in programming ELT/ETL processes.Understanding of CI/CD principles and practices.Proficiency in using a version control system (Git).Strong self-organization and problem-solving skills.
About the job
Key Responsibilities:
Design, develop, and implement efficient ELT/ETL processes to handle extensive datasets.
Construct and enhance data processing workflows utilizing Apache Spark.
Employ Python for data manipulation, transformation, and analysis tasks.
Create and oversee data pipelines with Apache Airflow.
Write and refine SQL queries for data extraction, transformation, and loading.
Collaborate with data scientists, analysts, and fellow engineers to comprehend data requirements and deliver impactful solutions.
Operate within an on-premise computing environment for data processing and storage.
Guarantee data quality, integrity, and performance throughout the data lifecycle.
Engage in the implementation and maintenance of CI/CD pipelines for data processes.
Utilize Git for version control and collaborative development.
Diagnose and rectify issues related to data pipelines and infrastructure.
Contribute to the documentation of data processes and systems.
About Inetum
Inetum Polska is a prominent member of the global Inetum Group, dedicated to facilitating the digital transformation of businesses and public institutions. With a presence in cities such as Warsaw, Poznan, Katowice, Lublin, Rzeszow, and Lodz, the company offers a diverse array of IT services. Inetum Polska is committed to employee development, fully financing training, certifications, and attendance at technology conferences. The company also engages in local social initiatives, including charitable projects and the promotion of an active lifestyle. It takes pride in creating a diverse and inclusive work environment, ensuring equal opportunities for all.Globally, Inetum operates across 19 countries and employs over 28,000 professionals, focusing on four key areas:Consulting (Inetum Consulting): Providing strategic advisory services that assist organizations in defining and implementing innovative solutions.Infrastructure and Application Management: Delivering robust infrastructure services to support business operations.Digital Services: Offering a suite of digital solutions tailored to enhance business efficiency.Data and Analytics: Empowering organizations with data-driven insights to facilitate informed decision-making.