Develop and migrate data processing, ETL processes, data pipelines and analytics / data science algorithms using Python, GBQ, Oracle.
Develop tools and co-create an internal framework to automate work in the DATA area
Collaborate with other DATA teams (Data Analytics Consulting, Data Science/MLOps, Datawarehouse Engineering) in the joint execution of projects
Active participation in the internal community of experts in the evaluation and co-creation of internal work standards
Substantive support of other team members
Your Profile:
A minimum of 3 years of Python programming experience in the creation, development and maintenance of high-quality applications / modules focused on Data Processing (ETL), Analysis and Data Modeling that utilizes object-oriented approach
Ability to develop, effectively test (pytest) and maintain clean, clear, and high-quality code
Willingness to acquire new knowledge and transfer knowledge to colleagues, skills in effective knowledge management and knowledge transfer (knowledge of Confluence and Mural is a plus)
Good working knowledge of ticket management environment (JIRA type) and distributed code management (git/BitBucket type).
Good knowledge of SQL (Oracle SQL or GBQ preferred) and PySpark
Fluency in spoken and written English:
-spoken and written English (B2) and Polish native-speaker or
-English (C1) (if not Polish native-speaker)
Your additional assets will be:
- Experience in Process Automation and maintenance (DevOps) using Airflow/CI-CD/Jenkins
- You are familiar with containerization techniques (preferred: Docker and Colima)
- Good knowledge of Google Cloud Platform (Terraform knowledge is a plus)