Skip to main content
circle
Copied to clipboard

Your responsibilities:

  • Develop and migrate data processing, ETL processes, data pipelines and analytics / data science algorithms using Python, GBQ, Oracle.
  • Develop tools and co-create an internal framework to automate work in the DATA area
  • Collaborate with other DATA teams (Data Analytics Consulting, Data Science/MLOps, Datawarehouse Engineering) in the joint execution of projects
  • Active participation in the internal community of experts in the evaluation and co-creation of internal work standards
  • Substantive support of other team members

Your Profile:

  • A minimum of 3 years of Python programming experience in the creation, development and maintenance of high-quality applications / modules focused on Data Processing (ETL), Analysis and Data Modeling that utilizes object-oriented approach
  • Ability to develop, effectively test (pytest) and maintain clean, clear, and high-quality code
  • Willingness to acquire new knowledge and transfer knowledge to colleagues, skills in effective knowledge management and knowledge transfer (knowledge of Confluence and Mural is a plus)
  • Good working knowledge of ticket management environment (JIRA type) and distributed code management (git/BitBucket type).
  • Good knowledge of SQL (Oracle SQL or GBQ preferred) and PySpark
  • Fluency in spoken and written English:
  • -spoken and written English (B2) and Polish native-speaker or
  • -English (C1) (if not Polish native-speaker)
  • Your additional assets will be:
  • - Experience in Process Automation and maintenance (DevOps) using Airflow/CI-CD/Jenkins
  • - You are familiar with containerization techniques (preferred: Docker and Colima)
  • - Good knowledge of Google Cloud Platform (Terraform knowledge is a plus)
dots

How about?

Don't wait send your cv!
Apply
circle
scroll-to-top