As a Data Engineer you will be responsible for designing, maintaining, and optimizing ELT processes that process many billions of records. Data Engineering applies our curiosity and passion to create data pipelines that deliver insights and help our clients make better decisions. We design, implement and deploy scalable pipelines that power 1010data’s products.

Our modern tech stack builds pipelines using 1010data’s industry-leading big data platform and open-source tools such as Airflow, Spark, and Python to analyze billions of rows of data every day. In this role, you will work closely with data scientists, product leaders, and client-facing teams as you build new capabilities.

Very well

  • End-to-end ownership of every step in a complex data pipeline, in multiple languages including our proprietary query language.
  • Understand the business logic of every step in the process, allowing you and other team members to make informed decisions about design changes and improvements.
  • Design and write Apache Spark scripts to process terabytes of data for ingestion into our pipeline.
  • Design, write, and maintain Apache Airflow jobs that coordinate between multiple technologies.
  • Learn, design, and write high-efficiency transformations in our proprietary query language and integrate those queries into the pipeline.
  • Ensuring quality, reliability, and uptime for critical automated processes, including by helping the data quality team diagnose and resolve issues in the pipeline and in the data.

What you already have:

  • 2+ years of professional experience programming in Python and/or Scala .
  • 1+ year of experience maintaining a client-facing pipeline in a production environment.
  • Fluent in SQL and communicates database concepts including joins, schemas, indexing.
  • Ability to communicate clearly in meetings and written documents, consider options presented by others, and reach a balanced technical opinion.
  • Ability to relate the needs of the customer to the data engineering solutions we can build with our products.

Preferred Skills:

  • Proficiency in NoSQL databases and database design, distributed systems and/or information retrieval.
  • Experience writing software using Apache Airflow.
  • Experience writing software using Apache Spark.
  • Experience writing high-performance queries in SQL or another query language.
  • Ability to independently plan and collect requirements for projects, and interact with the analyst and data science teams.


STEM Bachelor’s required; a graduate degree is a big plus.

Packages and extras

  • Healthcare package
  • Healthcare package for families
  • Leisure package
  • Leisure package for families
  • Language courses
  • Conferences
  • Trainings
  • Books
  • Equity

Relocation package

  • Adaptation tips
  • Help finding an apartment
  • Visa Services
  • Flight ticket
  • Open to Ukrainian candidates


  • Bicycle parking
  • Cold beverages
  • Hot beverages
  • Fruits
  • Snacks
  • Shower
  • Chill room
  • Integration events
  • Lunches

HR screening 30min zoom call.
1h technical interview.
General call with the team leader.
Making an offer meeting.



An extended tribe of Inuits Belgium, founded in 2019 in Kraków.  On the professional side – we support our clients by building them teams focused on all things Data. That’s the theory, but who we really are? In one word: people. Our slightly unconventional team consists of amazing individuals coming from different backgrounds, skilled in a wide spectrum of professions within IT and specifically Big Data. 

What we stand for? Diversity and inclusion play a significant role in our team. Why? As an Inuits family and group of professionals, we aim to remain open-minded in order to deliver the best solutions and quality to our clients. As such, we believe these are rooted in diverse backgrounds, experiences, and therefore points of view. 

What’s (less, but) also important? Sense of humor. Don’t have it? No worries, this comes eventually after spending some time in our Igloo/office.