Senior Data Engineer

2 weeks ago


Warsaw, Czech Republic Spyrosoft Full time

Requirements: Proven experience in owning and operating mission-critical data products in both batch and streaming environments.  Strong expertise in data modeling, schema management, data contracts, and observability practices.  Solid understanding of building systems optimised for AI/agent use cases (e.g., agent-friendly schemas, embeddings, vector search).  Hands-on experience with modern data infrastructure and AWS-based ecosystems.  Excellent communication skills and fluency in English.  Join our team in Warsaw, where we’re collaborating on a cutting-edge fintech venture with a global industry leader. Together with our Partner – Klarna, we’re building an IT hub designed to drive innovation in digital payment solutions. We’re on the lookout for top-tier engineers who thrive in dynamic, forward-thinking environments. Spyrosoft is leading the recruitment process, facilitating a seamless experience for candidates who are ready to shape the future of online shopping and payments. This opportunity is ideal for engineers who value independence, proactiveness, and flexibility. Our engagement begins with a B2B contract through Spyrosoft, transitioning to a direct contract with our Partner. We offer a hybrid work model in Warsaw’s vibrant Wola district. English fluency and eligibility to work in Poland are essential, as is the successful completion of a background check to meet the rigorous standards of the financial domain. Our process: CV selection  Initial recruitment screening  Online logic test  Cultural fit interview  Project description: Join Klarna’s global underwriting data team and take ownership of the data products that power decisioning, monitoring, and optimisation across the company. You will design and operate self-describing data products built for both AI agents and humans that serve as the single source of truth for underwriting globally. This is a hands-on engineering role with significant ownership and cross-functional collaboration, directly impacting Klarna’s risk, credit, and financial models. Tech stack: Languages: SQL, PySpark, Python  Frameworks: Apache Airflow, AWS Glue, Kafka, Redshift  Cloud & DevOps: AWS (S3, Lambda, CloudWatch, SNS/SQS, Kinesis), Terraform, Git, CI/CD  Collaborate closely with Credit Portfolio Management, Policy, Modelling, Treasury, and Finance teams to deliver data features supporting regulatory reporting, model development, and business optimisation. ,[Own and maintain Klarna’s global underwriting (UW) data tables, covering canonical facts and dimensions such as applications, decisions, features, repayments, and delinquency. , Ensure data quality and reliability through well-defined SLAs focused on freshness, completeness, accuracy, and lineage. , Design and maintain systems optimised for both AI agents and human users, including consistent IDs, canonical event structures, explicit metric definitions, and rich metadata (schemas, data dictionaries, machine-readable contracts). , Build and operate scalable data pipelines (batch and streaming) that feed underwriting scoring systems, real-time decisioning, monitoring, and optimisation processes. , Implement observability and quality monitoring - alerts, audits, reconciliations, and automated backfills and drive post-incident reviews. ] Requirements: AI, SQL, PySpark, Python, Apache Airflow, AWS, Glue, Kafka, Redshift, Cloud, AWS S3, AWS Lambda, CloudWatch, SNS, Amazon SQS, Kinesis, Terraform, Git, Data modeling, Use cases, Communication skills, Data pipelines, Audits Tools: .



  • Warsaw, Czech Republic Bayer Full time

    Bachelor’s degree in Computer Science, Data Science, Information Technology, or a related field. 5+ years of experience in data engineering, data operations, or a similar role, delivering production-grade data pipelines and services. Familiarity with data product lifecycle management across disciplines: Demonstrated expertise designing, implementing, and...


  • Warsaw, Czech Republic Experis Polska Full time

    Tech Stack Programming: Python, PySpark, SQL, SparkSQL, Bash Azure: Databricks, Data Factory, Delta Lake, Data Vault 2.0 CI/CD: Azure DevOps, GitHub, Jenkins Orchestration: Airflow, Azure Data Factory Databases: SQL Server, Oracle, PostgreSQL, Vertica Cloud: Azure (expert), AWS (intermediate) Tools: FastAPI, REST APIs, Docker, Unity Catalog Preferred...


  • Remote, Warsaw, Czech Republic hubQuest Full time

    What we expect 5+ years of professional experience as a Data Engineer or Software Engineer in data-intensive environments Strong Python development skills, with solid understanding of OOP, modular design, and testing (unit/integration) Experience with PySpark and distributed data processing frameworks Hands-on experience with Azure Data ecosystem,...


  • Warsaw, Czech Republic RemoDevs Full time

    3+ years of Python development experience, including Pandas 5+ years writing complex SQL queries with RDBMSes. 5+ years of Experience with developing and deploying ETL pipelines using Airflow, Prefect, or similar tools. Experience with cloud-based data warehouses in environments such as RDS, Redshift, or Snowflake. Experience with data warehouse design:...


  • Warsaw, Czech Republic Bayer Full time

    5+ years in analytics/BI roles, with a track record of delivering business impact in commercial or lifecycle environments. Expert in data wrangling and analysis skills; comfort with ambiguous questions and iterative problem framing. Good proficiency in Python, SQL, visualization tools such as PowerBI, Tableau, or ThoughtSpot, and data platforms such as...


  • Remote, Warsaw, Czech Republic hubQuest Full time

    What we expect 5+ years of professional experience in Data Science or ML Engineering, including production deployments MSc or PhD in Computer Science, Statistics, Mathematics, Physics, or related technical field Strong Python programming skills, including software engineering practices (OOP, modular code design, testing) Solid experience with ML frameworks...


  • Remote, Krakow, Wrocław, Warsaw, Czech Republic N-iX Full time

    Must-Have Technologies Experience with cloud data warehouse technologies, including Snowflake and AWS S3. Solid SQL skills; experience migrating legacy environments to cloud platforms. Familiarity with ETL frameworks and tools (preferably SSIS, dbt). Proficiency in Python for data pipeline development and automation. Strong documentation and communication...


  • Remote, Warsaw, Czech Republic KMD Poland Full time

    Ideal candidate:   Has 5+ years of commercial experience in implementing, developing, or maintaining data load systems (ETL/ELT).  Demonstrates strong programming skills in Python, with a deep understanding of data-related challenges.  Has hands-on experience with Apache Spark and Databricks.  Is familiar with MSSQL databases.  Has experience working...


  • Warsaw, Czech Republic Bayer Full time

    Bachelor’s degree in Computer Science, Information Systems, Data Science, or a related field. 5+ years in data management, including 2+ years with enterprise data cataloging or metadata platforms at scale in complex environments. Expert in solutions like Collibra; proven success integrating catalog standards (datasets, pipelines, metrics, lineage,...


  • Warsaw, Czech Republic ITMAGINATION Full time

    Strong SQL (performance, CTEs, window functions) and Python for data work.   Hands-on Snowflake (warehousing, performance, role/security basics) and dbt (models, tests, docs).   Proven experience delivering end-to-end pipelines (extract → transform → model → serve for analytics).   Ability to handle unstructured/inconsistent data and normalize...