Senior Data Engineer
2 weeks ago
Requirements: Proven experience in owning and operating mission-critical data products in both batch and streaming environments. Strong expertise in data modeling, schema management, data contracts, and observability practices. Solid understanding of building systems optimised for AI/agent use cases (e.g., agent-friendly schemas, embeddings, vector search). Hands-on experience with modern data infrastructure and AWS-based ecosystems. Excellent communication skills and fluency in English. Join our team in Warsaw, where we’re collaborating on a cutting-edge fintech venture with a global industry leader. Together with our Partner – Klarna, we’re building an IT hub designed to drive innovation in digital payment solutions. We’re on the lookout for top-tier engineers who thrive in dynamic, forward-thinking environments. Spyrosoft is leading the recruitment process, facilitating a seamless experience for candidates who are ready to shape the future of online shopping and payments. This opportunity is ideal for engineers who value independence, proactiveness, and flexibility. Our engagement begins with a B2B contract through Spyrosoft, transitioning to a direct contract with our Partner. We offer a hybrid work model in Warsaw’s vibrant Wola district. English fluency and eligibility to work in Poland are essential, as is the successful completion of a background check to meet the rigorous standards of the financial domain. Our process: CV selection Initial recruitment screening Online logic test Cultural fit interview Project description: Join Klarna’s global underwriting data team and take ownership of the data products that power decisioning, monitoring, and optimisation across the company. You will design and operate self-describing data products built for both AI agents and humans that serve as the single source of truth for underwriting globally. This is a hands-on engineering role with significant ownership and cross-functional collaboration, directly impacting Klarna’s risk, credit, and financial models. Tech stack: Languages: SQL, PySpark, Python Frameworks: Apache Airflow, AWS Glue, Kafka, Redshift Cloud & DevOps: AWS (S3, Lambda, CloudWatch, SNS/SQS, Kinesis), Terraform, Git, CI/CD Collaborate closely with Credit Portfolio Management, Policy, Modelling, Treasury, and Finance teams to deliver data features supporting regulatory reporting, model development, and business optimisation. ,[Own and maintain Klarna’s global underwriting (UW) data tables, covering canonical facts and dimensions such as applications, decisions, features, repayments, and delinquency. , Ensure data quality and reliability through well-defined SLAs focused on freshness, completeness, accuracy, and lineage. , Design and maintain systems optimised for both AI agents and human users, including consistent IDs, canonical event structures, explicit metric definitions, and rich metadata (schemas, data dictionaries, machine-readable contracts). , Build and operate scalable data pipelines (batch and streaming) that feed underwriting scoring systems, real-time decisioning, monitoring, and optimisation processes. , Implement observability and quality monitoring - alerts, audits, reconciliations, and automated backfills and drive post-incident reviews. ] Requirements: AI, SQL, PySpark, Python, Apache Airflow, AWS, Glue, Kafka, Redshift, Cloud, AWS S3, AWS Lambda, CloudWatch, SNS, Amazon SQS, Kinesis, Terraform, Git, Data modeling, Use cases, Communication skills, Data pipelines, Audits Tools: .
-
Senior Data Engineer @ Bayer
2 weeks ago
Warsaw, Czech Republic Bayer Full timeBachelor’s degree in Computer Science, Data Science, Information Technology, or a related field. 5+ years of experience in data engineering, data operations, or a similar role, delivering production-grade data pipelines and services. Familiarity with data product lifecycle management across disciplines: Demonstrated expertise designing, implementing, and...
-
Senior Azure Data Engineer @ Experis Polska
2 weeks ago
Warsaw, Czech Republic Experis Polska Full timeTech Stack Programming: Python, PySpark, SQL, SparkSQL, Bash Azure: Databricks, Data Factory, Delta Lake, Data Vault 2.0 CI/CD: Azure DevOps, GitHub, Jenkins Orchestration: Airflow, Azure Data Factory Databases: SQL Server, Oracle, PostgreSQL, Vertica Cloud: Azure (expert), AWS (intermediate) Tools: FastAPI, REST APIs, Docker, Unity Catalog Preferred...
-
Senior Data Engineer @ hubQuest
5 days ago
Remote, Warsaw, Czech Republic hubQuest Full timeWhat we expect 5+ years of professional experience as a Data Engineer or Software Engineer in data-intensive environments Strong Python development skills, with solid understanding of OOP, modular design, and testing (unit/integration) Experience with PySpark and distributed data processing frameworks Hands-on experience with Azure Data ecosystem,...
-
Senior Data Engineer @ RemoDevs
2 weeks ago
Warsaw, Czech Republic RemoDevs Full time3+ years of Python development experience, including Pandas 5+ years writing complex SQL queries with RDBMSes. 5+ years of Experience with developing and deploying ETL pipelines using Airflow, Prefect, or similar tools. Experience with cloud-based data warehouses in environments such as RDS, Redshift, or Snowflake. Experience with data warehouse design:...
-
Senior Data Analyst @ Bayer
2 weeks ago
Warsaw, Czech Republic Bayer Full time5+ years in analytics/BI roles, with a track record of delivering business impact in commercial or lifecycle environments. Expert in data wrangling and analysis skills; comfort with ambiguous questions and iterative problem framing. Good proficiency in Python, SQL, visualization tools such as PowerBI, Tableau, or ThoughtSpot, and data platforms such as...
-
Senior Data Scientist @ hubQuest
5 days ago
Remote, Warsaw, Czech Republic hubQuest Full timeWhat we expect 5+ years of professional experience in Data Science or ML Engineering, including production deployments MSc or PhD in Computer Science, Statistics, Mathematics, Physics, or related technical field Strong Python programming skills, including software engineering practices (OOP, modular code design, testing) Solid experience with ML frameworks...
-
Senior Data Engineer @ N-iX
1 week ago
Remote, Krakow, Wrocław, Warsaw, Czech Republic N-iX Full timeMust-Have Technologies Experience with cloud data warehouse technologies, including Snowflake and AWS S3. Solid SQL skills; experience migrating legacy environments to cloud platforms. Familiarity with ETL frameworks and tools (preferably SSIS, dbt). Proficiency in Python for data pipeline development and automation. Strong documentation and communication...
-
Remote, Warsaw, Czech Republic KMD Poland Full timeIdeal candidate: Has 5+ years of commercial experience in implementing, developing, or maintaining data load systems (ETL/ELT). Demonstrates strong programming skills in Python, with a deep understanding of data-related challenges. Has hands-on experience with Apache Spark and Databricks. Is familiar with MSSQL databases. Has experience working...
-
Senior Data Governance Specialist @ Bayer
2 weeks ago
Warsaw, Czech Republic Bayer Full timeBachelor’s degree in Computer Science, Information Systems, Data Science, or a related field. 5+ years in data management, including 2+ years with enterprise data cataloging or metadata platforms at scale in complex environments. Expert in solutions like Collibra; proven success integrating catalog standards (datasets, pipelines, metrics, lineage,...
-
Data Engineer @ ITMAGINATION
1 day ago
Warsaw, Czech Republic ITMAGINATION Full timeStrong SQL (performance, CTEs, window functions) and Python for data work. Hands-on Snowflake (warehousing, performance, role/security basics) and dbt (models, tests, docs). Proven experience delivering end-to-end pipelines (extract → transform → model → serve for analytics). Ability to handle unstructured/inconsistent data and normalize...