Data Engineer @

3 days ago


Kraków, Lesser Poland, Czech Republic ABB Full time
  • Advanced degree in Computer Science, Engineering, Data Science, or a related field (Master's preferred).
  • Proven experience (preferably 3+ years) as a Data Engineer with demonstrated expertise in building production-grade data pipelines and hands-on experience with Microsoft Fabric (Data Factory, Lakehouse, Dataflows).
  • Strong knowledge of ETL/ELT concepts, data pipeline design, and experience integrating data from diverse sources including APIs, databases (SQL Server), Snowflake, MuleSoft, and semi-structured formats.
  • Proficiency in SQL and Python, with experience in data processing frameworks and modern software development practices (Git, CI/CD, automated testing).
  • Familiarity with data modeling, data warehousing, domain-driven design, and experience with cloud platforms, ideally Azure.
  • Knowledge of data governance principles, Power BI semantic modeling, Delta Lake, or Synapse Analytics (preferred).
  • Experience with industrial data sources, time-series data, and IoT data streams.

At ABB, we help industries outrun - leaner and cleaner. Here, progress is an expectation - for you, your team, and the world. As a global market leader, we'll give you what you need to make it happen. It won't always be easy, growing takes grit. But at ABB, you'll never run alone. Run what runs the world.

 ABB's Service Division partners with our customers to improve the availability, reliability, predictability and sustainability of electrical products and installations. The Division's extensive service portfolio offers product care, modernization, and advisory services to improve performance, extend equipment lifetime and deliver new levels of operational and sustainable efficiency. We help customers keep resources in use for as long as possible, extracting the maximum value from them, and then recovering and regenerating products and materials at the end of their useful life.

We are seeking a skilled and detail-oriented Data Engineer to design and implement robust data infrastructure solutions that enable advanced analytics and AI-driven insights for industrial asset management. This role involves building scalable data pipelines using Microsoft Fabric to consolidate, transform, and model data from multiple heterogeneous sources. The primary objective is to provide reliable, efficient, and scalable access to high-quality data that supports predictive maintenance analytics, risk assessment models, and strategic decision-making. You will be responsible for creating the data foundation that empowers data scientists and analysts to deliver actionable insights for optimizing maintenance strategies and enhancing operational efficiency. 

The work model for the role is: hybrid #LI-hybrid

,[Design, develop, and maintain ETL/ELT pipelines in Microsoft Fabric for ingesting and transforming data from various sources including REST APIs, SQL Server, MuleSoft middleware, Snowflake, and file data sources (JSON, CSV, Excel, etc.). , Implement and manage dataflows, data pipelines, and Lakehouse models in Fabric to support advanced analytics and AI model development. , Develop and optimize data processing logic using PySpark within Microsoft Fabric notebooks for complex transformations and large-scale data processing tasks. , Build and maintain domain-driven data models that support analytics, reporting, self-service BI, and machine learning workflows. , Ensure data quality, integrity, and security across the entire data lifecycle, implementing robust data governance practices. , Collaborate with data scientists, analysts, software architects, and business stakeholders to understand requirements and deliver fit-for-purpose data solutions. , Monitor and troubleshoot pipeline performance, apply best practices in data architecture and performance optimization, and implement improvements as needed. , Document data processes, models, and technical decisions to ensure knowledge transfer and maintainability. ] Requirements: ETL, SQL Server, Snowflake, MuleSoft, SQL, Python, Git, CI/CD, Automated testing, Data modeling, Domain Driven Design, Cloud platform, IoT, Azure, Power BI, Delta Lake, Synapse Analytics

  • Kraków, Lesser Poland, Czech Republic ITDS Full time

    You're ideal for this role if you have:Strong experience in PySpark, Scala, or similar data engineering languages Hands-on experience building production data pipelines using Hadoop, Spark, and Hive Knowledge of cloud platforms and migrating on-premise solutions to the cloud Experience with scheduling tools such as Airflow and workflow...


  • Kraków, Lesser Poland, Czech Republic HSBC Technology Poland Full time

    Strong experience with database technologies (SQL, NoSQL), data warehousing solutions, and big data technologies (Hadoop, Spark).Proficiency in programming languages such as Python, Java, or Scala.Experience with cloud platforms (AWS, Azure, Google Cloud) and their data services. Certifications in cloud platforms (AWS Certified Data Analytics, Google...

  • Data Engineer @

    1 week ago


    Kraków, Lesser Poland, Czech Republic Antal Full time

    5+ years of IT experience, with 2+ years in software development using Big Data technologies, microservices, and event-driven cloud architectures.Hands-on experience with Apache NiFi, Kafka, Spark, Hive, HDFS, Oozie, SQL, Python, and Linux Shell scripting.Strong database skills: at least one SQL database (Oracle, PostgreSQL, MySQL, etc.) and one NoSQL...


  • Kraków, Lesser Poland, Czech Republic HSBC Technology Poland Full time

    What you need to have to succeed in this roleExcellent experience in the Data Engineering Lifecycle. You will have created data pipelines which take data through all layers from generation, ingestion, transformation and serving. Senior stakeholder management skills. Experience of modern Software Engineering principles and experience of creating well tested...


  • Kraków, Lesser Poland, Czech Republic beBeeDataEngineering Full time 105,000 - 135,000

    Job Title: Big Data Expertise EngineerJob DescriptionAs a skilled data engineer, you will play a pivotal role in creating innovative solutions for managing and analyzing large datasets. Your primary focus will be on utilizing Scala and Spark to design and implement scalable data pipelines that cater to the needs of our organization.Key responsibilities...


  • Kraków, Lesser Poland, Czech Republic beBeeDataEngineering Full time 5,400,000 - 7,800,000

    Unlock your potential as a Data and Financial Engineering Expert in our Product Control department. Collaborate with cross-functional teams to design, build, and maintain automated solutions that enhance controls and analytical processes.Key Responsibilities:Develop innovative solutions using Python, SQL, and VBA to improve data quality and reduce manual...

  • Data Engineer @

    3 days ago


    Kraków, Lesser Poland, Czech Republic HSBC Technology Poland Full time

    Experience with most of the following technologies (Apache Hadoop, Scala, Apache Spark, Sparkstreaming, YARN, Kafka, Hive, Python, ETL frameworks, Map Reduce, SQL, RESTful services). Sound knowledge on working Unix/Linux Platform. Hands-on experience building data pipelines using Hadoop components - Hive, Spark, Spark SQL. Experience with industry...

  • Data Engineer @

    5 days ago


    Kraków, Lesser Poland, Czech Republic Mindbox S.A. Full time

    Minimum 10 years of software development experience, including minimum 7 years of Python programming experience.Solid experience in Python, with knowledge of at least one Python web framework such as Django, Flask, etc.Have mindset of design thinking and ability to draw out the solution design.Experience of streaming data pipeline using PySpark, Apache Beam...

  • Data Engineer @

    2 weeks ago


    Kraków, Lesser Poland, Czech Republic Mindbox S.A. Full time

    Minimum 5 years of overall IT experience, including 2+ years in software development with Big Data technologies, microservices, and cloud-based event-driven architectures.Strong hands-on expertise with Apache NiFi, Apache Kafka, Apache Spark, Apache Hive, Apache HDFS, Apache Oozie, SQL, Python, Google SDK, REST API, Linux Shell Scripting.Solid database...

  • Data Engineer @

    5 days ago


    Kraków, Lesser Poland, Czech Republic Tesco Technology Full time

    This job requires to be based in/close to Kraków. We currently work in a hybrid model and meet in our office 3 days a week.QualificationsMandatory skills: Data Processing: Apache Spark - Scala or PythonData Storage: Apache HDFS or respective cloud alternativeResource Manager: Apache Yarn or respective cloud alternativeLakehouse: Apache Hive/Kyuubi or...