Lead Big Data Engineer @ KMD Poland

5 days ago


Remote Warsaw, Czech Republic KMD Poland Full time

Personal requirements:   Have 4+ years of Apache Spark experience and have faced various data engineering challenges in batch or streaming Have an interest in stream processing with Apache Spark Structured Streaming on top of Apache Kafka Have experience leading technical solution designs Have experience with distributed systems on a cloud platform Have experience with large-scale systems in a microservice architecture Are familiar with Git and CI/CD practices and can design or implement the deployment process for your data pipelines Possess a proactive approach and can-do attitude Are excellent in English and Polish, both written and spoken Have a higher education in computer science or a related field Are a team player with strong communication skills Nice to have requirements:  Apache Spark Structured Streaming Azure Domain Driven Development Docker containers and Kubernetes Message brokers (i.e. Kafka) and event-driven architecture Agile/Scrum #Data Engineer #Apache Spark #Databricks #Java #Apache Kafka #Batch Processing #Structured Streaming #Azure #SQL #Microservices #CI/CD #Docker #DDD   Are you ready to join our international team as a Lead Big Data Engineer? We shall tell you why you should... What product do we develop? We are building an innovative solution, KMD Elements, on Microsoft Azure cloud dedicated to the energy distribution market (electrical energy, gas, water, utility, and similar types of business). Our customers include institutions and companies operating in the energy market as transmission service operators, market regulators, distribution service operators, energy trading, and retail companies. KMD Elements delivers components allowing implementation of the full lifecycle of a customer on the energy market: meter data processing, connection to the network, physical network management, change of operator, full billing process support, payment, and debt management, customer communication, and finishing on customer account termination and network disconnection. The key market advantage of KMD Elements is its ability to support highly flexible, complex billing models as well as scalability to support large volumes of data. Our solution enables energy companies to promote efficient energy generation and usage patterns, supporting sustainable and green energy generation and consumption. We work with always up-to-date versions of:  Apache Spark on Azure Databricks Apache Kafka Delta Lake Java MS SQL Server and NoSQL storages like Elastic Search, Redis, Azure Data Explorer Docker containers Azure DevOps and fully automated CI/CD pipelines with Databricks Asset Bundles, ArgoCD, GitOps, Helm charts Automated tests How do we work? #Agile #Scrum #Teamwork #CleanCode #CodeReview #Feedback #BestPracticies    We follow Scrum principles in our work – we work in biweekly iterations and produce production-ready functionalities at the end of each iteration – every 3 iterations we plan the next product release We have end-to-end responsibility for the features we develop – from business requirements, through design and implementation up to running features on production More than 75% of our work is spent on new product features Our teams are cross-functional (7-8 persons) – they develop, test and maintain features they have built Teams’ own domains in the solution and the corresponding system components We value feedback and continuously seek improvements We value software best practices and craftsmanship Product principles: Domain model created using domain-driven design principles Distributed event-driven architecture / microservices Large-scale system for large volumes of data (>100TB data), processed by Apache Spark streaming and batch jobs powered by Databricks platform Our offer: Contract type: B2B Work Mode: Flexible — this role supports on-site, hybrid, and remote arrangements, depending on your individual preferences. Occasional on-site presence may be required — for example, onboard new team members, explore new business domains, or refine requirements in close collaboration with stakeholders or team building activities. What does the recruitment process look like?  Phone conversation with Recruitment Partner Technical interview with the Hiring Team Cognitive test Offer We are currently accepting CVs and expect to begin the recruitment process in January 2026, with initial contact and interviews taking place during the second week of the month. ,[   Develop and maintain the leading IT solution for the energy market using Apache Spark, Databricks, Delta Lake, and Apache Kafka,    Have end-to-end responsibility for the full lifecycle of features you develop,    Design technical solutions for business requirements from the product roadmap,    Maintain alignment with architectural principles defined on the project and organizational level,    Ensure optimal performance through continuous monitoring and code optimization.,    Refactor existing code and enhance system architecture to improve maintainability and scalability.,    Design and evolve the test automation strategy, including technology stack and solution architecture.,    Prepare reviews, participate in retrospectives, estimate user stories, and refine features ensuring their readiness for development. ] Requirements: Kafka, CD, Big Data, Microsoft Azure, Domain-driven design, Microservices, Databricks, Apache Spark, Apache Kafka, Data engineering, Git, Data pipelines, SQL, Azure Databricks, Java, MS SQL Server, NoSQL, Elasticsearch, Redis, Docker, DevOps, CI/CD Pipelines, ArgoCD, Helm, Kubernetes Tools: Agile, Scrum. Additionally: Remote work, Integration events, Sport subscription, Training budget, Flat structure, Private healthcare, International projects, Free coffee, Playroom, Free snacks, Free beverages, In-house trainings, No dress code.



  • Remote, Warsaw, Czech Republic KMD Poland Full time

    Ideal candidate:   Has 3+ years of commercial experience in implementing, developing, or maintaining data load systems (ETL/ELT).  Is proficient in Python, with a solid understanding of data processing challenges.  Has experience working with Apache Spark and Databricks.  Is familiar with MSSQL databases or other relational databases.  Has some...


  • Remote, Warsaw, Czech Republic KMD Poland Full time

    Ideal candidate:   Has 5+ years of commercial experience in implementing, developing, or maintaining data load systems (ETL/ELT).  Demonstrates strong programming skills in Python, with a deep understanding of data-related challenges.  Has hands-on experience with Apache Spark and Databricks.  Is familiar with MSSQL databases.  Has experience working...


  • Remote, Warsaw, Czech Republic KMD Poland Full time

    Ideal candidate:   Has 3+ years of experience in administering PostgreSQL and Microsoft SQL Server relational databases.  Is familiar with Microsoft Azure and its database services, such as: Azure SQL Database, Azure Database for PostgreSQL (Flexible Server), Azure Monitor, Log Analytics, Azure Backup.  Has experience with optimizing JSON and JSONB data...


  • Warsaw, Czech Republic KMD Poland Full time

    We are a GOOD match if you: Professionally: Have experience in Windows Server technologies. Have worked with Hyper-V virtualization. Have knowledge in managing physical server infrastructure. Are familiar with Kubernetes cloud deployments (AKS/EKS/GKE). Have DevOps and CI/CD Pipeline knowledge of principles and experience in implementation (preferably Azure...


  • Warsaw, Czech Republic KMD Poland Full time

    Key Responsibilities and examples of skills: Endpoint Management across macOS, iOS, Windows and Android using Microsoft Intune, JAMF Software packaging and deployment (macOS, Windows, custom scripts) Device compliance and security policies (Conditional Access, Defender for Endpoint, BitLocker, etc.) Automation and scripting (PowerShell or other relevant...


  • Warsaw, Czech Republic ITMAGINATION Full time

    7+ years of experience in Big Data systems and data management   3+ years of hands-on experience with Scala/Spark   Hands-on with Big Data technologies like Spark, Cloudera, Data Platform, Airflow, Hive   Excellent understanding of dimensional data and data modeling techniques   Experience implementing and deploying solutions in cloud environments...


  • Remote, Warsaw, Czech Republic KMD Poland Full time

    Personal Requirements: At least 5 years of experience as an SAP FI consultant, including full-cycle implementations and application maintenance or support projects Practical knowledge of SAP FI processes and configuration, including billing and settlements Excellent communication and documentation skills and the ability to work effectively in a team...


  • Warsaw, Czech Republic Innowise Full time

    What we expect: — Python knowledge: OOP basics, threads and GIL, working with the pandas library;— SQL knowledge: query language (DQL, DDL, DML, TCL), transactions and ACID principles, indexes;— Big Data knowledge: basic concepts — OLAP vs. OLTP, Data Warehouse and Data Lake, data normalization and denormalization;— Spoken English...

  • Lead Data Scientist

    1 week ago


    Remote, Warsaw, Wrocław, Białystok, Kraków, Gdańsk, Czech Republic Addepto Full time

    What you’ll need to succeed in this role: 5+ years of commercial experience in designing and implementing scalable AI solutions (Machine Learning, Predictive Modeling, Optimization, NLP, Computer Vision, GenAI, LLMs, Deep Learning). Proficiency in developing ML algorithms from scratch to production deployment. Strong programming skills in Python:...


  • Remote, Warsaw, Kraków, Czech Republic Elitmind Full time

    What makes you a great fit? Technical Expertise:  Deep experience with modern data platforms: Snowflake, Microsoft Fabric,  Databricks Strong proficiency in SQL and Python with a solid grasp of software engineering best practices Hands-on knowledge of cloud-native services, ETL/ELT frameworks, and orchestration tools (ADF, Airflow, dbt)...