Senior Data Engineer
2 weeks ago
Experience: A minimum of 5-7+ years in data engineering, preferably supporting AI/ML applications and hold B.Sc., B.Eng., or higher, or equivalent in Computer Science, Data Engineering or related fields Experience with Snowflake Programming: Proficiency in Python, SQL and vector database native languages Databases: Experience with relational databases Cloud Platforms: Hands-on experience with AWS (OpenSearch, S3, Lambda) or Azure (Azure AI Search, Azure Blob Storage, Azure Automation) ETL/ELT Pipelines: Experience building scalable ETL/ELT workflows using dbt, Apache Airflow, or similar APIs & Microservices: Ability to design and integrate RESTful APIs for data exchange Data Security & Governance: Understanding of encryption and role-based access controls Version Control & DevOps: Familiarity with Git, CI/CD, containerization (Docker, Kubernetes), and Infrastructure as Code (Terraform, CloudFormation) Generative AI Support: Experience working with AI-specific data needs, such as embeddings, RAG (Retrieval Augmented Generation), and LLM fine-tuning data preparation Nice to have: NoSQL, vector databases IoT data streaming (Kafka, Kinesis, PySpark etc). ,[Design, build, and maintain scalable data pipelines (ETL/ELT) leveraging Snowflake and Airflow, Implement optimized schemas, partitioning, and indexing strategies in Snowflake and relational databases, Develop data processing workflows and automation scripts in Python and SQL; integrate with APIs and microservices, Ensure scalability, performance, and resilience of pipelines; implement observability for jobs and data flows, Partner with data scientists and ML engineers to deliver high-quality datasets optimized for AI/ML workloads, Prepare, transform, and manage datasets for embeddings, RAG workflows, and LLM fine-tuning] Requirements: Snowflake, Python, AWS, ETL/ELT, CI/CD, Docker, Kubernetes, SQL, NoSQL, Vector Databases, Kafka, Kinesis, PySpark Additionally: Sport subscription, Training budget, Private healthcare, Flat structure, International projects, Free coffee, Playroom, Bike parking, Free snacks, Free beverages, In-house trainings, Modern office, No dress code.
-
Senior Data Engineer @ CommerzBank
1 week ago
Łódź, Czech Republic CommerzBank Full timeWhich technology & skills are important for us? 👌 Very good knowledge of data pipeline orchestration (design scalable, cloud-native data pipelines for data transformation and aggregation based on business use cases). Very good knowledge of GCP (or other Cloud) and creating Cloud based architecture (BigQuery, Dataproc/PySpark, Cloud Composer/Apache...
-
Senior Data Engineer @ Link Group
6 days ago
Remote, Czech Republic Link Group Full timeRequired Skills & Experience 5–8 years of hands-on experience in data engineering or similar roles. Strong knowledge of AWS services such as S3, IAM, Redshift, SageMaker, Glue, Lambda, Step Functions, and CloudWatch. Practical experience with Databricks or similar platforms (e.g., Dataiku). Proficiency in Python or Java, SQL (preferably Redshift), Jenkins,...
-
Senior Data Engineer @ 1dea
6 days ago
Remote, Czech Republic 1dea Full timemin 5 yrs of relevant experience Solid experience with AWS services (S3, IAM, Redshift, Sagemaker, Glue, Lambda, Step Functions, CloudWatch) Experience with platforms like Databricks, Dataiku Proficient in Python / Java, SQL – Redshift preferred, Jenkins, CloudFormation, Terraform, Git, Docker, 2-3 years of Spark – PySpark Good communication and SDLC...
-
Senior Azure Data Engineer @ Antal
2 weeks ago
Warszawa, Bydgoszcz, Wrocław, Łódź, Kraków, Czech Republic Antal Full timeWymagania: Minimum 5 lat doświadczenia w pracy z danymi w roli Data Engineer lub pokrewnej. Bardzo dobra znajomość SQL (głównie Oracle) oraz doświadczenie w pracy z dużymi zbiorami danych. Doświadczenie w pracy z rozwiązaniami chmurowymi Azure (Azure Data Factory, Synapse, Databricks, Delta Lake). Znajomość języków skryptowych (np. Python,...
-
Senior Data Scientist
2 weeks ago
Remote, Warsaw, Wrocław, Białystok, Kraków, Gdańsk, Czech Republic Addepto Full timeWhat you’ll need to succeed in this role: 5+ years of commercial experience designing and implementing scalable AI solutions (Machine Learning, Predictive Modeling, Optimization, NLP, Computer Vision, GenAI, LLMs, Deep Learning). Proficiency in developing ML algorithms from scratch to production deployment. Strong programming skills in Python: writing...
-
Senior Data Engineer @ AVENGA
2 weeks ago
Wrocław, Czech Republic AVENGA (Agencja Pracy, nr KRAZ: 8448) Full timeRequirements: min. 5 years of experience as Data Engineer proven experience in Azure Databricks (data engineering, pipelines, performance tuning). Azure DevOps (Repos, Pipelines, YAML) Python Pyspark SQL Streaming Workflows Unity Catalog SQL Server experience as desirable Excellent communication and stakeholder management abilities. We are looking for a...
-
Senior Data Engineer
2 weeks ago
Remote, Warsaw, Wrocław, Białystok, Kraków, Gdańsk, Czech Republic Addepto Full timeWhat you’ll need to succeed in this role: At least 5 years of commercial experience implementing, developing, or maintaining Big Data systems. Strong programming skills in Python: writing a clean code, OOP design. Strong SQL skills, including performance tuning, query optimization, and experience with data warehousing solutions. Experience in...
-
Senior Data Engineer
2 weeks ago
Remote, Warszawa, Wrocław, Białystok, Kraków, Gdańsk, Czech Republic Addepto Full timeWhat you’ll need to succeed in this role: At least 5 years of commercial experience implementing, developing, or maintaining Big Data systems, data governance and data management processes. Strong programming skills in Python (or Java/Scala): writing a clean code, OOP design. Hands-on with Big Data technologies like Spark, Cloudera, Data Platform,...
-
Data Engineer @ Ework Group
2 weeks ago
Remote, Wrocław, Czech Republic Ework Group Full timeAzure Databricks (PySpark, Spark SQL; Unity Catalog; Jobs/Workflows). Azure data services: Azure Data Factory, Azure Key Vault, storage (ADLS), fundamentals of networking/identities. Python for data engineering (APIs, utilities, tests). Azure DevOps (Repos, Pipelines, YAML) and Git-based workflows. Experience operating production pipelines (monitoring,...
-
Remote, Czech Republic INNOBO Full timeTo thrive and succeed, you are expected to have: Bachelor’s degree in computer science, engineering, or a related field, complemented by experience in data engineering. A master’s degree is preferred Extensive experience with Git and managing version control in a collaborative environment Proven track record of implementing and managing CI/CD pipelines...