Вакансии Data Engineer / Big Data Architect

Опыт от 5 лет Lead Full-time Не имеет значения Есть тестовое задание Remote Україна, Bulgaria, Portugal, Spain, Poland
15.07.2025
Подробнее
  • Python
  • SQL
  • AWS
  • AWS Glue
  • Amazon S3
  • Athena
  • AWS Lambda
  • DataZone
  • Apache Airflow
  • Apache Spark
  • Git
  • CI/CD
  • NoSQL

Join a transformative data and AI platform initiative aimed at modernizing enterprise-scale capabilities and enabling real-time decision-making. This project delivers a comprehensive roadmap covering AI, MLOps, data governance, and platform scalability, supporting a shift towards data-first operations and intelligent automation.

What you will do

  • Lead a data engineering team.
  • Collaborate with business stakeholders and technical teams to understand and analyze data requirements.
  • Lead the design and implementation of data models and database structures that meet business needs.
  • Profile, refactor, and tune performance in the database.
  • Design and implement complex ETL processes to extract, transform, and load data from various source systems into the data warehouse.
  • Ensure data integrity, consistency, and accuracy through robust data quality assurance measures.
  • Review and support team members, providing guidance and mentorship.
  • Supervise and contribute to the data-driven strategy for the project, aligning it with business objectives.

What you need for this

  • 5+ years of hands-on experience in data engineering with Python and SQL.
  • 3+ years of experience with AWS (Glue, S3, Athena, Lambda, DataZone)
  • Strong experience with Airflow and Spark
  • Drive the design and implementation of data warehouse and data lakes.
  • Proficient in code versioning (git) and building CI/CD for data projects.
  • Experience with requirement gathering and documentation.

Will be a plus:

  • Experience with NoSQL.

Информация о компании Intellias

Intellias – одна из крупнейших украинских ИТ компаний, объединяющая более 3000 специалистов. Компания занимается разработкой комплексного программного обеспечения и предоставлением профессиональных сервисов, специализируясь на автомобильной отрасли, навигационных системах, финансовых и телекоммуникационных технологиях.
Год основания: 2002
Количество сотрудников: 1001-5000
Сайт: intellias.ua

Преимущества сотрудникам

Откликнуться
Опыт от 5 лет Senior Full-time Не имеет значения Есть тестовое задание Office, Remote, Hybrid Україна, Poland
Бонус за рекомендацию: $1000
14.07.2025
Подробнее
* - разовый бонус кандидату посля найма
  • Python
  • AWS Glue
  • Azure Data Factory
  • Google Cloud Dataflow
  • ETL
  • ELT
  • PostgreSQl
  • MySQL
  • NoSQL
  • Docker
  • Kubernetes
  • TensorFlow
  • PyTorch
  • AWS SageMaker
  • Azure ML
  • Vertex AI
  • Lucene
  • Rest API
  • Veeva CRM
  • Reltio
  • SAP
  • Palantir Foundry
  • JavaScript
  • TypeScript

N-iX is seeking a proactive Senior Data Engineer to join our vibrant team. As a Senior Data Engineer, you will play a critical role in designing, developing, and maintaining sophisticated data pipelines, Ontology Objects, and Foundry Functions within Palantir Foundry. Your background in machine learning and data science will be valuable in optimizing data workflows, enabling efficient model deployment, and supporting AI-driven initiatives. The ideal candidate will possess a robust background in cloud technologies, data architecture, and a passion for solving complex data challenges.

Key Responsibilities:

  • Collaborate with cross-functional teams to understand data requirements, and design, implement, and maintain scalable data pipelines in Palantir Foundry, ensuring end-to-end data integrity and optimizing workflows.
  • Gather and translate data requirements into robust and efficient solutions, leveraging your expertise in cloud-based data engineering. Create data models, schemas, and flow diagrams to guide development.
  • Develop, implement, optimize and maintain efficient and reliable data pipelines and ETL/ELT processes to collect, process, and integrate data to ensure timely and accurate data delivery to various business applications, while implementing data governance and security best practices to safeguard sensitive information.
  • Monitor data pipeline performance, identify bottlenecks, and implement improvements to optimize data processing speed and reduce latency.
  • Collaborate with Data Scientists to facilitate model deployment and integration into production environments.
  • Support the implementation of basic ML Ops practices, such as model versioning and monitoring.
  • Assist in optimizing data pipelines to improve machine learning workflows.
  • Troubleshoot and resolve issues related to data pipelines, ensuring continuous data availability and reliability to support data-driven decision-making processes.
  • Stay current with emerging technologies and industry trends, incorporating innovative solutions into data engineering practices, and effectively document and communicate technical solutions and processes.

Tools and skills you will use in this role:

  • Palantir Foundry
  • Python
  • PySpark
  • SQL
  • TypeScript

Required:

  • 5+ years of experience in data engineering, preferably within the pharmaceutical or life sciences industry;
  • Strong proficiency in Python;
  • Hands-on experience with cloud services (e.g., AWS Glue, Azure Data Factory, Google Cloud Dataflow);
  • Expertise in data modeling, data warehousing, and ETL/ELT concepts;
  • Hands-on experience with database systems (e.g., PostgreSQL, MySQL, NoSQL, etc.);
  • Hands-on experience in containerization technologies (e.g., Docker, Kubernetes);
  • Familiarity with ML Ops concepts, including model deployment and monitoring.
  • Basic understanding of machine learning frameworks such as TensorFlow or PyTorch.
  • Exposure to cloud-based AI/ML services (e.g., AWS SageMaker, Azure ML, Google Vertex AI).
  • Experience working with feature engineering and data preparation for machine learning models.
  • Effective problem-solving and analytical skills, coupled with excellent communication and collaboration abilities;
  • Strong communication and teamwork abilities;
  • Understanding of data security and privacy best practices;
  • Strong mathematical, statistical, and algorithmic skills.

Nice to have:

  • Certification in Cloud platforms, or related areas;
  • Experience with search engine Apache Lucene, Web Service Rest API;
  • Familiarity with Veeva CRM, Reltio, SAP, and/or Palantir Foundry;
  • Knowledge of pharmaceutical industry regulations, such as data privacy laws, is advantageous;
  • Previous experience working with JavaScript and TypeScript.

Информация о компании N-iX

N-iX – это глобальная компания, предоставляющая программные решения и инженерные услуги. Имея более 2000 профессионалов в 25 странах Европы и Америки, компания предлагает экспертные решения в аналитике данных, встроенном программном обеспечении, IoT, искусственном интеллекте, машинном обучении и других технологических областях.
Год основания: 2002
Количество сотрудников: 1001-5000
Резидент Дія.City
Сайт: n-ix.com

Преимущества сотрудникам

Откликнуться
Опыт от 2 лет Senior, Lead Full-time Upper-Intermediate / B2 Есть тестовое задание Office, Remote, Hybrid Україна
Бонус за рекомендацию: $2000
14.07.2025
Подробнее
* - разовый бонус кандидату посля найма
  • SQL
  • PySpark
  • Python
  • AWS
  • GCP
  • Azure
  • DataBricks
  • Snowflake
  • Kubernetes
  • Delta Lake
  • Agile
  • Lucid
  • Jira
  • Confluence
  • Terraform
  • Ansible
  • SAFe
  • Apache Spark

N-iX is looking for a talented Big Data Architect to join our Technology Office team! Our client is one of Europe's fastest-growing mobile companies with almost 5 million active customers, 1,000 employees worldwide, and operations in five countries. The client is using the mobile virtual network operator business model.

Responsibilities:

  • Acting as a technical thought leader and as an architect for the Data Lake platform.
  • Building and extending the architecture Data platform as per the laid out guiding principles.
  • Transforming High level architecture into Technical Designs.
  • Transforming Functional & Non-functional requirements into designs.
  • Co-ordinating with product & business teams to close the loop for requirements & include into final solution proposal.
  • Assessing various available options, weighing their pros and cons, and documenting the results as an Architectural Decision Record (ADR).
  • Assisting and advising the engineering teams in selecting tools and evaluating multiple providers.
  • Outlining the Technical Architecture for each individual component in the system.
  • Facilitating and participating in design review sessions to ensure alignment with architectural standards and best practices.
  • Conducting risk assessments of new & existing technologies and implementations, ensuring compliance with security, data protection and regulatory requirements.

Requirements:

  • At least 2 years of experience in a similar Data Architect role with minimum 10+yrs of total experience.
  • Proven experience in SQL, PySpark and Python.
  • Experience in architecting, designing and developing Big Data platforms.
  • Proven experience in working in complex enterprise solution with terabytes of data processed daily.
  • Deep understanding of architectural patterns in Big Data like data mesh, data vault, medallion.
  • Experience with real-time stream processing.
  • Solid experience with Architectural patterns, design reviews and usage of best practices in arriving at the right solution.
  • Proven experience with technical impact assessment, assessing various options thoroughly and documenting as an Architectural Decision Record (ADR).
  • Working experience of building solutions in any big data cloud technology (AWS, GCP, Azure) and platforms (Databricks, Snowflake).
  • Experience in working with real-time and batch-processing architecture.
  • Experience with Kubernetes, Delta Lake formatting.
  • Experience to articulate the design preferences & choices to senior stakeholders using documented analytical approach.
  • Familiarity with Agile working principles.
  • Experience of working with tools like Lucid, JIRA, Confluence etc.
  • Excellent problem-solving skills.
  • Strong commitment to quality, documentation, and teamwork.
  • Upper intermediate level of English.
  • Ukrainian language Advanced or higher level.

Nice to have:

  • Experience with Infrastructure as Code (Terraform, Ansible).
  • Certification in architecture from Azure or AWS.
  • Experience of migration from cloud to hosted Spark platform.
  • Working experience with SAFE framework.

Информация о компании N-iX

N-iX – это глобальная компания, предоставляющая программные решения и инженерные услуги. Имея более 2000 профессионалов в 25 странах Европы и Америки, компания предлагает экспертные решения в аналитике данных, встроенном программном обеспечении, IoT, искусственном интеллекте, машинном обучении и других технологических областях.
Год основания: 2002
Количество сотрудников: 1001-5000
Резидент Дія.City
Сайт: n-ix.com

Преимущества сотрудникам

Откликнуться
Опыт от 4 лет Senior Full-time Не имеет значения Есть тестовое задание Hybrid Tel Aviv
07.07.2025
Подробнее
  • Java
  • Go
  • Spring
  • Kubernetes
  • Terraform
  • Helm
  • Argo CD

We're looking for an experienced Backend Developer to join our growing data platform team.
As a Backend Developer, you'll work on a massive data processing pipeline, ingesting over a billion daily events from multiple sources. You'll also create the next-generation pipeline and help us scale from a billion events a day to tens of billions of events.

Responsibilities

  • Own projects from initial discussions to release, including data exploration, architecture design, benchmarking new technologies, and product feedback.
  • Work with massive amounts of data from different sources using state-of-the-art technology to make big data accessible in real-time.
  • Develop and deploy real-time and batch data processing infrastructures.
  • Manage the development of distributed data pipelines and complex software designs to support high data rates (millions of daily active users) using cloud-based tools.
  • Work closely with company stakeholders on data-related issues.
  • Develop unit, integration, end-to-end (e2e), and load tests.

Requirements

  • 4+ years of experience as a Software Engineer, including design & development.
  • Proven experience with Java or Go.
  • Experience in the design and development of scalable big data solutions.
  • Experience working in a cloud-based environment.
  • Passionate about technologies, frameworks, and best practices.
  • Ability to work in a fast-paced environment.

Advantages

  • Experience with Spring / Kubernetes.
  • Experience with Terraform / Helm / Argo.

Информация о компании Moon Active

Moon Active – одна из самых быстрорастущих IT-компаний, которая занимается разработкой высококачественных мобильных игр на платформах Facebook, iOS и Android. Компания создает увлекательные и интересные развлечения, которыми пользуются миллионы игроков во всем мире.
Год основания: 2011
Количество сотрудников: 101-250
Сайт: moonactive.com

Преимущества сотрудникам

Откликнуться
Опыт от 3 лет Middle, Senior Full-time Upper-Intermediate / B2 Есть тестовое задание Remote Україна, Poland
04.07.2025
Подробнее
  • SQL
  • ETL
  • DataBricks
  • Python
  • AWS
  • Amazon S3
  • Amazon Redshift
  • Athena
  • AWS Glue
  • AWS Lambda

ELEKS Software Engineering and Development Office is looking for a Senior/Middle Data Engineer in Ukraine, Poland or Croatia.

About client

The customer is a British company producing electricity with zero carbon emissions.

Requirements

  • Experience in Data Engineering, SQL, ETL(data validation + data mapping + exception handling) 3+ years
  • Hands-on experience with Databricks 2+ years
  • Experience with Python
  • Experience with AWS (e.g. S3, Redshift, Athena, Glue, Lambda, etc.)
  • Knowledge of the Energy industry (e.g. energy trading, utilities, power systems etc.) would be a plus
  • Experience with Geospatial data would be a plus
  • At least an Upper-Intermediate level of English

Responsibilities

  • Building Databases and Pipelines: Developing databases, data lakes, and data ingestion pipelines to deliver datasets for various projects
  • End-to-End Solutions: Designing, developing, and deploying comprehensive solutions for data and data science models, ensuring usability for both data scientists and non-technical users. This includes following best engineering and data science practices
  • Scalable Solutions: Developing and maintaining scalable data and machine learning solutions throughout the data lifecycle, supporting the code and infrastructure for databases, data pipelines, metadata, and code management
  • Stakeholder Engagement: Collaborating with stakeholders across various departments, including data platforms, architecture, development, and operational teams, as well as addressing data security, privacy, and third-party coordination

Информация о компании Eleks

ELEKS является избранным партнером для ведущих мировых предприятий, малого и среднего бизнеса и технологических конкурентов. Компания помогает предприятиям повысить свою ценность посредством специальной разработки программного обеспечения, дизайна продуктов, контроля качества и консультационных услуг.
Год основания: 1991
Количество сотрудников: 1001-5000
Сайт: eleks.com

Преимущества сотрудникам

Откликнуться
Опыт не имеет значения Senior Full-time Не имеет значения Есть тестовое задание Office Львів
25.06.2025
Подробнее
  • Apache Airflow
  • Python
  • SQL
  • dbt
  • Snowflake
  • AWS
  • Docker

We are seeking a Senior Data Engineer to join our growing team and help design, build, and maintain robust data infrastructure for our client — a leader in digital banking solutions. This role is ideal for someone who is passionate about building scalable and efficient data pipelines and enjoys working with modern data tools in a cloud environment.

This is the job

As a Senior Data Engineer, you will play a key role in the development of a modern data platform. You will collaborate closely with data scientists, analysts, and other engineers to ensure high data availability, quality, and performance.

This is you

  • You have proven experience with Airflow, Python programming, and SQL / dbt.
  • You’ve worked with Snowflake or similar cloud-based data warehouses.
  • You have a solid understanding of AWS services and Docker.
  • You hold a Bachelor’s degree in Mathematics, Computer Science, or another relevant quantitative field.
  • You approach problems with strong analytical and problem-solving skills.
  • You’re familiar with Data Engineering best practices, including Data Quality and Monitoring/Observability.
  • You’re comfortable working in a dynamic, fast-paced environment, and take ownership of your work.
  • You have a growth mindset and are eager to learn and grow through hands-on experience.

This is your role

  • Design and develop scalable data pipelines to efficiently process and analyze large volumes of data, utilizing Snowflake, Looker, Airflow, and dbt.
  • Collaborate with stakeholders to translate their requirements into technical steps and coordinate the projects you drive with them.
  • Monitor and improve the health of our data pipelines.
  • Promote knowledge sharing within the team to foster collaboration and continuous learning, and mentor junior colleagues.
  • Stay updated on emerging technologies and best practices in data engineering, and bring new ideas to enhance the technical setup.

Информация о компании Avenga

Avenga – это организация, созданная объединением четырех ИТ-компаний из Малайзии, Германии, Польши и Украины с целью предоставления лучшей помощи партнерам в создании индивидуального программного обеспечения. Предлагаемые компанией аутсорсинговые услуги уже в течение десятилетий приносят ценность фармацевтической и биологической отраслям, страховой, финансовой и автомобильной промышленности. Компания работает над разработкой и дизайном программного обеспечения на заказ, управлением продуктами, проектами технологического консалтинга и цифровой трансформацией под ключ.
Год основания: 2019
Количество сотрудников: 1001-5000
Сайт: avenga.com

Преимущества сотрудникам

Откликнуться
Product / Startup, Outsource
Опыт от 2 лет Junior, Middle Full-time Upper-Intermediate / B2 Есть тестовое задание Remote Україна
24.06.2025
Подробнее
  • Python
  • SQL
  • Apache Spark
  • Pandas
  • ETL
  • Apache Airflow
  • Azure Cloud
  • DataBricks
  • dbt
  • CI/CD
  • Terraform
  • GCP
  • AWS

What you will do:

  • Develop internal tools and data services using Python;
  • Design, build, and maintain ETL pipelines and data workflows using Python, Airflow, Azure, Databricks, DBT, Pandas, Spark, and DBT;
  • Contribute to infrastructure development using Terraform on Azure cloud;
  • Support CI/CD processes;
  • Participate in the ongoing improvement of data platforms and pipelines;
  • Collaborate with team members across teams to drive best practices and solutions in data engineering.

Must haves:

  • 2+ years of professional experience with Python;
  • 2+ years of professional experience in a Data Engineering role;
  • Proficiency in programming languages commonly used in data engineering such as Python, SQL for working with data processing frameworks like Spark and libs like Pandas;
  • Ability to build and support services on Python and other programming languages;
  • Ability in designing, deploying, and managing ETL data pipelines using Apache Airflow for workflow orchestration and scheduling;
  • Experience with Azure Cloud Platform, Databricks, DBT;
  • Experience with CI/CD pipelines in a modern DevOps environment;
  • Upper-intermidiate English level.

Nice to haves:

  • Familiarity with Terraform;
  • Familiarity with GCP, AWS;
  • Experience developing internal tools for data teams.

Информация о компании AgileEngine

AgileEngine – международная компания, разработчик программного обеспечения на заказ. Работает с венчурными стартапами и известными технологическими брендами, предлагает полный спектр услуг от выделенных удаленных команд до консультаций на месте.
Год основания: 2010
Количество сотрудников: 1001-5000
Сайт: agileengine.com

Преимущества сотрудникам

Откликнуться
Outsource, Outstaff
Опыт от 5 лет Senior Full-time Upper-Intermediate / B2 Есть тестовое задание Office, Remote Україна, Львів, Poland
23.06.2025
Подробнее
  • Python
  • PySpark
  • DataBricks
  • Microsoft SQL Server
  • AWS
  • Azure
  • GCP
  • Docker
  • Kubernetes

We’re opening the position of a Senior Data Engineer to work with our Client – a leading energy sector analytics company, as they expand their innovative data products that revolutionize document analysis. As a Data Engineer, you will focus on executing defined tasks to support our data product coverage. You’ll work closely with a talented team of Engineers to ensure high-quality data solutions. Here, you’ll find a culture that values collaboration and encourages learning and growth.

Responsibilities:

  • Develop and maintain scalable data pipelines using Python, SQL, AWS services (Amazon Bedrock, S3), and Databricks
  • Build and optimize ETL jobs in Databricks using PySpark, ensuring efficient processing of large-scale distributed datasets
  • Play a pivotal role in enhancing the breadth and depth of our courthouse data products
  • Utilize your Python expertise to parse complex datasets, manipulate intricate image data, and craft innovative data products that meet our customers’ evolving needs
  • Champion data quality, consistency, and reliability throughout our product lifecycle
  • Contribute to the development of new features and the continuous improvement of existing data systems
  • Design and implement distributed Data Engineering solutions in Databricks, leveraging PySpark for optimized workflows

Requirements:

  • At least 5 years of experience as a Data Engineer, with a proven track record of successful projects
  • Advanced proficiency in Python and PySpark – the languages of data manipulation and analysis
  • Expertise in Databricks as a distributed Data Engineering platform
  • Solid experience with relational database systems, particularly MS SQL Server
  • Ability to collaborate effectively within a team environment and meet project deadlines
  • Strong communication skills and fluency in English
  • English level – Upper-Intermediate or higher

Will be a plus:

  • Knowledge of cloud platforms such as AWS, Azure, or Google Cloud
  • Experience with containerization technologies like Docker and Kubernetes

Информация о компании Edvantis

Edvantis – аутсорсинговая компания по разработке ПО с более чем 400 реализуемыми проектами. Компания помогает достичь большего: выходит за пределы решения первоначальной проблемы с помощью разработки программного обеспечения как услуги и помогает клиентам получить максимальную отдачу от инновационных технологий и процессов разработки.
Год основания: 2005
Количество сотрудников: 251-500
Резидент Дія.City
Сайт: edvantis.com

Преимущества сотрудникам

Откликнуться
Outsource, Consulting / Integrator
Опыт от 3 лет Middle Full-time Upper-Intermediate / B2 Есть тестовое задание Office, Remote Україна, Київ, Львів
19.06.2025
Подробнее
  • Python
  • Apache Airflow
  • Kubeflow
  • Kafka
  • HTML
  • JSON
  • API
  • AWS
  • GCP
  • Azure

About the project

Generative AI technologies are rapidly changing how digital content is created and consumed. However, many of these systems are trained on vast amounts of data, including articles, videos, and other creative works – often without the knowledge or consent of the original creators. As a result, publishers, journalists, and content producers face the risk of losing both visibility and critical revenue streams such as advertising, subscriptions, and licensing. Our project addresses this issue by developing a system that allows AI platforms to identify when specific content has influenced a generated result. This enables transparent attribution and the possibility for content creators to receive compensation based on how often their work is used. The goal is to build a sustainable ecosystem where creators are fairly rewarded, while AI-generated content remains trustworthy and ethically grounded.

Requirements

  • 3+ years of experience in Data Engineering;
  • Solid Python programming skills, especially in data processing and system automation;
  • Proven experience with Airflow, Kubeflow, or Kafka for orchestrating data workflows;
  • Familiarity with search engine concepts and indexing;
  • Experience working with structured and semi-structured web data (HTML, JSON, APIs);
  • Ability to work with large-scale distributed systems and cloud platforms (e.g., AWS, GCP, Azure);
  • English: Upper-Intermediate+.

Информация о компании Geniusee

Geniusee – компания по разработке программного обеспечения и продуктов, ориентированная на удовлетворение потребностей своих клиентов. Команда уже реализовала более 150 выдающихся проектов в областях FinTech, EdTech и MarTech. Компания разрабатывает продукты и услуги, предоставляет консультации и поддержку разным клиентам и партнерам.
Год основания: 2017
Количество сотрудников: 101-250
Резидент Дія.City
Сайт: geniusee.com

Преимущества сотрудникам

Откликнуться
Опыт от 5 лет Lead Full-time Не имеет значения Есть тестовое задание Remote Україна
11.06.2025
Подробнее
  • Apache Airflow
  • dbt
  • SQL
  • ETL
  • AWS

Our Client is the Enterprise Worldwide Company. The product you will be working with, provides management and data processing/handling capabilities for networks of the clients scientific lab equipment such as microscopes, etc. The main goals are:

  • Collection and centralized management of data outputs (measurement results, etc.) provided by clients devices
  • Outdated data utilization
  • Managing large volumes of data acquired from measurement devices in the cloud securely and reliably
  • Seamless sharing of measurement data with collaborators
  • The ability to share measurement results and accelerate customer service.

Requirements

We are looking for a Lead Data Engineer with at least 6 years of commercial experience in development of data platforms for enterprise applications. With the experience to Lead a team of engineers and take responsiility for the technical solution.

  • Proficiency in Airflow for workflow orchestration, dbt for data transformation, and SQL for data querying and manipulation.
  • Experience in data modeling, ETL (Extract, Transform, Load) processes, and data warehousing concepts.
  • Familiarity with cloud platforms (AWS) and their data services.
  • Excellent analytical and problem-solving skills with meticulous attention to detail.
  • Strong communication and collaboration skills with the ability to lead and motivate cross-functional teams.Good to have ability to participate onsite meeting.

Job responsibilities

  • Implement new solutions into the current system with the refactoring and from scratch methods;
  • Preparing the technical documentation;
  • Participating in client meetings to understand business and user requirements and estimate tasks;
  • Collaborating closely with other engineers, product owners and testers to identify and solve challenging problems;
  • Taking part in defect investigation, bug fixing, troubleshooting;

Информация о компании GlobalLogic

GlobalLogic, компания Hitachi Group, является лидером в области разработки продуктов полного жизненного цикла, сочетающего опыт разработки программного обеспечения Chip-to-Cloud и опыт вертикальной отрасли, чтобы помочь нашим клиентам проектировать, создавать и поставлять продукты следующего поколения и цифровой опыт. Компания умело интегрирует дизайн, комплексное проектирование и возможности гибкой доставки, чтобы добиться отличных бизнес-результатов для мировых брендов.
Год основания: 2000
Количество сотрудников: 1001-5000
Резидент Дія.City
Сайт: globallogic.com

Преимущества сотрудникам

Откликнуться
Outsource, Outstaff
Опыт не имеет значения Senior Full-time Upper-Intermediate / B2 Есть тестовое задание Remote
30.05.2025
Подробнее
  • AWS
  • SQL
  • Python
  • Tableau
  • Apache Airflow
  • dbt
  • Java
  • Docker
  • Kubernetes
  • BigQuery
  • Amazon Redshift

We are looking for a Senior Data Engineer for our US-based client – a global leader in connectivity intelligence, delivering data-driven insights and solutions that optimize networks, enhance digital experiences, and drive innovation for service providers, enterprises, governments, and individuals worldwide.
Project – advanced SaaS analytics platform for mobile operators, which collects, analyzes, and correlates massive volumes of network performance and user experience data. This gives the ability to help make the internet better, faster and more accessible for everyone.

Requirements:

  • Proven experience building both streaming and batch data pipelines
  • Hands-on experience with at least one cloud platform (preferably AWS)
  • Proficient in SQL and comfortable working with Python or Java
  • Experience designing and optimizing data models tailored for analytics and used in BI tools like Tableau
  • Familiar with data orchestration tools such as Apache Airflow
  • Experience using DBT or Dataform for transformation workflows
  • Comfortable working with Docker and Kubernetes
  • Solid understanding of data warehouse solutions like BigQuery or Amazon Redshift
  • Ability to thrive in a fast-paced, start-up environment
  • Background in or familiarity with the telecom industry is a plus
  • Machine learning experience is a bonus
  • Strong communication skills in English (Upper-Intermediate level or higher)

Responsibilities:

  • Participate in the design and help drive the implementation of our data platform
  • Design, implement, and operate streaming and batch pipelines that scale
  • Partner with both engineers and data analysts to build reliable datasets that can be trusted, understood, and used by the rest of the company
  • Work with data pipelines taking our raw data, delivered through APIs, to enhanced datasets that can be used by data analysts and data scientists

Информация о компании NCube

NCube – компания, предоставляющая удаленные команды высококвалифицированных разработчиков программного обеспечения, которые могут легко вписаться в рабочую/техническую и командную среду клиента либо на основе проекта, либо на основе расширения команды. Навыки решения технических проблем, глубокое понимание принципов разработки программного обеспечения, качество кода и окончательных решений, поток новых идей, энтузиазм в повседневной работе и общие затраты – основные сильные стороны команд, предоставляемых компанией.
Год основания: 2008
Количество сотрудников: 101-250
Сайт: ncube.com

Преимущества сотрудникам

Откликнуться
Product / Startup, Investor / Accelerator
Опыт не имеет значения Middle, Senior Full-time Не имеет значения Есть тестовое задание Remote
26.05.2025
Подробнее
  • SQL
  • Python
  • ETL
  • ELT
  • BigQuery
  • GCP
  • Apache Airflow
  • Prefect
  • Dagster
  • Looker Studio
  • Microsoft Power BI
  • Tableau
  • Grafana
  • Git
  • CI/CD
  • GitHub Actions
  • Docker

We’re looking for a Data Engineer to join our team in the iGaming industry, where real-time insights, affiliate performance, and marketing analytics are at the center of decision-making. In this role, you’ll own and scale our data infrastructure, working across affiliate integrations, product analytics, and experimentation workflows.
Your primary responsibilities will include building and maintaining data pipelines, implementing automated data validation, integrating external data sources via APIs, and creating dashboards to monitor data quality, consistency, and reliability. You’ll collaborate daily with the Affiliate Management team, Product Analysts, and Data Scientists to ensure the data powering our reports and models is clean, consistent, and trustworthy.

What you’ll do:

  • Design, develop, and maintain ETL/ELT pipelines to transform raw, multi-source data into clean, analytics-ready tables in Google BigQuery, using tools such as dbt for modular SQL transformations, testing, and documentation.
  • Emphasize data quality, consistency, and reliability by implementing robust validation checks, including schema drift detection, null/missing value tracking, and duplicate detection using tools like Great Expectations or Soda.
  • Integrate and automate affiliate data workflows, replacing manual processes in collaboration with the related stakeholders.
  • Proactively monitor and manage data pipelines using tools such as Airflow, Prefect, or Dagster, with proper alerting and retry mechanisms in place.
  • Build a Data Consistency Dashboard (in Looker Studio, Power BI, Tableau or Grafana) to track schema mismatches, partner anomalies, and source freshness, with built-in alerts and escalation logic.
  • Ensure timely availability and freshness of all critical datasets, resolving latency and reliability issues quickly and sustainably.
  • Control access to cloud resources, implement data governance policies, and ensure secure, structured access across internal teams.
  • Monitor and optimize data infrastructure costs, particularly related to BigQuery usage, storage, and API-based ingestion.
  • Document all pipelines, dataset structures, transformation logic, and data contracts clearly to support internal alignment and knowledge sharing.
  • Build and maintain postback-based ingestion pipelines to support event-level tracking and attribution across the affiliate ecosystem.
  • Collaborate closely with Data Scientists and Product Analysts to deliver high-quality, structured datasets for modeling, experimentation, and KPI reporting.
  • Act as a go-to resource across the organization for troubleshooting data discrepancies, supporting analytics workflows, and enabling self-service data access.

What we expect from you:

  • Strong proficiency in SQL and Python.
  • Proven ability to design, deploy, and scale ETL/ELT pipelines.
  • Experience with Google BigQuery and other GCP tools (e.g., Cloud Storage, Cloud Functions, Composer).
  • Hands-on experience integrating and automating data from various platforms.
  • Familiarity with postback tracking, attribution logic, and affiliate data reconciliation.
  • Skilled in orchestration tools like Airflow, Prefect, or Dagster.
  • Experience with Looker Studio, Power BI, Tableau, or Grafana for building dashboards for data quality monitoring.
  • Use of Git for version control and experience managing CI/CD pipelines (e.g., GitHub Actions).
  • Experience with Docker to build isolated and reproducible environments for data workflows.
  • Strong sense of data ownership, documentation, and operational excellence.
  • Exposure to iGaming data structures and KPIs is a strong advantage.

Информация о компании Boosta

Boosta – международная IT-компания, которая создает и развивает разные бизнесы в цифровых областях. Компания имеет проекты по performance-маркетингу, собственный инвестиционный фонд, а также более 10 успешных IT-продуктов, которыми пользуются десятки тысяч человек в Европе, Азии, Австралии, Северной и Южной Америке. Boosta открыла R&D отдел для реализации внешних идей, запуска новых продуктов, активно поддерживает стартапы и помогает развиваться перспективным проектам.
Год основания: 2014
Количество сотрудников: 501-1000
Сайт: boosta.biz

Преимущества сотрудникам

Откликнуться
Outsource, Outstaff
Опыт от 5 лет Lead Full-time Upper-Intermediate / B2 Есть тестовое задание Office, Remote Україна, Львів, Poland
16.05.2025
Подробнее
  • Snowflake
  • SQL
  • Dataiku
  • Azure Data Factory

We’re opening the position of a Lead Data Engineer to join our team aimed at developing a data warehouse solution for a big US food company. You will play a key role in shaping our data architecture, integrating data from multiple sources, and powering business-critical dashboards used across the company.
As a Lead Data Engineer, you will also be expected to guide the Development Team, making technical decisions and ensuring best practices are followed across data architecture, pipeline design, and performance optimization.

Responsibilities:

  • Build a data mart design and documentation of sophisticated data architecture solutions, including data models, data integration patterns, and data storage mechanisms
  • Collaborate with the Product Owner and the team to implement data solutions aligned with the organization’s strategic goals
  • Work alongside Data Engineers to ensure seamless integration of data pipelines and data products with the overall data architecture
  • Optimize data storage and retrieval processes to fully leverage Snowflake capabilities, focusing on performance, scalability, and cost efficiency
  • Provide technical leadership, mentorship, and support to the Data Engineering Team
  • Perform code reviews and ensure adherence to coding standards and best practices

Requirements:

  • 5+ years in data engineering, data modelling, database design, data warehousing, and data governance practices
  • Production experience with building, maintaining, and performance optimization of Snowflake DWH
  • Proven experience in leading data projects, architecting and designing new solutions
  • Experience in building and maintaining data pipelines
  • Proficiency in SQL
  • Ability to conduct code reviews and ensure adherence to coding standards
  • Excellent problem-solving and analytical skills
  • Strong communication and collaboration abilities
  • English level – Upper-Intermediate or higher

Will be a plus:

  • Experience with Dataiku for pipeline orchestration
  • SnowPro® Certifications
  • Experience with Azure data factory

Информация о компании Edvantis

Edvantis – аутсорсинговая компания по разработке ПО с более чем 400 реализуемыми проектами. Компания помогает достичь большего: выходит за пределы решения первоначальной проблемы с помощью разработки программного обеспечения как услуги и помогает клиентам получить максимальную отдачу от инновационных технологий и процессов разработки.
Год основания: 2005
Количество сотрудников: 251-500
Резидент Дія.City
Сайт: edvantis.com

Преимущества сотрудникам

Откликнуться
Опыт не имеет значения Middle, Senior Full-time Не имеет значения Есть тестовое задание Remote Київ
12.05.2025
Подробнее
  • GitOps
  • IaC
  • Kubernetes
  • Helm
  • ArgoCD
  • Prometheus
  • Grafana
  • Grafana Loki
  • HashiCorp Vault
  • Apache Spark
  • Kafka
  • MinIO
  • Amazon S3
  • Apache Airflow
  • Docker
  • GitLab
  • GitLab CI
  • Python
  • Golang
  • Java
  • Scala
  • Kerberos
  • Active Directory
  • ClickHouse
  • Elasticsearch
  • oAuth
  • OpenID
  • Keycloak
  • Azure
  • Azure Fabric

Our project is to develop the data platform, where all the analytical data of the company will be stacked. This is a great opportunity to participate in the launch and operation of a large Kubernetes/Spark/S3 cluster and one of the most interesting BI practices in Eastern Europe.

Responsibilities:

  • Support and active development of the Data Platform and Hybrid Data Platform (on-prem + Azure Fabric, in progress)
  • Support for a team of data engineers and analysts

Skills:

  • Understanding the advantage of GitOps/IaC over manual work
  • Kubernetes, Helm, ArgoCD, Prometheus, Grafana, Loki, HashiCorp Vault
  • Apache Spark in Kubernetes, Apache Kafka, Minio/S3, Apache Airflow
  • Docker (BuildKit), Gitlab, Gitlab CI
  • Experience with at least one of the most popular programming languages, such as Python, Golang, Java, Scala, etc.Be able writing a code

Will be a plus:

  • Kerberos, Active directory
  • Clickhouse
  • Datahub
  • Elasticsearch
  • Experience with any OLAP DB, support, optimize
  • Security in K8s, HashiCorp Vault, Oauth, OpenID, Keycloak

Will be a significant plus:

  • Azure, Azure Fabric.

Technologies that we use:

  • Kubernetes RKE2 1.31. Cilium 1.17. Gitops, Argocd, Helm, Kustomize, Kyverno.
  • Gitlab, Gitlab CI, Gitlab Kubernetes Runner, Docker, Buildkit.
  • Apache Airflow.
  • Apache Spark, Apache Kyuubi, Hive Metastore
  • Minio, Redis, PostgreSQL (Cloudnative-PG), Elasticsearch, Apache Kafka, Clickhouse.
  • Datahub.
  • Prometheus Stack, Grafana, Grafana Loki.
  • Python, Golang metrics exporters, Datadog Vector, Fluent-bit.
  • PowerBI, Azure Fabric.
  • Ansible.

Информация о компании TemaBit

TemaBit – украинская IT-компания, является частью Fozzy Group, которая разрабатывает уникальные решения для более 40 компаний. TemaBit специализируется на веб- и мобильных разработках, программном обеспечении для управления бизнес-процессами и инфраструктурными проектами, поддержке разнообразных программных продуктов в разных сферах, а также работает над глобальной миграцией от крупных локальных сервисов до собственной AWS.
Год основания: 2017
Количество сотрудников: 1001-5000
Резидент Дія.City
Сайт: temabit.com

Преимущества сотрудникам

Откликнуться
Опыт от 5 лет Senior, Lead Full-time Upper-Intermediate / B2 Есть тестовое задание Remote Україна, Київ
09.05.2025
Подробнее
  • T-SQL
  • Python
  • Azure SQL
  • Azure Data Factory
  • Snowflake
  • dbt
  • C#
  • .NET

SimCorp is an independent subsidiary of the Deutsche Börse Group. Following the recent merger with Axioma, we leverage the combined strength of our brands to provide an industry-leading, full, front-to-back offering for our clients.  
SimCorp is an equal-opportunity employer. We are committed to building a culture where diverse perspectives and expertise are integrated into our everyday work. We believe in the continual growth and development of our employees, so that we can provide best-in-class solutions to our clients.
You will be joining an innovative application development team within SimCorp’s Product Division. As a primary provider of SaaS offerings based on next-generation technologies, our Digital Engagement Platform is a cloud-native data application developed on Azure, utilizing SRE methodologies and continuous delivery. Your contribution to evolving DEP’s data platform will be vital in ensuring we can scale to future customer needs and support future analytics requirements. Our future growth as a SaaS product is rooted in a cloud-native strategy that emphasizes adopting a modern data platform tool stack and the application of modern engineering principles as essential components.
We are looking into a technology shift from Azure SQL to SnowFlake in order to meet new client demands for scalability. You will be an important addition to the team for achieving this goal.

Requirements

Most importantly, you can see yourself contributing and thriving in the position described above. How you gained the skills needed for doing that is less important.
We expect you to be good at and have had hands-on experience with the following:

  • Expert in T-SQL
  • Proficiency in Python
  • Experience in Microsoft cloud technologies data services including but not limited to Azure SQL and Azure Data Factory
  • Experience with Snowflake and star schema and data modeling – experience with migrations to Snowflake will be an advantage
  • Experience or strong interest with DBT (data build tool) for transformations, test. Validation, data quality etc.
  • English – Upper Intermediate

On top of that, it would an advantage to have knowledge / interest in the following: 

  • Some proficiency in C# .NET
  • Security first mindset, with knowledge on how to implement row level security etc.
  • Agile development methodologies and DevOps / DataOps practices such as continuous integration, continuous delivery, and continuous deployment. For example, automated DB validations and deployment of DB schema using DACPAC.

As a person, you have following traits:

  • Strong collaborator with team mates and stakeholders
  • Clear communicator who speaks up when needed.

Job responsibilities

  • Ensure quality in our data solutions and that we can ensure good data quality across multiple customer tenants every time we release.
  • Work together with the Product Architect on defining and refining the data architecture and roadmap.
  • Facilitate the migration of our current data platform towards a more modern tool stack that can be easier maintained by both data engineers and software engineers.
  • Ensure that new data entities get implemented in the data model using schemas that are appropriate for their use, facilitating good performance and analytics needs.
  • Guide and support people of other roles (engineers, testers, etc.), to ensure the spread of data knowledge and experience more broadly in the team

Информация о компании GlobalLogic

GlobalLogic, компания Hitachi Group, является лидером в области разработки продуктов полного жизненного цикла, сочетающего опыт разработки программного обеспечения Chip-to-Cloud и опыт вертикальной отрасли, чтобы помочь нашим клиентам проектировать, создавать и поставлять продукты следующего поколения и цифровой опыт. Компания умело интегрирует дизайн, комплексное проектирование и возможности гибкой доставки, чтобы добиться отличных бизнес-результатов для мировых брендов.
Год основания: 2000
Количество сотрудников: 1001-5000
Резидент Дія.City
Сайт: globallogic.com

Преимущества сотрудникам

Откликнуться
Опыт от 7 лет Senior Full-time Не имеет значения Есть тестовое задание Remote Україна
07.05.2025
Подробнее
  • Java
  • Scala
  • Python
  • Go
  • C++
  • Rust
  • Kafka
  • Apache Flink
  • Apache Spark
  • Apache Beam
  • NoSQL
  • Cassandra
  • MongoDB
  • OLAP
  • ClickHouse
  • StarRocks
  • Doris
  • SQL
  • Kubernetes
  • Helm
  • ArgoCD
  • Iceberg
  • Delta Lake
  • Apache Hudi
  • GCP
  • AWS
  • Azure

We are seeking an experienced developer to create a high-performance, scalable, and flexible behavioral analytics engine platform.
You will be a key member of our team, responsible for the architecture, development, and optimization of core components for processing and analyzing large volumes of data [terrabytes].

Required professional experience:

  • 7+ years of experience in developing analytics platforms or big data processing systems.
  • Deep knowledge of programming languages such as Java, Scala, Python, Go, C++, or Rust.
  • Experience with distributed systems and big data technologies [Kafka, Flink, Spark, Apache BEAM].
  • Understanding of scalable system design principles and architectures for real-time data processing.
  • Experience with NoSQL databases [Cassandra, MongoDB].
  • Experience with OLAP databases [ClickHouse, StarRocks, Doris].
  • Knowledge of SQL.
  • Understanding of statistical methods and principles of data analysis.
  • Experience with Kubernetes [Helm, ArgoCD].

Desired Skills:

  • Experience with open table format [Apache Iceberg/Delta Lake/Hudi].
  • Experience with cloud platforms [Google Cloud, AWS, Azure].
  • Knowledge of data security methods and compliance with regulatory requirements [GDPR, CCPA].

Key Responsibilities:

  • Design and develop the architecture of an behavioral analytics platform for real-time big data processing.
  • Implement key engine systems [data collection, event processing, aggregation, prepare data for visualization].
  • Optimize the platform performance and scalability for handling large data volumes.
  • Develop tools for user behavior analysis and product metrics.
  • Collaborate with data analysts and product managers to integrate the engine into analytics projects.
  • Research and implement new technologies and methods in data analysis.

Информация о компании Burny Games

Burny Games – украинская продуктовая компания, которая занимается созданием мобильных игр в жанре Casual Puzzles. Компания разрабатывает первоклассные игры, чтобы обеспечить массовую аудиторию инновационным и качественным геймплеем, используя отлично разработанные технические системы.
Год основания: 2022
Количество сотрудников: 51-100
Сайт: burny.games

Преимущества сотрудникам

Откликнуться
Outstaff, Consulting / Integrator
Опыт от 5 лет Senior Full-time Не имеет значения Есть тестовое задание Remote Україна, Bulgaria, Portugal, Romania, Poland
06.05.2025
Подробнее
  • Python
  • SQL
  • Apache Spark
  • AWS Glue
  • Athena
  • Apache Airflow
  • ETL
  • ELT
  • Amazon S3
  • AWS Lambda
  • AWS RDS
  • Amazon API Gateway
  • CI/CD
  • FastAPI
  • Great Expectations

Our client is a leading SaaS company offering pricing optimization solutions for e-commerce businesses. Its advanced technology utilizes big data, machine learning, and AI to assist customers in optimizing their pricing strategies and maximizing their profits.

About the Role:

As a data engineer you’ll have end-to-end ownership – from system architecture and software development to operational excellence.

Key Responsibilities:

  • Design and implement scalable machine learning pipelines with Airflow, enabling efficient parallel execution.
  • Enhance our data infrastructure by refining database schemas, developing and improving APIs for internal systems, overseeing schema migrations, managing data lifecycles, optimizing query performance, and maintaining large-scale data pipelines.
  • Implement monitoring and observability, using AWS Athena and QuickSight to track performance, model accuracy, operational KPIs and alerts.
  • Build and maintain data validation pipelines to ensure incoming data quality and proactively detect anomalies or drift.
  • Collaborate closely with software architects, DevOps engineers, and product teams to deliver resilient, scalable, production-grade machine learning pipelines.

Required Competence and Skills:

To excel in this role, candidates should possess the following qualifications and experiences:

  • A Bachelor’s or higher in Computer Science, Software Engineering or a closely related technical field, demonstrating strong analytical and coding skills.
  • At least 5 years of experience as a data engineer, software engineer, or similar role and using data to drive business results.
  • At least 5 years of experience with Python, building modular, testable, and production-ready code.
  • Solid understanding of SQL, including indexing best practices, and hands-on experience working with large-scale data systems (e.g., Spark, Glue, Athena).
  • Practical experience with Airflow or similar orchestration frameworks, including designing, scheduling, maintaining, troubleshooting, and optimizing data workflows (DAGs).
  • A solid understanding of data engineering principles: ETL/ELT design, data integrity, schema evolution, and performance optimization.
  • Familiarity with AWS cloud services, including S3, Lambda, Glue, RDS, and API Gateway.

Nice-to-Have:

  • Experience with MLOps practices such as CI/CD, model and data versioning, observability, and deployment.
  • Familiarity with API development frameworks (e.g., FastAPI).
  • Knowledge of data validation techniques and tools (e.g., Great Expectations, data drift detection).
  • Exposure to AI/ML system design, including pipelines, model evaluation metrics, and production deployment.

Информация о компании Adaptiq

Adaptiq – это технологическая консалтинговая компания, специализирующаяся на создании и масштабировании R&D команд для высококлассных, быстрорастущих продуктовых компаний в различных отраслях.
Год основания: 2020
Количество сотрудников: 51-100
Сайт: adaptiq.co

Преимущества сотрудникам

Откликнуться
Outsource, Outstaff
Опыт от 3 лет Middle Full-time Не имеет значения Есть тестовое задание Remote
23.04.2025
Подробнее
  • Python
  • Kafka
  • ClickHouse
  • Data lake
  • Argo Workflows
  • Apache Airflow
  • Prefect
  • CI/CD
  • Docker
  • Kubernetes

Boosty Labs is one of the most prominent outsourcing companies in the blockchain domain. Among our clients are such well-known companies as Ledger, Consensys, Storj, Animoca brands, Walletconnect, Coinspaid, Paraswap, and others.
About project: Advanced blockchain analytics and on-the-ground intelligence to empower financial institutions, governments & regulators in the fight against cryptocurrency crime

Requirements:

  • 3+ years of experience in data engineering or a similar role
  • Strong programming skills in Python
  • Solid hands-on experience with Apache Kafka for real-time data streaming
  • Experience working with ClickHouse or other columnar databases
  • Understanding of Data Lake architecture and cloud data storage solutions
  • Familiarity with Argo Workflows or similar workflow orchestration tools (e.g., Airflow, Prefect)
  • Experience with CI/CD processes and containerization (Docker, Kubernetes) is a plus
  • Strong problem-solving skills and the ability to work independently

Responsibilities:

  • Design and implement scalable, efficient, and reliable data pipelines
  • Work with real-time and batch data processing using Kafka and ClickHouse
  • Develop and maintain ETL/ELT processes using Python
  • Manage and optimize data storage in cloud-based Data Lake environments
  • Use Argo Workflows to orchestrate complex data workflows
  • Collaborate with data scientists, analysts, and engineering teams to support their data needs
  • Ensure data quality, consistency, and governance throughout the pipeline

Информация о компании Boosty Labs

Boosty Labs – крупнейший магазин и венчурная студия блокчейн-разработчиков в Европе. Компания предлагает как комплексную разработку программного обеспечения по фиксированной цене, так и специальные команды, полностью интегрирующиеся в операционные процессы бизнеса клиента и сосредотачивающиеся исключительно на его потребностях. Эксперты создают инновационные продукты для быстрорастущих стартапов, малого и среднего бизнеса в области блокчейна, DeFi, NFT, облачных технологий и искусственного интеллекта. Boosty Labs предоставляет услуги по техническому консалтингу, продуктовому менеджменту, инженерным и дизайнерским решениям.
Год основания: 2017
Количество сотрудников: 101-250
Сайт: boostylabs.com

Преимущества сотрудникам

Откликнуться
Recruitment Agency, Consulting / Integrator
Опыт не имеет значения Middle, Senior Full-time Upper-Intermediate / B2 Есть тестовое задание Office, Remote, Hybrid Київ
18.04.2025
Подробнее
  • ETL
  • ELT
  • AWS Glue
  • Apache Spark
  • Python
  • SQL
  • MWAA
  • Apache Airflow
  • IAM
  • Amazon S3
  • AWS
  • Aurora
  • PostgreSQl
  • Kafka
  • AWS Lake Formation

On behalf of our Client, a well-established financial institution from the Caribbean region Mobilunity is looking for a Data Engineer.
Our Client is the largest bank in the Caribbean region that serves 14 countries/territories. The aim is to make this organization from a traditional bank into a new era of fintech, working on the edge of what current fintech may offer.

Requirements:

  • Experience with ETL/ELT
  • Proficiency in Glue and Spark
  • Strong programming skills in Python and SQL
  • Hands-on experience with MWAA / Airflow
  • Good understanding of AWS Basics (IAM, S3)
  • Experience working with Aurora and PostgreSQL
  • Knowledge of Kafka / MSK, including Kafka Connect and Debezium
  • Familiarity with Lake Formation
  • Experience using Glue Data Catalog
  • Solid understanding of data modeling principles
  • Experience with Glue Streaming
  • Level of English – Upper-Intermediate and higher

Nice to have:

  • Previous experience working in the fintech industry

Информация о компании Mobilunity

Mobilunity – украинская компания nearshoring является глобальным поставщиком украинских команд разработчиков. Модель компании предоставляет доступ к кадровому резерву из более чем 200 000 украинских инженеров программного обеспечения, позволяя клиенту полностью контролировать процесс разработки.
Год основания: 2010
Количество сотрудников: 101-250
Сайт: mobilunity.com

Преимущества сотрудникам

Откликнуться
Опыт от 5 лет Lead Full-time Upper-Intermediate / B2 Есть тестовое задание Remote Україна
17.04.2025
Подробнее
  • Apache Spark
  • PySpark
  • Spark SQL
  • AWS
  • EC2
  • IAM
  • Amazon S3
  • AWS Glue
  • EMR
  • SQL
  • NoSQL
  • Apache Airflow
  • Terraform
  • Jira
  • Slack
  • JetBrains
  • Git
  • GitLab
  • Github
  • Docker
  • Jenkins
  • Scala
  • Python

We’re looking for a Lead Big Data Engineer to join our client, a prominent American ebook and audiobook subscription service with a vast library of over 200 million titles.
As a Lead Big Data Engineer, you will play a key role in designing, developing, and maintaining large-scale data processing systems, while providing technical leadership and mentoring to other teams.
This remote position, ideally suited for candidates located in Europe or the US time zone, is perfect for someone with a deep technical background in big data technologies, looking to drive innovation in data engineering.

About the project

Client is an American e-book and audiobook subscription service that includes one million titles. The platform hosts 60 million documents on its open publishing platform.
Core Platform provides robust and foundational software, increasing operational excellence to scale apps and data. We are focused on building, testing, deploying apps and infrastructure which will help other teams rapidly scale, inter-operate, integrate with real-time data, and incorporate machine learning into their products. Working with our customers in the Data Science and Content Engineering, and our peers in Internal Tools and Infrastructure teams we bring systems-level visibility and focus to our projects.
Client’s goal is not total architectural or design perfection, but rather choosing the right trade-offs to strike a balance between speed, quality and cost.

  • Tech Stack: Scala and/or Python, Apache Spark/PySpark/Spark SQL DSL, AWS (EC2, IAM, S3, Glue, EMR), AWS data platform services, Datalake/Lakehouse, SQL and NoSQL, Apache Airflow, Terraform, Jenkins
  • Team Composition: A distributed team consisting of 9 KITRUM Big Data Engineers (Middle/Senior)

Must-have for the position

  • 5+ years Scala and/or Python;
  • 3+ years Spark/PySpark/Spark SQL;
  • Experience with AWS (EC2, IAM, S3, Glue, EMR);
  • Fluency in at least one dialect of SQL and NoSQL;
  • Strong technical and team leadership experience;
  • English Level: Upper-Intermediate English or higher.

Will be a strong plus

  • Experience with Apache Airflow, including writing plugins and custom operators.
  • Experience with Terraform;
  • Strong grasp of AWS data platform services and their strengths/weaknesses;
  • Strong experience using Jira, Slack, JetBrains IDEs, Git, GitLab, GitHub, Docker, Jenkins;

Responsibilities

  • Lead the design, development, and maintenance of scalable and efficient data pipelines;
  • Conduct code reviews and create design documents for new features;
  • Provide technical support and mentorship to other teams within the organization;
  • Participate in management meetings to discuss priorities, scopes, deadlines, and cross-team dependencies;
  • Develop, optimize, and tune Apache Spark jobs (Scala, PySpark, Spark SQL).
  • Implement CI/CD processes using Jenkins and GitHub Actions;
  • Design and implement data processing pipelines using AWS and GCP services;
  • Work with Apache Airflow to manage and automate workflows, including custom plugin development;
  • Utilize Databricks for job scheduling, SQL warehousing, and data visualization;
  • Integrate with 3rd party services to download and upload data as required.

Информация о компании KitRUM

KitRUM – это компания по разработке программного обеспечения и преданные команды, обеспечивающие первоклассные результаты на любом этапе вашего проекта. Компания предоставляет необходимые клиенту конкретные технические знания в предметной отрасли, доступ к командам разработчиков программного обеспечения мирового класса и технологическим консультантам, обеспечивает гибкое масштабирование бизнеса и прочее.
Год основания: 2014
Количество сотрудников: 101-250
Сайт: kitrum.com

Преимущества сотрудникам

Откликнуться

Страницы

Читайте нас в Telegram, чтобы не пропустить анонсы новых курсов.