Вакансії Data Engineer / Big Data Architect
- Python
- SQL
- NoSQL
- Firestore
- BigQuery
- Bigtable
- Redis
- Kafka
- OLAP
- Snowflake
- ClickHouse
- Apache Airflow
- Google Cloud Dataflow
- Hadoop
- Apache Spark
- TDD
- Machine learning
- Docker
- Kubernetes
- GCP
- IaC
- Scala
The company is the first Customer-Led Marketing Platform. Its solutions ensure that marketing always starts with the customer instead of a campaign or product.
It is powered by the combination of:
- rich historical, real-time, and predictive customer data;
- AI-led multichannel journey orchestration;
- statistically credible multitouch attribution of every marketing action.
Requirements:
- At least 5 years of experience with Python.
- At least 3 years of experience in processing structured terabyte-scale data (processing structured data of several hundreds of gigabytes).
- Solid experience in SQL and NoSQL (ideally GCP storages Firestore, BigQuery, BigTable and/or Redis, Kafka).
- Hands-on experience with OLAP storage (at least one of Snowflake, BigQuery, ClickHouse, etc.).
- Deep understanding of data processing services (Apache Airflow, GCP Dataflow, Hadoop, Apache Spark).
- Proven experience in DevOps.
- Experience in automated test creation (TDD).
- Freely spoken English.
Advantages:
- Being fearless of mathematical algorithms (part of our team’s responsibility is developing ML models for data analysis; although knowledge of ML is not required for the current position, it would be awesome if a person felt some passion for algorithms).
- Experience in any OOP language.
- Familiarity with Docker and Kubernetes.
- Experience with GCP services would be a plus.
- Experience with IaC would be a plus.
- Experience in Scala.
Інформація про компанію Gemicle
Переваги співробітникам
- English Courses
- Team buildings
- Гнучкий графік роботи
- Кава, фрукти, перекуси
- Оплачувані лікарняні
- Оплачувана відпустка
- Регулярний перегляд зарплатні
- SQL
- Python
- ETL
- GCP
- BigQuery
- Pub/Sub
- Cloud Storage
- Logging
- Cloud Functions
- Datastream
- Dataform
- dbt
- Microsoft Power BI
- Jira
- Microsoft Fabric
- PySpark
- AWS
- Microsoft Azure
- MySQL
All-in-one SEO and digital marketing platform SE Ranking is looking for a Data Engineer to join its team. We are looking for a highly motivated and proactive professional who is excited to take on new challenges.
Skills and experience:
- Minimum 3 years of experience as a data engineer.
- At least an upper-intermediate level of English.
- Bachelor’s degree in Computer Science, Engineering, or another related technical field.
- Extensive knowledge of SQL and Python.
- Experience in organizing, monitoring, and maintaining data infrastructure: managing ETL processes, writing tests for data analytics code, and monitoring data consistency.
- Practical experience with GCP (Google BigQuery, Logging, Cloud Storage, Cloud Functions, Cloud Run, Datastream, Pub/Sub).
- Practical experience with Dataform (or dbt).
- Understanding of the needs for data structures and data modeling for business intelligence reporting, including Microsoft Power BI.
- Experience with Jira or similar project management tools.
Will be a plus:
- Experience with Microsoft Fabric, PySpark.
- Azure, AWS experience as a plus.
- Experience with MySQL
Duties and responsibilities:
- Work with large, complex data sets and high-throughput data pipelines that meet business requirements.
- Build and maintain the infrastructure required for optimal extraction, transformation, and loading of data from a variety of data sources.
- Conduct SQL performance tuning for existing queries within cloud databases.
- Work with stakeholders to assist with data-related technical issues and support their data infrastructure needs.
- Perform analysis on existing relational and non-relational structures from various sources and business groups.
- Evaluate data for integrity and accuracy; provide feedback and work with business and technical teams to backfill gaps, if any.
- Create and maintain optimized data models and schemas (including star schema) for further usage in analytics and reporting tools like Power BI.
- Design and implement data security and governance measures to ensure data integrity and compliance.
- Collaborate with data scientists and analysts to improve data collection and processing methods.
- Actively participate in strategic business decision-making processes
For this position, you must be able to work independently while also being a dependable team player, highly functional, and capable of efficiently handling complex tasks. Beyond possessing extensive technical expertise, you must also have superior communication abilities to effectively work alongside different teams, swiftly resolving complex challenges.
Інформація про компанію SE Ranking
Переваги співробітникам
- English Courses
- Team buildings
- Бухгалтерський супровід
- Гнучкий графік роботи
- Компенсація навчання
- Медичне страхування
- ETL
- SQL
- Python
- Snowflake
- BigQuery
- DataBricks
- Apache Spark
- Terraform
Sigma Software is seeking an experienced Data Engineer to join our growing team.
If you thrive on solving complex problems with advanced cloud-based data collection, transformation, analysis, and monitoring tools, this opportunity is for you!
Teamwork is one of our core values, so we would like to welcome specialists who excel in collaboration, contribute to a positive and friendly atmosphere, and are passionate about their work.
Requirements
- At least 5 years of experience in data engineering and with cloud computing services solutions in the area of data and analytics, Azure is preferred one.
- Conceptual knowledge of data analytic fundamentals, e.g., dimensional modeling, ETL, reporting tools, data governance, data warehousing, and structured and unstructured data
- Knowledge of SQL and experience with at least one programming language (Python)
- Excellent communication skills and business fluency in English
- Understanding Big Data DB like Snowflake, BigQuery, etc
- Experience in database development and data modeling, ideally with Databricks/Spark
- Experience in the pre sales activities
- Excellent interpersonal skills
- Strong problem-solving and decision-making abilities, with a focus on driving results and meeting deadlines
- Strong passion for learning/practicing new techniques
Would be a plus:
- A subject matter expert in a specific domain: Logistics, Aviation, or Sustainability
- Experience with Infra as code (Terraform or similar)
- Good communication skills in German
Responsibilities
- Work closely with clients and other stakeholders to clarify the tech requirements and expectations
- Implement architectures based on Azure Cloud platforms (Data Factory, Databricks, etc.)
- Design, develop, optimize, and maintain squad-specific data architecture and pipelines that adhere to defined ETL and Data Lake principles
- Discover, understand, and organize disparate data sources and structure them into clean data models with clear and understandable schemas
- Contribute to the evaluation of new tools for analytical data engineering
- Participate in evaluating and preparing proposals for new opportunities
- Suggest and contribute to training and improvement plans regarding analytical data engineering skills, standards, and processes
Інформація про компанію Sigma Software
Переваги співробітникам
- Work-life balance
- Гнучкий графік роботи
- Медичне страхування
- Освітні програми, курси
- Юридичний супровід
- Snowflake
- DWH
- CI/CD
- Dataiku
- Azure Data Factory
We’re opening the position of a Middle / Middle Strong Data Engineer to our team aimed at developing a data warehouse solution which includes business planning & execution, process automation, and visualization of data. The solution will allow streamlining data processes, gaining valuable insights, and making informed business decisions based on the integrated data.
Responsibilities:
- Build a data mart design and documentation of sophisticated data architecture solutions, including data models, data integration patterns, and data storage mechanisms
- Collaborate with the Product Owner and the team to implement data solutions aligned with the organization’s strategic goals
- Work alongside Data Engineers to ensure seamless integration of data pipelines and data products with the overall data architecture
- Guide the migration from our current in-house data warehouse to Snowflake, focusing on strategic planning, risk management, and execution
- Optimize data storage and retrieval processes to fully leverage Snowflake capabilities, focusing on performance, scalability, and cost efficiency
Requirements:
- A minimum of 4 years of experience in data engineering, data modelling, database design, data warehousing, and data governance practices
- Production experience with building, maintaining, and performance optimization of Snowflake DWH
- Strong knowledge of integration with monitoring and observability capabilities
- Experience in performance testing and optimization
- Solid understanding of standard engineering practices such as testing and CI/CD
- Ability to conduct code reviews and ensure adherence to coding standards
- Excellent problem-solving and analytical skills
- Strong communication and collaboration abilities
- English level – Upper-Intermediate or higher
Will be a plus:
- SnowPro® Certifications
- Experience with Dataiku for pipeline orchestration
- Production experience with Azure data factory
Інформація про компанію Edvantis
Переваги співробітникам
- English Courses
- Допомога психотерапевта
- Медичне страхування
- Надається ноутбук
- Оплачувані лікарняні
- Освітні програми, курси
- Регулярний перегляд зарплатні
- Java
- Prometheus
- Grafana
- Zabbix
- Datadog
- ETL
- Apache Airflow
- Kubernetes
- ClickHouse
- Snowflake
- Redis
- SQL
- RDBMS
- NoSQL
Must haves
- BSc. in Computer Sciences from a top university, or equivalent;
- 5+ years in data engineering, and data pipeline development in high-volume production environments;
- 2+ years experience with Java;
- 2+ years experience with monitoring systems (Prometheus, Grafana, Zabbix, Datadog);
- Ability to develop, design, and maintain end-to-end ETL workflows, including data ingestion and transformation logic, involving different data sources;
- Experience with data-engineering cloud technologies as: Apache Airflow, K8S, Clickhouse, Snowflake, Redis and cache technologies;
- Experience with relational and non-relational DBs. Proficient in SQL and queries optimizations;
- Experience with designing infrastructure to maintain high availability SLAs;
- Experience with monitoring and managing production environments;
- Upper-intermediate English level.
Інформація про компанію AgileEngine
Переваги співробітникам
- Гнучкий графік роботи
- Зарплатня вище ринку
- Регулярний перегляд зарплатні
- AWS
- GCP
- NoSQL
- Microsoft Azure
- Hadoop
- Kafka
- Apache Airflow
- Redshift
- BigQuery
- Snowflake
- HBase
- DynamoDB
- Apache Spark
- Apache Flink
- Lambda
- Data Vault
- Data Fabric
- DWH
- GDPR
- ETL
- ELT
- CI/CD
- SQL
- Python
- Java
- Scala
Що потрібно робити:
- Розробка архітектурних рішень у сфері Big Data на основі сервісів від хмарних провайдерів AWS та/або GCP
- Менторинг технічних команд. Підтримка розробки спроектованих рішень
- Розробка процесів та політик Data Governance компанії, їх затвердження та контроль виконання
- Проектування рішень для real-time та батчингової аналітики та систем зберігання та обробки Big Data
- Участь у проектуванні нового сховища даних (DWH)
- Управління потоками даних всередині організації
- Проектування MDM та RDM систем
- Врядування (governance) MDM та RDM даних та Data Security
- Розробка ефективної гібридної хмарної архітектури даних
- Проектування ефективної взаємодії рішень і сервісів між хмарами
- Оптимізація Платформи даних компанії з точки зору cost/perf підходів
- Розробка конвеншенів та правил
- Розробка міграційних стратегій
- Розробка функціональних та нефункціональних вимог для вибору рішень, інструментів, технологій
- Проведення пілотів (PoC) для різних технологій, рішень та підходів
- Запровадження best practices в сферах інженерії даних, Big Data, аналітики на великих даних
- Проведення внутрішніх навчальних вебінарів та воркшопів для технічної (в першу чергу) та бізнес аудиторії
Необхідні знання, досвід та особисті якості:
- Вища освіта (технічна, IT)
- Загальний досвід в ІТ індустрії 7+ років
- Досвід роботи у Big Data сфері 3+ років
- 1+ роки досвіду роботи на позиції Data Architect у великих компаніях з Big Data/або досвід в ролі Tech Lead з архітектурними обов’язками 1.5+ років (або на подібній позиції)
- Досвід у розробці архітектурних рішень у сфері даних (batching, streaming) на основі AWS та/або GCP стеку
- Вміння логічно мислити, формувати і довести свою думку, використовуючи різні канали комунікації (усний, візуальний). Вміння розробляти презентації, будувати діаграми та mind maps
- Знання та практичне застосування однієї з архітектурних методологій або бажання їх швидко освоїти
- Досвід роботи з вимогами та стейкхолдерами. Вміння відстоювати правильну точку зору
- Розуміння Big Data стеків та технологій, включаючи хмари, MPP системи, NoSQL бази даних: GCP, AWS, Azure, Hadoop, Apache Kafka, Apache Airflow, Redshift, Bigquery, Snowflake, Hbase, DynamoDB, Spark, Flink та ін. Вміння оптимально вибирати технологію під конкретне рішення
- Шаблони проектування та підходи у сфері Великих Даних (Lambda та Kappa архітектури, Data Vault, Data Fabric та ін.)
- Розуміння підходів до моделювання корпоративного DWH
- Розуміння GDPR чи подібних законодавчих норм
- Компетентність в Data Security, інформаційній безпеці. Досвід розробки або проектування захищених рішень
- Досвід та розуміння принципів оркестрування обробки Великих Даних, володіння відповідними найкращими практиками, інструментами та підходами. Розуміння та досвід побудови ETL/ELT процесів
- Досвід та практичне застосування сучасних практик розробки, таких як Code Review, CI/CD, IaC та ін.
- Відмінні знання SQL та однієї з імперативних мов: Python, Java, Scala та ін.
- Англійська Upper-Intermediate B2+
Додатковою перевагою буде:
- Наявність сертифікації AWS/GCP
Інформація про компанію ПриватБанк
Переваги співробітникам
- English Courses
- Fitness Zone
- Gaming room
- Кімната відпочинку
- Кава, фрукти, перекуси
- Компенсація навчання
- Медичне страхування
- TypeScript
- Node.js
- Python
- ETL
- GCP
- AWS
- Apache Spark
- BigQuery
- Fivetran
- dbt
Svitla Systems Inc. is looking for a Senior Data Engineer for a full-time position (40 hours per week) in Ukraine. Our client is a renewable energy company that offers solutions for organizations transitioning to sustainable energy sources. Based in Portland, Oregon, the company provides a comprehensive platform that helps businesses reduce their carbon footprints by sourcing renewable energy across their supply chains. Their services include financial, logistical, and technical support for integrating renewable energy solutions, such as Power Purchase Agreements (PPAs), energy credits, and carbon offsets. It unlocks cost savings and the ability to meet carbon commitments in time. They identify energy needs within a company’s operations using data-driven insights and remote sensing technology and create tailored solutions that optimize costs and environmental impact.
Requirements:
- A Bachelor’s or Master’s degree in Computer Science, Software Engineering, Data Engineering, or a related field.
- 7+ years of experience in data engineering, particularly in designing and implementing scalable, high-performance, cloud-based software solutions using TypeScript, Node.js, Python, and related frameworks.
- 7+ years of experience in building and scaling complex software systems in fast-paced environments, focusing on performance optimization, data throughput, and fault-tolerant architectures.
- Extensive knowledge in data architecture, including database design, data modeling, and the development of data pipelines and ETL processes.
- Expertise in designing and deploying reliable, multi-tenant cloud services focusing on data integrity, performance, and scalability.
- Expertise in designing distributed data systems that handle massive datasets, with in-depth knowledge of batch and real-time processing frameworks.
- Advanced understanding of data modeling principles, relational and non-relational databases, and data warehousing solutions.
- Strong knowledge of cloud infrastructure platforms (e.g., Google Cloud Platform, AWS) and experience with infrastructure-as-code tools to manage data-centric environments.
- Hands-on experience with performance tuning and scaling data systems to meet the demands of enterprise-grade SaaS products.
- Understanding of designing and optimizing analytics engines, particularly leveraging tools like Spark, BigQuery, debt, and Fivetran to build robust data processing pipelines.
- A proven track record of leading large-scale data engineering projects, improving data infrastructure, and driving technical strategies for multi-tenant SaaS platforms.
- Demonstrated ability to lead and mentor other engineers in best practices for data management, data engineering, and backend development.
- Exceptional problem-solving skills with a keen eye for optimizing data architectures and backend performance.
- A mindset of continuous improvement and a passion for driving data-driven decisions and innovations within your team.
- Experience leading cross-functional initiatives focusing on data governance, security, and compliance in cloud-based environments.
Responsibilities
- Play a crucial role in architecting and building a highly scalable, multi-tenant, data-driven system.
- Leverage your expertise in open-source technologies like Apache Spark, Google BigQuery, dbt, and Fivetran to handle large-scale data processing and analytics.
- Collaborate with other senior engineers to design and implement innovative solutions to complex challenges around data architecture, high-throughput data pipelines, concurrency, and system reliability.
- Excel at balancing high-level architectural decisions with hands-on implementation and leading technical initiatives that shape the future of the data platform.
- Create rock-solid reports and clear, exciting data visualizations.
- Develop a deep understanding of the data architecture, related datasets, and interactions.
- Collaborate across multi-functional teams.
- Automate frequent reporting requests to maximize efficiency.
Інформація про компанію Svitla
Переваги співробітникам
- English Courses
- Pet-friendly
- Team buildings
- Work-life balance
- Відпустка по догляду за дитиною
- Гнучкий графік роботи
- Кава, фрукти, перекуси
- Компенсація витрат на спорт
- Компенсація навчання
- Медичне страхування
- Оплачувані державні свята
- Оплачувані лікарняні
- Регулярний перегляд зарплатні
- Azure Data Factory
- DataBricks
- Python
- SQL
- Azure DevOps
- Parquet
- SFTP
- Scrum
- ETL
- Informatica PowerCenter
- Talend
- SSIS
- Apache NiFi
- Jira
- Confluence
- Control-M
We are seeking a proactive Mid-Level Data Engineer with a minimum of 3 years of experience in data ingestion, processing, and management within cloud environments, particularly the Azure ecosystem. The ideal candidate will collaborate with agile teams using methodologies like Scrum to develop and optimize our data infrastructure.
Responsibilities:
- Develop and manage data pipelines using Azure Data Factory (ADF).
- Process, transform, and optimize data with Databricks, Python, and SQL.
- Collaborate with Product Owners and Tech Leaders to implement data solutions during agile sprints.
- Store and organize data in Azure Storage Accounts.
- Automate workflows to enhance data ingestion and processing efficiency.
- Document processes and implement best practices across the data pipeline.
Required Skills:
- Bachelor's degree in Computer Science, Information Systems, or a related field.
- Minimum of 2 years of experience with Azure Data Factory (ADF) for orchestrating data pipelines.
- Proficiency in Databricks, Python, and SQL.
- Familiarity with Azure DevOps and data formats like Parquet.
- Understanding of data transfer protocols such as SFTP.
- Experience working in agile teams using Scrum methodologies.
- Bachelor's degree in Computer Science, Information Systems, or a related field.
- English proficiency at B2 level or higher.
Nice to Have:
- Experience with data integration or ETL tools such as Informatica PowerCenter, Talend, SSIS, Nifi or similar platforms.
- Knowledge of ingesting data from SAP R/3 systems via IDocs.
- Familiarity with tools like Jira, Confluence, and schedulers like Control-M.
- Understanding of data security practices in cloud environments.
Additional Skills:
- Strong communication and collaboration abilities.
- Proactive in documenting and improving data workflows.
- Team-oriented with a commitment to excellence.
Інформація про компанію Capgemini Engineering
Переваги співробітникам
- English Courses
- Відпустка по догляду за дитиною
- Компенсація витрат на спорт
- Медичне страхування
- Надається ноутбук
- Освітні програми, курси
- Парковка для авто
- PostgreSQl
- Microsoft SQL Server
- Oracle
- GCP
- BigQuery
- Hadoop
- Apache Spark
- Hive
- HBase
- NoSQL
- Elasticsearch
- Apache NiFi
- Apache Airflow
- Kafka Streams
- Storm
- Python
- Scala
We are looking for a savvy Data Engineer to join our growing team of analytics experts. The hire will be responsible for expanding and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection for cross-functional teams. The ideal candidate is an experienced data pipeline builder and data wrangler, who enjoys optimizing data systems and building them from the ground up.
The Data Engineer will support our software developers, database architects, data analysts, and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects. They must be self-directed and comfortable supporting the data needs of multiple teams, systems, and products. The right candidate will be excited about the prospect of optimizing or even re-designing our company's data architecture to support our net generation of products and data initiatives.
Main tasks and responsibilities:
- Create and maintain optimal data pipeline architecture,
- Assemble large, complete data sets that meet functional / non-functional business requirements. Identified, designed, and implemented internal process improvements: automating manual processes,
- optimizing data delivery, re-designing infrastructure for greater scalability, performance, etc.
- Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and big data technologies. Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, revenue management, operational efficiencies, and other key business performance metrics. Work with stakeholders including the Executive, Product, Data, and Design teams to assist with
- data-related technical issues and support their data infrastructure needs.
- Keep our data separated and secure across national boundaries through multiple data centers and regions.
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our offerings.
- Work with data and analytics experts to strive for greater functionality in our data systems.
- Other duties as assigned
Must have:
- 5+ years of experience in a Data Engineer role
- Must have Experience working with relational databases and data warehouses, including Postgres, Microsoft SQL Server, Oracle, GCP BigQuery, etc.
- Experience with big data tools: Hadoop, Spark, Hive, HBase, etc.
- Experience with NoSQL data repositories, ideally, Elastic Search.
- Experience building and optimizing big data pipelines, architectures, and data sets. Familiar with workflow management tools: NiFi, Airflow, etc.
- Must have experience with stream-processing systems, such as Apache Spark and/or Kafka Streaming, Storm, etc.
- Must have Experience with programming and scripting languages: Python, Scala, OOP
Інформація про компанію Capgemini Engineering
Переваги співробітникам
- English Courses
- Відпустка по догляду за дитиною
- Компенсація витрат на спорт
- Медичне страхування
- Надається ноутбук
- Освітні програми, курси
- Парковка для авто
- Python
- ETL
- Apache Airflow
- Grafana
- Redash
- Plotly
- AWS Glue
- Athena
- AWS
- Microsoft Excel
- SDLC
- SQL
- API
Customer is one of the biggest companies on the market of home entertainment consumer electronics devices that strives to provide their clients with high-quality products and services.
This position collaborates with a geographically diverse team to develop, deliver, and maintain systems for digital subscription and transactional products across the Customer’ SVOD portfolio.
Requirements:
- 1+ years of python development
- Experience building ETLs in python
- Experience with data tools (ex.: Airflow, Grafana, Redash, Plotly, AWS Glue, AWS Athena)
- Other AWS experience
- Advanced skills in Excel
- Agile SDLC knowledge
- Detail oriented
- Data-focused
- Strong verbal/written communication, including an ability to effectively communicate with both business and technical teams
- An ability and interest in working in a fast-paced and rapidly changing environment
- Be self-driven and show ability to deliver on ambiguous projects with incomplete or dirty data
Would be a plus:
- 1+ years of experience with SQL
- Understanding of basic SVOD store purchase workflows
- Background in supporting data scientists in conducting data analysis / modelling to support business decision making
- Expereince with mixpanel, mparticle, youbora and similar systems
Responsibilities:
- Building Pyhton APIs as a data instrumentation
- ETL buildouts for data reconciliation
- Creation of automatically-running audit tools
- Interactive log auditing to look for potential data problems
- Help in troubleshooting customer support team cases
- Troubleshooting and analyzing subscriber reporting issues:
- Answer management questions related to subscriber count trends
- App purchase workflow issues
- Audit/reconcile store subscriptions vs userdb
- Assist PSVOD BI team with questions related to subscriber status/counts
Інформація про компанію GlobalLogic
Переваги співробітникам
- Relocation assistance
- Б'юті послуги
- Допомога психотерапевта
- Компенсація витрат на спорт
- Медичне страхування
- Освітні програми, курси
- AWS
- Elasticsearch
- OpenSearch
- Enterprise Search Platforms
- Python
- Kubernetes
- NLP
- Terraform
Наш клієнт – це компанія, яка займається аналітичними дослідженнями фінансового ринку. Наш клієнт належить до трійки найвпливовіших міжнародних рейтингових агентств, і є провайдером потужних рішень, пов’язаних з великими даними.
Навички
- Досвід роботи в якості Big Data Engineer від 3 років
- Знання та досвід роботи з:
- Amazon Web Services
- Elasticsearch
- Enterprise Search Platforms
- OpenSearch
- Python
- Semantic Web and Text Analytics
- Знання письмової та усної англійської мови на рівні Upper-Intermediate і вище (В2+)
Обов'язки
- Створення аналітичної платформи, що працює з декількома джерелами даних, збирає дані, обробляє їх за допомогою NLP-пайплайнів
- Ефективний пошук даних в AWS Opensearch, а також надання даних для використання внутрішнім клієнтам та потоку ML на проєкті
Буде перевагою
- Data Science
- Kubernetes
- NLP Labelling
- NLP Preprocessing
- Terraform
Інформація про компанію EPAM
Переваги співробітникам
- English Courses
- Relocation assistance
- Гнучкий графік роботи
- Допомога психотерапевта
- Компенсація домашнього офісу
- Компенсація навчання
- Медичне страхування
- Оплачувані лікарняні
- Освітні програми, курси
- Python
- SQL
- AWS
- Apache Spark
- DataBricks
- Apache Airflow
- MWAA
- Kafka
- AWS MSK
- AWS Kinesis
Big Data & Analytics is the Center of Excellence's data consulting and data engineering branch. Hundreds of data engineers and architects nowadays build Data & Analytics end-to-end solutions from strategy through technical design and proof of concepts to full-scale implementation. We have customers in the healthcare, finance, manufacturing, retail, and energy domains.
We hold top-level partnership statuses with all the major cloud providers and collaborate with many technology partners like AWS, GCP, Microsoft, Databricks, Snowflake, Confluent, and others.
If you are
- A Big Data engineer specializing in data pipeline creation
- Adept at batch processing and streaming processing
- Proficient in Python and SQL
- Experienced in developing data solutions on the AWS cloud platform
- Skilled in utilizing Apache Spark, Databricks
- Familiar with orchestration tools like Apache Airflow or Managed Apache Airflow (MWAA)
- Knowledgeable about Apache Kafka, Amazon MSK (Managed Streaming for Apache Kafka), or Kinesis for data streaming
And you want to
- Be part of a team of data-focused engineers dedicated to continuous learning, improvement, and daily knowledge-sharing
- Work with a cutting-edge technology stack, including services from major cloud providers that are at the forefront of innovation and may not yet be generally available
- Engage with customers from diverse backgrounds, including large global corporations to emerging crypto startups preparing to launch their first product
- Be involved in the entire project lifecycle, from initial design and proof of concepts (PoCs) to minimum viable product (MVP) development and full-scale implementation
Інформація про компанію SoftServe
Переваги співробітникам
- Fitness Zone
- Гнучкий графік роботи
- Компенсація витрат на спорт
- Медичне страхування
- Оплачувані лікарняні
- Python
- Snowflake
- SAP HANA
- SQL
We are looking for a seasoned Senior Data Engineer to help us shape Emergn’s exciting future and play an important role in our growth.
We want you to:
- Work with stakeholders including data, design, product and executive teams and assist with data-related technical issues.
- Identify, design and implement process improvements including infrastructure re-design for greater scalability, data delivery optimization, and automation of manual processes.
- Build and support required infrastructure for optimal extraction, transformation and loading of data from various data sources.
- Design and assemble large, complex sets of data that meet non-functional and functional business requirements.
- Willingness to continuously learn & share learnings with others.
This job might be for you if you have:
- 3+ years of experience in Data Engineering or Data Science;
- Experience with Python;
- Experience with Snowflake or SAP Hana;
- Experienced with SQL;
- Highly experienced and skilled at collaborating with business clients;
- An analytical and creative approach to design and problem-solving;
- Excellent communication skills, including presentation skills;
- Must be fluent in English.
Інформація про компанію Emergn
Переваги співробітникам
- Team buildings
- Компенсація навчання
- Медичне страхування
- Освітні програми, курси
- Регулярний перегляд зарплатні
- dbt
- Python
- Flask
- Pandas
- AWS
- Boto3
- Amazon S3
- DataBricks
- Apache Airflow
- scikit-learn
- Snowflake
- VS Code
- Terraform
- GitHub Actions
- EKS
We are looking for a passionate and motivated Data Engineer to join our team.
Our customer is a leading global provider of high-quality licensed images, videos, and music. Our customer helps inspire graphic designers, creative directors, video editors, filmmakers, web developers, and other creative professionals by providing diverse content to businesses, marketing agencies and media organizations around the world. Content creators contribute their work to our client, where end users buy and use it in a diverse array of creative personal and business projects. It's a two-sided marketplace that empowers the world's storytellers.
Essential requirements:
- Strong communication skills
- DBT (core) Proficiency (Ideally also dbt architecture experience)
- Experience creating production quality models.
- Python Expert – writing production level microservices and large scale data wrangling.
- OOP, Flask, pandas
- AWS experience
- Boto3, S3
Preferred:
- Databricks
- Airflow
- ML experience – sklearn
- Snowflake
- VS code
Nice to have:
- Terraform
- Git actions
- EKS
Responsibilities:
- Prepare and deliver specific datasets to meet detailed client requirements, supporting sales teams in their data needs.
- Process complex data requests through both manual and semi-automated methods, ensuring high accuracy.
- Utilize various data tools and techniques, including vector search and data science methods, for efficient dataset preparation.
- Collaborate with team members to address unique data requests.
- Contribute to automation efforts, refining workflows and processes where possible.
Soft skills:
- Strong communication
- Good organizational skills
- Ability to work in a dynamic environment with shifting priorities and tight timelines
- Experience training/ mentoring junior team members
Інформація про компанію N-iX
Переваги співробітникам
- English Courses
- Гнучкий графік роботи
- Компенсація витрат на спорт
- Компенсація навчання
- Медичне страхування
- PySpark
- DataBricks
- Cloudera
- Apache Spark
- Python
- SQL
- Azure DevOps
- Microsoft Azure
- CI/CD
- Git
- Data lake
- Agile
- Scrum
We are hiring for projects in the Middle East. The team comprises front-end and back-end developers, data analysts and data scientists, architects, analysts, and project managers. You will work with modern technologies and be an integral part of continuing to evolve our high-performing platforms. If you are serious about being a part of this fast-paced and exciting journey, this project should be top of the list!
This role will be primarily responsible for closely collaborating with a business SME to implement data validation rules.
Responsibilities
- Collaborating with a business SME to implement data validation rules.
- Understand business requirements and actively provide input from a data perspective.
- Understand the underlying data and flow of data.
- Understand overall platform architecture and capabilities and how to load data effectively.
- Recognize and adapt to the changes in processes.
Must have skills
- 7 years + of relevant experience.
- Proficient hands-on experience with PySpark-based data platforms:
- Databricks, Cloudera, Apache Spark, etc. (3y+).
- Proficiency in Python and PySpark (5y+).
- Advanced knowledge in SQL (5y+).
- Should be able to analyse and understand complex data.
- Knowledge of Azure data/delta lake, Azure DevOps, CI/CD, and Git is essential.
- Should clearly understand the data platform lifecycle and contribute to preparing Design documents, Unit Test plans, and Code review reports.
- Knowledge of data Lake house and business intelligence architecture.
- Experience working in an Agile environment (Scrum, Agile) is a plus.
- Interpersonal/communication skills (both oral/written) with the ability to communicate at various levels with clarity & precision.
- Bachelor's degree in IT/Computer Science or related field
Nice to have
- Knowledge of the financial markets, portfolio theory, and risk management is a plus
Інформація про компанію Luxoft
Переваги співробітникам
- Relocation assistance
- Team buildings
- Багатонаціональна команда
- Велика стабільна компанія
- Освітні програми, курси
- PySpark
- Azure Databricks
- Azure Data Factory
- ELT
- ETL
- SQL
- Scala
- Python
- Agile
- Scrum
Our client is a home improvement company with over 2,000 stores operating in 8 countries across Europe. The company offers a wide range of home improvement products and services to both consumers and trade professionals, available in-store and through our e-commerce channels.
Project overview
Working in cross functional team using Agile framework to design, build and maintain home improvement solution.
The successful candidate will be involved into work within IT teams to support developers in designing and building of Data Products and Solutions.
Position overview
Providing engineering support and data insights by collecting, analyzing, and summarizing development and service issues.
Technology stack:
- Pyspark
- Azure Databricks
- Azure Data Factory
Responsibilities
- Design and develop new framework with modern cloud DWH architecture and ETL processes
- Defining principles for acquisition, storage, corporate data services, and data accesses
- Evolve and adapt the data platform to the latest open-source technologies
- Help with building scalable ETL pipelines
- Solve data-related business objectives using existing tools
- Integrate easily with APIs
- Perform data preparation for Data Scientists
- Provide direction to data engineers working on the data platform, including best practices, review and testing processes
- Provide information by collecting, analysing, and summarising development and service issues
Requirements
- 3-5 years of experience or 3-5 completed projects
- Experience with Databricks and DBT
- Knowledge of data management fundamentals (data modeling, ELT/ETL, data quality, metadata management, data warehouse/lakes patterns, distributed systems)
- Implement cloud data technologies on Azure
- Strong SQL skills
- Knowledge of programming language (Scala, Python, etc.)
- Hands-on experience with data processing software and algorithms
- Good analytical background and the ability to perform data preparation for Data Scientists
Nice to have
- Understanding of Agile / Scrum approaches
Інформація про компанію DataArt
Переваги співробітникам
- English Courses
- Fitness Zone
- Gaming room
- Paid overtime
- Team buildings
- Work-life balance
- Без дрес-коду
- Відпустка по догляду за дитиною
- Велика стабільна компанія
- Велопарковка
- Гнучкий графік роботи
- Довгострокові проекти
- Кімната відпочинку
- Кава, фрукти, перекуси
- Медичне страхування
- Оплачувані лікарняні
- Освітні програми, курси
- AWS
- SaaS
- Fivetran
- Looker
- dbt
- DataBricks
- Apache Airflow
- API
- SQL
- Python
Our client is a successful, self-funded specialty food and wellness company, rapidly growing with a team of over 550 people. Originally established in Newark, New Jersey during the Great Depression, they began by selling premium nuts at an open-air market.
Position overview
The ideal candidate will play a key role in establishing the groundwork for diverse types of analysis. This includes conducting comprehensive cross-channel analysis of media interaction logs, optimizing paid media efforts, and distilling purchase data to gain deeper insights into customer behavior. Additionally, the candidate will evaluate warehouse and sourcing data to enhance operational efficiency. You’ll also be expected to brainstorm innovative methods to improve our data systems, supporting better business decisions and enhancing the overall shopping experience.
Responsibilities
- Apply your engineering expertise to solve large-scale business problems. Drive design and architecture discussions and build alignment on complex technical decisions
- Design, implement, validate, and deploy software solutions to enhance our reporting and analytics systems. This includes new features and architectural improvements for our data pipeline and analytics tools, such as Looker, dbt, Databricks, Fivetran, Airbyte, and AWS Redshift
- Provide technical leadership in data architecture and help establish best practices for data modeling, storage, and retrieval
- Collaborate with cross-functional teams to understand data needs and design solutions that meet business requirements
- Drive the adoption of advanced analytics techniques and technologies to unlock insights from large datasets
- Perform design and code reviews to ensure quality and scalability
- Lead the design of data governance policies and procedures. Help establish and enforce data security and privacy protocols to safeguard sensitive information
- Further agile and DevOps practices to ensure efficient and high-quality data engineering processes
- Support on-call assignments and operational management
- Collaborate with business analysts and product owners to assist with measuring experiments to vet hypotheses
- Mentor, lead, and train other members of the data engineering team
Requirements
- Bachelor's degree required in Computer Science, Software Engineering, or related field
- 3+ years of experience in data engineering working with big data systems that can process and transform data at scale
- Deep knowledge of analytics technologies, modeling techniques, data streaming and other ingestion approaches, data warehousing concepts, cloud platforms (AWS preferred), and proven experience with data architecture including data pipelines, dimensional modeling, and dependency management
- Experience developing with data SaaS and BI tooling (Fivetran, Looker, dbt, Databricks, Apache Airflow, etc.)
- Experience automating data integration using APIs
- Proficiency in programming languages such as SQL and Python in an analytics context
- Experience handling operational aspects of large-scale systems focusing on observability, tech debt management, code refactoring, and robust software development and release processes
- Experience articulating technically complex topics in an easy-to-understand way via written and verbal communication methods
- Proven track record of mentorship, leadership, and ownership within teams and organizations
Nice to have
- Master's degree preferred in a related field
Інформація про компанію DataArt
Переваги співробітникам
- English Courses
- Fitness Zone
- Gaming room
- Paid overtime
- Team buildings
- Work-life balance
- Без дрес-коду
- Відпустка по догляду за дитиною
- Велика стабільна компанія
- Велопарковка
- Гнучкий графік роботи
- Довгострокові проекти
- Кімната відпочинку
- Кава, фрукти, перекуси
- Медичне страхування
- Оплачувані лікарняні
- Освітні програми, курси
- SQL
- Java
- Python
- Scala
- Hadoop
- MapReduce
- Pig
- Hive
- Impala
- Apache Spark
- Kafka
- Storm
- NoSQL
- HBase
- Cassandra
- AWS
- GCP
- Microsoft Azure
- Kubernetes
- Vertica
- Netezza
- Greenplum
- Aster Data
- PowerBI
- Tableau
- Agile
- Scrum
- CI/CD
- DataBricks
- Snowflake
Ciklum is looking for a Lead Data Engineer to join our team full-time in Ukraine.
About the role:
As a Lead Data Engineer, become a part of a cross-functional development team engineering experiences of tomorrow.
Responsibilities
- Responsible for the building, deployment, and maintenance of mission critical analytics solutions that process data quickly at big data scales
- Contributes design, code, configurations, and documentation for components that manage data ingestion, real time streaming, batch processing, data extraction, transformation, and loading across multiple data storages
- Owns one or more key components of the infrastructure and works to continually improve it, identifying gaps and improving the platform’s quality, robustness, maintainability, and speed
- Cross-trains other team members on technologies being developed, while also continuously learning new technologies from other team members
- Interacts with engineering teams and ensures that solutions meet customer requirements in terms of functionality, performance, availability, scalability, and reliability
- Performs development, QA, and dev-ops roles as needed to ensure total end to end responsibility of solutions
- Works directly with business analysts and data scientists to understand and support their use-cases
- Contribute in the Unit’s activities and community building, participate in conferences, provide excellence in exercise and best practices
- Help in sales activities, customer meetings and digital services
Requirements
We know that sometimes, you can’t tick every box. We would still love to hear from you if you think you’re a good fit!
- 5+ years of experience coding in SQL, Java, Python, Scala, with solid CS fundamentals including data structure and algorithm design
- 3+ years contributing to production deployments of large backend data processing and analysis systems as a team lead
- 2+ years of hands-on implementation experience working with a combination of the following technologies: Hadoop, Map Reduce, Pig, Hive, Impala, Spark, Kafka, Storm, SQL and NoSQL data warehouses such as Hbase and Cassandra
- 3+ years of experience in cloud data platforms (AWS, Azure, GCP)
- Experience with Kubernetes
- Knowledge of SQL and MPP databases (e.g. Vertica, Netezza, Greenplum, Aster Data)
- Knowledge of professional software engineering best practices for the full software
- Knowledge of Data Warehousing, design, implementation and optimization
- Knowledge of Data Quality testing, automation and results visualization
- Knowledge of BI reports and dashboards design and implementation (PowerBI, Tableau)
- Knowledge of development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations
- Experience participating in an Agile software development team, e.g. SCRUM
- Experience designing, documenting, and defending designs for key components in large distributed computing systems
- A consistent track record of delivering exceptionally high-quality software on large, complex, cross-functional projects
- Demonstrated ability to learn new technologies quickly and independently
- Ability to handle multiple competing priorities in a fast-paced environment
- Undergraduate degree in Computer Science or Engineering from a top CS program required. Masters preferred
- Experience with supporting data scientists and complex statistical usecases highly desirable
Desirable
- Understanding of cloud infrastructure design and implementation
- Experience in data science and machine learning
- Experience in backend development and deployment
- Experience in CI/CD configuration
- Good knowledge of data analysis in enterprises
- Experience with Databricks, Snowflake
Інформація про компанію Ciklum
Переваги співробітникам
- Team buildings
- Англомовне середовище
- Бухгалтерський супровід
- Компенсація домашнього офісу
- Надається ноутбук
- Освітні програми, курси
- Microsoft Power BI
- DAX
- Power Query
- ETL
- SQL
- Azure SQL
- MSSQL
- Azure Data Factory
- Azure Synapse
- Agile
- Scrum
- Power BI Embedded
- REST
- JS API
- CI/CD
- Azure DevOps
- Tabular
- DAX Studio
- SQL Profiler
- PowerShell
- Swagger
- Postman
Ciklum is looking for a Senior Data Engineer to join our team full-time in Ukraine.
About the role:
As a Senior Data Engineer, become a part of a cross-functional development team engineering experiences of tomorrow.
Responsibilities
- Designing, developing, and deploying Power BI solutions, providing advanced data analytics, and creating dashboards and reports that deliver key insights to the business
- Data models design and implementation
- Optimize reports and dashboards for better performance by improving query speed and ensuring efficient data model design
- Contributes design, code, configurations, manage data ingestion, closed to real-time processing, batch processing, ETL across multiple data storages
- Responsible for performance tuning of complicated SQL and DAX queries
- Identify gaps and improve the platform’s quality, robustness, maintainability, and speed
- Cross-train other team members on technologies being developed, while also continuously learning new technologies from other team members
- Contribute in the Unit’s activities and community building, participate in conferences, and provide excellence in exercise and best practices
Requirements
We know that sometimes, you can’t tick every box. We would still love to hear from you if you think you’re a good fit!
- 3+ years of experience in developing BI solutions with Power BI
- Proficient in Power BI Desktop and Power BI Service, including report and dashboard creation, data transformations, and DAX formulas
- Strong expertise in DAX (Data Analysis Expressions) for complex calculations
- Experience with Power Query for data transformations and ETL processes
- Solid understanding of data modeling principles and best practices
- 3+ years of experience coding in SQL(Azure SQL DB or MSSQL preferable), solid CS fundamentals including data structure and algorithm design
- 2+ years contributing to production deployments of large backend data processing and analysis systems
- Experience in data integration and ETL processes
- Proficient with Azure Data Services (Azure Data Factory, Azure Synapse Analytics, etc.)
- Strong problem-solving and analytical skills, with a keen attention to detail
- Excellent communication skills, both written and verbal, with the ability to present complex data in a clear and concise manner
- Experience in working with cross-functional teams and collaborating with stakeholders at various levels of the organization
- Knowledge of professional software engineering best practices for the full software
- Knowledge of Data Warehousing, design, implementation and optimization
- Knowledge of Data Quality testing, automation and results visualization
- Knowledge of the development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations
- Experience participating in an Agile software development team, e.g. SCRUM
- Experience designing, documenting, and defending designs for key components in large distributed computing systems
- A consistent track record of delivering exceptionally high-quality software on large, complex, cross-functional projects
- Demonstrated ability to learn new technologies quickly and independently
- Ability to handle multiple competing priorities in a fast-paced environment
Desirable
- Experience with Power BI Embedded
- Experience with Power BI REST or JS APIs
- Power BI Premium Capacity Management
- Experience in backend development and deployment
- Experience in CI/CD configuration
- Experience with Azure DevOps for CICD
- Experience with Tabular Editor, DAX Studio, SQL Profiler
- Experience with Powershell
- Experience with Swagger, Postman
Інформація про компанію Ciklum
Переваги співробітникам
- Team buildings
- Англомовне середовище
- Бухгалтерський супровід
- Компенсація домашнього офісу
- Надається ноутбук
- Освітні програми, курси
- Salesforce
- SQL
- Python
- AWS Glue
- Transact-SQL
We are looking for a Senior Data Engineer to join one of our data migration projects for a leader in global information technology research and consulting. We are looking for a candidate with at least 7 years of hands-on experience in Salesforce, SQL & Data migration. Strong technical skills, excellent communication, and adaptability are essential for success in this dynamic role.
Your role
- Migrate and integrate data from different platforms into Salesforce
- Develop procedures and scripts for data migration
- Develop, implement, and optimize stored procedures and functions using TSQL
- Prepare data migration plan and handle kick outs/fallouts
- Perform SQL database partitioning and indexing procedures as required to handle heavy traffic loads
Your skills
- Strong understanding of Salesforce objects (such as accounts, contacts, cases, etc), custom objects, fields & restrictions
- Understanding of Salesforce architecture
- Experience in designing, creating, and maintaining databases
- Hands-on experience in data migration & integration from different platforms into Salesforce
- Ability to create fast and efficient database queries, including joins with several tables
- Good knowledge of SQL optimization techniques
- Previous work with Python
- Experience in AWS GLUE
- Good written and verbal English skills
Інформація про компанію Sii Ukraine
Переваги співробітникам
- English Courses
- Гнучкий графік роботи
- Довгострокові проекти
- Регулярний перегляд зарплатні
Сторінки
Читайте нас в Telegram, щоб не пропустити анонси нових курсів.