Collaborate with product owners and team leads to identify, design, and implement new features to support the growing data needs.
Build and maintain optimal architecture to extract, transform, and load data from a wide variety of data sources, including external APIs, data streams and data lakes.
Implement data privacy and data security requirements to ensure solutions stay compliant with security standards and frameworks.
Monitor and anticipate trends in data engineering and propose changes in alignment with organizational goals and needs.
Share knowledge with other teams on various data engineering or project-related topics.
Collaborate with the team to decide which tools and strategies to use within specific data integration scenarios.
What we expect from you
-
Commercial experience 5+ years in data engineering.
-
Strong programming skills in Python.
-
Solid with distributed computing approaches, patterns, and technologies (PySpark).
-
Experience working with any cloud platform (GCP, AWS, Azure) and its data-oriented components.
-
Proficiency in SQL and query tuning.
-
Understanding of data warehousing principles and modeling concepts (e.g., knowledge of data model types and terminology including OLTP/OLAP, SCD, (de)normalization, dimensional, star/snowflake modeling, etc.).
-
Expertise in the use of any listed relational databases (PostgreSQL, MSSQL or MySQL).
-
Experience with the orchestration of any data flows (e.g., Apache Airflow, Prefect, Glue, Azure Data Factory).
-
A team player with excellent collaboration skills.
-
Minimum English level B2.
Expertise in stream processing using the current industry standards (e.g., AWS Kinesis, Kafka streams, Spark/PySpark, etc.).
Expertise in data storage design principles. Understanding of pros and cons of SQL/NoSQL solutions, their types, and configurations (standalone/cluster, column/row-oriented, key-value/document stores).
Experience in modern data warehouse building using Snowflake, AWS Redshift or BigQuery.
Deep knowledge of Spark internals (tuning, query optimization).
Experience with data integration and business intelligence architecture.
Experience with data lakes and lake-houses (Azure Data Lake, Apache Hudi, Apache Iceberg, Delta Lake).
Experience with containerized (Docker, ECS, Kubernetes) or serverless (Lambda) deployment.
Good knowledge of popular data standards and formats (e.g., JSON, XML, Proto, Parquet, Avro, ORC, etc.).
Experience with Platforms: Informatica, Databricks, Talend, Fivetran or similar.
Experience in data science and machine learning with building Machine Learning models.
NIX is a global supplier of software engineering and IT outsourcing services
NIX teams collaborate with partners from different countries. Our specialists have experience in developing innovative projects from ecommerce to cloud for some of the largest companies in the world, including from the Fortune 500. The teams are focused on stable development of the international IT market, business, and their own professional skills.
What we offer
-
Competitive compensation packages.
-
Stable employment, based on a full-time employment contract.
-
Private health insurance (Medicover Сlinic).
-
AYCM sport pass, providing discounts at various sports facilities in Hungary.
-
Interesting tasks and diverse opportunities for developing your skills.
-
Free training courses, including English.
-
Participation in internal and external thematic events, technical conferences.
-
A spacious office in the heart of Budapest (13th district).
-
All necessary devices and tools for your work.
-
Active corporate life.