Junior Data Engineer role at Allegro, focusing on developing ETL/ELT pipelines and processing large datasets. Collaborate with cross-functional teams for data quality and reporting.
Responsibilities
You will develop and maintain ETL/ELT data pipelines processing massive datasets from the Allegro platform.
You will expand your expertise in the Google Cloud ecosystem, working with tools like BigQuery and Cloud Composer (Airflow)
You will be part of a cross-functional team, collaborating closely with Data Scientists and Analysts to provide high-quality, structured data for ML models and business reporting.
You will verify data quality and implement basic monitoring solutions to ensure the reliability of our data warehouse.
You will optimize existing data processes, write clean code, and learn best practices in DataOps and software engineering.
You will be part of the Infrastructure department, learning how we manage and analyze vast amounts of technical data across our data centers and the public cloud.
You will be responsible for assisting in building a seamless data process flow, cooperating with various teams to understand existing metrics and company targets.
Requirements
Are familiar with Python (or Java/Scala) and know how to write clean, maintainable code.
Are familiar with the basics of SQL and want to learn how to use it in practice for processing large-scale technical datasets.
Like analyses and can apply critical thinking to identify basic trends and patterns, treating complex data as a puzzle waiting to be solved.
Are eager to continuously develop your skills and expand knowledge in a highly technical infrastructure environment, tackling the challenges of large-scale datasets in one of Europe's leading e-commerce platforms.
Know English at B2+ level
Benefits
Flexible working hours in the hybrid model (4/1) - working hours start between 7:00 a.m. and 10:00 a.m.
The opportunity to learn, work on exciting challenges, collaborate with amazing people and have an unforgettable adventure
Mentorship and support from your buddy throughout your entire program
Additionally, you will be part of a supportive, inclusive culture that fosters personal growth, career development, and the building of meaningful connections with colleagues
A wide selection of fringe benefits in a cafeteria plan - you choose what you like (e.g. medical, sports or lunch packages, insurance, purchase vouchers)
The necessary tools for work
Working in a team you can always count on - we have on board top-class specialists and experts to learn from
Hackathons/Open days, workshops, guilds, meetups and internal knowledge sharing
Internal learning platform (including training courses on work organization, means of communication, motivation to work and various technologies and subject-matter issues)
Job title
Junior Data Engineer – e-Xperience Program Associate
Technical Lead for data engineering and reporting in healthcare technology at Dedalus. Shaping innovative software solutions and leading cross - functional technical teams in Australia.
Senior ML Data Engineer working on data pipeline curation for Mobileye's autonomous vehicle dataset. Collaborating across teams to enhance ML engineering and vision model applications.
Data Engineer managing customer datasets to enhance industrial research and development. Responsible for ETL pipelines and data ingestion for the Uncountable Web Platform.
Data Engineer designing and maintaining scalable data solutions on Databricks for clinical trials. Collaborating with teams to overcome data challenges and ensure the smooth logistics of clinical supplies.
Senior Manager leading a team of database engineers to manage CCC's data platform. Overseeing mission - critical applications and collaborating with cross - functional teams in a hybrid environment.
As a Principal Data Architect at Solstice, lead the design and implementation of data architecture solutions. Ensure data integrity, security, and accessibility to meet strategic organizational goals.
Data Platform Specialist overseeing data workflows and enhancing data quality for Stackgini's AI - driven IT solutions. Collaborating with teams to drive improvements and stakeholder support.
Data Engineer designing data pipelines in Python for a major railway industry client. Collaborate with Data Scientists and ensure code quality with agile methodologies.
Senior Data Engineer responsible for building and optimizing data pipelines for banking analytics initiatives. Collaborating with data teams to ensure data quality and readiness for enterprise use.
Senior Data Engineer developing scalable data solutions on Databricks for analytics and operational workloads. Collaborating with cross - functional teams to modernize the data ecosystem.