Data Engineer building and operating data pipelines for Qloo's platform. Collaborating with teams on data integrity and accessibility processes.
Responsibilities
Design, develop, and maintain batch data pipelines using Python, Spark (EMR), and AWS Glue, loading data from S3, RDS, and external sources into Hive/Athena tables.
Model datasets in our S3/Hive data lake to support analytics (Hex), API use cases, Elasticsearch indexes, and ML models.
Implement and operate workflows in Airflow (MWAA), including dependency management, scheduling, retries, and alerting via Slack.
Build robust data quality and validation checks (schema validation, freshness/volume checks, anomaly detection) and ensure issues are surfaced quickly with monitoring and alerts.
Optimize jobs for cost and performance (partitioning, file formats, join strategies, proper use of EMR/Glue resources).
Collaborate closely with data scientists, ML engineers, and application engineers to understand data requirements and design schemas and pipelines that serve multiple use cases.
Contribute to internal tooling and shared libraries that make working with our data platform faster, safer, and more consistent.
Document pipelines, datasets, and best practices so the broader team can easily understand and work with our data.
Requirements
Bachelor’s degree in Computer Science, Software Engineering, or a related field, or equivalent practical experience.
Experience with Python and distributed data processing using Spark (PySpark) on EMR or a similar environment.
Hands-on experience with core AWS data services, ideally including:
• S3 (data lake, partitioning, lifecycle management)
• AWS Glue (jobs, crawlers, catalogs)
• EMR or other managed Spark platforms
• Athena/Hive and SQL for querying large datasets
• Relational databases such as RDS (PostgreSQL/MySQL or similar)
Experience building and operating workflows in Airflow (MWAA experience is a plus).
Strong SQL skills and familiarity with data modeling concepts for analytics and APIs.
Solid understanding of data quality practices (testing, validation frameworks, monitoring/observability).
Comfortable working in a collaborative environment, managing multiple projects, and owning systems end-to-end.
Benefits
Competitive salary and benefits package, including health insurance, retirement plan, and paid time off.
The opportunity to shape a modern cloud-based data platform that powers real products and ML experiences.
A collaborative, low-ego work environment where your ideas are valued and your contributions are visible.
Flexible work arrangements (remote and hybrid options) and a healthy respect for work-life balance.
Senior Data Engineer at Technis developing scalable data pipelines and solutions for innovative connected spaces products. Collaborating within a cross - functional team to deliver high - quality data - driven outcomes.
Data Architect designing and implementing data architectures supporting analytics and ML for federal clients. Collaborating with teams to translate mission needs into robust data solutions.
IT Data Engineer developing data pipelines and integrations for Scanfil Group's global IT organization. Collaborating across teams to enhance data solutions and reporting capabilities.
Data Engineer developing Azure data solutions at PwC New Zealand. Responsibilities include data quality monitoring, pipeline development, and collaboration with stakeholders in a supportive environment.
Senior Data Engineer designing and implementing the Enterprise Data Platform at Stellix. Focusing on analytics and insights with a growth path to Principal Data Engineer or Data Architect.
R&D Data Engineer at DXC, transforming complex data into digital assets for global analytics and Smart Lab solutions. Collaborating on ELN and LIMS tools for enhanced data management.
Senior Data Engineer at mobility AI company designing large - scale data processing pipelines. Leading technical decisions and mentoring junior engineers in data architecture.
Data Engineer role focusing on data pipelines and processing at 42dot, a mobility AI company. Responsibilities include data collection, schema management, and pipeline monitoring.
Senior Data Engineer at Booz Allen building advanced tech solutions for mission - driven projects. Utilizing data engineering activities, pipelines, and platforms for impactful data insights.