Data Platform Engineer at Taxfix managing data infrastructures and pipelines to support analytics and AI-driven product features. Collaborating with cross-functional teams to ensure data reliability.
Responsibilities
Build and maintain ingestion pipelines that capture changes from application databases, APIs, SaaS and deliver clean, analytics-ready tables to our cloud data warehouse
Design data models with proper layering that handle real-world data complexity: out-of-order events, schema evolution, late arrivals, and backfills
Own and evolve cloud platform infrastructure — manage GCP resources (GCS, Dataflow, Dataproc), provision and maintain environments with Terraform, and ensure the platform is cost-efficient and scalable
Own data quality monitoring — build validation, monitoring, and alerting that catches problems before downstream consumers do
Implement privacy and compliance controls — anonymization, pseudonymization, access policies, and deletion propagation (GDPR right-to-be-forgotten) across raw and derived layers
Prepare data for ML and AI use cases — build governed, privacy-safe datasets and feature pipelines that ML engineers and data scientists can use for model training, evaluation, and production inference
Operate and improve our orchestration layer — scheduling, retries, SLA tracking, and observability for data pipelines
Define and raise the bar on engineering standards — code quality, testing, CI/CD, documentation, and infrastructure-as-code
Evaluate and adopt new technologies that help the team achieve its goals across data management, analytics, and machine learning
Incorporate AI into platform services — enable AI-assisted development workflows and build internal AI backend services as part of the data platform offering
Communicate across domains — work closely with analytics, product, compliance, and engineering teams; translate between technical and business language
Mentor and grow with the team — share what you learn, support others, and contribute to a culture of honest technical discussion
Requirements
6+ years of experience in Data Engineering or a similar role (backend engineer working on data-intensive systems counts)
Strong Python skills for data pipeline development — you write production code, not just scripts
Strong SQL skills — window functions, CTEs, query optimization are second nature
Experience with event-driven data pipelines — CQRS, event ordering, idempotency, and the difference between initial load and incremental processing
Expert with Airflow — you’ve built DAGs with proper task dependencies, retries, and monitoring
Experience with Snowflake or/and BigQuery — you understand their architecture, performance characteristics, and how they differ from each other and from other analytical or operational tools
Cloud platform experience — you’ve worked with GCP (GCS, Dataflow, Dataproc etc) or equivalent AWS/Azure services and understand how to manage cloud resources at scale
Infrastructure-as-code — experience with Terraform, Helm, or similar tools for provisioning and managing cloud environments
K8S and Docker containerization — you package and deploy your own work
Data quality mindset — you profile data, validate assumptions, build checks, and don’t trust that “the data looked clean”
Data for AI readiness — you understand what it takes to prepare data for ML and AI: governance, lineage, privacy controls, and reproducibility
Awareness of data privacy requirements — you can identify PII, understand GDPR, and know how to implement anonymization and deletion across multiple data layers
AI-enabled engineering practices — you actively use AI assistants and code generation tools to accelerate development and deliver and you can establish standards for their effective use across the team.
Benefits
A chance to do meaningful, people-centric work with an international team of passionate professionals.
Holistic well-being with free mental health coaching sessions and yoga.
A monthly allowance to spend on an extensive range of services that you can use and roll over as flexibly as you like.
Employee stock options for all employees—because everyone deserves to benefit from the success they help to create.
30 annual vacation days and flexible working hours.
Work from abroad for up to six weeks every year. Just align with your team, and then enjoy your trip.
Plenty of opportunities to socialise as a team. In addition to internal tech meetups, our international team hosts regular get-togethers—virtually and in person when possible.
Free tax declaration filing, of course, through the Taxfix app—and internal support for all personal tax-related questions.
Have a four-legged friend in your life? We’re happy to have dogs join us in the office.
Data Engineer focused on analytics and data pipeline development for network optimisation. Collaborating with teams to deliver high - quality data solutions with Python and SQL.
Senior Product Manager defining platform capabilities for Data Cloud in Salesforce. Collaborating with R&D teams while shaping product strategy for Data 360 integration.
Senior Data Engineer at Goodwin enhancing data platforms and fostering data - driven culture across teams. Collaborating with IT and Finance on technology solutions and data governance practices.
Director, Data Platform Design and Strategy at MedImpact leading data platform and AI innovations to enhance healthcare services. Overseeing enterprise projects and managing teams to meet strategic goals.
Data Engineer delivering AI - and data - driven solutions for Honeywell’s industrial customers. Architecting and implementing scalable data pipelines and platforms focused on IoT and real - time data processing.
Data Engineering Associate focusing on data quality control and management for distribution platform. Collaborates on large scale data projects to ensure data accuracy and availability for users.
Data Architect managing enterprise data platform built on Microsoft Fabric at Johnstone Supply. Leading architectural standards and collaborating with business and IT leaders for strategic data - driven insights.
Data Engineer at Studyportals responsible for data pipelines and infrastructure. Join a team ensuring accurate and trustworthy data for analytics and business decisions.
AI/ML Engineer designing and refining prompts and workflows using large language models. Responsible for developing data pipelines and delivering scalable AI solutions in a hybrid work environment.
AWS Data Architect at Fractal designing and operationalizing AWS data solutions at enterprise scale. Collaborating with clients and mentoring engineers in best practices.