Data Engineer leading data foundation architecture and optimization for a Kenyan startup. Constructing data pipelines that fuel machine learning models and internal analytics.
Responsibilities
Architect and sustain scalable ETL workflows, guaranteeing consistency and accuracy across diverse data origins.
Refine and optimize data models and database structures specifically tailored for reporting and analytics.
Enforce industry best practices regarding data warehousing and storage methodologies.
Fine-tune data systems to handle the demands of both real-time streams and batch processing.
Manage the cloud data environment, utilizing platforms such as AWS, Azure, or GCP.
Coordinate with software engineers to embed data solutions directly into our product suite.
Design robust processes for ingesting both structured and unstructured datasets.
Script automated quality checks and deploy monitoring instrumentation to instantly detect data anomalies.
Build APIs and services that ensure seamless data interoperability between systems.
Continuously monitor pipeline health, troubleshooting bottlenecks to maintain an uninterrupted data flow.
Embed data governance and security protocols that meet rigorous industry standards.
Collaborate with data scientists and analysts to maximize the usability and accessibility of our data assets.
Maintain comprehensive documentation covering schemas, transformations, and pipeline architecture.
Keep a pulse on emerging trends in cloud tech, analytics, and data engineering to drive continuous improvement.
Requirements
A minimum of 3 years of professional experience in Data Engineering or a similar technical role.
Bachelor’s or Master’s degree in Engineering, Computer Science, Data Science, or a relevant discipline.
Expert-level command of SQL and management systems like PostgreSQL or MySQL.
Hands-on proficiency with pipeline tools such as Luigi, DBT, or Apache Airflow.
Practical experience with heavy-lifting technologies like Hadoop, Spark, or Kafka.
Proven skills with cloud data stacks, specifically Google BigQuery, AWS Redshift, or Azure Data Factory.
Strong programming logic in Java, Scala, or Python for data processing tasks.
Familiarity with data integration frameworks and API utilization.
Understanding of security best practices and compliance frameworks.
Senior Data Engineer at Goodwin enhancing data platforms and fostering data - driven culture across teams. Collaborating with IT and Finance on technology solutions and data governance practices.
Director, Data Platform Design and Strategy at MedImpact leading data platform and AI innovations to enhance healthcare services. Overseeing enterprise projects and managing teams to meet strategic goals.
Data Engineer delivering AI - and data - driven solutions for Honeywell’s industrial customers. Architecting and implementing scalable data pipelines and platforms focused on IoT and real - time data processing.
Data Engineering Associate focusing on data quality control and management for distribution platform. Collaborates on large scale data projects to ensure data accuracy and availability for users.
Data Architect managing enterprise data platform built on Microsoft Fabric at Johnstone Supply. Leading architectural standards and collaborating with business and IT leaders for strategic data - driven insights.
Data Engineer at Studyportals responsible for data pipelines and infrastructure. Join a team ensuring accurate and trustworthy data for analytics and business decisions.
AI/ML Engineer designing and refining prompts and workflows using large language models. Responsible for developing data pipelines and delivering scalable AI solutions in a hybrid work environment.
AWS Data Architect at Fractal designing and operationalizing AWS data solutions at enterprise scale. Collaborating with clients and mentoring engineers in best practices.
Senior Data Engineer driving data - driven success at Pacific Life. Collaborating with a team to build scalable and secure data solutions in Newport Beach, CA or Charlotte, NC.
Data Architect managing Commercial Data architecture initiatives for Valmet's sales and service team. Leading AI - driven data integrity and quality efforts in a global context.