Data Engineer building and maintaining data infrastructure for AI life science products. Collaborating with cross-functional teams to deliver impactful solutions while participating in Agile processes.
Responsibilities
Develop, maintain and update ETL catalog.
Build, maintain and update repeatable and trackable data pipelines.
Collect, analyze and organize raw data in collections of datasets.
Design databases and data stores for big datasets with performance considerations in mind.
Collaborate with engineers and data scientists to develop and productize prototypes.
Participate in architectural decisions regarding our data architecture.
Participate in our Agile processes like planning and daily stand-up meetings.
Requirements
3+ years of professional experience as a data engineer or in a similar role.
Proven experience with Relational and NoSQL data stores.
Proven experience with data models, data mining, and segmentation techniques.
Experience with big data tools, specifically Spark (preferably with Databricks).
Experience with data pipeline / workflow orchestration tools (Azure DataFactory, Prefect).
Experience with cloud compute frameworks like Azure Batch.
Experience with Python.
Experience with SQL query authoring.
Experience with Git and collaborative development workflows.
Experience with data manipulation and transformation using Pandas and/or Polars.
Experience working with cross-functional teams in agile environments.
Excellent verbal and written communication in English.
Ability to provide clear and concise step-by-step technical help, verbally and in writing.
Ability to interact effectively with audiences of varying technical backgrounds & seniority.
BS/MS degree in Computer Science, Information Technology or a related field (nice to have).
Familiarity with Atlassian stack of tools(Jira, Jira service management & Confluence)(nice to have).
Familiarity with microservices-based architectures (nice to have).
Familiarity with bioinformatics or life sciences data (nice to have).
Benefits
Competitive compensation packages based on qualifications.
Flexible work schedule.
Professional and personal development opportunities.
Private life & health insurance.
Room to experiment, learn and have fun.
Peers with big smiles and fascinating ideas.
A multi-disciplinary, multinational team that values trust and autonomy.
Data Engineer Tech Lead developing data solutions at Carelon. Leading a cross - functional team to optimize data workflows and maintain data integrity.
Lead Data Engineer responsible for evolving Manna’s data infrastructure for drone delivery. Overseeing data architecture and analytics while building scalable data pipelines.
Data Engineer designing, implementing, and optimizing data pipelines for DeepLight AI. Collaborating closely with a multidisciplinary team to analyze large - scale data.
Data Engineer designing and maintaining scalable ETL pipelines at Satori Analytics. Collaborating with teams to deliver high - quality analytics solutions across various industries.
Data Architect responsible for defining enterprise data architecture on AWS and Databricks Lakehouse platforms. Enabling scalable data lakes and enterprise analytics for financial services organizations.
Data Platform Operations Support leading data engineering strategy across projects for EXL. Driving innovation and optimization while collaborating with various teams in the organization.
Manager II leading data engineering projects at Navy Federal Credit Union. Overseeing data governance and quality initiatives while managing engineering teams in a hybrid work environment.
Senior Data Engineer building and maintaining data pipelines for cloud and AI solutions at Qodea. Collaborating with ML engineers and focusing on reliability and performance in a cloud - native environment.
Principal Data Engineer responsible for architecting scalable data pipelines and building high - quality data foundations. Collaborating closely with experts to ensure data readiness for advanced analytics.
Senior Data Engineer at Qodea designing scalable data pipelines and infrastructure. Delivering solutions utilizing cutting - edge tools and collaborating closely with teams for impactful results.