Data Engineer responsible for driving technical implementation of data products on Azure Databricks. Collaborates with teams in Midland, MI or Houston, TX for critical data initiatives.
Responsibilities
Collaborate with senior data engineers to translate complex business requirements and ambiguous problem statements into clear, robust, and scalable technical designs and data models (e.g., dimensional modeling, star schemas), and independently drive the implementation of these designs.
Design, build, and deploy high-volume data transformation logic using highly optimized PySpark.
Contribute significantly to the design and improvement of CI/CD pipelines in Azure DevOps/Git, ensuring reliable, automated, and secure deployment of data solutions across environments.
Deeply understand and connect to various source systems, demonstrating proficiency in managing data persistence and query performance across diverse technologies like SQL Server, Neo4j, and CosmosDB.
Proactively implement and maintain advanced data quality frameworks (e.g., Delta Live Tables, Great Expectations) and monitoring solutions to ensure data reliability for mission-critical applications.
Serve as a go-to technical resource for peers, conducting technical code reviews and informally mentoring Associate Data Engineers on PySpark and Databricks best practices.
Requirements
A minimum of a bachelor’s degree or relevant military experience at or above a U.S. E5 ranking or Canadian Petty Officer 2nd Class or Sergeant OR 5 years relevant experience in lieu of a Bachelor’s degree.
Minimum of 2 years of professional experience in Data Engineering, Software Engineering, or a closely related field.
A minimum requirement for this U.S. based position is the ability to work legally in the United States.
No visa sponsorship/support is available for this position, including for any type of U.S. permanent residency (green card) process.
Proven ability to write highly optimized, production-grade PySpark/Spark code.
Experience identifying and resolving performance bottlenecks in a distributed computing environment.
Practical experience designing and implementing analytical data models (e.g., dimensional modeling, star/snowflake schemas) and handling Slowly Changing Dimensions (SCDs).
Expertise in using Azure Data Factory (ADF), Databricks Workflows, or equivalent tools (e.g., Airflow) for complex dependency management, error handling, and end-to-end pipeline orchestration.
Demonstrated experience with advanced SQL and hands-on experience querying and integrating data from at least one non-relational or Graph database (e.g., CosmosDB, Neo4j).
Benefits
Equitable and market-competitive base pay and bonus opportunity across our global markets, along with locally relevant incentives.
Benefits and programs to support your physical, mental, financial, and social well-being, to help you get the care you need...when you need it.
Competitive retirement program that may include company-provided benefits, savings opportunities, financial planning, and educational resources to help you achieve your long term financial-goals.
Employee stock purchase programs (availability varies depending on location).
Student Debt Retirement Savings Match Program (U.S. only).
Robust medical and life insurance packages that offer a variety of coverage options to meet your individual needs.
Opportunities to learn and grow through training and mentoring, work experiences, community involvement and team building.
Competitive yearly vacation allowance.
Paid time off for new parents (birthing and non-birthing, including adoptive and foster parents).
Paid time off to care for family members who are sick or injured.
Paid time off to support volunteering and Employee Resource Group’s (ERG) participation.
Wellbeing Portal for all Dow employees, our one-stop shop to promote wellbeing, empowering employees to take ownership of their entire wellbeing journey.
On-site fitness facilities to help stay healthy and active (availability varies depending on location).
Senior Data Engineer at Goodwin enhancing data platforms and fostering data - driven culture across teams. Collaborating with IT and Finance on technology solutions and data governance practices.
Director, Data Platform Design and Strategy at MedImpact leading data platform and AI innovations to enhance healthcare services. Overseeing enterprise projects and managing teams to meet strategic goals.
Data Engineer delivering AI - and data - driven solutions for Honeywell’s industrial customers. Architecting and implementing scalable data pipelines and platforms focused on IoT and real - time data processing.
Data Engineering Associate focusing on data quality control and management for distribution platform. Collaborates on large scale data projects to ensure data accuracy and availability for users.
Data Architect managing enterprise data platform built on Microsoft Fabric at Johnstone Supply. Leading architectural standards and collaborating with business and IT leaders for strategic data - driven insights.
Data Engineer at Studyportals responsible for data pipelines and infrastructure. Join a team ensuring accurate and trustworthy data for analytics and business decisions.
AI/ML Engineer designing and refining prompts and workflows using large language models. Responsible for developing data pipelines and delivering scalable AI solutions in a hybrid work environment.
AWS Data Architect at Fractal designing and operationalizing AWS data solutions at enterprise scale. Collaborating with clients and mentoring engineers in best practices.
Senior Data Engineer driving data - driven success at Pacific Life. Collaborating with a team to build scalable and secure data solutions in Newport Beach, CA or Charlotte, NC.
Data Architect managing Commercial Data architecture initiatives for Valmet's sales and service team. Leading AI - driven data integrity and quality efforts in a global context.