Senior Data Engineer delivering scalable data solutions in data engineering team at fintech startup. Building and maintaining data pipelines, collaborating with cross-functional teams for accurate data delivery.
Responsibilities
Design, build, and maintain scalable data pipelines and infrastructure across batch and streaming systems.
Own core components of Imprint’s data stack, including Snowflake, dbt Cloud, Change Data Capture frameworks, and reverse ETL integrations.
Develop and enforce best practices in data modeling, testing, observability, and governance.
Partner with stakeholders across Product, Analytics, Finance, and Engineering to ensure timely and accurate data delivery.
Work on external data integrations, such as partner-facing data shares (e.g., S3, SFTP, Snowflake) and financial reporting pipelines (e.g., with Netsuite).
Contribute to architectural decisions for how we scale data infrastructure, including schema design, orchestration, and data lineage.
Champion clear documentation, reproducibility, and reliability for critical datasets and workflows.
Stay informed about modern data tools and trends and help drive their adoption when appropriate.
Requirements
6+ years of experience in data engineering, analytics engineering, or related roles.
Expertise in Snowflake and dbt Cloud, with a strong understanding of dimensional modeling and data warehouse best practices.
Experience working with Change Data Capture (e.g., Fivetran, Hevo), ETL/ELT pipelines, and orchestration frameworks (e.g., dbt Cloud, Airflow).
Familiarity with reverse ETL tools like Hightouch or Segment, and operational analytics use cases.
Strong SQL skills and proficiency in Python or a similar programming language.
A track record of technical ownership and shipping production-grade data systems.
A detail-oriented mindset and a passion for building clean, maintainable, and observable data systems.
Strong communication skills and the ability to collaborate effectively with cross-functional partners.
Benefits
Competitive compensation and equity packages
Leading configured work computers of your choice
Flexible paid time off
Fully covered, high-quality healthcare, including fully covered dependent coverage
Additional health coverage includes access to One Medical and the option to enroll in an FSA
20 weeks of paid parental leave for the primary caregiver and 8 weeks for all new parents
Access to industry-leading technology across all of our business units, stemming from our philosophy that we should invest in resources for our team that foster innovation, optimization, and productivity
Data Engineering Intern at Efficy supporting data management and ETL pipeline development. Collaborate with teams and contribute to the enhancement of data architecture.
Senior Data Engineer building and optimizing data pipelines for Garner Health. Seeking a candidate with experience in AWS, SQL, and Python with a mission - driven mindset.
Data Engineer (GCP) designing and maintaining scalable data platforms at LUZA Group in Portugal. Collaborating and ensuring data integrity across multiple complex datasets.
Data Architect at Integrant responsible for designing and building data solutions for analytical purposes. Involves eliciting requirements, data pipelines, and coaching teams on methodologies.
Senior Data Engineer developing and maintaining data pipelines for clients in an Agile setting. Collaborating with teams to enhance data quality and mentoring junior engineers.
Senior Data Engineer designing and maintaining scalable data pipelines using modern technologies. Collaborating with cross - functional teams and providing mentorship in a dynamic environment.
Data Architect leading design and implementation of cloud data platforms for digital transformation. Collaborating with stakeholders to define data strategies and governance models.
Data Engineer Consultant designing and optimizing data infrastructure for clients' business needs. Working with SQL and data visualization tools in a mainly remote role with some onsite responsibilities in Denver.
Data Engineer creating Real - Time Data Processing applications for a leading iGaming operator. Work involves stream data manipulation and collaboration in an Agile environment.