Specialist in Data Engineering leading pipeline optimization at Inmetrics. Collaborating in innovative data-driven projects within a hybrid work environment.
Responsibilities
Lead the development and optimization of data pipelines, architectures, and platforms.
Ensure high performance, governance, and availability of information across the company's ecosystem.
Work in a collaborative, innovative, and results-oriented environment where data is central to the strategy and the driver of intelligent, sustainable decisions.
Collaborate with the Card team's Data Engineering group to create data pipelines for ingestion and provisioning of card-domain data into Santander Brazil's corporate data lake.
Master and promote institution-wide concepts, tools, and technologies related to Data Analysis.
Technically structure and update the current Guidelines and Policies for the Data Analysis area.
Participate in defining principles for control, management, and specification of various data methods and models.
Assist in designing, structuring, and optimizing databases.
Requirements
Databricks skills: Experience working with Apache Spark on Databricks, including building and optimizing data pipelines.
Experience with PySpark, Python, and Kedro: Strong programming skills in PySpark, Python, and Kedro to develop, debug, and maintain data transformation code.
Batch and streaming data processing: Knowledge of batch and streaming (messaging) data processing, with the ability to design, implement, and maintain data processing pipelines.
DevOps knowledge: Familiarity with using Jenkins for continuous integration and delivery (CI/CD), as well as automation of deployment tasks and pipeline management.
Git: Proficiency in Git for source code version control and effective collaboration in development teams.
Agile methods: Understanding of principles and practices of agile methodologies, such as Kanban and Scrum, for effective collaboration and project management.
Orchestration (e.g., Control-M or others): Knowledge of workflow orchestration tools, important for scheduling and controlling workflows.
Microsoft Azure knowledge: Experience with key Microsoft Azure data services, including Azure Databricks, Azure Data Factory, and Azure Storage Accounts.
Experience in on-premises environments (Cloudera): Previous experience with the Cloudera platform or other on-premises big data solutions, including Hadoop, HBase, and Hive, is desirable.
Object-oriented development knowledge: Familiarity with Java is helpful (not required to code, but to interpret).
Optional certifications: AZ-900 (Microsoft Azure Fundamentals) and DP-900 (Microsoft Azure Data Fundamentals) certifications are preferred and demonstrate solid knowledge of the Azure platform and data concepts.
Benefits
Bradesco Health Plan (30% copay)
Bradesco Dental Plan (no employee contribution)
Life Insurance
Wellhub (Gympass)
Childcare assistance
Assistance for children with special needs
Payroll-deductible loan
Private pension plan
Pet benefit plan
SESC membership
Conexa telemedicine
Expense allowance
Meal and food vouchers
Multi-benefit card
Medical plan upgrade
Extended maternity and paternity leave
Support program for pregnant employees
Newborn gift basket and the book "Acontecia quando eu nascia"
Professional development: courses available through the internal university
100% remote or hybrid work, depending on project requirements.
Data Engineer focused on analytics and data pipeline development for network optimisation. Collaborating with teams to deliver high - quality data solutions with Python and SQL.
Senior Product Manager defining platform capabilities for Data Cloud in Salesforce. Collaborating with R&D teams while shaping product strategy for Data 360 integration.
Senior Data Engineer at Goodwin enhancing data platforms and fostering data - driven culture across teams. Collaborating with IT and Finance on technology solutions and data governance practices.
Director, Data Platform Design and Strategy at MedImpact leading data platform and AI innovations to enhance healthcare services. Overseeing enterprise projects and managing teams to meet strategic goals.
Data Engineer delivering AI - and data - driven solutions for Honeywell’s industrial customers. Architecting and implementing scalable data pipelines and platforms focused on IoT and real - time data processing.
Data Engineering Associate focusing on data quality control and management for distribution platform. Collaborates on large scale data projects to ensure data accuracy and availability for users.
Data Architect managing enterprise data platform built on Microsoft Fabric at Johnstone Supply. Leading architectural standards and collaborating with business and IT leaders for strategic data - driven insights.
Data Engineer at Studyportals responsible for data pipelines and infrastructure. Join a team ensuring accurate and trustworthy data for analytics and business decisions.
AI/ML Engineer designing and refining prompts and workflows using large language models. Responsible for developing data pipelines and delivering scalable AI solutions in a hybrid work environment.
AWS Data Architect at Fractal designing and operationalizing AWS data solutions at enterprise scale. Collaborating with clients and mentoring engineers in best practices.