AI Engineer enhancing data processing with generative AI solutions. Working with the Azure ecosystem to automate and optimize data workflows.
Responsibilities
Design and implement GenAI-based solutions that optimize critical stages of the DDLC (Data Development Life Cycle).
Develop agents that act as assistants in creating Databricks notebooks, pipelines in ADF (Azure Data Factory), and structures in Microsoft Fabric.
Build workflows that integrate LLMs into our ecosystem (SQL Server and Power BI), automating tasks such as data documentation and creation of DAX measures.
Apply advanced prompt engineering techniques to ensure the AI generates optimized, secure, and high-performance PySpark and T-SQL code.
Define quality metrics to validate AI-generated artifacts (e.g., ensuring the efficiency of generated queries).
Ensure solutions align with Data Governance practices, security, and privacy.
Stay up to date on AI capabilities within Microsoft Fabric and Databricks to maximize the use of native tools versus custom developments.
Requirements
Degree in Computer Science, Data Engineering, Statistics, or a related field.
Strong experience in Data Engineering and a deep understanding of CI/CD pipelines applied to data.
Required experience with AI agent orchestration frameworks (e.g., LangChain, Semantic Kernel, or AutoGen).
Technical expertise in the Azure Data ecosystem (Databricks, Data Factory, and Fabric) and in SQL and Python/PySpark.
Hands-on experience building GenAI solutions (OpenAI API, Azure OpenAI, or open-source models) with a strong focus on prompt engineering is essential.
Ability to translate business challenges into agents that can assist in data modeling and dashboard creation.
DataOps mindset: understanding that data requires testing, versioning, and monitoring.
Benefits
Health and dental insurance;
Meal and food allowance;
Childcare assistance;
Extended parental leave;
Partnerships with gyms and health & wellness professionals via Wellhub (Gympass) / TotalPass;
Profit Sharing (PLR);
Life insurance;
Continuous learning platform (CI&T University);
Discount club;
Free online platform dedicated to physical and mental health and wellbeing;
Data Management professional at Kyndryl involved in creating innovative data solutions and ensuring the seamless operation of complex data systems. Collaborating with teams to transform requirements into scalable database solutions.
Software Engineer designing and developing scalable data processing applications on cloud infrastructure for Thomson Reuters. Collaborating with Data Analysts on AI - enabled solutions for data management and insight generation.
Manager of Data Platform overseeing AWS cloud infrastructure and Snowflake data warehouses for Thomson Reuters. Leading the design and implementation of data processing applications in a hybrid role located in Bengaluru.
Senior Data Engineer designing scalable data pipelines and solutions for Enterprise Data Lake at Thomson Reuters. Collaborating across teams to ensure efficient data ingestion and accessibility.
Senior Data Engineer at Technis developing scalable data pipelines and solutions for innovative connected spaces products. Collaborating within a cross - functional team to deliver high - quality data - driven outcomes.
Data Architect designing and implementing data architectures supporting analytics and ML for federal clients. Collaborating with teams to translate mission needs into robust data solutions.
IT Data Engineer developing data pipelines and integrations for Scanfil Group's global IT organization. Collaborating across teams to enhance data solutions and reporting capabilities.
Data Engineer developing Azure data solutions at PwC New Zealand. Responsibilities include data quality monitoring, pipeline development, and collaboration with stakeholders in a supportive environment.
Senior Data Engineer designing and implementing the Enterprise Data Platform at Stellix. Focusing on analytics and insights with a growth path to Principal Data Engineer or Data Architect.