Role Summary :
The role will entail managing a team of data engineers responsible for implementing machine learningmodels and creating highly resilient data pipelines. It will also entail leading various projects includingmigration activities and POCs.
Job Description :
- Manage big data initiatives for the Data Science and AI Group while working closely with the technology team, vendors, and consultants.
- Coordinate with business and technology stakeholders to ensure that requirements are implemented as expected.
- Facilitate implementation of analytical tools.
- Recommend and maintain the data model that will support business intelligence, advanced analytics, and campaign management initiatives.
- Implement enhancements in the existing MLOps pipeline.
- Establish data quality checks to identify integrity issues and report findings to technology management for resolutions.
- Establish and implement data governance guidelines to ensure delivery of high quality and secure data to meet regulatory requirements and promote efficiency and revenue growth.
- Establish a catalog of commonly used data fields to aid data users in exploring the banks data, improve the understanding of data across the bank, and promote consistency of KPI definition.
- Conduct best practice research to continuously improve current data management and analytic processes and ensure data management and analytics tools / processes are at par with global standards.
- Conduct POC on new solutions.
Soft Skills Required :
Leadership : Experience in mentoring and coaching team members, fostering a culture of growth and improvement, and proactively learning new tools, technologies, and best practices to drive team success.Strong Communication & Collaboration : Proven ability to build relationships across functions, communicate effectively with both business and technical stakeholders, and manage conflicts through proactive stakeholder engagement.Analytical & Strategic Thinking : Demonstrates sound decision-making by involving the right stakeholders, understanding interdependencies, and planning accordingly with a big-picture mindset.Technical Skills Required :
Proficient in data processing tools (SQL, Python, R).Strong knowledge of batch and stream data processing techniques (Spark, Kafka).Strong knowledge in MLOps and related concepts.Experienced in data modelling and data warehousing techniques.Experienced in using Git for version control and CI / CD.Experienced in using business intelligence tools (Power BI, Tableau).Preferably with experience in Cloudera and Snowflake.Preferably with experience in the banking and financial services industry.#J-18808-Ljbffr