Assistant Manager - SA2 (AI Hub GTS)
KPMG India.com
Office
Bangalore, Karnataka, India
Full Time
Roles & Responsibilities
·Design, develop, and implement data pipelines for scalable solutions using various technologies – Databricks or Microsoft Fabric ·Conduct Code reviews and recommend best coding practices ·Provide effort estimates to implement solutions, development effort for the proposed solutions ·Develop and maintain comprehensive documentation for all proposed solutions. models, including detailed technical specifications, results of testing and evaluation, and datasets used for developing the solutions ·Lead Architect and design efforts for Product development and application development for relevant use cases providing guidance and support to team members and clients ·Implement best practices of Data Engineering and Architectural solution design, development, Testing and documentation ·Participate in team meetings, brainstorming sessions, and project planning activities ·Stay up-to-date with the latest advancements in Data Engineering area, to drive innovation and maintain a competitive edge ·Stay hands-on with the design, development and validation of systems and models deployed ·Collaborate with audit professionals to understand business, regulatory and risk requirements and key alignment considerations for Audit ·Drive efforts in the Data Engineering and Architecture practice areaMandatory technical & functional skills
•Databricks/Fabric/Spark Notebooks •SQL/NoSQL databases, Redis, •Full-Stack Development - React, Angular •Data Management: Design, implement, and manage AI-driven data solutions on the Microsoft Azure cloud platform, ensuring scalability and performance. •Data Integration: Develop and maintain data pipelines for AI applications, ensuring efficient data extraction, transformation, and loading (ETL) processes using Azure Data Factory. •Big Data Processing: Utilize big data technologies such as Azure Databricks and Apache Spark to handle, analyze, and process large datasets for machine learning and AI applications. •machine learning frameworks -TensorFlow, PyTorch •Backend frameworks - FastAPI, DjangoPreferred technical & functional skills
•Experience working with frameworks LangChain/LlamaIndex /LlamaPrase/LlamaCloud/Semantic Kernel etc. •Develop real-time data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark (SQL, Scala, Java), Python and Hadoop Platform and any Cloud Data Platform. •Certifications: Relevant certifications such as Microsoft Certified: AI 102, DP 700, DP 900 or AWS certiicationsKey Behavioral Attributes/Requirements
·Strong analytical, problem-solving, and critical-thinking skills ·Excellent collaboration skills, with the ability to work effectively in a team-oriented environment ·Excellent written and verbal communication skills, with the ability to present complex technical concepts to non-technical audiences ·Willingness to learn new technologies and work on themRoles & Responsibilities
·Design, develop, and implement data pipelines for scalable solutions using various technologies – Databricks or Microsoft Fabric ·Conduct Code reviews and recommend best coding practices ·Provide effort estimates to implement solutions, development effort for the proposed solutions ·Develop and maintain comprehensive documentation for all proposed solutions. models, including detailed technical specifications, results of testing and evaluation, and datasets used for developing the solutions ·Lead Architect and design efforts for Product development and application development for relevant use cases providing guidance and support to team members and clients ·Implement best practices of Data Engineering and Architectural solution design, development, Testing and documentation ·Participate in team meetings, brainstorming sessions, and project planning activities ·Stay up-to-date with the latest advancements in Data Engineering area, to drive innovation and maintain a competitive edge ·Stay hands-on with the design, development and validation of systems and models deployed ·Collaborate with audit professionals to understand business, regulatory and risk requirements and key alignment considerations for Audit ·Drive efforts in the Data Engineering and Architecture practice areaMandatory technical & functional skills
•Databricks/Fabric/Spark Notebooks •SQL/NoSQL databases, Redis, •Full-Stack Development - React, Angular •Data Management: Design, implement, and manage AI-driven data solutions on the Microsoft Azure cloud platform, ensuring scalability and performance. •Data Integration: Develop and maintain data pipelines for AI applications, ensuring efficient data extraction, transformation, and loading (ETL) processes using Azure Data Factory. •Big Data Processing: Utilize big data technologies such as Azure Databricks and Apache Spark to handle, analyze, and process large datasets for machine learning and AI applications. •machine learning frameworks -TensorFlow, PyTorch •Backend frameworks - FastAPI, DjangoPreferred technical & functional skills
•Experience working with frameworks LangChain/LlamaIndex /LlamaPrase/LlamaCloud/Semantic Kernel etc. •Develop real-time data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark (SQL, Scala, Java), Python and Hadoop Platform and any Cloud Data Platform. •Certifications: Relevant certifications such as Microsoft Certified: AI 102, DP 700, DP 900 or AWS certiicationsKey Behavioral Attributes/Requirements
·Strong analytical, problem-solving, and critical-thinking skills ·Excellent collaboration skills, with the ability to work effectively in a team-oriented environment ·Excellent written and verbal communication skills, with the ability to present complex technical concepts to non-technical audiences ·Willingness to learn new technologies and work on themThis role is for you if you have the below
Educational Qualifications
•Minimum qualification required: BTech in Computer Science/ MTech/ MCA - Fulltime education.Work Experience
•6 to 8 years of experience in design, develop data centric applications using various tools and technologies eg: Databases,reporting,ETL/ELT,NoSQL etc. •5+ years of experience in designing, architecting solutions using Microsoft Data technologies like ADF/SYNAPSE •Relevant Data Professional certifications – AWS, GCP or AzureAssistant Manager - SA2 (AI Hub GTS)
Office
Bangalore, Karnataka, India
Full Time
August 20, 2025