KPMG Global Services logo
Engineer (A2 DES, Databricks, PySpark, Python)
full-timeBengaluru

Summary

Location

Bengaluru

Type

full-time

Explore Jobs

About this role

Roles & responsibilities

Role Overview: The Associate 2 - “Data Engineer with Databricks/Python skills” will be part of the GDC Technology Solutions (GTS) team, working in a technical role in the Audit Data & Analytics domain that requires developing expertise in KPMG proprietary D&A (Data and analytics)) tools and audit methodology. He/she will be a part of the team responsible for extracting and processing datasets from client ERP systems (SAP/Oracle/Microsoft Dynamics) or other sources to provide insights through data warehousing, ETL and dashboarding solutions to Audit/internal teams and be involved in developing solutions using a variety of tools & technologies

The Associate 2 - “Data Engineer” will be predominantly responsible for:

Data Engineering

·Understand requirements, validate assumptions, and develop solutions using Azure Databricks, Azure Data Factory or Python. Able to handle any data mapping changes and customizations within Databricks using PySpark
·Build Azure Databricks notebooks to perform data transformations, create tables, and ensure data quality and consistency. Leverage Unity Catalog for data governance and maintaining a unified data view across the organization
·Analyze enormous volumes of data using Azure Databricks and Apache Spark. Create pipelines and workflows to support data analytics, machine learning, and other data-driven applications
·Able to integrate Azure Databricks with ERP systems or third part systems using APIs and build Python or PySpark notebooks to apply business transformation logic as per the common data model
·Debug, optimize and performance tune and resolve issues, if any, with limited guidance, when processing large data sets and propose possible solutions
·Must have experience in concepts like Partitioning, optimization, and performance tuning for improving the performance of the process
·Implement best practices of Azure Databricks design, development, Testing and documentation
·Work with Audit engagement teams to interpret the results and provide meaningful audit insights from the reports
·Participate in team meetings, brainstorming sessions, and project planning activities
·Stay up-to-date with the latest advancements in Azure Databricks, Cloud and AI development, to drive innovation and maintain a competitive edge
·Enthusiastic to learn and use Azure AI services in business processes.
·Work experience on using Microsoft Fabric is an added advantage
·Write production ready code
·Design, develop, and maintain scalable and efficient data pipelines to process large datasets from various sources using Azure Data Factory (ADF).
·Integrate data from multiple data sources and ensure data consistency, quality, and accuracy, leveraging Azure Data Lake Storage (ADLS).
·Design and implement ETL (Extract, Transform, Load) processes to ensure seamless data flow across systems using Azure
·Work experience on Microsoft Fabric is an added advantage
·Enthusiastic to learn, adapt and integrate Gen AI into the business process and should have experience working with Azure AI services
·Optimize data storage and retrieval processes to enhance system performance and reduce latency.

Technical Skills

Primary Skills:

Ø2-4 years of experience in data engineering, with a strong focus on Databricks, PySpark, Python and Spark SQL.
ØProven experience in implementing ETL processes and data pipelines
ØHands-on experience with Azure Databricks, Azure Data Factory (ADF), Azure Data Lake Storage (ADLS)
ØAbility to write reusable, testable, and efficient code
ØDevelop low-latency, high-availability, and high-performance applications
ØUnderstanding of fundamental design principles behind a scalable application
ØGood knowledge of Azure cloud services
ØFamiliarity with Generative AI and its applications in data engineering
ØKnowledge of Microsoft Fabric and Azure AI services is an added advantage

 Enabling Skills

·Excellent analytical, and problem-solving skills
·Quick learning ability and adaptability
·Effective communication skills
·Attention to detail and good team player
·Willingness and ability to deliver within tight timelines
·Flexible to work timings and willingness to work on different projects/technologies

Other facts

Tech stack
Data Engineering,Databricks,PySpark,Python,Spark SQL,ETL,Azure Data Factory,Azure Data Lake Storage,Data Governance,Machine Learning,Data Quality,Data Transformation,Debugging,Performance Tuning,Generative AI,Microsoft Fabric,Cloud Services

About KPMG Global Services

KPMG entities in India are established under the laws of India and are owned and managed (as the case may be) by established Indian professionals. Established in September 1993, the KPMG entities have rapidly built a significant competitive presence in the country. Today we operate from offices across 14 cities including in Ahmedabad, Bengaluru, Chandigarh, Chennai, Gurugram, Hyderabad, Jaipur, Kochi, Kolkata, Mumbai, Noida, Pune, Vadodara and Vijayawada.

KPMG entities have a domestic client base of over 2700 companies. Our global approach to service delivery helps provide value-added services to clients.

Our differentiation is derived from a rapid performance-based, industry-tailored and technology-enabled business advisory services delivered by some of the leading talented professionals in the country. KPMG professionals are grouped by industry focus and our clients are able to deal with industry professionals who speak their language. Our internal information technology and knowledge management systems enable the delivery of informed and timely business advice to clients.

Team size: 10,001+ employees
LinkedIn: Visit
Industry: Business Consulting and Services

What you'll do

  • The Associate 2 - Data Engineer will be responsible for extracting and processing datasets from client ERP systems and developing solutions using Azure Databricks and Python. They will also analyze large volumes of data and create pipelines to support data analytics and machine learning applications.

Ready to join KPMG Global Services?

Take the next step in your career journey

Frequently Asked Questions

What does a Engineer (A2 DES, Databricks, PySpark, Python) do at KPMG Global Services?

As a Engineer (A2 DES, Databricks, PySpark, Python) at KPMG Global Services, you will: the Associate 2 - Data Engineer will be responsible for extracting and processing datasets from client ERP systems and developing solutions using Azure Databricks and Python. They will also analyze large volumes of data and create pipelines to support data analytics and machine learning applications..

Why join KPMG Global Services as a Engineer (A2 DES, Databricks, PySpark, Python)?

KPMG Global Services is a leading Business Consulting and Services company.

Is the Engineer (A2 DES, Databricks, PySpark, Python) position at KPMG Global Services remote?

The Engineer (A2 DES, Databricks, PySpark, Python) position at KPMG Global Services is based in Bengaluru, India. Contact the company through Clera for specific work arrangement details.

How do I apply for the Engineer (A2 DES, Databricks, PySpark, Python) position at KPMG Global Services?

You can apply for the Engineer (A2 DES, Databricks, PySpark, Python) position at KPMG Global Services directly through Clera. Click the "Apply Now" button above to start your application. Clera's AI-powered platform will help match your profile with this opportunity and guide you through the application process. You can also learn more about KPMG Global Services on their website.