The Job logo

What

Where

Data Engineer

ApplyJoin for More Updates

You must Sign In before continuing to the company website to apply.

Smart SummaryPowered by Roshi
KPMG in India is seeking a Data Engineer with experience in Data and Analytics, implementing data pipelines on cloud-based platforms, programming in Python, Pyspark, SQL, and working with stream-processing applications. The role involves maintaining CI/CD pipelines, handling large datasets, debugging systems, and ensuring code quality. Full-time, On-site opportunity in Bengaluru, Karnataka, India.

JOB DESCRIPTION

About KPMG in India

KPMG entities in India are professional services firm(s). These Indian member firms are affiliated with KPMG International Limited. KPMG was established in India in August 1993. Our professionals leverage the global network of firms, and are conversant with local laws, regulations, markets and competition. KPMG has offices across India in Ahmedabad, Bengaluru, Chandigarh, Chennai, Gurugram, Jaipur, Hyderabad, Jaipur, Kochi, Kolkata, Mumbai, Noida, Pune, Vadodara and Vijayawada. 

KPMG entities in India offer services to national and international clients in India across sectors. We strive to provide rapid, performance-based, industry-focused and technology-enabled services, which reflect a shared knowledge of global and local industries and our experience of the Indian business environment.

  • Should have experience in Data and Analytics and overseen end-to-end implementation of data pipelines on cloud-based data platforms.
  • Strong programming skills in Python, Pyspark and some combination Java, Scala (good to have)
  • Experience writing SQL, Structuring data, and data storage practices.
  • Experience in Pyspark for Data Processing and transformation.
  • Experience building stream-processing applications (Spark steaming, Apache-Flink, Kafka, etc.)
  • Maintaining and developing CI/CD pipelines based on Gitlab.
  • You have been involved assembling large, complex structured and unstructured datasets that meet functional/non-functional business requirements.
  • Experience of working with cloud data platform and services.
  • Conduct code reviews, maintain code quality, and ensure best practices are followed.
  • Debug and upgrade existing systems.
  • Nice to have some knowledge in Devops.
Set alert for similar jobsData Engineer role in Bengaluru, India
KPMG Logo

Company

KPMG

Job Posted

8 months ago

Job Type

Full-time

WorkMode

On-site

Experience Level

0-2 Years

Category

Engineering and Data & Analytics

Locations

Bengaluru, Karnataka, India

Qualification

Bachelor

Applicants

53 applicants

Related Jobs

Capgemini Logo

Data Engineer

Capgemini

Bengaluru, Karnataka, India

Posted: 7 months ago

At Capgemini Engineering, the world leader in engineering services, we bring together a global team of engineers, scientists, and architects to help the world’s most innovative companies unleash their potential. From autonomous cars to life-saving robots, our digital and software technology experts think outside the box as they provide unique R&D and engineering services across all industries. Join us for a career full of opportunities. Where you can make a difference. Where no two days are the same. Job Description:                Expert knowledge in Python                Expert knowledge in popular machine learning libraries and frameworks, such as TensorFlow, Keras, scikit-learn.                Proficient understanding and application of clustering algorithms (e.g., K-means, hierarchical clustering) for grouping similar data points.                Expertise in classification algorithms (e.g., decision trees, support vector machines, random forests) for tasks such as image recognition                Natural language Processing and recommendation systems.                Proficiency in working with databases, both relational and non-relational like MySQL with experience in designing database schemas                And optimizing queries for Efficient data retrieval.                Strong knowledge in areas like Object Oriented Analysis and Design, Multi-threading, Multi process handling and Memory management.                Good knowledge model evaluation metrics and techniques.                Experience in deploying machine learning models to production environments.                Currently working in an Agile scrum team and proficient in using version control systems (e.g., Git) for collaborative development.  Primary Skill:              Excellent in Python Coding              Excellent in Communication Skills              Good in Data modelling, popular machine learning libraries and framework

NatWest Group Logo

Data Engineer

NatWest Group

Bengaluru, Karnataka, India

Posted: 7 days ago

Job description Join us as a Data Engineer This is an exciting opportunity to use your technical expertise to collaborate with colleagues and build effortless, digital first customer experiences You’ll be simplifying the bank by developing innovative data driven solutions, using insight to be commercially successful, and keeping our customers’ and the bank’s data safe and secure Participating actively in the data engineering community, you’ll deliver opportunities to support the bank’s strategic direction while building your network across the bank We're offering this role at associate level What you'll do As a Data Engineer, you’ll play a key role in driving value for our customers by building data solutions. You’ll be carrying out data engineering tasks to build, maintain, test and optimise a scalable data architecture, as well as carrying out data extractions, transforming data to make it usable to data analysts and scientists, and loading data into data platforms. You’ll also be: Developing comprehensive knowledge of the bank’s data structures and metrics, advocating change where needed for product development Practicing DevOps adoption in the delivery of data engineering, proactively performing root cause analysis and resolving issues Collaborating closely with core technology and architecture teams in the bank to build data knowledge and data solutions Developing a clear understanding of data platform cost levers to build cost effective and strategic solutions Sourcing new data using the most appropriate tooling and integrating it into the overall solution to deliver for our customers The skills you'll need To be successful in this role, you’ll need a good understanding of data usage and dependencies with wider teams and the end customer, as well as experience of extracting value and features from large scale data. You'll have experience of data warehouse and data lake projects and strong knowledge of data engineering tech stack like Spark architecture, SQL, Python, Pyspark. You'll also need a good understanding of cloud tech stack and AWS services like EMR, IAM, S3 and Devops like Ci/CD, gitlab, gitlab runners. You’ll also demonstrate: Experience of ETL technical design, including data quality testing, cleansing and monitoring, and data warehousing and data modelling capabilities Experience of using programming languages alongside knowledge of data and software engineering fundamentals Good knowledge of modern code development practices Strong communication skills with the ability to proactively engage with a wide range of stakeholders

Zebra Technologies Logo

Data Engineer, I

Zebra Technologies

Bengaluru, Karnataka, India

Posted: 3 months ago

Job Description Remote Work: Hybrid Overview: At Zebra, we are a community of innovators who come together to create new ways of working to make everyday life better. United by curiosity and care, we develop dynamic solutions that anticipate our customer’s and partner’s needs and solve their challenges. Being a part of Zebra Nation means being seen, heard, valued, and respected. Drawing from our diverse perspectives, we collaborate to deliver on our purpose. Here you are a part of a team pushing boundaries to redefine the work of tomorrow for organizations, their employees, and those they serve. You have opportunities to learn and lead at a forward-thinking company, defining your path to a fulfilling career while channeling your skills toward causes that you care about – locally and globally. We’ve only begun reimaging the future – for our people, our customers, and the world. Let’s create tomorrow together. A Data Engineer will be responsible for understanding the client's technical requirements, design and build data pipelines to support the requirements. In this role, the Data Engineer, besides developing the solution, will also oversee other Engineers' development. This role requires strong verbal and written communication skills and effectively communicate with the client and internal team. A strong understanding of databases, SQL, cloud technologies, and modern data integration and orchestration tools like Azure Data Factory (ADF), Informatica, and Airflow are required to succeed in this role. Responsibilities:   • Play a critical role in the design and implementation of data platforms for the AI products. • Develop productized and parameterized data pipelines that feed AI products leveraging GPUs and CPUs. • Develop efficient data transformation code in spark (in Python and Scala) and Dask. • Build workflows to automate data pipeline using python and Argo. • Develop data validation tests to assess the quality of the input data. • Conduct performance testing and profiling of the code using a variety of tools and techniques. • Build data pipeline frameworks to automate high-volume and real-time data delivery for our data hub. • Operationalize scalable data pipelines to support data science and advanced analytics. • Optimize customer data science workloads and manage cloud services costs/utilization. Qualifications:   • Minimum Education: o Bachelors, Master's or Ph.D. Degree in Computer Science or Engineering. • Minimum Work Experience (years): o 1+ years of experience programming with at least one of the following languages: Python, Scala, Go. o 1+ years of experience in SQL and data transformation o 1+ years of experience in developing distributed systems using open source technologies such as Spark and Dask. o 1+ years of experience with relational databases or NoSQL databases running in Linux environments (MySQL, MariaDB, PostgreSQL, MongoDB, Redis). • Key Skills and Competencies: o Experience working with AWS / Azure / GCP environment is highly desired. o Experience in data models in the Retail and Consumer products industry is desired. o Experience working on agile projects and understanding of agile concepts is desired. o Demonstrated ability to learn new technologies quickly and independently. o Excellent verbal and written communication skills, especially in technical communications. o Ability to work and achieve stretch goals in a very innovative and fast-paced environment. o Ability to work collaboratively in a diverse team environment. o Ability to telework o Expected travel: Not expected.