The Job logo

What

Where

Data Engineer

ApplyJoin for More Updates

You must Sign In before continuing to the company website to apply.

Smart SummaryPowered by Roshi
As a Data Engineer at Nielsen, you will be responsible for accelerating data acquisition, ensuring data reliability, enhancing data accessibility, and collaborating with teams. This role involves designing and automating data pipelines, monitoring data systems, and implementing data governance processes. The ideal candidate should have a Bachelor's degree in a relevant field with 3-5 years of experience in Python, PySpark, SQL, AWS, EC2, GitLab, and Airflow. Strong problem-solving skills, attention to detail, and effective communication are essential for this full-time hybrid opportunity in Bengaluru, Karnataka, India.

Job description 

Role Details

  • Analytics team specific: This role drives the acceleration of data acquisition, ensures the reliability and accuracy of data outputs and enhances the accessibility and discovery of data.  They offer technical guidance and mentorship, uphold data integrity and reusability, and collaborate with diverse teams to meet our extensive data needs.
  • Meters team specific: In this role, you will learn and become an expert in Nielsen’s systems and TV panel data, with a focus on underlying meter hardware and infrastructure used for data collection and crediting. You will develop and test software and related database infrastructure used to support both research and production pipelines, all while working within the Software Development Life Cycle framework and applying software development best practices.

Responsibilities

  • Design and automate essential data pipelines and inputs, ensuring seamless integration with downstream analytics and production systems. 
  • Collaborate with cross-functional teams to integrate new functionalities into existing data pipelines, including lower test environments to help validate and assess impact prior to Production integration, where applicable. 
  • Implement data governance and quality processes to ensure the integrity and accuracy of data throughout its lifecycle.
  • Monitor data systems and processes to identify issues and proactively implement improvements to prevent future problems.
  • Participate in code reviews with senior developers prior to pushing the code into production to ensure meeting accuracy and best practice standards.
  • Implement data pipeline Directed Acyclic Graphs (DAGs) and maintenance DAGs. Configure and setup DAGs based on the data to run Spark commands in parallel and sequential. 
  • Perform unit testing using test cases and fix any bugs. 
  • Optimize code to meet product SLAs
  • Support multiple projects and communicate with stakeholders in various organizations. This includes regularly providing status updates, developing timelines, providing insights, etc.

Key Skills

  • Bachelor’s Degree in Computer Science, Data Science, Analytics or related field 
  • 3-5 years of experience with the following:
  • Coding in Python, PySpark, and SQL
  • Hive data storage technologies
  • Working within cloud-based infrastructures and tools such as AWS, EC2, GitLab, and Airflow. 
  • Working within the Software Development Life Cycle framework and applying software development best practices
  • Building monitoring checks and tools to ensure infrastructure and related processes are working as expected
  • Solid understanding of system design, data structures and performance optimization techniques
  • Excellent problem solving skills and attention to detail
  • Well-organized and able to handle and prioritize multiple assignments
  • Able to communicate effectively both orally and in writing
  • (Preferred) 2+ years experience with visualization and reporting tools, e.g. Tableau
  • (Preferred) Experience deploying and maintaining Machine Learning models within Production environments
  • (Preferred) Experience working with Jira, Confluence, and Smartsheets
  • (Preferred) Experience with Alteryx, Databricks platforms
Set alert for similar jobsData Engineer role in Bengaluru, India
Nielsen Logo

Company

Nielsen

Job Posted

8 months ago

Job Type

Full-time

WorkMode

Hybrid

Experience Level

3-7 Years

Category

Data & Analytics

Locations

Bengaluru, Karnataka, India

Qualification

Bachelor or Master

Applicants

Be an early applicant

Related Jobs

Nielsen Logo

Data Scientist and Business Analyst, Content Analytics

Nielsen

Bengaluru, Karnataka, India

Posted: 8 months ago

Join Nielsen as a Data Scientist and Business Analyst focused on audience measurement and media behavior analytics. Design and deliver global analytics solutions and insights for big brands. Requires expertise in SQL, Python, and Spark, with a focus on data synthesis, storytelling, and innovation.

Zebra Technologies Logo

Data Engineer, I

Zebra Technologies

Bengaluru, Karnataka, India

Posted: 23 days ago

Job Description Remote Work: Hybrid Overview: At Zebra, we are a community of innovators who come together to create new ways of working to make everyday life better. United by curiosity and care, we develop dynamic solutions that anticipate our customer’s and partner’s needs and solve their challenges. Being a part of Zebra Nation means being seen, heard, valued, and respected. Drawing from our diverse perspectives, we collaborate to deliver on our purpose. Here you are a part of a team pushing boundaries to redefine the work of tomorrow for organizations, their employees, and those they serve. You have opportunities to learn and lead at a forward-thinking company, defining your path to a fulfilling career while channeling your skills toward causes that you care about – locally and globally. We’ve only begun reimaging the future – for our people, our customers, and the world. Let’s create tomorrow together. A Data Engineer will be responsible for understanding the client's technical requirements, design and build data pipelines to support the requirements. In this role, the Data Engineer, besides developing the solution, will also oversee other Engineers' development. This role requires strong verbal and written communication skills and effectively communicate with the client and internal team. A strong understanding of databases, SQL, cloud technologies, and modern data integration and orchestration tools like Azure Data Factory (ADF), Informatica, and Airflow are required to succeed in this role. Responsibilities:   • Play a critical role in the design and implementation of data platforms for the AI products. • Develop productized and parameterized data pipelines that feed AI products leveraging GPUs and CPUs. • Develop efficient data transformation code in spark (in Python and Scala) and Dask. • Build workflows to automate data pipeline using python and Argo. • Develop data validation tests to assess the quality of the input data. • Conduct performance testing and profiling of the code using a variety of tools and techniques. • Build data pipeline frameworks to automate high-volume and real-time data delivery for our data hub. • Operationalize scalable data pipelines to support data science and advanced analytics. • Optimize customer data science workloads and manage cloud services costs/utilization. Qualifications:   • Minimum Education: o Bachelors, Master's or Ph.D. Degree in Computer Science or Engineering. • Minimum Work Experience (years): o 1+ years of experience programming with at least one of the following languages: Python, Scala, Go. o 1+ years of experience in SQL and data transformation o 1+ years of experience in developing distributed systems using open source technologies such as Spark and Dask. o 1+ years of experience with relational databases or NoSQL databases running in Linux environments (MySQL, MariaDB, PostgreSQL, MongoDB, Redis). • Key Skills and Competencies: o Experience working with AWS / Azure / GCP environment is highly desired. o Experience in data models in the Retail and Consumer products industry is desired. o Experience working on agile projects and understanding of agile concepts is desired. o Demonstrated ability to learn new technologies quickly and independently. o Excellent verbal and written communication skills, especially in technical communications. o Ability to work and achieve stretch goals in a very innovative and fast-paced environment. o Ability to work collaboratively in a diverse team environment. o Ability to telework o Expected travel: Not expected.