The Job logo

What

Where

Data Engineer, Analytics and Data Science, YouTube

ApplyJoin for More Updates

You must Sign In before continuing to the company website to apply.

Smart SummaryPowered by Roshi
As a Data Engineer at YouTube Analytics and Data Science, you will collaborate with a team of analytics professionals on data pipelines, tools for content analysis, and business reporting. Responsibilities include requirements gathering, data architecture design, ETL pipeline optimization, and ensuring data integrity and governance.

Minimum qualifications:

  • Bachelor's degree in a quantitative field (e.g., Statistics, Computer Science, Math, Physics, Engineering), or equivalent practical experience.
     
  • 1 year of experience in data engineering or business intelligence roles.
     
  • Experience with relational databases, including SQL queries, database definition, and schema design.
     
  • Experience with one or more programming languages (e.g., Python, Java, C++, etc.).
     

Preferred qualifications:

  • Master’s degree in a quantitative field (e.g., Computer Science, Engineering, Statistics, Math).
     
  • Experience with data warehouses, distributed data platforms, and data lakes.
     
  • Ability to navigate ambiguity and work in a fast-moving environment with multiple stakeholders.
     
  • Excellent business and technical communication, organizational, and problem-solving skills.
     
  • Excellent structured thinking skills, with the ability to break down multi-dimensional problems.
     

About the job

As a Data Engineer within YouTube Analytics and Data Science, you will be part of a community of analytics professionals who work on projects ranging from developing data pipelines that help run the business, and build tools to analyze the content partnerships and creator ecosystem that guide business leadership on the effectiveness of partner facing business teams.
 

The team uses SQL and YouTube’s ETL systems to produce useful datasets, establish best practices for data sets and reporting, and develop expertise in various data domains.

At YouTube, we believe that everyone deserves to have a voice, and that the world is a better place when we listen, share, and build community through our stories. We work together to give everyone the power to share their story, explore what they love, and connect with one another in the process. Working at the intersection of cutting-edge technology and boundless creativity, we move at the speed of culture with a shared goal to show people the world. We explore new ideas, solve real problems, and have fun — and we do it all together.

Responsibilities

  • Conduct requirements gathering and project scoping sessions with subject matter experts, business users, and executive stakeholders to discover and define business data needs.
  • Design, build, and optimize the data architecture and extract, transform, and load (ETL) pipelines to make them accessible for Business Data Analysts, Data Scientists, and business users.
  • Work with analysts to scale value-creating capabilities, including data integrations and transformations, model features, as well as statistical and machine learning models.
  • Drive standards in data reliability, data integrity, and data governance, enabling accurate, consistent, and trustworthy data sets, business intelligence products, and analyses.
  • Engage with the analyst community, communicate with analysts to understand critical user journeys and data sourcing inefficiencies, advocate best practices, and lead analyst trainings.
Set alert for similar jobsData Engineer, Analytics and Data Science, YouTube role in Bengaluru, India
Google Logo

Company

Google

Job Posted

7 months ago

Job Type

Full-time

WorkMode

On-site

Experience Level

0-2 Years

Category

Data & Analytics

Locations

Bengaluru, Karnataka, India

Qualification

Bachelor or Master

Applicants

Be an early applicant

Related Jobs

Google Logo

Data Engineer, YouTube

Google

Bengaluru, Karnataka, India

Posted: a year ago

Join the YouTube Business Strategy and Operations team to drive go-to-market functions and shape priorities for growth. We are looking for someone with data engineering and business intelligence experience to design and optimize data architecture, build ETL pipelines, and enable data-driven decision-making. Come work with a team that combines strategic thinking, operational skills, and a pragmatic approach to drive change and make an impact in a global organization.

Capgemini Logo

Data Engineer

Capgemini

Bengaluru, Karnataka, India

Posted: 3 months ago

At Capgemini Engineering, the world leader in engineering services, we bring together a global team of engineers, scientists, and architects to help the world’s most innovative companies unleash their potential. From autonomous cars to life-saving robots, our digital and software technology experts think outside the box as they provide unique R&D and engineering services across all industries. Join us for a career full of opportunities. Where you can make a difference. Where no two days are the same. Job Description:                Expert knowledge in Python                Expert knowledge in popular machine learning libraries and frameworks, such as TensorFlow, Keras, scikit-learn.                Proficient understanding and application of clustering algorithms (e.g., K-means, hierarchical clustering) for grouping similar data points.                Expertise in classification algorithms (e.g., decision trees, support vector machines, random forests) for tasks such as image recognition                Natural language Processing and recommendation systems.                Proficiency in working with databases, both relational and non-relational like MySQL with experience in designing database schemas                And optimizing queries for Efficient data retrieval.                Strong knowledge in areas like Object Oriented Analysis and Design, Multi-threading, Multi process handling and Memory management.                Good knowledge model evaluation metrics and techniques.                Experience in deploying machine learning models to production environments.                Currently working in an Agile scrum team and proficient in using version control systems (e.g., Git) for collaborative development.  Primary Skill:              Excellent in Python Coding              Excellent in Communication Skills              Good in Data modelling, popular machine learning libraries and framework

GE Appliances Logo

Intern - Data Analytics

GE Appliances

Bengaluru, Karnataka, India

Posted: a month ago

At GE Appliances, a Haier company, we come together to make “good things, for life.”  As the fastest-growing appliance company in the U.S., we’re powered by creators, thinkers and makers who believe that anything is possible and that there’s always a better way.  We believe in the power of our people and in giving them the freedom to explore, discover and build good things, together.     The GE Appliances philosophy, backed by three simple commitments defines the way we work, invent, create, do business, and serve our communities:  we come together ,  we always look for a better way , and  we create possibilities .    Interested in joining us on our journey?    The Intern will be working on Python scripting for automation of data, Machine Learning (Training the data for automating), building user interface to access SQL database. Position Intern - Data Anaytics Location IND, Bangalore, KA   How You'll Create Possibilities Essential Responsibilities  Understand Data Automation Requirements : Quickly assimilate and comprehend data automation requirements to ensure alignment with project goals and objectives. Develop and Maintain Automation Scripts : Create and maintain Python scripts to automate data collection, cleaning, and processing tasks, ensuring efficiency and accuracy. Convert R Scripts to Python : Analyze existing automation scripts written in R and convert them to Python, optimizing for performance and maintainability. Leverage Advanced AI and NLP Technologies : Utilize Large Language Models (LLMs), OpenAI (Generative AI) API, and advanced Natural Language Processing (NLP) techniques to enhance data relevancy and insights. Create User Interface to retrieve information from databases   What You'll Bring to Our Team Qualifications/Requirements  B.Tech final year or B.Tech completed. Computer Science engineering Technical Skills  Python Programming: Advanced proficiency in Python, with experience in libraries such as Pandas, NumPy, Scikit-learn, NLTK or SpaCy, OpenAI, and Sentence-transformers. Skilled in implementing machine learning algorithms such as Gradient Boosting, ADASYN, and cosine similarity. Large Language Models (LLMs): Expertise in implementing, leveraging, and comparing BERT and other LLMs for advanced NLP tasks, with a thorough understanding of their advantages and limitations. Natural Language Processing (NLP): Proficient in utilizing advanced NLP techniques and integrating them into Python scripts for comprehensive data analysis. Full-Stack Development: Proficiency in both front-end (HTML, CSS, JavaScript) and back-end (Python, Flask, Django) languages for developing user-friendly and interactive scripts and applications. R Programming: Advanced proficiency in the R programming language for statistical analysis and data visualization. SQL Databases : Strong understanding of relational databases and SQL for efficient data storage and retrieval. NoSQL Databases : Familiarity with NoSQL databases such as MongoDB for handling unstructured data. Jupyter Notebooks : Proficiency in using Jupyter Notebooks for interactive data analysis, visualization, and model development. Integrated Development Environments (IDEs): Experience with IDEs like PyCharm and VS Code for efficient coding, debugging, and project management. Prompt Engineering : Skilled in designing and optimizing prompts for various NLP tasks and applications. Version Control : Basic knowledge of version control systems like Git for effective code management and collaboration. Data Security : Understanding of data security and cybersecurity policies for secure data handling and compliance. OpenAI API: Experience in using the OpenAI API and its applications for advanced AI and NLP tasks. MS Office Suite : Proficient in handling MS Office Suite, particularly Excel, for data manipulation and reporting. Passion for Data Science: Keen interest in data science, with a commitment to continuous learning and professional development.