At GE Appliances, a Haier company, we come together to make “good things, for life.” As the fastest-growing appliance company in the U.S., we’re powered by creators, thinkers and makers who believe that anything is possible and that there’s always a better way. We believe in the power of our people and in giving them the freedom to explore, discover and build good things, together.
The GE Appliances philosophy, backed by three simple commitments defines the way we work, invent, create, do business, and serve our communities: we come together, we always look for a better way, and we create possibilities.
Interested in joining us on our journey?
The Intern will be working on Python scripting for automation of data, Machine Learning (Training the data for automating), building user interface to access SQL database.
Position
Intern - Data Anaytics
Location
IND, Bangalore, KA
How You'll Create Possibilities
Essential Responsibilities
Understand Data Automation Requirements: Quickly assimilate and comprehend data automation requirements to ensure alignment with project goals and objectives.
Develop and Maintain Automation Scripts: Create and maintain Python scripts to automate data collection, cleaning, and processing tasks, ensuring efficiency and accuracy.
Convert R Scripts to Python: Analyze existing automation scripts written in R and convert them to Python, optimizing for performance and maintainability.
Leverage Advanced AI and NLP Technologies: Utilize Large Language Models (LLMs), OpenAI (Generative AI) API, and advanced Natural Language Processing (NLP) techniques to enhance data relevancy and insights. Create User Interface to retrieve information from databases
What You'll Bring to Our Team
Qualifications/Requirements
- B.Tech final year or B.Tech completed.
- Computer Science engineering
Technical Skills
- Python Programming: Advanced proficiency in Python, with experience in libraries such as Pandas, NumPy, Scikit-learn, NLTK or SpaCy, OpenAI, and Sentence-transformers. Skilled in implementing machine learning algorithms such as Gradient Boosting, ADASYN, and cosine similarity.
- Large Language Models (LLMs): Expertise in implementing, leveraging, and comparing BERT and other LLMs for advanced NLP tasks, with a thorough understanding of their advantages and limitations.
- Natural Language Processing (NLP): Proficient in utilizing advanced NLP techniques and integrating them into Python scripts for comprehensive data analysis.
- Full-Stack Development: Proficiency in both front-end (HTML, CSS, JavaScript) and back-end (Python, Flask, Django) languages for developing user-friendly and interactive scripts and applications.
- R Programming: Advanced proficiency in the R programming language for statistical analysis and data visualization.
- SQL Databases: Strong understanding of relational databases and SQL for efficient data storage and retrieval.
- NoSQL Databases: Familiarity with NoSQL databases such as MongoDB for handling unstructured data.
- Jupyter Notebooks: Proficiency in using Jupyter Notebooks for interactive data analysis, visualization, and model development.
- Integrated Development Environments (IDEs): Experience with IDEs like PyCharm and VS Code for efficient coding, debugging, and project management.
- Prompt Engineering: Skilled in designing and optimizing prompts for various NLP tasks and applications.
- Version Control: Basic knowledge of version control systems like Git for effective code management and collaboration.
- Data Security: Understanding of data security and cybersecurity policies for secure data handling and compliance.
- OpenAI API: Experience in using the OpenAI API and its applications for advanced AI and NLP tasks.
- MS Office Suite: Proficient in handling MS Office Suite, particularly Excel, for data manipulation and reporting.
- Passion for Data Science: Keen interest in data science, with a commitment to continuous learning and professional development.