Job description
About the role - you will:
• Be part of 10-12 Platform Engineers that are the crux for developing and maintaining Big Data (Data Lake, Data Warehouse and Data Integration) and advanced analytics platforms at Seagate
• Apply hands-on subject matter expertise in the Architecture of and administration of Big Data platforms - Data Warehouse Appliances , Open Data Lakes (AWS EMR, HortonWorks), Data Lake Technologies (AWS S3/Databricks/Other) and experience with ML and Data Science platforms (Spark ML , H2O , KNIME)
• Develop and manage SPARK ETL Frameworks, Data orchestration with Airflow and support building PRESTO/Trino queries for the key stakeholders
• Design, scale and deploy Machine Learning pipelines
• Collaborate with Application Architects and Business SMEs to design and develop end-to-end data pipelines and supporting infrastructure
• Establish and maintain productive relationships with peer organizations, partners, and software vendors
• Work with customers and business stakeholders to understand the needs and deliver on their highest priority features and requirements incrementally
• Work with PMO and effectively communicate with management team on the platform charter and scope
• Manage budget responsibilities, development of staff, resource planning and goal setting for the team
• Provide guidance, career development, and mentoring to technical team members, and help achieve maximum potential
About you:
• You’re a passionate professional who is up to the challenge of blending the fast-changing technology landscape of Big Data analytics with the complex and high-impact space of HiTech and Manufacturing analytics
• As a motivated self-starter, you have the experience working in a dynamic environment
• Exceptional data engineering experience in building large, high-scale Data platforms and applications using cloud and big data technologies like Hadoop ecosystem and Spark
• Strong appetite for constant learning, thinking out of the box, questioning the problems & solutions with the intent to understand and solve better
• As well, you’re uncompromisingly detail oriented, well organized with solid time management skills, and you have solid, effective verbal and written communications abilities as well
• Excellent interpersonal skills to develop relationships with different teams and peers in the organization
Your experience includes:
• Excellent technical skills with a proven and successful history working with data at scale and empowering organizations through data
• Big data processing frameworks: Spark, Hadoop, Hive, Kafka, EMR
• Architecting and developing big data solutions on cloud (AWS or Other)
• Advanced experience and hands-on architecture and administration experience on big data platforms
• Data Warehouse Appliances, Hadoop (AWS EMR), Data Lake Technologies (AWS S3/GCS/Other) and experience with ML and Data Science platforms (Spark ML , H2O , KNIME )
• Python, Java, Scala
• DevOps, Continuous Delivery, and Agile development
• Creating a culture of technical excellence by leading code and design reviews, promoting mentorship, and identifying and promoting educational opportunities for engineers
• Strong understanding of Micro-services and container-based development using Docker and Kubernetes ecosystem is a plus
• Experience working in a Software Product Development environment is a plus