Posted On Jul 30

This job posting has expired

Lead Machine Learning Ops Engineer


Location : Hadapsar, Pune, Maharashtra

Headquarters : Ireland

Hiring Mode : Full Time

Hiring Role : DevOps Engineer

Experience : Mid Level

Hadapsar, Pune, Maharashtra

If you desire to be part of something special, to be part of a winning team, to be part of a fun team – winning is fun. We are looking forward to Lead Machine Learning Ops Engineer based in Pune, India. In Eaton, making our work exciting, engaging, meaningful; ensuring safety, health, wellness; and being a model of inclusion & diversity are already embedded in who we are - it’s in our values, part of our vision, and our clearly defined aspirational goals. 

This exciting role offers opportunity to:
  • Demonstrate exceptional impact in delivering projects, products and/or platforms in terms of scalable data processing and application architectures, technical deliverables and delivery throughout the project lifecycle.
  • Responsible for ensuring that ML deployment pipelines can maintain and monitor model accuracy and ensure application code quality.
  • Work on developing the Intelligence layer MLOps pipelines, ensuring that infrastructure and code is delivered incrementally and according to our internal security deployment policies.
  • Responsible for ensuring model development is deployable and sustainable at scale and meets the business needs.
  • Author high-quality, high-performance, unit-tested code to extract and transform data based on business and data science needs
  • Must be conversant with Agile methodologies and tools and have a track record of delivering products in a production environment.
  • Work directly with stakeholders, engineering, and test to create high quality solutions that solve end-user problems.
  • Develop and execute agile work plans for iterative and incremental project delivery
  • Explore and recommend new tools and processes which can be leveraged across the data preparation pipeline for capabilities and efficiencies.
  • Ensure that our development and deployment are tightly integrated to each other to maximize the deployment user experience.
  • Curator for all code and binary artifact repositories (containers, compiled code)

  • Requires a minimum of a Bachelor level degree in computer science or equivalent software engineering discipline.
  • 7+ years of experience in a relvant and comparable role in the software industry, with a proven track record of success.
  • Proficient in Python and Linux environments to support model training, analysis and prediction pipelines for our advanced analytics services.
  • Experienced in designing and building highly distributed and scalable Analytics/Data Science/Machine Learning pipeplines hosted in major public and/or private cloud providers such as Azure, AWS or Google Cloud.
  • Proficient in developing and maintaining AI & ML DevSecOps pipelines to support Model training and prediction requirements.
  • Experience with Docker & Docker orchestration technologies (e.g. Kubernetes (K8S), Docker compose).
  • Ability to develop Machine learning and validation pipelines in a secure and enterprise grade environment.
  • Strong problem solving and software debugging skills with a track record of delivery and a willingness to roll up sleeves to get the job done.
  • Excellent verbal and written communication skills including the ability to effectively explain technical concepts.
  • Masters or equivalent degree in computer science or comparable software engineering discipline.
  • Experience in deployment of Infrastructure-as-Code using Terraform.
  • Proficiency with cloud technologies (IaaS, PaaS, serverless technology) micro-service design, CI/CD and DevOps technologies.
  • Experience in Knowledge Graphs and integration this technology into intelligence pipelines.
  • Experience in building and deploying AI/ML models in large scale production projects with modern light-weight design (API, Microservices etc.)
  • Experience with ML frameworks and libraries such as TensorFlow, Keras, Pytorch and Kubeflow.
  • Experienced in big data orchestration frameworks such as Apache Airflow.
  • Knowledgeable in leveraging data transit protocols and technologies (AMQP, MQTT, Rest API, JDBC, etc).
  • Knowledge of multiple cloud and edge technologies, especially Microsoft Azure e.g. Databricks, Azure IoT Edge.
  • Knowledge of cloud development platforms such as Azure or AWS and their associated data storage options.
  • Experience in integrating Artifact repositries (eg JFrog Artifactory) into CI/CD pipeline to manage and control access to binary dependancies
  • Experience with Agile development methodologies and concepts.
  • Keeps abreast of upcoming software development/engineering tools, trends, and methodologies.
  • Good judgment, time management, and decision-making skills.
Yes! Because you are the one we are looking for, we hope to hear from you now!
We make what matters work. Everywhere you look—from the technology and machinery that surrounds us, to the critical services and infrastructure that we depend on every day—you’ll find one thing in common. It all relies on power. That’s why Eaton is dedicated to improving people’s lives and the environment with power management technologies that are more reliable, efficient, safe and sustainable. Because this is what matters.

We are confident we can deliver on this promise because of the attributes that our employees embody. We’re ethical, passionate, accountable, efficient, transparent and we’re committed to learning. These values enable us to tackle some of the toughest challenges on the planet, never losing sight of what matters.

Job: Engineering
Region: Asia Pacific
Organization: INNOV Innovation Center
Job Level: Team Leader/Supervisor
Schedule: Full-time
Is remote work (i.e. working from home or another Eaton facility) allowed for this position?: No
Does this position offer relocation?: Relocation from within hiring country only
Travel: Yes, 10 % of the Time