Principal Cloud Infrastructure Engineer
Stem
About Stem - Driven by human and artificial intelligence – Stem is unlocking energy intelligence.
Stem is a global leader reimagining technology to support the energy transition. Turning complexity into clarity, and potential into performance.
We help asset owners, operators and stakeholders benefit from the full value of their energy portfolio by enabling the intelligent development, deployment, and operation of clean energy assets. Our integrated software suite, PowerTrack, is the industry standard and best-in-class for asset monitoring, supported by professional and managed services, under one roof. Meant to tackle challenges as seamlessly as possible, Stem shows the information needed clearly and accurately and helps harness raw data to inform actionable insight. With global projects managed in 55 countries – from Germany to Japan and across North America – customers have relied on Stem for nearly 20 years to maximize the value of their clean energy projects.
Stem’s culture embodies diversity & inclusion beyond the traditional facets of gender, ethnicity, age, disabilities, and sexual orientation to include experience, personality, communication, workstyles, and more. At our core, Stem is at the momentous intersection of clean energy and software technology where diverse ideas, experiences, and professional skills converge to make the inclusive culture we have today. Together, we are turning old school thoughts about software and energy into progressive, collaborative, and innovative solutions. By joining our team, you will be collaborating with data scientists, energy experts, skilled salespeople, thought-leading executives and more from a range of backgrounds. This intersection of ideas, beliefs, and skills is what makes us unique enough to lead the world’s largest network of digitally connected energy storage systems.
Role Overview
We are seeking a Principal Cloud Infrastructure Engineer to provide technical leadership and ownership of Stem’s core data platform and pipeline architecture. Approximately 70% of this role will focus on designing, scaling, and operating high-performance data pipelines and distributed backend systems. The remaining 30% will support microservices development.
This is a technical leadership position responsible for defining platform architecture, setting engineering standards, and driving reliability, scalability, and operational excellence across the organization.
Key Responsibilities
Architect, design, and own large-scale data pipelines for real-time and batch ingestion, transformation, and delivery
Provide technical leadership and architectural direction for distributed backend and data platform systems
Establish platform-wide standards for performance, reliability, security, and maintainability
Lead technical design reviews and mentor senior engineers on data platform and infrastructure best practices
Drive operational excellence, including observability, alerting, incident response, and post-incident improvements
Own the reliability and correctness of systems powering analytics, forecasting, and control platforms
Influence long-term technical roadmap and platform evolution in partnership with engineering leadership
Collaborate closely with US-based engineering, product, and operations teams
Support OpenADR integrations and utility data import services
Required Qualifications
12+ years of experience in distributed systems, cloud infrastructure, or data platform engineering
Expert-level programming skills in Java and Python (both required)
Deep expertise in designing and operating data pipelines (streaming and batch)
Strong understanding of distributed systems principles: fault tolerance, scalability, consistency, and performance
Proven experience operating Kubernetes-based systems in production environments
Strong experience with real-time processing frameworks such as Spark or Flink
Solid experience with:
SQL and NoSQL datastores
Search platforms such as Elasticsearch
Proven leadership in building and operating monitoring, observability, and alerting platforms
Demonstrated ability to influence architecture across multiple teams and services
Preferred Qualifications
Bachelor's degree in computer science, Engineering, or a related technical field is required
Experience with OpenADR or demand response / energy market integrations
Experience designing utility data ingestion pipelines (e.g., meter data, market data, settlement files)
Experience working in regulated, compliance-driven, or reliability-critical environments
Strong hands-on experience with AWS
Nice to have: Experience building or integrating AI agents or intelligent automation systems, such as LLM-based agents, workflow orchestration, or AI-driven operational tooling
Core Technologies and Platforms
Java, Python
Kubernetes and cloud-native infrastructure
Large-scale data pipelines (streaming and batch)
Spark, Flink, or similar processing frameworks
SQL databases, NoSQL datastores, Elasticsearch
Observability platforms (metrics, logging, tracing, alerting)
Stem, Inc. is an equal opportunity employer committed to diversity in the workplace and does not discriminate against any employee or applicant for employment because of race, color, sex, pregnancy, religion, national origin, ethnicity, citizenship, sexual orientation, gender identity, age, marital status, disability, genetic information, military status, protected veteran status or any other factor protected by applicable federal, state or local laws.