Associate Data & ML Engineer
Globeleq
Globeleq’s various generation technologies include gas, wind, solar PV, battery energy storage (BESS), and geothermal. The company is also actively pursuing new opportunities which are emerging from the energy transition. In South Africa, Globeleq owns and operates renewable energy (RE) power plants throughout the country.
The Associate Data & ML Engineer is responsible for executing the technical implementation of Globeleq’s Data Transformation initiative on behalf of, and under the direction of the Data Engineering Manager. The role focuses on the Globeleq’s data development journey through data ingestion, processing, automated data pipelines, and the establishment of an integrated Single Source of Truth (SSOT)database, cross-functional system integrations and AI/ML-ready data structures.
- Application Deadline
- March 15, 2026
- Department
- Operations
- Employment Type
- Permanent
- Location
- South Africa
- Workplace type
- Hybrid
- Reporting To
- Data Engineering Manager
Key Responsibilities
- Design, build and maintain end-to-end automated data pipelines from internal and external sources into a central data platform including scheduling, monitoring and resolving issues for assigned workflows:
- Platform maintenance and development of Globeleq’s Central Asset Management system (CAMs), including calculation development, on-boarding assets and equipment, and a continuous improvement focus.
- Develop, implement and extend the integrated SSOT database, consolidating data from CAMs, ERP, OT/IoT, SharePoint and other platforms according to agreed standards ensuring scalable data flows.
- Develop modular and scalable data ingestion using API integrations, SQL stored procedures and ETL frameworks, maintaining reliable, automated data ingestion between internal and external platforms to the SSOT database.
- Support and provide data development solutions. Be a practical driver and enforcer of the Data Transformation plan on behalf of the Data Engineering Manager; escalating risks and developing solutions.
- Design and develop scalable data models (staging, core, marts, feature sets) that support strategic reporting, advanced analytics and ML.
- Develop scalable data models across clearly defined business layers, including staging (raw landed data), core (cleaned and standardised single source of truth), marts (business-ready views for specific domains), and feature sets (model-ready tables for machine learning and advanced analytics).
- Develop and support workflow automation and lightweight data applications using tools such as Power Apps and Power Automate, integrating these solutions with the core data platform to enable efficient business processes.
- Develop ML algorithms, dedicated toward anomaly detection, prediction and neural networks. Ensuring all data models, pipelines and storage approaches are AI/ML-ready, including feature-ready datasets for pattern recognition, prediction and anomaly detection
- Own end-to-end development and processing (data algorithms and ML solutions)
- Technical platform development, data orchestration and data management
- Adhere to governing data management, security and governance standards in line with the Data Governance Policy (Data Engineering, Audit and Risk, Cyber Security and IT requirements).
- Maintain, document and track comprehensive technical documentation and change management records for architectures, pipelines, automations, environments and access.
- Work with divisional data owners to reduce data silos, standardise data flows and ensure adherence to agreed standards and timelines.
Skills and Competencies
- Full-stack data engineering competence:
- API integration (REST/JSON, auth, pagination, error handling)
- ETL/ELT orchestration and job scheduling (Automated workflows)
- Data modelling (staging, core, marts, feature sets); production operations (monitoring, alerting, incident response)
- Strong SQL; proficiency in Python for data engineering and ML-enabling tasks; and solid programming foundations in Python, SQL and/or C#
- Ability to make scalable architectural decisions and prepare data for ML and model integration into workflows.
- Solid ML foundations (feature engineering, evaluation, overfitting, drift) and ability to design data pipelines that are fit for ML.
- Automated workflow development (Power Automate and PowerApps)
- Strong, hands-on engineering mindset; comfortable taking technical ownership of assigned work:
- Proven ability to design and lead data platform or data product builds.
- Identifies problems, proposes solutions and drives implementation.
- Clear, structured communication skills; and can explain technical aspects to non-technical stakeholders and leadership.
- Strong systems thinking and architecture skills: designs for scalability, maintainability and AI/ML-readiness from the outset.
- Strong engineering discipline: version control, testing, deployment processes, documentation and incident handling.
- Enjoys building automation, integrations and ML-ready datasets.
- Cross-team coordination. Strong ability to work with multiple divisions, follow up with stakeholders and enforce agreed standards and timelines.
Experience, Knowledge and Qualifications
- Degree in Computer Science, Information Systems, Engineering, Mathematics or a related field. Proficient in data engineering.
- 2-4 years in data engineering/data platform development or similar technical role, with hands on responsibility and clear evidence of data engineering development.
- Programming foundations in Python, SQL and/or C#, with clear evidence of making scalable architectural decisions.
- Hands-on responsibility for API integration (REST/JSON, auth, pagination, error handling); ETL/ELT orchestration and job scheduling; data modelling (staging, core, marts/feature sets); and production operations (monitoring, alerting, incident response).
- Strong SQL skills (DDL/DML, performance tuning, stored procedures, views, functions).
- Proficiency in Python for data engineering and ML-enabling tasks, plus experience with ETL tooling and automation frameworks.
- Strong foundations in ML concepts and practical experience preparing data for ML and integrating models into data workflows (even if not a pure data scientist).
- Power Engineering experience.
- Hands-on experience with ML models in production (e.g. forecasting, classification, anomaly detection) and associated MLOps tooling.
- Exposure to neural networks/deep learning (e.g. TensorFlow, PyTorch) and modern ML pipelines.
- Experience with data lakes/big data architectures and orchestration tools (e.g. Airflow, Prefect, Azure Data Factory or similar).
- Familiarity with AI/ML governance, model risk and secure data handling.
- Experience working with industrial/IoT or energy sector data.
About Globeleq
Not quite right? Register your interest to be notified of any roles that come along that meet your criteria.

