Job Description
This
role requires expertise in data modeling, ETL development, stream processing, and distributed data systems. The ideal candidate will work closely with developers, DevOps, and analytics teams to transform raw network data into clean, structured, and query–ready datasets that power dashboards, machine learning models, and business logic. The candidate will provide technical leadership, optimize data workflows for performance and reliability, and drive best practices in data engineering methodologiesResponsibilities:
- Build ETL/ELT pipelines for ingesting, cleansing, and transforming CDRs and telecommunications logs
from multiple network elements (5G/4G/3G/2G). - Design and maintain real-time data flows using Kafka, Apache NiFi and Apache Flink.
- Work with large-scale distributed file systems for batch and streaming ingestion.
- Integrate and structure data for analytics platform like Apache Druid, Hudi and Superset.
- Develop CI/CD pipelines for deploying data workflows and transformation logic.
- Ensure data quality, schema validation and compliance with retention and security policies.
- Monitor data pipeline health and optimize performance throughput and cost efficiency.
- Write complex and performant queries for data validation, transformation, aggregation, and analytics
across relational and distributed platforms. - Develop and optimize big data processing workflows in platforms like Apache Spark, Hive and Druid.
- Establish efficient issue tracking and workflow processes, enhancing productivity and collaboration
across engineering teams. - Implement security best practices and compliance frameworks to safeguard infrastructure, data, and
applications from vulnerabilities and threats. - Maintain secure role-based access control mechanisms, encryption strategies, and identity management
solutions to protect sensitive data and ensure regulatory compliance. - Mapping data flows from source to transformation to consumption.
- Design and implement full-text search and indexing solutions for querying and retrieval of structured and unstructured telecommunications data using Apache tools or similar search engines.
- Analyze and estimate and implement storage requirements and strategies for large-scale CDR datasets
and real-time data streams, ensuring optimal resource allocation and scalability across environments. - Ensure data integrity and consistency across ingestion, transformation and storage layers through
validation checks, schema enforcement and robust error handling mechanisms. - Develop and maintain quality monitoring tools to proactively detect anomalies, missing records or data
corruption across pipelines. - Perform other duties as assigned
Requirements:
- Bachelors degree in Computer Engineering, Software Engineering, Computer Science, or a related field.
- Strong experience in building data pipelines using tools like Apache NiFi, Kafka, Airflow or similar.
- Proficiency in SQL, Python, database administration and management like PostgreSQL, MySQL.
-
Solid understanding of distributed data systems like Hive and Hudi and Spark.
-
Experience with streaming frameworks like Kafka Streams, Apache Flink and Apache Beam.
-
Familiarity with data serialization formats like JSON.
- Knowledge of SFTP and secure data transfer mechanisms for ingesting remote files.
- Proficient with Linux environments, shell scripting and storage systems like Ceph.
- Experience with data governance, including data privacy and regulatory compliance like GDPR and
implementing access control, auditing and data usage policies. - Experience in maintaining central inventory of data assets, managing metadata and enabling searchable
discovery across structured and unstructured datasets. - Experience in data lineage tracking to map data flows, visualize and track dependencies.
- Experience with OLAP systems, analytical modelling and columnar databases and designing and
querying multidimensional cubes. - Strong problem–solving skills, ability to work in a fast–paced environment, and manage multiple projects
efficiently. - Strong collaboration skills, adaptability, and a commitment to continuous learning
How to Apply:
If you would like to apply for this role, kindly forward your CV to Gaby Turner at *****@*****.co.za or you may forward your CV to *****@*****.co.za
GO APPLY NOW
Safe & secure application process
Explore More Opportunities
Get Similar Job Alerts
Job Seeker Tip
Always customize your CV and cover letter for each job application. Generic applications are often ignored.
How to Apply
Click “GO APPLY NOW” to visit the company’s application page.
Follow their instructions carefully.
JVR Jobs connects you with employers – we don’t process applications directly.
Latest Job Opportunities
Boksburg: Internal Sales Consultant posted by Kenesis
About the Company:We are a dynamic marketing company helping businesses grow online through Google Ads, Facebook Ads, and website design.…
View JobJohannesburg South: Actuarial Analyst posted by Network Finance
Key Responsibilities:Build, nurture, and expand strong, mutually beneficial client partnerships.Develop compelling product positioning strategies and...
View JobJohannesburg North: Actuary Manager posted by Network Finance
Key Responsibilities:- Assist with the development, testing, and validation of market risk models.- Support consulting projects across banking clients...
View JobWestern Cape: Client Relationship Coordinator posted by Manpower Group
Duties & ResponsibilitiesLead Screening & QualificationContact inbound and outbound leads to assess interest and fit.Ask qualifying questions to determine...
View JobGauteng: Sales Representative posted by Phakisa Holdings
Sales RepresentativeType: Temporary Location: JHB - BenoniDuties and Responsibilities:Sell products and services to potential customersBuild relationships...
View JobHotazel: Mine Technical Services Manager posted by Hire Resolve
Description:Hire Resolve's Client is currently looking for an experienced Mine Technical Services Manager to join their mining company based in…
View Job
Browse Employers
Job Alerts