Job Description
ESSENTIAL SKILLS:
- Expertise with Apache Spark (PySpark), Databricks notebooks, Delta Lake, and SQL
- Strong programming skills in Python for data processing
- Experience with cloud data platforms (Azure) and their Databricks offerings; familiarity with object storage (ADLS)
- Proficient in building and maintaining ETL/ELT pipelines, data modeling, and performance optimization
- Knowledge of data governance, data quality, and data lineage concepts
- Experience with CI/CD for data pipelines, and orchestration tools (GitHub Actions, Asset Bundles or Databricks’ jobs)
- Strong problem-solving skills, attention to detail, and ability to work in a collaborative, cross-functional team
ADVANTAGEOUS SKILLS:
- Experience with streaming data (Structured Streaming, Kafka, Delta Live Tables).
- Familiarity with materialized views, streaming tables, data catalogs and metadata management.
- Knowledge of data visualization and BI tools (Splunk, Power BI, Grafana).
- Experience with data security frameworks and compliance standards relevant to the industry.
- Certifications in Databricks or cloud provider platforms.
QUALIFICATIONS/EXPERIENCE:
Bachelor’s or Master’s degree in Computer Science, Data Engineering, Information Systems, or a related field.
3+ years of hands-on data engineering experience.
Key Responsibilities:
- Design, develop, test, and maintain robust data pipelines and ETL/ELT processes on Databricks (Delta Lake, Spark, SQL, Python/Scala/SQL notebooks)
- Architect scalable data models and data vault/ dimensional schemas to support reporting, BI, and advanced analytics
- Implement data quality, lineage, and governance practices; monitor data quality metrics and resolve data issues proactively
- Collaborate with Data Platform Engineers to optimize cluster configuration, performance tuning, and cost management in cloud environments (Azure Databricks)
- Build and maintain data ingestion from multiple sources (RDBMS, SaaS apps, files, streaming queues) using modern data engineering patterns (CDC, event-driven pipelines, change streams, Lakeflow Declarative Pipelines)
- Ensure data security and compliance (encryption, access controls) in all data pipelines
- Develop and maintain CI/CD pipelines for data workflows; implement versioning, testing, and automated deployments
GO APPLY NOW
Safe & secure application process
Explore More Opportunities
Get Similar Job Alerts
Job Seeker Tip
Before the interview, research the company thoroughly and prepare thoughtful questions to ask.
How to Apply
Click “GO APPLY NOW” to visit the company’s application page.
Follow their instructions carefully.
JVR Jobs connects you with employers – we don’t process applications directly.
Latest Job Opportunities
Alberton: Senior Quantity Surveyor posted by Hire Resolve
A leading construction contractor active across major commercial and infrastructure projects in South Africa, is currently seeking a Senior Quantity...
View JobWelkom: Engineering Supervisor (Mechanical) posted by Hire Resolve
Description:Hire Resolve's Client is currently looking for an experienced Engineering Supervisor (Mechanical) to join their mining company based in the…
View JobSandton: Credit Controller posted by Hire Resolve
A prominent Fashion/Retail company is seeking an experienced Credit Controller with a strong background in financial reconciliation, preferably within the...
View JobSandton: Credit Controller posted by Hire Resolve
A prominent Fashion/Retail company is seeking an experienced Credit Controller with a strong background in financial reconciliation, preferably within the...
View JobSandton: Credit Controller posted by Hire Resolve
A prominent Fashion/Retail company is seeking an experienced Credit Controller with a strong background in financial reconciliation, preferably within the...
View JobSandton: Credit Controller posted by Hire Resolve
A prominent Fashion/Retail company is seeking an experienced Credit Controller with a strong background in financial reconciliation, preferably within the...
View Job
Browse Employers
Job Alerts