Job Title: Data Engineer
 Location: Johannesburg
 Job Type: Contract (8-month contract)
 Hybrid  
Our client is looking for a skilled Data Engineer to join the team on a contract basis, working within a dynamic and fast-paced banking/financial services environment. The ideal candidate will have solid experience in designing and building robust data pipelines and architectures to support advanced analytics, reporting, and data-driven decision-making.
Key Responsibilities
- Design, develop, and maintain scalable and efficient data pipelines and ETL processes.
 - Ingest, transform, and load data from various internal and external sources into data warehouses or data lakes.
 - Ensure data quality, integrity, and consistency across all systems and processes.
 - Collaborate with business stakeholders, data analysts, and BI teams to understand data requirements and deliver appropriate solutions.
 - Implement data solutions that comply with governance, regulatory, and security standards.
 - Perform root cause analysis on internal and external data and processes to answer specific business questions.
 - Monitor and troubleshoot data pipeline issues and performance bottlenecks.
 - Document data flow processes and contribute to maintaining best practices for data engineering across the team.
 
Experience and Qualifications
- Bachelor’s degree in computer science, Information Systems, Engineering, or a related field.
 - 3-5+ years of proven experience as a Data Engineer in a banking or financial services environment.
 - Relevant certifications in data engineering, cloud platforms, or database management are advantageous.
 - Strong proficiency in SQL and experience with large-scale data systems.
 - Hands-on experience with ETL tools and frameworks (e.g., Apache NiFi, Talend, SSIS).
 - Experience with cloud platforms such as Azure (preferred), AWS, or GCP.
 - Proficiency with Azure Data Factory, Synapse Analytics, and Data Lake solutions.
 - Solid experience with Python or Scala for data processing.
 - Familiarity with big data tools such as Databricks, Spark, or Kafka.
 - Exposure to CI/CD tools and DevOps practices for data pipelines.
 - Strong understanding of data modelling, data warehousing, and data governance principles.
 - Knowledge of APIs and integrating data from multiple source systems.
 
Join us in shaping the future of client solutions! If you’re ready to take on a new challenge and make an impact, we want to hear from you. Apply now!
  
Desired Skills:
- Azure
 - AWS
 - GCP
 - CI/CD
 - DevOps
 - Python
 - Scala
 - Spark
 - Kafka
 - ETL Tools
 - Data engineering
 - Big data
 - Hadoop