Position Purpose:
The Big Data Systems Engineer II role aims to build, test, and maintain data processing systems involving large data sets. The role assumes responsibility for deploying data transformation code into production and, with the assistance of appropriate monitoring, proactively troubleshooting any issues that may arise. The role is well suited to an individual who has progressed in their career journey from a Data Engineer or Systems Engineer into a Big Data DevOps Engineer role in a Linux-based environment. Individuals who enjoy working at scale within a large, diverse team of specialists, and supporting a 24/7 operation will thrive in this role. An understanding of at least one DevOps tools chain is required along with exposure to scheduled data transformation on clustered compute (e.g., Hadoop/PySpark) or query engines.

Qualifications:

  • Degree or Diploma in a Technology-related field.
  • Associate level Linux Certification.
  • Azure Associate Certification.

Experience:

  • +4 years’ experience in a technology-related field as a Data or Systems Engineer for Linux hosted data management application managed through scripted automated deployment
  • Experience deploying and managing applications and databases in the cloud
  • Experience implementing SDLC automation and testing for data transformation or data query jobs
  • Experience configuring and deploying infrastructure using Terraform
  • Experience implementing logging and monitoring data applications using the environment and application-specific logging
  • Experience implementing high availability, and disaster recovery for big data jobs

Job objectives:
Work with a team of specialists in operations to ensure data applications and the jobs running on them are secure and available in production:

  • Implement appropriate monitoring and logging to ensure potential failure can be detected, diagnosed, and remediated before services are impacted.
  • Deploy highly available disaster recovery infrastructure as appropriate for cluster-based solutions.
  • Work with the security team to ensure the application and its infrastructure is secure.
  • Work with development team leads to ensure security is built into their development efforts and deploy additional security measures such as WAFs into the production environment.
  • Exploit security services available from the cloud provider to monitor and ensure the security of the environment.

Work with the ETL development teams and Data Engineers to deploy and ensure the deploy ability of infrastructure, package applications, and data transformation jobs:

  • Use existing SDLC toolchains to deploy cluster-based data applications and the data transformation and queries that run on them.
  • Maintain and upgrade existing SDLC toolchains.
  • Work with the development managers to support them in SDLC automation and developing code to deploy infrastructure using existing SDLC toolchains.

Work with application owners to manage the cost infrastructure deployed for applications both in development and production:

  • Tag resources appropriately so that the system can monitor their cost.
  • Tactically work with the application and data artifact owners to implement tactical cost savings where possible while maintaining required performance.
  • Support system and data artifact owners in monitoring, predicting, and optimising the cost of operating their infrastructure relative to the required performance.

Knowledge & Skills:

  • Knowledge of big data/data warehousing/business intelligence application patterns
  • Knowledge of application troubleshooting, the root cause of failure analysis, and incident handling
  • Knowledge of agile methodologies and practices

Position Purpose:
The Big Data Systems Engineer II role aims to build, test, and maintain data processing systems involving large data sets. The role assumes responsibility for deploying data transformation code into production and, with the assistance of appropriate monitoring, proactively troubleshooting any issues that may arise. The role is well suited to an individual who has progressed in their career journey from a Data Engineer or Systems Engineer into a Big Data DevOps Engineer role in a Linux-based environment. Individuals who enjoy working at scale within a large, diverse team of specialists, and supporting a 24/7 operation will thrive in this role. An understanding of at least one DevOps tools chain is required along with exposure to scheduled data transformation on clustered compute (e.g., Hadoop/PySpark) or query engines.

Qualifications:

  • Degree or Diploma in a Technology-related field.
  • Associate level Linux Certification.
  • Azure Associate Certification.

Experience:

  • +4 years’ experience in a technology-related field as a Data or Systems Engineer for Linux hosted data management application managed through scripted automated deployment
  • Experience deploying and managing applications and databases in the cloud
  • Experience implementing SDLC automation and testing for data transformation or data query jobs
  • Experience configuring and deploying infrastructure using Terraform
  • Experience implementing logging and monitoring data applications using the environment and application-specific logging
  • Experience implementing high availability, and disaster recovery for big data jobs

Job objectives:
Work with a team of specialists in operations to ensure data applications and the jobs running on them are secure and available in production:

  • Implement appropriate monitoring and logging to ensure potential failure can be detected, diagnosed, and remediated before services are impacted.
  • Deploy highly available disaster recovery infrastructure as appropriate for cluster-based solutions.
  • Work with the security team to ensure the application and its infrastructure is secure.
  • Work with development team leads to ensure security is built into their development efforts and deploy additional security measures such as WAFs into the production environment.
  • Exploit security services available from the cloud provider to monitor and ensure the security of the environment.

Work with the ETL development teams and Data Engineers to deploy and ensure the deploy ability of infrastructure, package applications, and data transformation jobs:

  • Use existing SDLC toolchains to deploy cluster-based data applications and the data transformation and queries that run on them.
  • Maintain and upgrade existing SDLC toolchains.
  • Work with the development managers to support them in SDLC automation and developing code to deploy infrastructure using existing SDLC toolchains.

Work with application owners to manage the cost infrastructure deployed for applications both in development and production:

  • Tag resources appropriately so that the system can monitor their cost.
  • Tactically work with the application and data artifact owners to implement tactical cost savings where possible while maintaining required performance.
  • Support system and data artifact owners in monitoring, predicting, and optimising the cost of operating their infrastructure relative to the required performance.

Knowledge & Skills:

  • Knowledge of big data/data warehousing/business intelligence application patterns
  • Knowledge of application troubleshooting, the root cause of failure analysis, and incident handling
  • Knowledge of agile methodologies and practices

Desired Skills:

  • Data or Systems Engineer
  • Linux hosted data management
  • SDLC automation

Learn more/Apply for this position