Position Summary:
Load the data from local and group sources onto the shared platforms that are necessary for insight, analysis, and commercial actions.
Building applications that make use of large volumes of data and generate outputs that allow commercial actions that generate incremental value.
Support local markets and group functions in obtaining business value from the data.
Key Responsibilities:
- Build transformations to produce enriched data insights using automated testing and deployment by defined standards.
- Make contributions to the process of defining best practices for the agile development of applications to run on the Big Data Platform.
- Build batch and real-time data pipelines, using automated testing and deployment by defined standards.
- Source data from a variety of different sources, in the correct format, meeting data quality standards and assuring timeous access to data and analytical insights according to module designs.
- Develop highly performant, scalable, and stable Big Data cloud-native applications using defined patterns and methods.
What you’ll have:
- Working experience with other distributed technologies such as Cassandra DB, MongoDB, Elastic Search, and Flink would be desirable.
- Working experience with AWS or GCP would be desirable.
- Java and Python programming ability would be an advantage.
- Experience working with serverless services such as Openshift, GCP, or AWS.
- Experience with performance tuning streaming-based applications for real-time data processing using Kafka, Confluent Kafka, AWS Kinesis, GCP pub/sub, or similar.
- Relevant cloud certification at the professional or associate level.
- 2 to 5 years BI or related software development experience.
- 3 year IT or IS degree or diploma or related field is essential.
- Agile exposure, Kanban, or Scrum.
- Matric/ Grade 12 is essential.