Posted on: 03 December 2023
ID 894370

Senior Data Engineer

Kick-start your career in the online gaming world and experience the very latest in technology and innovation.

Purpose Of The Role

This role requires a motivated, energetic, inquisitive, and highly numerate person who has a solid technical base with an appreciation of the value that can be derived from well-designed data systems.

The role will be responsible for building and maintaining highly performant data pipelines that facilitate data management and data analytics solutions against very large data sets.

The ideal candidate is a very experienced and highly technical data delivery specialist who enjoys optimising data systems and building them from the ground up. The Data Engineer role will be responsible for ingesting and maintaining large data sets in an optimal manner that is aligned to the reference architecture and solution design provided by the solution architects. The person must be self-managed and comfortable supporting the data needs of multiple teams, systems, and products. The right candidate will be excited by the prospect of playing a key role in modernising our data processing systems.

Duties Include, But Not Limited To
  • Work as part of a multi-disciplinary data focussed team.
  • Develop greenfield projects using the both our Azure cloud platform and on-prem setups.
  • Assemble large, complex data sets that meet functional and non-functional business requirements.
  • Develop and maintain operational and analytical data systems.
  • Create robust ELT/ETL products for batch, micro-batch and near real-time data pipelines.
  • Develop and release data components in a DataOps environment.
  • Follow test-driven development practices.
  • Demo work to both technical and non-technical stakeholders.
  • Create documentation and training material for the solutions being delivered.
  • Guide and mentor junior team members
This job description is not intended to be an exhaustive list of responsibilities. The job holder may be required to complete any other reasonable duties in order to achieve business objectives.

Essential Criteria
  • Working knowledge of event-driven systems, message queuing, stream processing, and highly scalable big data pipelines and data stores (e.g., Confluent Cloud, Apache Kafka, Apache Flink, RabbitMQ, Azure Event Hubs, Kinesis etc.).
  • Skilled and experienced in the Azure or AWS cloud platforms.
  • Advanced working knowledge of SQL (DDL, DML, JSON, XML) and extensive experience in dealing with large datasets and managing incremental batch loading methodologies (CDC, CT, CDO).
  • Advanced understanding of relational data structures including keys, constraints, and triggers.
  • Performance tuning and optimization of RDBMS
  • Highly skilled and experienced in using relational or NoSQL database technologies (MS SQL Server, MongoDB, CosmosDB, etc.) in an environment with high data volumes and many Transactional systems.
  • Understand how to design and implement a conceptual, logical and physical data model that supports the needs of the organization.
  • Solid understanding and experience in data modelling, data management and governance methodologies
  • Good understanding of data-related frameworks, methodologies, and patterns
  • Experience in designing and developing ETL/ELT processes and pipelines for large data sets
  • Strong analytic skills related to working with structured, semi-structured and unstructured data sets
  • Proficiency in Python, Java, or Scala
  • Experienced in implementation of CICD pipelines through technologies such as GitLab, Azure DevOps etc
  • Experience deploying data systems in an Infrastructure-as-Code (IaC) manner, preferably using Terraform.
  • Experience supporting and working with cross-functional teams in a dynamic environment.
  • Communicate effectively with both technical and non-technical stakeholders.
  • Demonstrates consistent behavior aligned to the Combiined F and Organizational Culture
Desirable Criteria
  • Implementing security, disaster recovery, high availability, auditing, monitoring and alerting solutions in Azure.
  • Understand costing and optimizing spending in cloud platforms such as Azure, AWS, GCP etc
  • Deep understanding of C# concepts and advanced programming techniques.
  • Understanding of advance software development principles and patterns such as SOLID
  • Deep understanding of Python concepts and advanced programming techniques.
  • Understanding of implementation of alternative data modelling theories such as Data Vault
  • Test/Business-Driven Development
  • Experience with MS Power BI, Tableau or any similar data visualization tool that enables self service data analysis
Person Specifications
  • Adaptability
  • Ownership & Accountability
  • Initiating Action
  • Resilience
  • Team Orientation
  • Integrity
  • Innovative
  • Please note we will apply relevance to our Talent Management and Talent Development Programme as part of our recruitment process.
  • Shortlisted candidates may need to complete an assessment.
This position requires trust and honesty it has access to customers financial details - therefore a credit and criminal record check will be conducted. The qualifications identified herein are an inherent job requirement; therefore, a qualification verification check will be done. By applying for this role, and supplying the necessary details, you hereby grant us permission to apply for these checks. This will be done in a confidential manner, and solely for the purposes of verification.

Should you not hear from us within 2 weeks, please deem your application as unsuccessful.

The perfect place to work, play and grow!
Occupation:
IT, computing jobs


This job offer is not active at the moment.
Apply for a job
You have already applied to this job position
Save ad