DATA ENGINEER (Real Time)
About client:
At RemoteStar we are currently hiring for a client who is a world-class iGaming operator offering various online gaming products across multiple markets, both through their proprietary gaming sites and partner brands.
Their iGaming platform is central to their strategy, supporting over 25 online brands and growing, and it's used by hundreds of thousands of users worldwide. Our client embraces a Hybrid work-from-home model, with the flexibility of working three days in the office and two days from home.
About the Data Engineer role:
In this role, you will contribute to the design and development of Real-Time Data Processing applications to fulfil business needs.
For any Technical Data wiz out there, This is the perfect environment to put your skills to the test by building a consolidated Data Platform with innovative features and most importantly joining a bunch of talented and fun group of people.
What you will be involved in:
Development and Maintenance of Real-Time Data Processing Applications by using frameworks and libraries such as: Spark Streaming, Spark Structured Streaming, Kafka Streams and Kafka Connect.
Manipulation of Streaming Data: Ingestion, Transformation and Aggregation.
Keeping up to date on Research and Development of new Technologies and Techniques to enhance our applications.
Collaborating closely with the Data DevOps, Data-Oriented streams and other multi-disciplined teams.
Comfortable working in an Agile Environment involving SDLC.
Familiar with the Change and Release Management Process.
Have an investigative mindset to be able to troubleshoot – thinking outside the box when it comes to troubleshooting problems and incident management.
Full ownership of Projects and Tasks assigned together with being able to work within a team.
Able to document well processes and perform Knowledge Sharing sessions with the rest of the team.
You’re good with:
Have strong knowledge in Scala.
Knowledge or familiarity of Distributed Computing like Spark/KStreams/Kafka
Connect and Streaming Frameworks such as Kafka.
Knowledge on Monolithic versus Microserivce Architecture concepts for building large-scale applications.
Familiar with the Apache suite including Hadoop modules such as HDFS, Yarn, HBase, Hive, Spark as well as Apache NiFi.
Familiar with containerization and orchestration technologies such as Docker, Kubernetes.
Familiar with Time-series or Analytics Databases such as Elasticsearch.
Experience with Amazon Web Services using services such as S3, EC2, EMR, Redshift.
Familiar with Data Monitoring and Visualisation tools such as Prometheus and Grafana.
Familiar with software versioning tools like Git.
Comfortable working in an Agile environment involving SDLC.
Have a decent understanding of Data Warehouse and ETL concepts – familiarity with Snowflake is preferred
Have strong analytical and problem-solving skills.
Good learning mindset.
Can effectively prioritize and handle multiple tasks and projects.