Spark Scala Developer

Job Type: Permanent
Posted: over 2 years ago
Contact: Sue Templeton
Discipline:
Reference: 237869_1628136436

Acting as the software engineer of the group to perform design and implementation tasks for our various platforms. Main accountabilities are

  • Designing and implementing ingest, processing and curation solution using Big data technologies
  • Identify, resolve, and/or escalate all issues that threaten the successful delivery of assigned initiatives
  • Utilizing open source tool sets related to the Hadoop ecosystem.
  • Create high quality API's for data accessibility on the Hadoop platform that ultimately integrate with the broader environment.
  • Using big data analytics techniques provide data and analytics output as required to the broader community including business partners in a manner they can understand to ensure communication and buy in.
  • Producing high quality code with excellent test coverage
  • Code reviews and Efficient and effective code maintenance
  • Analysing product requirements and breaking down into user stories and tasks
  • Estimating and planning to assist with the development of the product roadmap
  • Mentoring junior data engineers
  • Developing build, test and deployment automation
  • Research and development of potential new technologies
  • Defining processes and standards within the team and contributing to the continuous improvement of the product development process
  • Producing a high standard of technical documentation
  • Working with customers to understand requirements and technical issues
  • Ability to provide technical expertise for new and ongoing initiatives to work with project stakeholder to come up most feasible technical solution.

Technology skills

  • Proficiency in Scala and Java is essential.
  • Must have experience working with real-time ingestion, processing, and curation of data at big data scale.
  • Big Data developer experience with hortonworks or Cloudera would be beneficial.
  • Fundamental knowledge of Hadoop components such as Kafka, Spark, Hive, HBase etc. is mandatory.
  • Previous experience developing data pipelines using Spark.
  • Past experience with NoSQL data stores such as ElasticSearch, MongoDB, Couchbase etc would be beneficial.
  • Experience on data analytics using SparkML or other ML tools such as R is a good to have.

Qualifications/experiences

  • Demonstrated technology subject matter expertise.
  • Familiarity with technology hands on configuration.
  • Able to communicate effectively with stakeholders.
  • Able to comprehend and deliver to a schedule.
  • Ability to work under pressure to deliver to near-term deadlines.
  • Ability to work in a matrix organisation, taking direction from both line and project management.
  • Self-starter, able to get up to speed quickly and work independently.
  • Solid experience working in software development
  • Experience working on Big data technologies