Full stack engineer datarole purposethe gcdr application (hosted on hadoop platform) is primarily concerned with entity resolution (matching internal and external party data together to form single view of the customer).gcdr covers all four lines of business across 57+ markets. It also supports other services such as ui, reporting, network generation, apis etc.the entity resolution and network generation services use quantexa software, however many other elements are hsbc build.main activities:develop, program, and maintain applications, creating new features, enhancing existing ones, etcwork with different aspects of apache spark ecosystem, including etl pipelines, data transformations and optimisationbecome familiar with the multiple applications and use cases that are supported on this platformassist in the root cause analysis of production issuesrequirementsstrong communication skills, with the ability to convey technical detail in a non technical languagebackground in hands-on technical development, with at least three years of industry experience in a data engineering role or engineering equivalentproficiency in developing large scale data processing applications using apache spark with scala including etl pipelines, data transformations and optimization techniquesexpertise building and deploying production level data processing batch systems maintained by application support teams.experience with a variety of modern development tooling (e.g. Git, gradle, nexus) and technologies supporting automation and dev ops (e.g. Jenkins, docker)experience working in an agile environmenta strong technical communication ability with demonstrable experience of working in rapidly changing client environments.knowledge of testing libraries of common programming languages (such as scala test or equivalent). Knows the difference between different test types (unit test, integration test) and can cite specific examples of what they have written themselves.optional/ nice to have: