Job summary
Are you passionate about harnessing technology to deliver business outcomes? Would you like to be part of the data transformation for a world class university with a mission to benefit society through excellence in science, engineering, medicine and business?
Working within a collaborative and dynamic team composed of Data Engineers and Business Intelligence Analysts you will be instrumental in enabling our data platform to be scaled across the organisation. This role enables you to be hands on working with Azure Data Factory and Databricks, crafting enterprise-level Data Engineering solutions from a variety of different data sources. Responsibilities include developing data-pipelines, supporting the Azure data platform and developing Imperial College London's data and analytics capabilities.
Imperial College London is one of world's top universities, the UKs top research university (2022 Research Excellence Framework) and an outstanding place to work. As a Data Engineer there is the potential to be part of the delivery of an industry changing modern data platform.
Duties and responsibilities
- Drive development of Imperial College’s cloud data platform
- Work within empowered, cross functional teams to deliver data solutions that drive value across Imperial College
- Be a lead for data security expertise, sharing knowledge across the data and analytics team and wider
- Continuously improve and develop our data platform, ensuring Imperial College remains on the forefront of data and analytics technologies
Essential requirements
Essential requirements
- Clear communicator, both written and verbal
- Problem solving seeking to find the right solution to deliver value
- Knowledge of designing data pipelines
- Experienced in working with Python and PySpark programming language
- SQL experience in a working environment
- Worked with Azure Data Factory
- Utilised Databricks to deliver data solutions/products
- Exposure to RDBMS (SQL Server, Oracle.)
- Exposure and understanding of dimensional modelling (star schema)
- Experience of embedding security into data solutions
- Willingness to continually learn and improve, including prototyping and testing new technologies and ways of working
- Experience with Agile delivery methods
Desirable
- Exposure to Lakehouse Architecture
- Experience working with APIs
- Experience working with DevOps solutions
- NoSQL database
Further information
Candidates will need to complete an online application.
Should you require any further details on the role please contact: Jose Vidal Fidel, [email protected]
Hybrid working may be considered for this role. Staff working in roles that are suitable for hybrid working will normally be expected to work 60% of their time onsite. The opportunity for hybrid working will be discussed at interview.
More information is available on the following web page: Work Location Framework | Administration and support services | Imperial College London