ETL Developer – Software Engineer

Full time
Permanent
Leeds
Competitive salary

Your new purpose (in a nutshell)?
To support the Data Science & Strategy (DSS) function in delivering numerous data initiatives supporting workstreams relating to IPF group change and governance processes and system developments as a key part of IPF’s data programme.

Working as an experienced ETL developer implementing strategic data initiatives. Using your experience and expertise to provide complex data transformations to drive the development of data science, business intelligence and insights, ensuring data sets are extracted from a wide range of business systems, transformed and loaded in a data lake - implementing plans, with project and data teams delivering data requirements to improve insights via data analysis and modelling.

What you will be required to deliver in your new role?
• Match business requirements and technical specifications, developing appropriate solutions across multiple systems personally delivering solutions.
• Provision of estimates and timings to Delivery Manager for the creation and support of structured code and test criteria within work packages.
• Translation of functional specifications and story elaboration to for the development of code/packages, based on in-house standards supporting data programme initiatives achieving agreed scope, data conformance and delivery plans.
• Critical evaluation of information gathered from multiple sources; reconcile, conflict, classifying information into logical categories.
• Documentation of sources to target mappings for both data integration as well as web services.
• Documentation and coding data quality and/or transformation rules, that can be easily understood by data science team members.
• Documentation and contribution using online collaboration tools JIRA/Confluence.
• Installation and configuration of Cloud server environments for the purposes of development/testing, including supporting the creation of test data sets.
• Contributions to the development of coding standards, development techniques and data compliance standards.
• Collaboration with data scientists and data analysts in data profiling and data source transformation processes enabling and/or enhancing advanced analytical modelling.
• Collaboration with third party suppliers and products, as well as in-country IT teams to identify and deliver appropriate technologies and solutions.
• Joint accountability with architects and data analysts to ensure conformance to group data governance policies regarding risk and data protection guidelines.
• Willingness to learn new technologies and move between different technology stacks



What skills are we looking for?

Qualifications:
• AWS Certified Developer (Desirable)
• Pentaho ETL Data Integration certified (Desirable)
• Degree qualified (2:2 or above), or equivalent (Desirable)

Skills:
• Agile/iterative development frameworks.
• Fluent English speaking.
• Excellent verbal & written communication.
• Analytically minded, abilities to develop insights from data sets.

Knowledge:
• Deep understanding of data and analytics, gained within data warehouse and/or data lake environments
• Demonstrable knowledge of Amazon Web Services and/or Pentaho Data Integration – data management technologies
• Deep knowledge of T-SQL and one or more languages of choice (Java, Scala, Ruby, Python, Javascript, C#)
• Structured, semi-structured and unstructured data types and functional strengths/weaknesses
• Working knowledge of Change Management processes and Project Management methodologies
• Online collaboration tools JIRA/Confluence.
• Financial services IT solutions
• Data modelling and mapping techniques
• Data quality and cleansing processes
• IT industry and ways of working within multi-disciplinary teams

Experience:
• Working within agile/lean delivery teams and frameworks.
• Code development using T-SQL, version control and change management process controls.
• With processes enabling data masking and anonymization.
• Data source identification, data profiling, interpretation of patterns and functional specifications.
• Assessment and improvement of data quality and latency factors.
• Demonstrable experience within data warehouse and/or data lake solution environments.
• Data quality, MDM, ETL, ELT and CDC processes.
• On-premise and cloud based data management solutions.
• Demonstrable experience with Amazon Web Services data management technologies (Aurora, Glue, Kinesis, SageMaker, RDS) MongoDB.
• Demonstrable experience with Data Integration patterns and/or Cloud Data Warehouse platforms, AWS Redshift, Snowflake.
• Operating in an international environment.
• Financial services would be highly beneficial.

Other: You have a desire to work internationally and are willing and able to travel overseas

If this sounds like the role for you then please press the APPLY button below.


Your application will be treated with absolute confidentiality.