Junior Data Engineer

  • Creditsafe
  • Cardiff Bay, United Kingdom
  • 07/05/2022
Full time Data Engineering Big Data Statistics Software Engineering jobs

Job Description

Job details

Job type

Full Job Description

Here at Creditsafe, we are looking for a Junior Data Engineer, to join the Data engineering team.
Our success over the last 25 years and our ongoing growth can be attributed to our people and our strong culture. Culture and engagement really are part of our DNA here at Creditsafe and we take pride in making Creditsafe a great place to work. It’s important to us that people can be themselves, feel a sense of professional and personal growth and feel part of a global community. We offer a varied range of benefits that support a good work-life balance, including a hybrid approach to work, which enables you the flexibility needed to thrive.
The Data Engineering department at Creditsafe Group comprises four delivery teams covering over 1000 data pipelines, scorecard processing, and also data services to facilitate access to the data via API. Reporting to the Director of Data Engineering, we are a friendly and supportive team working across Creditsafe’s data universe, building out and managing our centralised data warehouse, and delivering processed “business ready” data to the downstream products and services used by Creditsafe’s customers. We engage with stakeholders at all levels of the organisation to achieve the business’s objectives.
We use modern tools and methodologies, leveraging cloud services (AWS), Apache Airflow, DBT (Data Build Tool) and of course, SQL and Python. We work in an agile manner, delivering iteratively through a metadata-driven approach which allows us to generate and deploy consistent, repeatable code.”
You will be working closely with the data vault team building specific systems facilitating the transition from traditional data processing and approaches to a Data Vault based approach. The role will define and build data pipelines that will improve data-informed decision-making within the business. This is an opportunity to work with large volumes of data and gain exposure to big data architectures.

  • Play a hands on role as part of an Agile team to develop, test and maintain high quality systems that fulfil business needs.
  • Extracting data from various files, systems, cloud sources, databases and APIs through writing and executing code (SQL, Python and similar)
  • Cleaning and combining offline, online or mixed sources into datasets. Building in manual or automatic validation and accuracy checks. Making use of Python, SQL or specialist Big Data frameworks
  • Help support the team in maintaining existing software and data infrastructure
  • Strong focus on quality. Execute practices such as continuous integration and test driven development to enable the rapid delivery of working code.
  • Write documentation of new processes and products you’ve developed so that knowledge is shared
  • Create pattern based data pipeline using Python and SQL using industry standard loading patterns in accordance with guidelines set by the Senior Data Engineers
  • Help to design, build and launch new data models
  • The responsibilities detailed above are not exhaustive and you may be requested to take on additional responsibilities deemed as reasonable by their direct line manager.


  • Development experience within a commercial environment
  • Knowledge of Agile development methodologies
  • Some experience of working with data sources and Python
  • Knowledge of SQL programming and code optimisation.
  • Awareness of cloud technology particularly
  • Knowledge of automated delivery processes
  • Some experience designing and building data pipelines
  • Understanding of best engineering practices (handling and logging errors, system monitoring and building human-fault-tolerant applications)
  • Ability to write efficient code and comfortable undertaking system optimisation and performance tuning tasks
  • Experience working within a unix based environment
  • Comfortable working with relational databases such as PostgreSQL, MySQL, MariaDB or Redshift
  • Experience of reading and writing data using Python and SQL
  • Understanding of Agile development methodologies
  • Python unit testing Frameworks such as pytest and nose
  • Familiarity with cloud technology, preferably AWS
  • Good understanding of GIT
  • Knowledge of automated delivery processes

Desired Skills

  • Implemented data pipelines using Apache Airflow
  • Execute Data Transformations in SQL via DBT
  • Understanding of MPP data platforms such as Apache Hive, Presto, Spark, Redshift
  • Experience of working with large datasets

Creditsafe is an equal opportunities employer that values diversity. Please contact Creditsafe if there is any support you need with your application.

Job Type: Full-time