Strong proficiency in Python and pandas (or Polars) with a proven track record in data systems., Experience with data formats like JSON, XML, and CSV, and transforming unstructured data., Familiarity with cloud-native tools, particularly AWS services like Lambda and Step Functions., Interest in using large language models for data enrichment and transformation..
Key responsibilities:
Build and maintain reliable and scalable data pipelines in Python.
Utilize LLMs for data cleansing, enrichment, and classification tasks.
Collaborate with product and AI engineering teams to provide trustworthy data for prototypes.
Design workflows to convert noisy, semi-structured data into actionable insights.
Report This Job
Help us maintain the quality of our job listings. If you find any issues with this job post, please let us know.
Select the reason you're reporting this job:
Arrow Global Group Financial Services Large
1001 - 5000
Employees
About Arrow Global Group
Established in 2005, Arrow Global is a leading European investor and asset-manager in non-performing and non-core assets. We identify, acquire and manage secured and unsecured loan and real estate portfolios from financial institutions, such as banks and credit card companies.
We play an active role in helping financial institutions reduce their balance sheets and recapitalise in order to increase mainstream lending. By purchasing and managing non-performing loans and other non-core assets, we provide valuable capital and expertise to a growing European market.
We are a regulated business in all of our European markets, managing over £70 BN assets under management across five countries with over 2,500 employees.
Our Purpose is to help all of our stakeholders Build Better Financial Futures and our Vision is to be the Innovative and Valued Partner in Credit and Asset Management.
Our Values, we're trusted and valued, we succeed together, we do the right thing and we're brave and creative, bind our organisation together and guide our behaviours and decision making at all levels of the company.
Our Data Engineer will help to build and scale the data infrastructure that powers our AI products. This role is hands-on and technically deep — ideal for someone who cares about data quality, robustness, and automation. Working closely with AI engineers to design pipelines that do more than move data — they clean, enrich, and understand it, increasingly using large language models and agents to automate complex steps in the process.
About the Team
This role will be part of a team that is flat-structured, best-idea-wins culture and where engineers shape product direction. We operate a supportive culture that values ownership as we want people who take responsibility but aren’t afraid to ask for help where needed. Whilst our offices and extend teams are based in Manchester and London, we also offer flexibility to work from anywhere (UK) for this role — though we’re Europe-focused and love getting together for hackathons and team problem-solving when it matters.
About the role
Building and maintaining data pipelines in Python, with a focus on reliability, transparency, and scale.
Using LLMs to assist with data cleansing, enrichment, classification, and contextual tagging.
Experimenting with AI agents to automate complex research tasks and structured data extraction.
Working with product and AI engineering teams to feed trustworthy data into fast-moving prototypes.
Designing workflows that transform noisy, semi-structured data into actionable insight.
Supporting experimentation and iteration — shipping fast and learning from what works.
What we're looking for & more
Strong proficiency in Python and pandas (or Polars), and a track record of delivering working data systems.
Experience with common data formats (JSON, XML, CSV) and transforming unstructured data.
Familiarity with modern cloud-native tooling (we use AWS — especially Lambda and Step Functions).
Interest or experience in using LLMs for tasks like data enrichment or transformation.
A mindset that treats pipelines as products — robust, debuggable, and always improving.
Curiosity about how AI can go beyond the model — helping automate research and discovery.
It would be beneficial (not essential) if you have experience with tools like LangChain, Haystack, Pandas AI, or vector databases as well as any prior projects involving agents for data understanding or research automation.
Sound like you? Great! Whilst a CV tells us part of your story, we would love to see a short summary about you with any relevant links to Loom - following this we will reach out to you for a teams interview if suitable.
Whilst this position can be done from anywhere in the UK,you must already hold the relevant right to work in the UK as we unfortunately can't provide sponsorship for the role.
Required profile
Experience
Industry :
Financial Services
Spoken language(s):
English
Check out the description to know which languages are mandatory.