Join Our Client as an SC Cleared DSR Chief Software Engineer & Team Leader Location: Milton Keynes typically 2 days per week with some travel to London Salary: 88,000 and 95,000 + discretionary bonuses Must be willing and eligible to go through the SC Clearance process Are you a seasoned software engineering professional with a passion for leading innovative solutions in the public sector? Our client, a reputable organisation supporting UK Government departments, is seeking a proactive and strategic Chief Software & Systems Engineer to oversee critical digital record and sensitivity review programmes. What you'll be doing: Leading the delivery and evolution of Digital Records & Archive Management (DRAM) and Digital Sensitivity Review (DSR) systems. Managing and maintaining system architecture, ensuring security, compliance, and operational integrity. Overseeing a talented team of Software and Secure Cloud Engineers, fostering efficiency and continuous improvement. Collaborating with project teams, partners, and stakeholders to deliver secure, robust, and user-friendly solutions. Developing strategies, policies, and procedures aligned with government security standards. Supporting multiple projects, managing budgets, and ensuring timely delivery. Ensuring systems meet legislative and security requirements. What you'll bring: Extensive knowledge of software engineering principles, especially within secure, on-premise, and cloud environments. Proven leadership skills to manage and motivate technical teams. Strong project management capabilities across multiple initiatives. Excellent problem-solving, communication, and stakeholder engagement skills. Strategic thinker with a focus on security, compliance, and innovation. Familiarity with industry trends such as AI, Machine Learning, and cloud technologies Experience of Managing cloud services (e.g., AWS, Azure, GCP). Experience overseeing integration, and end-to-end tests using tools such as Junit, Test Containers, Vitetest and StoryBook. Expertise in developing and deploying onto a Windows and Linux environment. Desirable skills: Experience with data science, analytics, and AI solutions. Knowledge of secure private/public cloud platforms, virtualisation, and containerisation. Background working within the public sector or government agencies. Join a forward-thinking organisation committed to innovation, security, and public service excellence. We look forward to hearing from you!
Apr 29, 2026
Full time
Join Our Client as an SC Cleared DSR Chief Software Engineer & Team Leader Location: Milton Keynes typically 2 days per week with some travel to London Salary: 88,000 and 95,000 + discretionary bonuses Must be willing and eligible to go through the SC Clearance process Are you a seasoned software engineering professional with a passion for leading innovative solutions in the public sector? Our client, a reputable organisation supporting UK Government departments, is seeking a proactive and strategic Chief Software & Systems Engineer to oversee critical digital record and sensitivity review programmes. What you'll be doing: Leading the delivery and evolution of Digital Records & Archive Management (DRAM) and Digital Sensitivity Review (DSR) systems. Managing and maintaining system architecture, ensuring security, compliance, and operational integrity. Overseeing a talented team of Software and Secure Cloud Engineers, fostering efficiency and continuous improvement. Collaborating with project teams, partners, and stakeholders to deliver secure, robust, and user-friendly solutions. Developing strategies, policies, and procedures aligned with government security standards. Supporting multiple projects, managing budgets, and ensuring timely delivery. Ensuring systems meet legislative and security requirements. What you'll bring: Extensive knowledge of software engineering principles, especially within secure, on-premise, and cloud environments. Proven leadership skills to manage and motivate technical teams. Strong project management capabilities across multiple initiatives. Excellent problem-solving, communication, and stakeholder engagement skills. Strategic thinker with a focus on security, compliance, and innovation. Familiarity with industry trends such as AI, Machine Learning, and cloud technologies Experience of Managing cloud services (e.g., AWS, Azure, GCP). Experience overseeing integration, and end-to-end tests using tools such as Junit, Test Containers, Vitetest and StoryBook. Expertise in developing and deploying onto a Windows and Linux environment. Desirable skills: Experience with data science, analytics, and AI solutions. Knowledge of secure private/public cloud platforms, virtualisation, and containerisation. Background working within the public sector or government agencies. Join a forward-thinking organisation committed to innovation, security, and public service excellence. We look forward to hearing from you!
Genie + AI/BI Sales Specialist While candidates in the listed location(s) are encouraged for this role, candidates in other locations will be considered. London Munich Amsterdam Paris Databricks is seeking a Genie + AI/BI Sales Specialist to help enterprise customers transform how business users interact with data. This high-impact role sits within the AI Go-To-Market team and partners closely with Enterprise Account Executives to drive adoption of Databricks AI/BI and Genie. You will help organizations move beyond static dashboards to governed, conversational, AI-powered analytics at the center of the convergence of business intelligence, data platforms, and generative AI. Enterprise analytics is rapidly evolving from dashboards and static reporting to conversational, AI-driven decision platforms. Databricks AI/BI and Genie empower business users to securely interact with governed data using natural language, transforming the data platform into a true decision platform. If you want to be at the forefront of AI-powered analytics transformation at one of the fastest-growing data and AI companies in the world, this is your opportunity. The impact you will have Partner with Enterprise AEs to identify, qualify, and close AI/BI opportunities Engage C-level, analytics, and line-of-business leaders to modernize analytics strategies Displace or expand legacy BI platforms with AI-powered, governed analytics solutions Lead conversations around semantic governance, self-service analytics, and natural language data access Drive proof-of-value engagements and scale enterprise-wide adoption Align AI/BI initiatives to measurable business outcomes (productivity, speed to insight, revenue impact) Enable field teams and serve as a subject matter expert on modern analytics architectures What we look for Enterprise sales experience in BI, analytics, data platforms, or AI/ML Strong understanding of modern analytics architectures and data governance Ability to sell to both technical and business stakeholders Executive presence and experience navigating complex buying cycles Passion for AI and the impact of GenAI on enterprise analytics Experience operating in a specialist or overlay sales model Ability to translate technical capabilities into clear business value 7+ years of Enterprise Sales experience, exceeding quotas in larger accounts Bachelors Degree or equivalent experience About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide - including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 - rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark , Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone.
Apr 29, 2026
Full time
Genie + AI/BI Sales Specialist While candidates in the listed location(s) are encouraged for this role, candidates in other locations will be considered. London Munich Amsterdam Paris Databricks is seeking a Genie + AI/BI Sales Specialist to help enterprise customers transform how business users interact with data. This high-impact role sits within the AI Go-To-Market team and partners closely with Enterprise Account Executives to drive adoption of Databricks AI/BI and Genie. You will help organizations move beyond static dashboards to governed, conversational, AI-powered analytics at the center of the convergence of business intelligence, data platforms, and generative AI. Enterprise analytics is rapidly evolving from dashboards and static reporting to conversational, AI-driven decision platforms. Databricks AI/BI and Genie empower business users to securely interact with governed data using natural language, transforming the data platform into a true decision platform. If you want to be at the forefront of AI-powered analytics transformation at one of the fastest-growing data and AI companies in the world, this is your opportunity. The impact you will have Partner with Enterprise AEs to identify, qualify, and close AI/BI opportunities Engage C-level, analytics, and line-of-business leaders to modernize analytics strategies Displace or expand legacy BI platforms with AI-powered, governed analytics solutions Lead conversations around semantic governance, self-service analytics, and natural language data access Drive proof-of-value engagements and scale enterprise-wide adoption Align AI/BI initiatives to measurable business outcomes (productivity, speed to insight, revenue impact) Enable field teams and serve as a subject matter expert on modern analytics architectures What we look for Enterprise sales experience in BI, analytics, data platforms, or AI/ML Strong understanding of modern analytics architectures and data governance Ability to sell to both technical and business stakeholders Executive presence and experience navigating complex buying cycles Passion for AI and the impact of GenAI on enterprise analytics Experience operating in a specialist or overlay sales model Ability to translate technical capabilities into clear business value 7+ years of Enterprise Sales experience, exceeding quotas in larger accounts Bachelors Degree or equivalent experience About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide - including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 - rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark , Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit Our Commitment to Diversity and Inclusion At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics. Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone.
Purpose of the role To support the design, prototype and delivery of innovative, data-led products by combining market insight, advanced analytics and modern data platforms. Working in closely with Product Managers and Data Analysts, this role will convert data assets and models into scalable, commercially viable analytical insights, data visualisations and product features. The Product Innovation Analyst bridges commercial opportunity and technical feasibility, ensuring new products are designed with platform capabilities, data quality, performance and scalability in mind. Principal responsibilities Identify and define new data product opportunities enabled by advanced analytics, machine learning and large-scale data processing Evaluate and test emerging AI technologies and analytical techniques and their suitability for use within our Customer Intelligence Platform to unlock new product capabilities Drive product innovation from concept to launch, translating business and customer needs into technical product requirements and delivery specifications Lead rapid prototyping and proof-of-concept development using Databricks notebooks, analytical outputs and machine learning techniques to validate product concepts Own technical product definition, including data structures, feature sets, scoring methodologies, model architectures and delivery formats in collaboration with the Product Team Define and monitor technical success metrics (data coverage, refresh latency, model stability) alongside commercial KPIs to optimise product performance Product Innovation & Technical Design responsibilities: Identify new product opportunities enabled by advanced analytics, machine learning and large-scale data processing Identify new usage of existing attributes and products to create more value in existing data Translate business and customer needs into technical product requirements Support rapid prototyping and proof-of-concept development using Databricks notebooks and analytics outputs Define product-level data structures, feature sets, scoring outputs and delivery formats in collaboration with the Product Team Product Launch & Performance responsibilities: Define technical success metrics (data coverage, refresh latency, model stability) alongside commercial KPIs Support internal enablement by translating technical product detail into usable sales and client-facing materials Drive continuous optimisation using usage analytics, customer feedback and platform performance insights What success looks like in the role Clear, concise and insightful data analytics which enable sound business decisions based on fact Ability to translate data analysis into targeted information which can be converted into actionable improvements, based on specific client, sector, internal product need Cross functional collaboration to enable continued improvement of Sagacity's Product Suite through the delivery of robust data insights Ability to take accountability and ownership for client and internal deliverables Your efforts result in streamlined data analysis, product builds and reduced time to market Competencies and Behaviours 1 -3 years analytics / data science experience Practical knowledge of; Delta Lake architecture and versioned datasets, Data pipelines, orchestration and scheduling concepts Proficiency in analytical programming language such as python and/or SQL, with the ability to interrogate datasets and validate analytical outputs Experience designing data products using large-scale transactional, behavioural or marketing datasets Understanding of data modelling concepts (fact/dimension models, feature engineering, aggregations) Can balance time across multiple projects. Plans ahead working backwards from deadlines with all necessary steps e.g. testing, QA. Proactively identifies risk and suggests mitigation Is curious, sceptical, inquisitive, suggests 'next steps' analysis and translates analytical findings to actionable insight Flexible, self-motivated, good under pressure, has a commitment to personal development Excellent communication skills, both written and verbal, with a willingness to engage and influence others Commercial experience within Telecoms, Banking or Utilities industries; or within a data related consultancy company would be beneficial Able to travel throughout the UK Can be based at our London Office (min 2 days per week on site) Have the right to work in the UK
Apr 29, 2026
Full time
Purpose of the role To support the design, prototype and delivery of innovative, data-led products by combining market insight, advanced analytics and modern data platforms. Working in closely with Product Managers and Data Analysts, this role will convert data assets and models into scalable, commercially viable analytical insights, data visualisations and product features. The Product Innovation Analyst bridges commercial opportunity and technical feasibility, ensuring new products are designed with platform capabilities, data quality, performance and scalability in mind. Principal responsibilities Identify and define new data product opportunities enabled by advanced analytics, machine learning and large-scale data processing Evaluate and test emerging AI technologies and analytical techniques and their suitability for use within our Customer Intelligence Platform to unlock new product capabilities Drive product innovation from concept to launch, translating business and customer needs into technical product requirements and delivery specifications Lead rapid prototyping and proof-of-concept development using Databricks notebooks, analytical outputs and machine learning techniques to validate product concepts Own technical product definition, including data structures, feature sets, scoring methodologies, model architectures and delivery formats in collaboration with the Product Team Define and monitor technical success metrics (data coverage, refresh latency, model stability) alongside commercial KPIs to optimise product performance Product Innovation & Technical Design responsibilities: Identify new product opportunities enabled by advanced analytics, machine learning and large-scale data processing Identify new usage of existing attributes and products to create more value in existing data Translate business and customer needs into technical product requirements Support rapid prototyping and proof-of-concept development using Databricks notebooks and analytics outputs Define product-level data structures, feature sets, scoring outputs and delivery formats in collaboration with the Product Team Product Launch & Performance responsibilities: Define technical success metrics (data coverage, refresh latency, model stability) alongside commercial KPIs Support internal enablement by translating technical product detail into usable sales and client-facing materials Drive continuous optimisation using usage analytics, customer feedback and platform performance insights What success looks like in the role Clear, concise and insightful data analytics which enable sound business decisions based on fact Ability to translate data analysis into targeted information which can be converted into actionable improvements, based on specific client, sector, internal product need Cross functional collaboration to enable continued improvement of Sagacity's Product Suite through the delivery of robust data insights Ability to take accountability and ownership for client and internal deliverables Your efforts result in streamlined data analysis, product builds and reduced time to market Competencies and Behaviours 1 -3 years analytics / data science experience Practical knowledge of; Delta Lake architecture and versioned datasets, Data pipelines, orchestration and scheduling concepts Proficiency in analytical programming language such as python and/or SQL, with the ability to interrogate datasets and validate analytical outputs Experience designing data products using large-scale transactional, behavioural or marketing datasets Understanding of data modelling concepts (fact/dimension models, feature engineering, aggregations) Can balance time across multiple projects. Plans ahead working backwards from deadlines with all necessary steps e.g. testing, QA. Proactively identifies risk and suggests mitigation Is curious, sceptical, inquisitive, suggests 'next steps' analysis and translates analytical findings to actionable insight Flexible, self-motivated, good under pressure, has a commitment to personal development Excellent communication skills, both written and verbal, with a willingness to engage and influence others Commercial experience within Telecoms, Banking or Utilities industries; or within a data related consultancy company would be beneficial Able to travel throughout the UK Can be based at our London Office (min 2 days per week on site) Have the right to work in the UK
Reporting Analyst - SQL - SSRS - Power BI We are working exclusively with our client who are based in the heart of Bristol. We have an open vacancy for a Reporting Analyst - SQL - SSRS - Power BI. The role is also HYBRID. ROLE SUMMARY Supporting the Reporting & Analytics Manager with the design and development of the BI platform and advanced analytics, ensuring it: Is trusted and fit for purpose Is scalable and performant Leverages contemporary advanced analytics capabilities and visualisations Provides simple and efficient BI self-service reporting Supports a single analytical view of our data and information Uses agreed and defined metrics Is built in line with our information management strategy and guiding principles Delivers business value through action oriented insights Ensures "explainability" fit for audit in a regulatory controlled environment Working as part of the Reporting & Analytics team in close collaboration with the Data Engineering team and business areas to take responsibility and accountability for collaborative design and build of end-user visualisations, reports, and ad-hoc analysis ensuring all required documentation is complete prior to release and all standards and best practices are adhered to. Taking a lead in driving the progression of advanced analytics capabilities and visualisations in step with technology changes and business needs. Supporting the Reporting & Analytics team through technical peer reviews and the end-user community by providing clear explanation of interpreting provided information. KEY SKILLS Passionate about data and information with a strong understanding of data architecture principles and information "story telling" in order to maximise the value of our raw data Technology agnostic and able to adapt with agility as we evolve our information estate and technology stack towards contemporary technologies which may comprise tools such as PowerBI, SSRS, Tableau, Cognos 11, Qlik, MicroStrategy, and cloud based services within Amazon AWS and/or Microsoft Azure Good understanding of SQL & DAX Highly adept at visualising information through good use of chart types best suited to the situation Demonstrable understanding of advanced analytics, including predictive, with supporting capabilities such as 'R', Python desirable Strong problem solving and planning skills with a "can do" attitude to manage and mitigate risks to maintain delivery commitments Knowledge of the insurance industry and working within regulated environments would be desirable Demonstrable knowledge of Kimball methodologies, MI standards and best practices and standards Knowledge of different data storage paradigms such as Kimball, Data Vault 2.0, Data Lakes, and beyond desirable Due to the volume of applications received for positions, it will not be possible to respond to all applications and only applicants who are considered suitable for interview will be contacted. Proactive Appointments Limited operates as an employment agency and employment business and is an equal opportunities organisation We take our obligations to protect your personal data very seriously. Any information provided to us will be processed as detailed in our Privacy Notice, a copy of which can be found on our website
Apr 29, 2026
Full time
Reporting Analyst - SQL - SSRS - Power BI We are working exclusively with our client who are based in the heart of Bristol. We have an open vacancy for a Reporting Analyst - SQL - SSRS - Power BI. The role is also HYBRID. ROLE SUMMARY Supporting the Reporting & Analytics Manager with the design and development of the BI platform and advanced analytics, ensuring it: Is trusted and fit for purpose Is scalable and performant Leverages contemporary advanced analytics capabilities and visualisations Provides simple and efficient BI self-service reporting Supports a single analytical view of our data and information Uses agreed and defined metrics Is built in line with our information management strategy and guiding principles Delivers business value through action oriented insights Ensures "explainability" fit for audit in a regulatory controlled environment Working as part of the Reporting & Analytics team in close collaboration with the Data Engineering team and business areas to take responsibility and accountability for collaborative design and build of end-user visualisations, reports, and ad-hoc analysis ensuring all required documentation is complete prior to release and all standards and best practices are adhered to. Taking a lead in driving the progression of advanced analytics capabilities and visualisations in step with technology changes and business needs. Supporting the Reporting & Analytics team through technical peer reviews and the end-user community by providing clear explanation of interpreting provided information. KEY SKILLS Passionate about data and information with a strong understanding of data architecture principles and information "story telling" in order to maximise the value of our raw data Technology agnostic and able to adapt with agility as we evolve our information estate and technology stack towards contemporary technologies which may comprise tools such as PowerBI, SSRS, Tableau, Cognos 11, Qlik, MicroStrategy, and cloud based services within Amazon AWS and/or Microsoft Azure Good understanding of SQL & DAX Highly adept at visualising information through good use of chart types best suited to the situation Demonstrable understanding of advanced analytics, including predictive, with supporting capabilities such as 'R', Python desirable Strong problem solving and planning skills with a "can do" attitude to manage and mitigate risks to maintain delivery commitments Knowledge of the insurance industry and working within regulated environments would be desirable Demonstrable knowledge of Kimball methodologies, MI standards and best practices and standards Knowledge of different data storage paradigms such as Kimball, Data Vault 2.0, Data Lakes, and beyond desirable Due to the volume of applications received for positions, it will not be possible to respond to all applications and only applicants who are considered suitable for interview will be contacted. Proactive Appointments Limited operates as an employment agency and employment business and is an equal opportunities organisation We take our obligations to protect your personal data very seriously. Any information provided to us will be processed as detailed in our Privacy Notice, a copy of which can be found on our website
Data Solutions Architect Salary: 90K - 100K London (Hybrid) At Data Idols, we are working with a large-scale, data-driven organisation investing heavily in its data platform, analytics and AI capabilities. They are now looking for a Data Solutions Architect to provide design and architecture guidance across a broad portfolio of data initiatives. The Opportunity This is a senior architecture role focused on shaping how data solutions are designed and delivered across a complex enterprise data estate. You will work closely with product owners, engineering leads, data teams and senior stakeholders to understand business requirements and translate them into clear, robust high-level designs for engineering teams to build. The role is not focused on hands-on delivery. Instead, you will be responsible for articulating solution designs, documenting key decisions, defining reusable patterns and ensuring consistency across a shared data platform. You will support a modern cloud data platform, with exposure to areas such as analytics, reporting, data science, AI use cases and increasingly real-time data capabilities. Skills and experience Strong experience designing secure, scalable data solutions in complex environments Experience working with Azure-based data platforms Experience with Databricks or similar modern data platforms Good understanding of data warehouse, data lake or lakehouse architectures If you are looking for a new challenge and want to help shape data solutions across a large-scale cloud data platform, please submit your CV for initial screening and further details. Data Solutions Architect
Apr 29, 2026
Full time
Data Solutions Architect Salary: 90K - 100K London (Hybrid) At Data Idols, we are working with a large-scale, data-driven organisation investing heavily in its data platform, analytics and AI capabilities. They are now looking for a Data Solutions Architect to provide design and architecture guidance across a broad portfolio of data initiatives. The Opportunity This is a senior architecture role focused on shaping how data solutions are designed and delivered across a complex enterprise data estate. You will work closely with product owners, engineering leads, data teams and senior stakeholders to understand business requirements and translate them into clear, robust high-level designs for engineering teams to build. The role is not focused on hands-on delivery. Instead, you will be responsible for articulating solution designs, documenting key decisions, defining reusable patterns and ensuring consistency across a shared data platform. You will support a modern cloud data platform, with exposure to areas such as analytics, reporting, data science, AI use cases and increasingly real-time data capabilities. Skills and experience Strong experience designing secure, scalable data solutions in complex environments Experience working with Azure-based data platforms Experience with Databricks or similar modern data platforms Good understanding of data warehouse, data lake or lakehouse architectures If you are looking for a new challenge and want to help shape data solutions across a large-scale cloud data platform, please submit your CV for initial screening and further details. Data Solutions Architect
Hybrid - within commuting distance of our Redcar Campus The National Institute of Teaching (NIoT) has an unswerving commitment to high-quality, evidence-informed teacher education, and is on a mission to improve teacher and leader development across the education system. A key part of that mission is our Teacher Education Dataset (TED), an ambitious data and research initiative focused on building better evidence about teacher development and what makes teaching impactful. By bringing together education, research and digital infrastructure, TED will support new insight into how teachers develop, improve and thrive across their careers. About the role We are seeking a motivated and enthusiastic Senior Software Engineer to join our TED team and play an important role in building and maintaining the technical infrastructure that supports TED. Working at the intersection of software engineering, data and research, the postholder will help develop secure, reliable and scalable systems that enable robust analysis and contribute to high-quality, evidence-informed improvement across the sector. Corporate responsibilities To ensure that the responsibilities of the role are carried out in a way which reflects the mission and the values of the NIoT. To be aware of and observe all policies, procedures, working practices and regulations, and in particular to comply with policies relating to Equal Opportunities, Health and Safety, Confidentiality, Data Protection and Financial Regulations, reporting any concerns to an appropriate person. To comply with all reasonable management requests. Key responsibilities Design, develop, maintain and improve software and technical infrastructure that supports TED and related education research activity. Build secure, scalable and reliable systems that enable robust analysis of education and teacher development data. Apply software engineering and DevOps best practice to deliver high-quality, well-tested and maintainable technical solutions. Contribute to reproducible, transparent and efficient research workflows, pipelines and supporting tools. Work closely with developers, researchers and other colleagues to understand requirements and translate them into effective technical delivery. Contribute to architectural decisions and the ongoing improvement of platform design, developer experience, monitoring and software quality. Support the operation and maintenance of production systems and help troubleshoot issues where required. Produce and maintain clear technical documentation and contribute to wider project communication where appropriate. Contribute to a collaborative team environment, including supporting and mentoring colleagues where appropriate. Essential criteria Proven experience in software engineering, including building and maintaining production systems. Strong coding skills in relevant programming languages and experience of modern software development practices. Experience with version control, code review, testing and continuous integration. Ability to take ownership of complex technical systems, including feature development, maintenance and support. Strong technical judgement and problem-solving ability, including awareness of architectural trade-offs. Ability to work effectively in a multidisciplinary environment and communicate clearly with technical and non-technical audiences. Commitment to quality, security, maintainability and continuous improvement. Interest in applying software engineering to education, data and evidence-informed improvement. Desirable criteria Experience with Python, JavaScript or similar languages. Experience of Linux, Docker, CI/CD and DevOps tooling. Experience of database design, optimisation or data pipeline development. Experience working with sensitive data or secure analytical environments. Experience of research platforms, reproducible analytics or trusted research infrastructure. Experience in education, public sector or data-rich research settings. Experience mentoring others and contributing to team-wide engineering practice. Key benefits Generous Annual Leave 27 days holiday a year (plus 8 bank holidays). Flexible Working we offer flexible start and end working times, with hybrid working in place for all roles. Pension - Entry to the Local Government Pension Scheme. Salary Sacrifice Schemes we offer salary sacrifice schemes for bikes, cars and tech! Lifestyle Benefits We offer discounts on gyms, cinema, retail and much more! Parenthood Leave We offer above the statutory minimum for maternity, adoption and paternity leave. Working Environment We have a stimulating, supportive and rewarding working environment with a dedicated team of likeminded professionals. Support: Our Employee Assistance Programme (EAP) provides confidential support for employees on personal or work-related matters. Development: We fully support the development of our staff and ensure that you have high level of continuous professional development. We think a wide range of different work and educational experiences could support you to be successful in this role. We encourage applications from all backgrounds, communities and industries, and are committed to employing a team that has diverse skills, experiences and abilities. If you have any queries about this role, please contact our recruitment team. For more information and to apply, please visit our vacancies page. Closing date: 10.00am on Monday, 1 June 2026. We reserve the right to close this vacancy early if we receive a high volume of applications. This post requires a satisfactory enhanced disclosure from the Disclosure and Barring Service (DBS) with a Children s Barred List Check and Occupational Health Check is required as a condition of employment.
Apr 29, 2026
Full time
Hybrid - within commuting distance of our Redcar Campus The National Institute of Teaching (NIoT) has an unswerving commitment to high-quality, evidence-informed teacher education, and is on a mission to improve teacher and leader development across the education system. A key part of that mission is our Teacher Education Dataset (TED), an ambitious data and research initiative focused on building better evidence about teacher development and what makes teaching impactful. By bringing together education, research and digital infrastructure, TED will support new insight into how teachers develop, improve and thrive across their careers. About the role We are seeking a motivated and enthusiastic Senior Software Engineer to join our TED team and play an important role in building and maintaining the technical infrastructure that supports TED. Working at the intersection of software engineering, data and research, the postholder will help develop secure, reliable and scalable systems that enable robust analysis and contribute to high-quality, evidence-informed improvement across the sector. Corporate responsibilities To ensure that the responsibilities of the role are carried out in a way which reflects the mission and the values of the NIoT. To be aware of and observe all policies, procedures, working practices and regulations, and in particular to comply with policies relating to Equal Opportunities, Health and Safety, Confidentiality, Data Protection and Financial Regulations, reporting any concerns to an appropriate person. To comply with all reasonable management requests. Key responsibilities Design, develop, maintain and improve software and technical infrastructure that supports TED and related education research activity. Build secure, scalable and reliable systems that enable robust analysis of education and teacher development data. Apply software engineering and DevOps best practice to deliver high-quality, well-tested and maintainable technical solutions. Contribute to reproducible, transparent and efficient research workflows, pipelines and supporting tools. Work closely with developers, researchers and other colleagues to understand requirements and translate them into effective technical delivery. Contribute to architectural decisions and the ongoing improvement of platform design, developer experience, monitoring and software quality. Support the operation and maintenance of production systems and help troubleshoot issues where required. Produce and maintain clear technical documentation and contribute to wider project communication where appropriate. Contribute to a collaborative team environment, including supporting and mentoring colleagues where appropriate. Essential criteria Proven experience in software engineering, including building and maintaining production systems. Strong coding skills in relevant programming languages and experience of modern software development practices. Experience with version control, code review, testing and continuous integration. Ability to take ownership of complex technical systems, including feature development, maintenance and support. Strong technical judgement and problem-solving ability, including awareness of architectural trade-offs. Ability to work effectively in a multidisciplinary environment and communicate clearly with technical and non-technical audiences. Commitment to quality, security, maintainability and continuous improvement. Interest in applying software engineering to education, data and evidence-informed improvement. Desirable criteria Experience with Python, JavaScript or similar languages. Experience of Linux, Docker, CI/CD and DevOps tooling. Experience of database design, optimisation or data pipeline development. Experience working with sensitive data or secure analytical environments. Experience of research platforms, reproducible analytics or trusted research infrastructure. Experience in education, public sector or data-rich research settings. Experience mentoring others and contributing to team-wide engineering practice. Key benefits Generous Annual Leave 27 days holiday a year (plus 8 bank holidays). Flexible Working we offer flexible start and end working times, with hybrid working in place for all roles. Pension - Entry to the Local Government Pension Scheme. Salary Sacrifice Schemes we offer salary sacrifice schemes for bikes, cars and tech! Lifestyle Benefits We offer discounts on gyms, cinema, retail and much more! Parenthood Leave We offer above the statutory minimum for maternity, adoption and paternity leave. Working Environment We have a stimulating, supportive and rewarding working environment with a dedicated team of likeminded professionals. Support: Our Employee Assistance Programme (EAP) provides confidential support for employees on personal or work-related matters. Development: We fully support the development of our staff and ensure that you have high level of continuous professional development. We think a wide range of different work and educational experiences could support you to be successful in this role. We encourage applications from all backgrounds, communities and industries, and are committed to employing a team that has diverse skills, experiences and abilities. If you have any queries about this role, please contact our recruitment team. For more information and to apply, please visit our vacancies page. Closing date: 10.00am on Monday, 1 June 2026. We reserve the right to close this vacancy early if we receive a high volume of applications. This post requires a satisfactory enhanced disclosure from the Disclosure and Barring Service (DBS) with a Children s Barred List Check and Occupational Health Check is required as a condition of employment.
Head of Engineering Department: Digital & IT Employment Type: Permanent - Full Time Location: London Compensation: £52,000 - £64,000 / year Description techUK Overview: techUK is the trade association which brings together people, companies and organisations to realise the positive outcomes of what digital technology can achieve. With over 1100 members (the majority of which are SMEs) across the UK, techUK creates a network for innovation and collaboration across business, government and stakeholders to provide a better future for people, society, the economy and the planet. By providing expertise and insight, we support our members, partners and stakeholders as they prepare the UK for what comes next in a constantly changing world. Role purpose: The Head of Engineering is responsible for designing, building, and maintaining techUK's core digital systems and engineering capability. This is a hands on technical leadership role that combines software development, platform architecture, and business systems integration. The role will lead the development of scalable systems across the Microsoft ecosystem, including Azure, Dynamics 365, and Power BI, while introducing AI driven workflows and automation to improve organisational efficiency. Working closely with teams across the organisation, the Head of Engineering will translate business needs into practical technical solutions and ensure systems are well integrated across the organisation's digital environment. The organisation's website operates on a headless architecture using Contentful CMS, with the front end hosted on Vercel, and integrates with CRM, analytics, and internal systems through API driven services. Operating within a small engineering team, the Head of Engineering will remain actively involved in development, prototyping, and technical implementation while shaping the organisation's engineering capability and digital architecture. This role reports into the Chief Technology and Transformation Officer and will have 1 2 direct reports. Key Responsibilities Platform Development & Systems Architecture - Design, build, and maintain techUK's core digital systems across the Microsoft ecosystem, including Azure hosted services, integrations with Dynamics 365, and supporting data infrastructure. Website & Digital Platform Engineering - Provide technical leadership for the organisation's digital platforms, including the headless website architecture built on Contentful and hosted via Vercel, ensuring reliable integration with CRM, analytics, and other internal systems. Data Architecture & Business Intelligence - Develop and maintain reliable data flows between organisational systems, supporting accurate reporting and insight generation through platforms such as Power BI and helping establish a clear single source of truth for organisational data. Systems Integration & APIs - Design and implement API driven integrations between core systems including CRM, website platforms, marketing tools, analytics environments, and internal databases, ensuring systems operate as a connected digital ecosystem. AI, Automation & Rapid Prototyping - Develop and implement AI enabled workflows and automation that improve operational efficiency and unlock organisational insight. Rapidly prototype new tools and workflows to test ideas and deliver practical solutions across the business. Engineering Leadership - Establish and maintain best practice in software development, platform architecture, security, and DevOps practices while providing hands on engineering leadership within a small team. Skills, Knowledge and Expertise Competencies Technical Leadership - Ability to design robust digital systems and guide engineering approaches while remaining hands on in development. Systems Thinking - Strong ability to understand complex technical environments and design integrations that improve organisational capability. Problem Solving - Ability to diagnose operational challenges and translate them into practical technical solutions. Communication - Ability to explain technical concepts clearly and work effectively with non technical stakeholders. Innovation & Curiosity - Interest in emerging technologies such as AI, automation, and data platforms, and the ability to apply them practically. Delivery & Execution - Comfortable moving quickly from concept to working solutions through prototyping, development, and iterative improvement. Essential Knowledge and Experience Strong hands on experience in software development and platform engineering. Experience working within the Microsoft technology ecosystem, including: Azure cloud services Dynamics 365 Power BI Microsoft Graph APIs Azure Functions or serverless architectures Power Platform or related automation tools Experience building API driven integrations between enterprise platforms including CRM systems, web platforms, marketing systems, and analytics environments. Strong programming capability in languages such as Python, JavaScript/TypeScript, or C#. Experience developing integrations, automation pipelines, and backend services that support operational workflows. Experience working with modern web architectures, ideally including headless CMS platforms. Experience operating in small technical teams where individuals remain hands on in engineering work. Line management experience is essential. Ability to translate business requirements into technical solutions and deliver working systems. Desired Knowledge and Experience Experience working with Contentful CMS and Vercel hosting environments. Experience implementing AI driven tools, automation workflows, or LLM based systems. Experience working with data platforms, analytics environments, or business intelligence systems. Familiarity with DevOps and CI/CD practices, including version control platforms such as GitHub. Experience working in a technology sector, membership organisation, or policy focused environment. Benefits This is a full time role based out of techUK's London offices, however techUK operates a flexible working policy. This role profile is a guide to the work that is required but does not form part of any contract of employment and may change from time to time to reflect changing circumstances. The successful candidate must have permission to work in the UK prior to the commencement of employment. The salary range for this role is £52,000-£64,000 per annum based upon experience plus discretionary bonus and comprehensive benefits. To apply for this role, please click below on the 'Apply Now' button.
Apr 28, 2026
Full time
Head of Engineering Department: Digital & IT Employment Type: Permanent - Full Time Location: London Compensation: £52,000 - £64,000 / year Description techUK Overview: techUK is the trade association which brings together people, companies and organisations to realise the positive outcomes of what digital technology can achieve. With over 1100 members (the majority of which are SMEs) across the UK, techUK creates a network for innovation and collaboration across business, government and stakeholders to provide a better future for people, society, the economy and the planet. By providing expertise and insight, we support our members, partners and stakeholders as they prepare the UK for what comes next in a constantly changing world. Role purpose: The Head of Engineering is responsible for designing, building, and maintaining techUK's core digital systems and engineering capability. This is a hands on technical leadership role that combines software development, platform architecture, and business systems integration. The role will lead the development of scalable systems across the Microsoft ecosystem, including Azure, Dynamics 365, and Power BI, while introducing AI driven workflows and automation to improve organisational efficiency. Working closely with teams across the organisation, the Head of Engineering will translate business needs into practical technical solutions and ensure systems are well integrated across the organisation's digital environment. The organisation's website operates on a headless architecture using Contentful CMS, with the front end hosted on Vercel, and integrates with CRM, analytics, and internal systems through API driven services. Operating within a small engineering team, the Head of Engineering will remain actively involved in development, prototyping, and technical implementation while shaping the organisation's engineering capability and digital architecture. This role reports into the Chief Technology and Transformation Officer and will have 1 2 direct reports. Key Responsibilities Platform Development & Systems Architecture - Design, build, and maintain techUK's core digital systems across the Microsoft ecosystem, including Azure hosted services, integrations with Dynamics 365, and supporting data infrastructure. Website & Digital Platform Engineering - Provide technical leadership for the organisation's digital platforms, including the headless website architecture built on Contentful and hosted via Vercel, ensuring reliable integration with CRM, analytics, and other internal systems. Data Architecture & Business Intelligence - Develop and maintain reliable data flows between organisational systems, supporting accurate reporting and insight generation through platforms such as Power BI and helping establish a clear single source of truth for organisational data. Systems Integration & APIs - Design and implement API driven integrations between core systems including CRM, website platforms, marketing tools, analytics environments, and internal databases, ensuring systems operate as a connected digital ecosystem. AI, Automation & Rapid Prototyping - Develop and implement AI enabled workflows and automation that improve operational efficiency and unlock organisational insight. Rapidly prototype new tools and workflows to test ideas and deliver practical solutions across the business. Engineering Leadership - Establish and maintain best practice in software development, platform architecture, security, and DevOps practices while providing hands on engineering leadership within a small team. Skills, Knowledge and Expertise Competencies Technical Leadership - Ability to design robust digital systems and guide engineering approaches while remaining hands on in development. Systems Thinking - Strong ability to understand complex technical environments and design integrations that improve organisational capability. Problem Solving - Ability to diagnose operational challenges and translate them into practical technical solutions. Communication - Ability to explain technical concepts clearly and work effectively with non technical stakeholders. Innovation & Curiosity - Interest in emerging technologies such as AI, automation, and data platforms, and the ability to apply them practically. Delivery & Execution - Comfortable moving quickly from concept to working solutions through prototyping, development, and iterative improvement. Essential Knowledge and Experience Strong hands on experience in software development and platform engineering. Experience working within the Microsoft technology ecosystem, including: Azure cloud services Dynamics 365 Power BI Microsoft Graph APIs Azure Functions or serverless architectures Power Platform or related automation tools Experience building API driven integrations between enterprise platforms including CRM systems, web platforms, marketing systems, and analytics environments. Strong programming capability in languages such as Python, JavaScript/TypeScript, or C#. Experience developing integrations, automation pipelines, and backend services that support operational workflows. Experience working with modern web architectures, ideally including headless CMS platforms. Experience operating in small technical teams where individuals remain hands on in engineering work. Line management experience is essential. Ability to translate business requirements into technical solutions and deliver working systems. Desired Knowledge and Experience Experience working with Contentful CMS and Vercel hosting environments. Experience implementing AI driven tools, automation workflows, or LLM based systems. Experience working with data platforms, analytics environments, or business intelligence systems. Familiarity with DevOps and CI/CD practices, including version control platforms such as GitHub. Experience working in a technology sector, membership organisation, or policy focused environment. Benefits This is a full time role based out of techUK's London offices, however techUK operates a flexible working policy. This role profile is a guide to the work that is required but does not form part of any contract of employment and may change from time to time to reflect changing circumstances. The successful candidate must have permission to work in the UK prior to the commencement of employment. The salary range for this role is £52,000-£64,000 per annum based upon experience plus discretionary bonus and comprehensive benefits. To apply for this role, please click below on the 'Apply Now' button.
Loan IQ Application support London 12-month contract Inside ir35 Purpose of Job The Loan IQ Support Lead (VP) is accountable for the strategic and operational leadership of the Loan IQ production environment, ensuring stability, resilience, and continual improvement of a mission critical lending platform. This role provides senior oversight of incident, problem, and change management activities while driving platform optimization, risk reduction, and partnership with business stakeholders across Lending Operations, Front Office, Finance, Risk, and Technology. The VP serves as a key liaison between Technology, Vendors, and enterprise functions to support complex commercial lending workflows and integrations across the broader loan servicing ecosystem The role requires a deep understanding of syndicated and bilateral loan processing, strong technical acumen and experience with Loan IQ. Background Loan IQ provides loan booking and servicing functionality across two of the three product pillars (Loans, Trade Finance) of the front office pillars: SFS (Structured Finance Solutions) SPD - Specialised Products Department ISD - International & Structural Finance Department TPD - Transportation Department CIB (Corporate & Investment Bank) CBDE 1 and 2 - Corporate Banking Japanese and Non-Japanese GTBD - Global Transaction Banking Department GFIG - Global Financial Institutions Group Department GTFD - Global Trade Finance Department LCME - Loan Capital Markets FMD - Financial Markets Department Facts / Scale Lending makes up c.60% of EMEA gross profit, contributing c.$1.2bn pa. Engaging with IT and non-IT colleagues at all levels in EMEA is key to being successful in this role. Engaging with external vendor Finastra for incident resolution Office environment minimum 5 days per fortnight (split 2-3 days per week). Accountabilities & Responsibilities Direct and govern Level 2/3 support activities across incident, problem, and change management, ensuring audit ready processes and documentation. Provide executive oversight of daily batch cycles and upstream/downstream interfaces, proactively addressing bottlenecks and performance risks. Review root-cause analyses for major incidents and champion long term remediation plans. Lead the roadmap for platform stability, resiliency upgrades, technical debt remediation, and infrastructure modernization. Drive strategic enhancements, system upgrades, and integration initiatives across the lending technology estate. Champion automation, monitoring improvements, and self service capabilities to reduce manual intervention and operational risk. Serve as the senior point of escalation for Operations, Front Office, Product Control, and Risk teams. Drive alignment between business workflows and platform capabilities, enabling efficient loan servicing, reporting, and control processes. Translate complex technology issues into executive friendly communication during incidents and governance forums. Provide governance across all Loan IQ-related integration points (e.g., GL feeds, data lakes, reporting platforms, payment systems). Oversee data integrity controls, reconciliation processes, and triage of high impact data defects. Guide usage of SQL and data diagnostic tools to support analytics and operational investigations. Ensure all support activities meet enterprise risk, audit, and regulatory standards. Work with L1 production support team for issue resolution Knowledge, Skills, Experience & Qualifications Required: Deep understanding of commercial lending workflows (syndicated, bilateral, agency, trading, servicing). Strong understanding of the Loan IQ data model and configuration tables Strong background in production support leadership, including major incident management and platform governance. Familiarity with ITIL processes and tools (e.g., ServiceNow, JIRA). Experience managing cross functional teams and influencing senior stakeholders. Solid grasp of SQL, batch processing, integration architectures, and vendor supported systems. Ability to work in a fast-paced, regulated environment with on-call responsibilities. Preferred: Experience with Loan IQ APIs, data extracts and reporting tools. Exposure to regulatory environments related to credit, operational risk, and financial reporting. If you believe you have the experience required, please apply with your CV now for instant consideration! TO APPLY - PLEASE APPLY WITH AN UP-TO-DATE CV Candidates will ideally show evidence of the above in their CV in order to be considered. Please be advised if you haven't heard from us within 48 hours then unfortunately your application has not been successful on this occasion, we may however keep your details on file for any suitable future vacancies and contact you accordingly. Pontoon is an employment consultancy. We put expertise, energy, and enthusiasm into improving everyone's chance of being part of the workplace. We respect and appreciate people of all ethnicities, generations, religious beliefs, sexual orientations, gender identities, and more. We do this by showcasing their talents, skills, and unique experience in an inclusive environment that helps them thrive. We use generative AI tools to support our candidate screening process. This helps us ensure a fair, consistent, and efficient experience for all applicants. Rest assured, all final decisions are made by our hiring team, and your application will be reviewed with care and attention.
Apr 28, 2026
Contractor
Loan IQ Application support London 12-month contract Inside ir35 Purpose of Job The Loan IQ Support Lead (VP) is accountable for the strategic and operational leadership of the Loan IQ production environment, ensuring stability, resilience, and continual improvement of a mission critical lending platform. This role provides senior oversight of incident, problem, and change management activities while driving platform optimization, risk reduction, and partnership with business stakeholders across Lending Operations, Front Office, Finance, Risk, and Technology. The VP serves as a key liaison between Technology, Vendors, and enterprise functions to support complex commercial lending workflows and integrations across the broader loan servicing ecosystem The role requires a deep understanding of syndicated and bilateral loan processing, strong technical acumen and experience with Loan IQ. Background Loan IQ provides loan booking and servicing functionality across two of the three product pillars (Loans, Trade Finance) of the front office pillars: SFS (Structured Finance Solutions) SPD - Specialised Products Department ISD - International & Structural Finance Department TPD - Transportation Department CIB (Corporate & Investment Bank) CBDE 1 and 2 - Corporate Banking Japanese and Non-Japanese GTBD - Global Transaction Banking Department GFIG - Global Financial Institutions Group Department GTFD - Global Trade Finance Department LCME - Loan Capital Markets FMD - Financial Markets Department Facts / Scale Lending makes up c.60% of EMEA gross profit, contributing c.$1.2bn pa. Engaging with IT and non-IT colleagues at all levels in EMEA is key to being successful in this role. Engaging with external vendor Finastra for incident resolution Office environment minimum 5 days per fortnight (split 2-3 days per week). Accountabilities & Responsibilities Direct and govern Level 2/3 support activities across incident, problem, and change management, ensuring audit ready processes and documentation. Provide executive oversight of daily batch cycles and upstream/downstream interfaces, proactively addressing bottlenecks and performance risks. Review root-cause analyses for major incidents and champion long term remediation plans. Lead the roadmap for platform stability, resiliency upgrades, technical debt remediation, and infrastructure modernization. Drive strategic enhancements, system upgrades, and integration initiatives across the lending technology estate. Champion automation, monitoring improvements, and self service capabilities to reduce manual intervention and operational risk. Serve as the senior point of escalation for Operations, Front Office, Product Control, and Risk teams. Drive alignment between business workflows and platform capabilities, enabling efficient loan servicing, reporting, and control processes. Translate complex technology issues into executive friendly communication during incidents and governance forums. Provide governance across all Loan IQ-related integration points (e.g., GL feeds, data lakes, reporting platforms, payment systems). Oversee data integrity controls, reconciliation processes, and triage of high impact data defects. Guide usage of SQL and data diagnostic tools to support analytics and operational investigations. Ensure all support activities meet enterprise risk, audit, and regulatory standards. Work with L1 production support team for issue resolution Knowledge, Skills, Experience & Qualifications Required: Deep understanding of commercial lending workflows (syndicated, bilateral, agency, trading, servicing). Strong understanding of the Loan IQ data model and configuration tables Strong background in production support leadership, including major incident management and platform governance. Familiarity with ITIL processes and tools (e.g., ServiceNow, JIRA). Experience managing cross functional teams and influencing senior stakeholders. Solid grasp of SQL, batch processing, integration architectures, and vendor supported systems. Ability to work in a fast-paced, regulated environment with on-call responsibilities. Preferred: Experience with Loan IQ APIs, data extracts and reporting tools. Exposure to regulatory environments related to credit, operational risk, and financial reporting. If you believe you have the experience required, please apply with your CV now for instant consideration! TO APPLY - PLEASE APPLY WITH AN UP-TO-DATE CV Candidates will ideally show evidence of the above in their CV in order to be considered. Please be advised if you haven't heard from us within 48 hours then unfortunately your application has not been successful on this occasion, we may however keep your details on file for any suitable future vacancies and contact you accordingly. Pontoon is an employment consultancy. We put expertise, energy, and enthusiasm into improving everyone's chance of being part of the workplace. We respect and appreciate people of all ethnicities, generations, religious beliefs, sexual orientations, gender identities, and more. We do this by showcasing their talents, skills, and unique experience in an inclusive environment that helps them thrive. We use generative AI tools to support our candidate screening process. This helps us ensure a fair, consistent, and efficient experience for all applicants. Rest assured, all final decisions are made by our hiring team, and your application will be reviewed with care and attention.
Role : Data Architect Location: Bournemouth Salary: £90,000 + Bonus WFH policy: One Day a month My client who is a leading insurance company is looking for a Data Architect to join their team, this is an exciting opportunity as you will be working on transformational project this is not a BAU role. Role and Responsibilities Lead solution-level data architecture across a complex data estate, acting as the key link between business stakeholders, analysts, and engineering teams to ensure practical, delivery-focused outcomes Own end-to-end design authority, shaping both high- and low-level data solutions while ensuring alignment with architectural standards and governance Drive the evolution towards a modern cloud data platform (Microsoft Fabric), while supporting a hybrid environment including Azure Databricks and legacy SQL Server systems Lead a major transformation of legacy MI & reporting platforms, including cloud migration (lift-and-shift) and redesign into scalable, modern analytics solutions Partner closely with finance and actuarial teams to replace spreadsheet-driven processes with automated, governed data pipelines, alongside broader enterprise data initiative If this role sounds of interest and you would like to know more do not hesitate to contact me on: Reasonable Adjustments: Respect and equality are core values to us. We are proud of the diverse and inclusive community we have built, and we welcome applications from people of all backgrounds and perspectives. Our success is driven by our people, united by the spirit of partnership to deliver the best resourcing solutions for our clients. If you need any help or adjustments during the recruitment process for any reason , please let us know when you apply or talk to the recruiters directly so we can support you.
Apr 28, 2026
Full time
Role : Data Architect Location: Bournemouth Salary: £90,000 + Bonus WFH policy: One Day a month My client who is a leading insurance company is looking for a Data Architect to join their team, this is an exciting opportunity as you will be working on transformational project this is not a BAU role. Role and Responsibilities Lead solution-level data architecture across a complex data estate, acting as the key link between business stakeholders, analysts, and engineering teams to ensure practical, delivery-focused outcomes Own end-to-end design authority, shaping both high- and low-level data solutions while ensuring alignment with architectural standards and governance Drive the evolution towards a modern cloud data platform (Microsoft Fabric), while supporting a hybrid environment including Azure Databricks and legacy SQL Server systems Lead a major transformation of legacy MI & reporting platforms, including cloud migration (lift-and-shift) and redesign into scalable, modern analytics solutions Partner closely with finance and actuarial teams to replace spreadsheet-driven processes with automated, governed data pipelines, alongside broader enterprise data initiative If this role sounds of interest and you would like to know more do not hesitate to contact me on: Reasonable Adjustments: Respect and equality are core values to us. We are proud of the diverse and inclusive community we have built, and we welcome applications from people of all backgrounds and perspectives. Our success is driven by our people, united by the spirit of partnership to deliver the best resourcing solutions for our clients. If you need any help or adjustments during the recruitment process for any reason , please let us know when you apply or talk to the recruiters directly so we can support you.
Data Architect Ncounter is supporting a specialist consultancy delivering into highly sensitive National Security and Defence programmes, now seeking a DV Cleared Data Architect to join a high-impact transformation initiative. This contract requires 3 days per week onsite on the South Coast. This role sits at the heart of complex, secure data environments, focusing on the design of scalable architectures that enable integration, governance, and analytics across mission-critical systems. You will work alongside senior stakeholders, optimisation leads and solution architects to ensure data structures are aligned to operational needs, while providing clear architectural direction within fast-paced programmes. Key Responsibilities • Design and implement secure, scalable data architectures across complex environments • Define data models, standards and integration patterns across platforms • Embed data governance, quality and compliance aligned to Defence standards • Align data structures with operational tools, systems and user requirements • Collaborate with solution architects and programme teams to deliver cohesive solutions • Support full lifecycle delivery, from requirements through to deployment What We're Looking For • Active DV clearance is essential • Strong experience delivering data architecture within UK Defence or MOD environments • Proven background operating across multiple security classifications • Experience within consultancy or client-facing delivery roles • Knowledge of cloud data platforms, particularly Azure and AWS • Familiarity with data visualisation and analytics tooling such as Power BI or Qlik • Strong stakeholder engagement skills across technical and non-technical audiences This is a contract opportunity offering £750-850 per day, working on programmes where data architecture is critical to operational success. You will play a key role in shaping secure data capabilities while influencing how organisations structure, govern and utilise data at scale. If you hold active DV clearance and are looking to take on a high-impact contract within Defence, please get in touch to discuss further.
Apr 28, 2026
Contractor
Data Architect Ncounter is supporting a specialist consultancy delivering into highly sensitive National Security and Defence programmes, now seeking a DV Cleared Data Architect to join a high-impact transformation initiative. This contract requires 3 days per week onsite on the South Coast. This role sits at the heart of complex, secure data environments, focusing on the design of scalable architectures that enable integration, governance, and analytics across mission-critical systems. You will work alongside senior stakeholders, optimisation leads and solution architects to ensure data structures are aligned to operational needs, while providing clear architectural direction within fast-paced programmes. Key Responsibilities • Design and implement secure, scalable data architectures across complex environments • Define data models, standards and integration patterns across platforms • Embed data governance, quality and compliance aligned to Defence standards • Align data structures with operational tools, systems and user requirements • Collaborate with solution architects and programme teams to deliver cohesive solutions • Support full lifecycle delivery, from requirements through to deployment What We're Looking For • Active DV clearance is essential • Strong experience delivering data architecture within UK Defence or MOD environments • Proven background operating across multiple security classifications • Experience within consultancy or client-facing delivery roles • Knowledge of cloud data platforms, particularly Azure and AWS • Familiarity with data visualisation and analytics tooling such as Power BI or Qlik • Strong stakeholder engagement skills across technical and non-technical audiences This is a contract opportunity offering £750-850 per day, working on programmes where data architecture is critical to operational success. You will play a key role in shaping secure data capabilities while influencing how organisations structure, govern and utilise data at scale. If you hold active DV clearance and are looking to take on a high-impact contract within Defence, please get in touch to discuss further.
Pontoon is an employment consultancy. We put expertise, energy, and enthusiasm into improving everyone's chance of being part of the workplace. We respect and appreciate people of all ethnicities, generations, religious beliefs, sexual orientations, gender identities, and more. We do this by showcasing their talents, skills, and unique experience in an inclusive environment that helps them thrive. Are you an innovative Endpoint Security Engineer looking for an exciting opportunity to enhance global security operations? Join our client's dynamic team, where your expertise will help shape the future of endpoint security! Role: Endpoint Security Engineer Duration: 12 Months (extension Options) Location: Chester (Hybrid 3 days in Office) Rate: £ 550 per day (umbrella) Why Join Us? At our organization, we believe in fostering a collaborative environment where creativity and security go hand in hand. You'll work alongside dedicated professionals who are passionate about protecting digital assets while supporting cloud migration initiatives. Key Responsibilities: Drive Cloud Migration: Contribute to a major cloud migration program, ensuring seamless transitions and robust security measures. Engineer Security Solutions: Design, implement, and support cutting-edge endpoint security solutions using Trellix, BeyondTrust, and CrowdStrike. Deploy Solutions: Utilize ITSM systems like Remedy and ServiceNow for effective enterprise security deployments. Documentation & Workflows: Produce high-quality technical documentation, workflows, and support materials to guide best practices. Collaborate & Optimize: Partner with security vendor consultants and Technical Account Managers to maximize product performance. Align Compliance Standards: Work closely with compliance, audit, and information security teams to maintain organizational standards. Research & Consultancy: Conduct thorough research and provide informed consultancy on the latest security solutions and trends. Manage Configurations: Define and maintain security baseline configurations, ensuring the highest standards are met. Rapid Response: Address operational issues promptly and efficiently, adapting to a fast-paced global environment. Required Technical Skills & Certifications: Proven experience in endpoint security, with a strong background in both engineering and support roles. Expertise in Application Control, particularly with BeyondTrust solutions. Proficient in scripting with PowerShell and Python-essential skills for this role! Familiarity with Trellix ePO, BeyondTrust EPM Cloud, and policy management in Azure and AWS. Advanced knowledge of Microsoft Windows operating systems. Strong analytical and problem-solving skills to tackle complex security challenges. Exposure to data analytics tools like Splunk and Power BI. Essential Skills: Excellent workflow management, communication, and interpersonal abilities. Experience with EPM Cloud Migration (this role focuses on EPM, not IAM ). Ability to create clear and effective documentation, including diagrams and technical architecture. Knowledge of governance, risk management, and administrative oversight of workstreams. Quick learner with a knack for sharing knowledge and applying research-driven solutions. Proven ability to thrive under pressure and respond swiftly to emerging challenges. Experience collaborating with architecture and engineering teams on Windows platform design. Desired Skills: Understanding of cloud technologies, especially Azure and AWS. Security certifications, preferably CISSP. Experience supporting macOS/iOS environments would be a plus! What's Next? If you're ready to take your career to the next level and make a significant impact in the world of endpoint security, we want to hear from you! Apply today to join our dedicated team and help shape the future of security! Apply Now! Let's secure the digital future together! Candidates will ideally show evidence of the above in their CV in order to be considered. Please be advised if you haven't heard from us within 48 hours then unfortunately your application has not been successful on this occasion, we may however keep your details on file for any suitable future vacancies and contact you accordingly. Pontoon is an employment consultancy and operates as an equal opportunities employer.
Apr 28, 2026
Contractor
Pontoon is an employment consultancy. We put expertise, energy, and enthusiasm into improving everyone's chance of being part of the workplace. We respect and appreciate people of all ethnicities, generations, religious beliefs, sexual orientations, gender identities, and more. We do this by showcasing their talents, skills, and unique experience in an inclusive environment that helps them thrive. Are you an innovative Endpoint Security Engineer looking for an exciting opportunity to enhance global security operations? Join our client's dynamic team, where your expertise will help shape the future of endpoint security! Role: Endpoint Security Engineer Duration: 12 Months (extension Options) Location: Chester (Hybrid 3 days in Office) Rate: £ 550 per day (umbrella) Why Join Us? At our organization, we believe in fostering a collaborative environment where creativity and security go hand in hand. You'll work alongside dedicated professionals who are passionate about protecting digital assets while supporting cloud migration initiatives. Key Responsibilities: Drive Cloud Migration: Contribute to a major cloud migration program, ensuring seamless transitions and robust security measures. Engineer Security Solutions: Design, implement, and support cutting-edge endpoint security solutions using Trellix, BeyondTrust, and CrowdStrike. Deploy Solutions: Utilize ITSM systems like Remedy and ServiceNow for effective enterprise security deployments. Documentation & Workflows: Produce high-quality technical documentation, workflows, and support materials to guide best practices. Collaborate & Optimize: Partner with security vendor consultants and Technical Account Managers to maximize product performance. Align Compliance Standards: Work closely with compliance, audit, and information security teams to maintain organizational standards. Research & Consultancy: Conduct thorough research and provide informed consultancy on the latest security solutions and trends. Manage Configurations: Define and maintain security baseline configurations, ensuring the highest standards are met. Rapid Response: Address operational issues promptly and efficiently, adapting to a fast-paced global environment. Required Technical Skills & Certifications: Proven experience in endpoint security, with a strong background in both engineering and support roles. Expertise in Application Control, particularly with BeyondTrust solutions. Proficient in scripting with PowerShell and Python-essential skills for this role! Familiarity with Trellix ePO, BeyondTrust EPM Cloud, and policy management in Azure and AWS. Advanced knowledge of Microsoft Windows operating systems. Strong analytical and problem-solving skills to tackle complex security challenges. Exposure to data analytics tools like Splunk and Power BI. Essential Skills: Excellent workflow management, communication, and interpersonal abilities. Experience with EPM Cloud Migration (this role focuses on EPM, not IAM ). Ability to create clear and effective documentation, including diagrams and technical architecture. Knowledge of governance, risk management, and administrative oversight of workstreams. Quick learner with a knack for sharing knowledge and applying research-driven solutions. Proven ability to thrive under pressure and respond swiftly to emerging challenges. Experience collaborating with architecture and engineering teams on Windows platform design. Desired Skills: Understanding of cloud technologies, especially Azure and AWS. Security certifications, preferably CISSP. Experience supporting macOS/iOS environments would be a plus! What's Next? If you're ready to take your career to the next level and make a significant impact in the world of endpoint security, we want to hear from you! Apply today to join our dedicated team and help shape the future of security! Apply Now! Let's secure the digital future together! Candidates will ideally show evidence of the above in their CV in order to be considered. Please be advised if you haven't heard from us within 48 hours then unfortunately your application has not been successful on this occasion, we may however keep your details on file for any suitable future vacancies and contact you accordingly. Pontoon is an employment consultancy and operates as an equal opportunities employer.
Your new role "We are seeking an AI Data & Systems Architect to strengthen our data foundations during the implementation of a new ERP system and shape, develop, and deliver our future AI capability.The initial focus of this role will be on building strong data governance, improving data quality, and ensuring the business is ready to maximise value from the new ERP platform. From there, the role will define and prioritise AI and analytics opportunities, develop robust business cases, and lead the delivery of AI tools and solutions with support from third-party partners where required." What you'll need to succeed Strong experience in data governance, data architecture, ERP transformation, AI/data strategy, or solution delivery roles. Experience improving data quality and governance in complex business environments. Experience supporting ERP implementation or transformation, ideally with exposure to ERP platforms. Strong understanding of data controls, master data, and cross-functional business processes. Advanced capability in Python or R for data analysis, automation, modelling, or AI tool development. What you need to do now If you're interested in this role, click 'apply now' to forward an up-to-date copy of your CV, or call us now.If this job isn't quite right for you, but you are looking for a new position, please contact us for a confidential discussion about your career. Hays Specialist Recruitment Limited acts as an employment agency for permanent recruitment and employment business for the supply of temporary workers. By applying for this job you accept the T&C's, Privacy Policy and Disclaimers which can be found at hays.co.uk
Apr 28, 2026
Full time
Your new role "We are seeking an AI Data & Systems Architect to strengthen our data foundations during the implementation of a new ERP system and shape, develop, and deliver our future AI capability.The initial focus of this role will be on building strong data governance, improving data quality, and ensuring the business is ready to maximise value from the new ERP platform. From there, the role will define and prioritise AI and analytics opportunities, develop robust business cases, and lead the delivery of AI tools and solutions with support from third-party partners where required." What you'll need to succeed Strong experience in data governance, data architecture, ERP transformation, AI/data strategy, or solution delivery roles. Experience improving data quality and governance in complex business environments. Experience supporting ERP implementation or transformation, ideally with exposure to ERP platforms. Strong understanding of data controls, master data, and cross-functional business processes. Advanced capability in Python or R for data analysis, automation, modelling, or AI tool development. What you need to do now If you're interested in this role, click 'apply now' to forward an up-to-date copy of your CV, or call us now.If this job isn't quite right for you, but you are looking for a new position, please contact us for a confidential discussion about your career. Hays Specialist Recruitment Limited acts as an employment agency for permanent recruitment and employment business for the supply of temporary workers. By applying for this job you accept the T&C's, Privacy Policy and Disclaimers which can be found at hays.co.uk
Data Engineer - Insurance & London Market - London (Hybrid) - Python, PySpark, Databricks, Snowflake - Up to £90,000 This is an opportunity to join a business right in the middle of a major data transformation within the London Insurance Market. They're investing in modern, cloud-native platforms and are looking for senior engineers who can shape solutions end-to-end, not just deliver pre-defined tasks. Data sits at the heart of underwriting, pricing, claims, and reinsurance decisions here, which means your work will be highly visible and genuinely impactful. The environment is collaborative, engineering-led and focused on doing things properly. If you're looking for a role where you can lead technically, stay hands-on, and work on complex London Market data, this is a great move. You Will Work With Designing and building cloud-based data platforms using medallion architecture (Bronze / Silver / Gold) Developing batch and near real-time data pipelines Engineering scalable pipelines using Python and PySpark on Databricks and/or Snowflake Integrating data from PAS, claims systems, broker platforms, third-party providers, and market feeds Ensuring high standards of data quality, reconciliation, lineage, and auditability Working closely with Underwriting, Actuarial, Finance, and Regulatory stakeholders Carrying out code reviews, setting engineering best practice, and mentoring junior engineers Supporting CI/CD pipelines and Git-based delivery models Contributing to modern engineering practices, including AI-assisted software development, in a governed way Benefits Salary up to £90,000 (depending on experience) Hybrid working with a central London office High-impact role with real ownership and autonomy Modern cloud and data technology stack Strong exposure to senior London Market stakeholders Supportive, collaborative engineering culture Clear long-term progression within the data function Key Experience Proven experience as a Senior Data Engineer within Insurance, ideally the London Market Hands-on expertise with: Python & PySpark Databricks and/or Snowflake Cloud platforms (Azure, AWS, or GCP) Strong understanding of: Medallion architecture Batch and streaming pipelines Data modelling for analytics and reporting Solid domain experience across: Lloyd's Syndicates Delegated Authority / Bordereaux Reinsurance and Ceded Reinsurance Underwriting, pricing, and claims data Confident engaging senior technical and non-technical stakeholders Interested? Apply now or send your CV directly
Apr 28, 2026
Full time
Data Engineer - Insurance & London Market - London (Hybrid) - Python, PySpark, Databricks, Snowflake - Up to £90,000 This is an opportunity to join a business right in the middle of a major data transformation within the London Insurance Market. They're investing in modern, cloud-native platforms and are looking for senior engineers who can shape solutions end-to-end, not just deliver pre-defined tasks. Data sits at the heart of underwriting, pricing, claims, and reinsurance decisions here, which means your work will be highly visible and genuinely impactful. The environment is collaborative, engineering-led and focused on doing things properly. If you're looking for a role where you can lead technically, stay hands-on, and work on complex London Market data, this is a great move. You Will Work With Designing and building cloud-based data platforms using medallion architecture (Bronze / Silver / Gold) Developing batch and near real-time data pipelines Engineering scalable pipelines using Python and PySpark on Databricks and/or Snowflake Integrating data from PAS, claims systems, broker platforms, third-party providers, and market feeds Ensuring high standards of data quality, reconciliation, lineage, and auditability Working closely with Underwriting, Actuarial, Finance, and Regulatory stakeholders Carrying out code reviews, setting engineering best practice, and mentoring junior engineers Supporting CI/CD pipelines and Git-based delivery models Contributing to modern engineering practices, including AI-assisted software development, in a governed way Benefits Salary up to £90,000 (depending on experience) Hybrid working with a central London office High-impact role with real ownership and autonomy Modern cloud and data technology stack Strong exposure to senior London Market stakeholders Supportive, collaborative engineering culture Clear long-term progression within the data function Key Experience Proven experience as a Senior Data Engineer within Insurance, ideally the London Market Hands-on expertise with: Python & PySpark Databricks and/or Snowflake Cloud platforms (Azure, AWS, or GCP) Strong understanding of: Medallion architecture Batch and streaming pipelines Data modelling for analytics and reporting Solid domain experience across: Lloyd's Syndicates Delegated Authority / Bordereaux Reinsurance and Ceded Reinsurance Underwriting, pricing, and claims data Confident engaging senior technical and non-technical stakeholders Interested? Apply now or send your CV directly
Data Engineer Location: Remote / monthly visits to Cardiff Salary: up to £72,000 with generous bonus We're looking for a Data Engineer to join a forward-thinking data team focused on enabling smarter decision-making across the business. This team plays a key role in shaping and delivering data platforms, pipelines, and analytics capabilities that empower users and unlock valuable insights. The Role As a Data Engineer, you'll design, build, and maintain scalable data solutions that meet both business and customer needs. You'll work within a collaborative, cross-functional team while also having the autonomy to take ownership of your work, contribute ideas, and drive innovation. You'll be involved throughout the full lifecycle of data engineering - from ingestion and transformation to delivery and ongoing support - in a modern, cloud-based environment. Key Responsibilities Develop and maintain robust data pipelines and data solutions across multiple sources Translate complex data into actionable insights that support business objectives Ensure solutions align with engineering standards, governance, and regulatory requirements Contribute to a culture of automation, quality, and continuous improvement Collaborate with stakeholders across IT and the wider business Participate in Agile ceremonies and contribute to team delivery goals Take ownership of solutions, including ongoing support and occasional on-call cover Continuously learn, experiment, and share knowledge within the team Skills & Experience (Essential) Strong experience with cloud-based data architectures (GCP experience essential) Proficiency in Python, SQL, (epsoure to Java, Kotlin, Bash) Hands-on experience with modern data tools such as BigQuery, Dataflow, Pub/Sub, dbt, Kubernetes, or similar Experience building and maintaining data pipelines Familiarity with CI/CD, Git, Terraform, and modern engineering practices Experience working in Agile/Scrum environments Exposure to tools such as Airflow/Composer and Azure DevOps Understanding of AI/ML tooling such as Vertex AI Desirable Experience with BI/visualisation tools (e.g. Looker, MicroStrategy) Knowledge of data modelling approaches such as Kimball Familiarity with emerging AI-assisted development tools What We're Looking For A proactive, curious mindset with a passion for data Strong problem-solving skills and attention to detail A collaborative team player who can also work independently Someone who embraces change, innovation, and continuous improvement Reasonable Adjustments: Respect and equality are core values to us. We are proud of the diverse and inclusive community we have built, and we welcome applications from people of all backgrounds and perspectives. Our success is driven by our people, united by the spirit of partnership to deliver the best resourcing solutions for our clients. If you need any help or adjustments during the recruitment process for any reason , please let us know when you apply or talk to the recruiters directly so we can support you.
Apr 28, 2026
Full time
Data Engineer Location: Remote / monthly visits to Cardiff Salary: up to £72,000 with generous bonus We're looking for a Data Engineer to join a forward-thinking data team focused on enabling smarter decision-making across the business. This team plays a key role in shaping and delivering data platforms, pipelines, and analytics capabilities that empower users and unlock valuable insights. The Role As a Data Engineer, you'll design, build, and maintain scalable data solutions that meet both business and customer needs. You'll work within a collaborative, cross-functional team while also having the autonomy to take ownership of your work, contribute ideas, and drive innovation. You'll be involved throughout the full lifecycle of data engineering - from ingestion and transformation to delivery and ongoing support - in a modern, cloud-based environment. Key Responsibilities Develop and maintain robust data pipelines and data solutions across multiple sources Translate complex data into actionable insights that support business objectives Ensure solutions align with engineering standards, governance, and regulatory requirements Contribute to a culture of automation, quality, and continuous improvement Collaborate with stakeholders across IT and the wider business Participate in Agile ceremonies and contribute to team delivery goals Take ownership of solutions, including ongoing support and occasional on-call cover Continuously learn, experiment, and share knowledge within the team Skills & Experience (Essential) Strong experience with cloud-based data architectures (GCP experience essential) Proficiency in Python, SQL, (epsoure to Java, Kotlin, Bash) Hands-on experience with modern data tools such as BigQuery, Dataflow, Pub/Sub, dbt, Kubernetes, or similar Experience building and maintaining data pipelines Familiarity with CI/CD, Git, Terraform, and modern engineering practices Experience working in Agile/Scrum environments Exposure to tools such as Airflow/Composer and Azure DevOps Understanding of AI/ML tooling such as Vertex AI Desirable Experience with BI/visualisation tools (e.g. Looker, MicroStrategy) Knowledge of data modelling approaches such as Kimball Familiarity with emerging AI-assisted development tools What We're Looking For A proactive, curious mindset with a passion for data Strong problem-solving skills and attention to detail A collaborative team player who can also work independently Someone who embraces change, innovation, and continuous improvement Reasonable Adjustments: Respect and equality are core values to us. We are proud of the diverse and inclusive community we have built, and we welcome applications from people of all backgrounds and perspectives. Our success is driven by our people, united by the spirit of partnership to deliver the best resourcing solutions for our clients. If you need any help or adjustments during the recruitment process for any reason , please let us know when you apply or talk to the recruiters directly so we can support you.
Technical Lead l Quantexa l Implementation l Banking l Finance l Data l Architecture l Stakeholder management l £700-£725 per day l Outside IR35 l London Hybrid l 12 months We are seeking an experienced Quantexa Certified technical Lead to lead a Quantexa implementation for a major banking client. This is a hands-on leadership role, responsible for driving the technical delivery of Quantexa within a regulated financial services environment. Essential Skills & Experience Proven experience implementing Quantexa in a production environment Strong banking or financial services background (retail, corporate, or investment banking) Experience operating as a Technical Lead or Senior Architect on complex data platforms Strong understanding of data engineering, big data platforms, and analytics Experience with cloud environments (AWS, Azure, or GCP) Excellent stakeholder management and communication skills Desirable Experience with financial crime, AML, KYC, or fraud use cases Prior experience working in regulated banking environments Contract: Rate: £700 per day Contract: Outside IR35 Location: London hybrid Duration: Initial 12 months Start date: Imminent so need to be available within the next week or two tops If you're interested, please apply or send your cv to and I'll be in touch! RSG Plc is acting as an Employment Business in relation to this vacancy.
Apr 28, 2026
Contractor
Technical Lead l Quantexa l Implementation l Banking l Finance l Data l Architecture l Stakeholder management l £700-£725 per day l Outside IR35 l London Hybrid l 12 months We are seeking an experienced Quantexa Certified technical Lead to lead a Quantexa implementation for a major banking client. This is a hands-on leadership role, responsible for driving the technical delivery of Quantexa within a regulated financial services environment. Essential Skills & Experience Proven experience implementing Quantexa in a production environment Strong banking or financial services background (retail, corporate, or investment banking) Experience operating as a Technical Lead or Senior Architect on complex data platforms Strong understanding of data engineering, big data platforms, and analytics Experience with cloud environments (AWS, Azure, or GCP) Excellent stakeholder management and communication skills Desirable Experience with financial crime, AML, KYC, or fraud use cases Prior experience working in regulated banking environments Contract: Rate: £700 per day Contract: Outside IR35 Location: London hybrid Duration: Initial 12 months Start date: Imminent so need to be available within the next week or two tops If you're interested, please apply or send your cv to and I'll be in touch! RSG Plc is acting as an Employment Business in relation to this vacancy.
Hays Specialist Recruitment Limited
West Drayton, Middlesex
Role Purpose The Data Analyst / Data Engineer will lead the discovery, interpretation, and presentation of data to enable AI-driven solutions for our client, embedded within operational teams. The role focusses on turning complex, often legacy, datasets into clear insight, narrative, and decision-ready outputs, while ensuring data foundations are sufficiently robust to scale across OpCos. This position requires strong analytical judgement, stakeholder consultancy skills, and hands-on capability to shape and evolve supporting data pipelines. Contract - 6 months (high possibility to extend further) Location - waterside (UB7 0GB) Hybrid - 2-3 days onsite Pay - Premium-level role; competitive rates (inside IR35) Key Responsibilities Discover, explore, and process data from various sources (relational databases, flat files such as CSV, YML, XLS), forming a deep understanding of content, limitations, and business relevance. Identify, investigate, and clearly articulate data quality, completeness, and consistency issues, including their downstream impact on analytics and AI use cases. Challenge data provenance and assumptions in legacy datasets, reframing against current needs Translate business questions and operational needs into meaningful KPIs, metrics, dashboards, and analytical narratives consumable by non-technical stakeholders. Create clear metadata and documentation that explains datasets, transformations, assumptions, and analytical outputs to support reuse and trust Partner closely with Data Scientists and Visualisation specialists to enable advanced analytics. Support the adoption of MRO AI Solutions within BA operational workflows by ensuring insights are actionable, timely, and well-embedded in decision-making processes. Design, build, and optimise data pipelines for ingestion, transformation, and storage. Ensure data quality, integrity, and security controls are applied across systems. Apply cloud and data-engineering best practices pragmatically to ensure solutions scale where needed across OpCos, without over-engineering Design data architectures and pipelines that support multi-OpCo deployment, ensuring modularity and interoperability. Required Skills & Experience Core Data Analytical Capabilities Strong experience in data analysis within a product or operational environment, with demonstrable impact on decision-making Advanced hands-on experience with data processing and analysis tools (SQL, Python, Pandas, etc), with a bias towards exploration, insight and explanation Proven ability to understand legacy datasets/pipelines and to evaluate their fitness for new use cases Comfortable working independently and communicating with non-technical stakeholders Supporting Data Engineering Capabilities Solid understanding of data modelling concepts and API-driven data integration to influence pipeline design Proven experience in developing, testing, and deploying data solutions into production environments, ensuring reliability, scalability, and maintainability beyond proof-of-concept or prototype stages. Practical expertise in Python, SQL, and modern ETL or orchestration frameworks. (Preferred) Skills in data visualisation (PowerBI, Tableau, and/or other dashboarding tools) (Preferred) Hands-on experience with cloud platforms, ideally AWS Consulting-Level Competencies Significant experience in similar roles, with a proven ability to integrate quickly into new teams and deliver immediate value. Ability to design enterprise-grade data solutions under tight timelines. Strong stakeholder engagement and solution-oriented mindset. Track record of creating high-impact outcomes and driving stakeholder satisfaction from day one. Ability to implement standards and frameworks for scalable data solutions across multiple operating companies. Familiarity with airline or logistics data domains is a plus. Location & Travel Initial co-location with client teams in London is essential to ensure close collaboration. Candidates must also be prepared to occasionally travel internationally during later stages to facilitate group-wide deployment. Compensation Premium-level role; competitive rates aligned with UK consultancy benchmarks. What you need to do now If you're interested in this role, click 'apply now' to forward an up-to-date copy of your CV, or call us now.If this job isn't quite right for you, but you are looking for a new position, please contact us for a confidential discussion about your career. Hays Talent Solutions is a trading division of Hays Specialist Recruitment Limited and acts as an employment agency for permanent recruitment and employment business for the supply of temporary workers. By applying for this job you accept the T&C's, Privacy Policy and Disclaimers which can be found at hays.co.uk
Apr 27, 2026
Contractor
Role Purpose The Data Analyst / Data Engineer will lead the discovery, interpretation, and presentation of data to enable AI-driven solutions for our client, embedded within operational teams. The role focusses on turning complex, often legacy, datasets into clear insight, narrative, and decision-ready outputs, while ensuring data foundations are sufficiently robust to scale across OpCos. This position requires strong analytical judgement, stakeholder consultancy skills, and hands-on capability to shape and evolve supporting data pipelines. Contract - 6 months (high possibility to extend further) Location - waterside (UB7 0GB) Hybrid - 2-3 days onsite Pay - Premium-level role; competitive rates (inside IR35) Key Responsibilities Discover, explore, and process data from various sources (relational databases, flat files such as CSV, YML, XLS), forming a deep understanding of content, limitations, and business relevance. Identify, investigate, and clearly articulate data quality, completeness, and consistency issues, including their downstream impact on analytics and AI use cases. Challenge data provenance and assumptions in legacy datasets, reframing against current needs Translate business questions and operational needs into meaningful KPIs, metrics, dashboards, and analytical narratives consumable by non-technical stakeholders. Create clear metadata and documentation that explains datasets, transformations, assumptions, and analytical outputs to support reuse and trust Partner closely with Data Scientists and Visualisation specialists to enable advanced analytics. Support the adoption of MRO AI Solutions within BA operational workflows by ensuring insights are actionable, timely, and well-embedded in decision-making processes. Design, build, and optimise data pipelines for ingestion, transformation, and storage. Ensure data quality, integrity, and security controls are applied across systems. Apply cloud and data-engineering best practices pragmatically to ensure solutions scale where needed across OpCos, without over-engineering Design data architectures and pipelines that support multi-OpCo deployment, ensuring modularity and interoperability. Required Skills & Experience Core Data Analytical Capabilities Strong experience in data analysis within a product or operational environment, with demonstrable impact on decision-making Advanced hands-on experience with data processing and analysis tools (SQL, Python, Pandas, etc), with a bias towards exploration, insight and explanation Proven ability to understand legacy datasets/pipelines and to evaluate their fitness for new use cases Comfortable working independently and communicating with non-technical stakeholders Supporting Data Engineering Capabilities Solid understanding of data modelling concepts and API-driven data integration to influence pipeline design Proven experience in developing, testing, and deploying data solutions into production environments, ensuring reliability, scalability, and maintainability beyond proof-of-concept or prototype stages. Practical expertise in Python, SQL, and modern ETL or orchestration frameworks. (Preferred) Skills in data visualisation (PowerBI, Tableau, and/or other dashboarding tools) (Preferred) Hands-on experience with cloud platforms, ideally AWS Consulting-Level Competencies Significant experience in similar roles, with a proven ability to integrate quickly into new teams and deliver immediate value. Ability to design enterprise-grade data solutions under tight timelines. Strong stakeholder engagement and solution-oriented mindset. Track record of creating high-impact outcomes and driving stakeholder satisfaction from day one. Ability to implement standards and frameworks for scalable data solutions across multiple operating companies. Familiarity with airline or logistics data domains is a plus. Location & Travel Initial co-location with client teams in London is essential to ensure close collaboration. Candidates must also be prepared to occasionally travel internationally during later stages to facilitate group-wide deployment. Compensation Premium-level role; competitive rates aligned with UK consultancy benchmarks. What you need to do now If you're interested in this role, click 'apply now' to forward an up-to-date copy of your CV, or call us now.If this job isn't quite right for you, but you are looking for a new position, please contact us for a confidential discussion about your career. Hays Talent Solutions is a trading division of Hays Specialist Recruitment Limited and acts as an employment agency for permanent recruitment and employment business for the supply of temporary workers. By applying for this job you accept the T&C's, Privacy Policy and Disclaimers which can be found at hays.co.uk
Are you a Detection Engineer ready to take on the fight against modern adversaries? Join a well-established SOC working with high-profile Defence clients, where your expertise genuinely matters. In this hands-on technical role, you'll own the end-to-end design, development and maturity of detection logic across SIEM platforms-engineering effective responses to real-world attacker techniques. You'll operate with a high degree of autonomy, acting as a trusted SME across multiple secure environments within a complex MSSP setting. This is a standout opportunity to advance your career at the sharp end of cyber defence. Location: Hybrid working - 2 days per week in our Farnborough office. Security: You must hold or be eligible for SC Clearance. What you'll be doing: ? Design, build, test and continuously refine advanced SIEM detection logic, including rules, correlations and analytics. Research emerging threats, vulnerabilities and adversary TTPs, mapping them to MITRE ATT&CK to close detection and visibility gaps. Tune and validate detections to minimise false positives and deliver high-fidelity alerts for SOC analysts. Act as a technical authority, providing expert guidance to SOC Analysts, Architects and Engineers to strengthen overall detection capability. Define, implement and maintain technical detection standards across environments. Clearly communicate complex technical risks and detection logic to both technical teams and non-technical stakeholders. What you'll bring: Deep SIEM expertise, building advanced detection logic, automation and complex queries in Splunk (SPL) and Microsoft Sentinel (KQL). A proven track record delivering complex detection engineering projects within enterprise or MSSP environments. Strong analytical skills, with the ability to break down sophisticated attacks into actionable detection patterns. Confidence to own technical delivery end-to-end, driving work through to completion with minimal escalation. Expert knowledge of MITRE ATT&CK, with real-world application in detection engineering. A BSc in Computer Science, IT, or a related discipline. Solid scripting skills in Python, PowerShell, or similar, supporting automation and data manipulation. Experience developing detections in QRadar and/or conducting EDR-focused threat hunting (e.g. CrowdStrike, Microsoft Defender for Endpoint). Broad infrastructure awareness across Cloud (Azure/AWS), on-prem, and SaaS / PaaS / IaaS environments. If you are interested in this role but not sure if your skills and experience are exactly what we're looking for, please do apply, we'd love to hear from you! Employment Type: Full-time, Permanent. Location: Hybrid: 2 days per week in Farnborough. Security Clearance Level: SC Cleared or eligible. Internal Recruiter: Jane. Salary: To £65,000. Benefits: 25 days annual leave with the choice to buy additional days, health cash plan, life assurance and pension. Sopra Steria: Our Aerospace, Defence and Security business designs, develops and deploys digital solutions to Central Government clients. The work we do makes a real difference to the client's goal of National Security, and we operate in a unique and privileged environment. We are given time for professional development activities, and we coach and mentor our colleagues, sharing knowledge and learning from each other. We foster a culture in which employees feel valued and supported and have pride in their work for the customer, delivering outstanding rates of customer satisfaction in the UK's most complex safety- and security-critical markets.
Apr 27, 2026
Full time
Are you a Detection Engineer ready to take on the fight against modern adversaries? Join a well-established SOC working with high-profile Defence clients, where your expertise genuinely matters. In this hands-on technical role, you'll own the end-to-end design, development and maturity of detection logic across SIEM platforms-engineering effective responses to real-world attacker techniques. You'll operate with a high degree of autonomy, acting as a trusted SME across multiple secure environments within a complex MSSP setting. This is a standout opportunity to advance your career at the sharp end of cyber defence. Location: Hybrid working - 2 days per week in our Farnborough office. Security: You must hold or be eligible for SC Clearance. What you'll be doing: ? Design, build, test and continuously refine advanced SIEM detection logic, including rules, correlations and analytics. Research emerging threats, vulnerabilities and adversary TTPs, mapping them to MITRE ATT&CK to close detection and visibility gaps. Tune and validate detections to minimise false positives and deliver high-fidelity alerts for SOC analysts. Act as a technical authority, providing expert guidance to SOC Analysts, Architects and Engineers to strengthen overall detection capability. Define, implement and maintain technical detection standards across environments. Clearly communicate complex technical risks and detection logic to both technical teams and non-technical stakeholders. What you'll bring: Deep SIEM expertise, building advanced detection logic, automation and complex queries in Splunk (SPL) and Microsoft Sentinel (KQL). A proven track record delivering complex detection engineering projects within enterprise or MSSP environments. Strong analytical skills, with the ability to break down sophisticated attacks into actionable detection patterns. Confidence to own technical delivery end-to-end, driving work through to completion with minimal escalation. Expert knowledge of MITRE ATT&CK, with real-world application in detection engineering. A BSc in Computer Science, IT, or a related discipline. Solid scripting skills in Python, PowerShell, or similar, supporting automation and data manipulation. Experience developing detections in QRadar and/or conducting EDR-focused threat hunting (e.g. CrowdStrike, Microsoft Defender for Endpoint). Broad infrastructure awareness across Cloud (Azure/AWS), on-prem, and SaaS / PaaS / IaaS environments. If you are interested in this role but not sure if your skills and experience are exactly what we're looking for, please do apply, we'd love to hear from you! Employment Type: Full-time, Permanent. Location: Hybrid: 2 days per week in Farnborough. Security Clearance Level: SC Cleared or eligible. Internal Recruiter: Jane. Salary: To £65,000. Benefits: 25 days annual leave with the choice to buy additional days, health cash plan, life assurance and pension. Sopra Steria: Our Aerospace, Defence and Security business designs, develops and deploys digital solutions to Central Government clients. The work we do makes a real difference to the client's goal of National Security, and we operate in a unique and privileged environment. We are given time for professional development activities, and we coach and mentor our colleagues, sharing knowledge and learning from each other. We foster a culture in which employees feel valued and supported and have pride in their work for the customer, delivering outstanding rates of customer satisfaction in the UK's most complex safety- and security-critical markets.
. Senior Data Modeller - Banking (Contract) Location: Glasgow (Hybrid - 2 days onsite per week)Start Date: ImmediateDuration: Initial 6 months (extension likely)Sector: Investment Banking Engagement Context Our client, a leading consultancy delivering strategic data transformation across the banking sector, is seeking a Senior Data Modeller to support a high-impact programme focused on post-bind insurance data. The successful candidate will play a critical role in designing scalable, compliant data models that underpin operational reporting, financial reconciliation, and regulatory alignment across multiple specialty lines. Key Responsibilities Lead the design and delivery of conceptual, logical, and physical data models across post-bind domainsCollaborate with data architects, business analysts, and actuarial teams to translate complex business requirements into structured data assetsApply Data Vault 2.0 methodology to support auditability, scalability, and lineage trackingModel data for ingestion into Snowflake, ensuring compatibility with cloud-native architecture and downstream analyticsEnsure models support reconciliation of premium and claims, aged debt tracking, reserve movements, and regulatory reportingDocument metadata, data dictionaries, and lineage to support governance and complianceEngage with stakeholders across finance, operations, and claims to validate model assumptions and ensure business alignmentRequired Experience & Skills Extensive experience as a Data Modeller within the FS Market, with demonstrable expertise in dataProficiency in Data Vault 2.0 and dimensional modelling techniquesHands-on experience with Snowflake and cloud-based data platformsAbility to model across multiple specialty lines (e.g., marine, aviation, cyber)Strong stakeholder engagement and documentation skills Clear documentation and stakeholder alignment across finance, operations, and actuarial domainsSeamless integration of models into Snowflake and downstream analytics pipelinesContribution to a broader data transformation programme with measurable impact on operational efficiency and regulatory compliance
Apr 27, 2026
Contractor
. Senior Data Modeller - Banking (Contract) Location: Glasgow (Hybrid - 2 days onsite per week)Start Date: ImmediateDuration: Initial 6 months (extension likely)Sector: Investment Banking Engagement Context Our client, a leading consultancy delivering strategic data transformation across the banking sector, is seeking a Senior Data Modeller to support a high-impact programme focused on post-bind insurance data. The successful candidate will play a critical role in designing scalable, compliant data models that underpin operational reporting, financial reconciliation, and regulatory alignment across multiple specialty lines. Key Responsibilities Lead the design and delivery of conceptual, logical, and physical data models across post-bind domainsCollaborate with data architects, business analysts, and actuarial teams to translate complex business requirements into structured data assetsApply Data Vault 2.0 methodology to support auditability, scalability, and lineage trackingModel data for ingestion into Snowflake, ensuring compatibility with cloud-native architecture and downstream analyticsEnsure models support reconciliation of premium and claims, aged debt tracking, reserve movements, and regulatory reportingDocument metadata, data dictionaries, and lineage to support governance and complianceEngage with stakeholders across finance, operations, and claims to validate model assumptions and ensure business alignmentRequired Experience & Skills Extensive experience as a Data Modeller within the FS Market, with demonstrable expertise in dataProficiency in Data Vault 2.0 and dimensional modelling techniquesHands-on experience with Snowflake and cloud-based data platformsAbility to model across multiple specialty lines (e.g., marine, aviation, cyber)Strong stakeholder engagement and documentation skills Clear documentation and stakeholder alignment across finance, operations, and actuarial domainsSeamless integration of models into Snowflake and downstream analytics pipelinesContribution to a broader data transformation programme with measurable impact on operational efficiency and regulatory compliance
Data Engineer Preston (Hybrid working) £50,000 Reliable, high-quality data underpins effective decision-making across the organisation. This Data Engineer role plays a critical part in ensuring that trusted data flows seamlessly from operational systems through to reporting and analytics platforms. Acting as the architect of the organisation's data foundations, you will design, build, and maintain robust data pipelines and models that turn raw information into meaningful, actionable insight. The role goes beyond writing code - it is about creating scalable, resilient data infrastructure that connects systems, supports regulatory requirements, and enables teams at all levels to make informed decisions. Data Engineer Key Responsibilities Design, build, and maintain scalable data pipelines using Microsoft Fabric. Ensure data is ingested, transformed, and stored efficiently, securely, and reliably. Implement robust error handling, logging, and monitoring across ETL/ELT processes. Develop and maintain test plans to validate data accuracy and pipeline performance. Monitor pipeline health and apply tuning techniques to optimise performance. Create and maintain dimensional and relational data models to support reporting and analytics. Collaborate closely with analysts and business stakeholders to ensure data models meet operational and strategic requirements. Document data structures, transformations, and lineage in a clear and accessible manner. Write, optimise, and maintain complex SQL queries for data extraction and transformation. Implement data quality checks and validation processes to ensure accuracy, reliability, and completeness. Work in partnership with data governance stakeholders to align engineering practices with governance frameworks, security standards, and GDPR requirements. Undertake additional duties aligned with the scope and seniority of the role as required. Key Skills & Experience Strong experience designing and maintaining scalable ETL/ELT data pipelines, ideally within a cloud-first environment. Hands-on experience with Microsoft Fabric, including data ingestion, transformation, and storage. Advanced SQL skills, with experience writing and optimising complex queries; Python experience advantageous. Solid understanding of relational and dimensional data modelling to support reporting and analytics. Experience migrating data workloads from on-premise solutions (e.g. SSIS or manual processes) to modern cloud platforms. Strong focus on data quality, including validation, monitoring, error handling, and documentation. Familiarity with Power BI and supporting analytics use cases. Awareness of data governance, security, and GDPR principles. Ability to work collaboratively with technical and non-technical stakeholders, translating business requirements into effective data solutions. Robert Walters Operations Limited is an employment business and employment agency and welcomes applications from all candidates
Apr 27, 2026
Full time
Data Engineer Preston (Hybrid working) £50,000 Reliable, high-quality data underpins effective decision-making across the organisation. This Data Engineer role plays a critical part in ensuring that trusted data flows seamlessly from operational systems through to reporting and analytics platforms. Acting as the architect of the organisation's data foundations, you will design, build, and maintain robust data pipelines and models that turn raw information into meaningful, actionable insight. The role goes beyond writing code - it is about creating scalable, resilient data infrastructure that connects systems, supports regulatory requirements, and enables teams at all levels to make informed decisions. Data Engineer Key Responsibilities Design, build, and maintain scalable data pipelines using Microsoft Fabric. Ensure data is ingested, transformed, and stored efficiently, securely, and reliably. Implement robust error handling, logging, and monitoring across ETL/ELT processes. Develop and maintain test plans to validate data accuracy and pipeline performance. Monitor pipeline health and apply tuning techniques to optimise performance. Create and maintain dimensional and relational data models to support reporting and analytics. Collaborate closely with analysts and business stakeholders to ensure data models meet operational and strategic requirements. Document data structures, transformations, and lineage in a clear and accessible manner. Write, optimise, and maintain complex SQL queries for data extraction and transformation. Implement data quality checks and validation processes to ensure accuracy, reliability, and completeness. Work in partnership with data governance stakeholders to align engineering practices with governance frameworks, security standards, and GDPR requirements. Undertake additional duties aligned with the scope and seniority of the role as required. Key Skills & Experience Strong experience designing and maintaining scalable ETL/ELT data pipelines, ideally within a cloud-first environment. Hands-on experience with Microsoft Fabric, including data ingestion, transformation, and storage. Advanced SQL skills, with experience writing and optimising complex queries; Python experience advantageous. Solid understanding of relational and dimensional data modelling to support reporting and analytics. Experience migrating data workloads from on-premise solutions (e.g. SSIS or manual processes) to modern cloud platforms. Strong focus on data quality, including validation, monitoring, error handling, and documentation. Familiarity with Power BI and supporting analytics use cases. Awareness of data governance, security, and GDPR principles. Ability to work collaboratively with technical and non-technical stakeholders, translating business requirements into effective data solutions. Robert Walters Operations Limited is an employment business and employment agency and welcomes applications from all candidates
Tenth Revolution Group
Stratford-upon-avon, Warwickshire
Contract Data Architect - Senior SME (Modern Data Platform) Contract Length: 12-18 months (strong potential to extend or move permanent) It will be Inside IR35 Day Rate: £600-£700 Location: UK-based, primarily remote Ad hoc travel to Stratford-upon-Avon & London for workshops and occasional travel to Bucharest the main office is in Stratford. Overview We are seeking a Senior Data Architect to act as a subject matter expert for the design and build of a modern, enterprise-scale data platform . This is a high-impact contract role with full ownership of the data domain , supporting a multi-year roadmap to transform an existing platform into a cloud-native, federated data architecture on Azure. The organisation is in the process of lifting and modernising a large-scale data landscape into Azure , with a strong likelihood of winning a major government contract , making this a long-term and strategically critical engagement. This role requires a hands-on architect - someone who can define strategy and architecture, but also work closely with delivery teams to ensure successful execution. Key Responsibilities Own and shape the end-to-end data architecture and data strategy , treating the data platform as a product Design and build a modern Azure-based data platform supporting large-scale data processing Define and implement modern federated data architecture principles Lead the adoption of Medallion Architecture (Bronze / Silver / Gold) Design and oversee robust data ingestion and data pipeline frameworks , including ingestion of poor-quality and complex data Provide architectural leadership for: Data lakes and lakehouse patterns Analytics and reporting platforms Master Data Management (MDM) Act as a hands-on technical authority , supporting and mentoring engineering teams Collaborate with business and technical stakeholders across the UK and near-shore teams in Bucharest Ensure platforms are scalable, secure, and fit for future government and enterprise needs Technical Environment You will work extensively with: Azure Data Platform Databricks (Lakehouse architecture) Cosmos DB Azure Data Factory Talend (ETL) Power BI & Qlik Large-scale data processing and analytics Master Data Management (MDM) concepts and tooling Required Experience Proven experience as a Senior / Lead Data Architect on complex data transformation programmes Strong background designing and delivering modern cloud data platforms on Azure Hands-on experience with Databricks, Azure data services, and modern ETL pipelines Deep understanding of data architecture patterns , including lakehouse and federated models Experience working with imperfect and complex data sources Comfortable operating as the single point of ownership for a data domain Ability to translate strategy into detailed, actionable architecture Strong stakeholder engagement skills Desirable Experience working in regulated or government-aligned environments Prior exposure to long-term platform rebuild programmes Mentoring or leadership of multi-disciplinary data teams Why Apply? Long-term 12-18 month contract with strong extension or permanent potential Opportunity to own and architect a modern data platform end-to-end Significant influence on a multi-year data transformation roadmap Flexible day rate High-profile programme with future government exposure
Apr 27, 2026
Contractor
Contract Data Architect - Senior SME (Modern Data Platform) Contract Length: 12-18 months (strong potential to extend or move permanent) It will be Inside IR35 Day Rate: £600-£700 Location: UK-based, primarily remote Ad hoc travel to Stratford-upon-Avon & London for workshops and occasional travel to Bucharest the main office is in Stratford. Overview We are seeking a Senior Data Architect to act as a subject matter expert for the design and build of a modern, enterprise-scale data platform . This is a high-impact contract role with full ownership of the data domain , supporting a multi-year roadmap to transform an existing platform into a cloud-native, federated data architecture on Azure. The organisation is in the process of lifting and modernising a large-scale data landscape into Azure , with a strong likelihood of winning a major government contract , making this a long-term and strategically critical engagement. This role requires a hands-on architect - someone who can define strategy and architecture, but also work closely with delivery teams to ensure successful execution. Key Responsibilities Own and shape the end-to-end data architecture and data strategy , treating the data platform as a product Design and build a modern Azure-based data platform supporting large-scale data processing Define and implement modern federated data architecture principles Lead the adoption of Medallion Architecture (Bronze / Silver / Gold) Design and oversee robust data ingestion and data pipeline frameworks , including ingestion of poor-quality and complex data Provide architectural leadership for: Data lakes and lakehouse patterns Analytics and reporting platforms Master Data Management (MDM) Act as a hands-on technical authority , supporting and mentoring engineering teams Collaborate with business and technical stakeholders across the UK and near-shore teams in Bucharest Ensure platforms are scalable, secure, and fit for future government and enterprise needs Technical Environment You will work extensively with: Azure Data Platform Databricks (Lakehouse architecture) Cosmos DB Azure Data Factory Talend (ETL) Power BI & Qlik Large-scale data processing and analytics Master Data Management (MDM) concepts and tooling Required Experience Proven experience as a Senior / Lead Data Architect on complex data transformation programmes Strong background designing and delivering modern cloud data platforms on Azure Hands-on experience with Databricks, Azure data services, and modern ETL pipelines Deep understanding of data architecture patterns , including lakehouse and federated models Experience working with imperfect and complex data sources Comfortable operating as the single point of ownership for a data domain Ability to translate strategy into detailed, actionable architecture Strong stakeholder engagement skills Desirable Experience working in regulated or government-aligned environments Prior exposure to long-term platform rebuild programmes Mentoring or leadership of multi-disciplinary data teams Why Apply? Long-term 12-18 month contract with strong extension or permanent potential Opportunity to own and architect a modern data platform end-to-end Significant influence on a multi-year data transformation roadmap Flexible day rate High-profile programme with future government exposure