Specialist Solutions Architect - DE/DWH

Databricks
London, United Kingdom
Last week
Posted
9 Apr 2026 (Last week)

Req:FEQ127R163

Location:London

Recruiter:Dina Hussain

Skills:Data Engineering/DWH

As a Specialist Solutions Architect (SSA) - Data Engineering, you will guide customers in building big data solutions on Databricks that span a large variety of use cases. You will be in a customer-facing role, working with and supporting Solution Architects, and will require hands-on production experience with Apache Spark™ and expertise in other data technologies. SSEs help customers through the design and successful implementation of essential workloads while aligning their technical roadmap to expand the use of the Databricks Data Intelligence Platform. As a deep go-to-expert reporting to the Senior Specialist Field Engineering Manager, you will continue to strengthen your technical skills through mentorship, learning, and internal training programs, and establish yourself in an area of speciality - whether that be streaming, performance tuning, industry expertise, or more.

The impact you will have:

  • Provide technical leadership to guide strategic customers to successful implementations on big data projects, ranging from architectural design to data engineering to model deployment
  • Architect production-level data pipelines, including end-to-end pipeline load performance testing and optimisation
  • Become a technical expert in an area such as data lake technology, big data streaming, or big data ingestion and workflows
  • Assist Solution Architects with more advanced aspects of the technical sale, including custom proof of concept content, estimating workload sizing, and custom architectures
  • Provide tutorials and training to improve community adoption (including hackathons and conference presentations)
  • Contribute to the Databricks Community

What we look for:

  • Extensive experience in a customer-facing technical role. Pre-sales or post-sales experience working with external clients across a variety of industry markets
  • Nice to have: Databricks Certification
  • Travelling approx. 20-30% of the time

Data Engineer Skills

  • Experience as aData Engineer: query tuning, performance tuning, troubleshooting, and debugging Spark or other big data solutions.
  • Extensive experience building big data pipelines
  • Experience in maintaining and extending production data systems to evolve with complex needs
  • Deep Speciality Expertise in at least one of the following areas:
  • Experience scaling big data workloads (such as ETL) that are performant and cost-effective
  • Experience migrating Hadoop workloads to the public cloud - AWS, Azure, or GCP
  • Experience with large-scale data ingestion pipelines and data migrations - including CDC and streaming ingestion pipelines
  • Expert with cloud data lake technologies - such as Delta and Delta Live
  • Bachelor's degree in Computer Science, Information Systems, Engineering, or equivalent experience through work experience
  • Production programming experience in SQL and Python, Scala, or Java
  • Professional experience with Big Data technologies (Ex: Spark, Hadoop, Kafka) and architectures

Data Warehousing, Database Skills

  • Experience with the design and implementation of a broad range of analytical and transactional data technologies such as Hadoop, Apache Spark™, NoSQL, OLTP, OLAP, and ETL/ELT.
  • Hands-on experience working with MPP data warehouse appliances (Oracle Exadata, Teradata, IBM Netezza) or cloud data warehouses (Amazon Redshift, Azure Synapse, Snowflake)
  • Hands-on experience with RDBMS systems (PostGres, MySQL, SQL Server, Oracle, MariaDB)
  • Experience in SQL language or any SQL dialect (PL/SQL, Transact-SQL or others)
  • Experience with BI tools such as Power BI, Tableau, Qlik, or others
  • Knowledge of development tools and best practices for data engineers, including CI/CD, unit and integration testing, plus automation and orchestration
  • Expertise in data warehousing - such as query tuning, performance tuning, troubleshooting, and debugging MPP data warehouses or other big data solutions. Maintained, extended, or migrated a production data warehouse system to evolve with complex customer needs.
  • Production programming experience in PySpark.

About Databricks

Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook.

Benefits

At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region click here.

Our Commitment to Diversity and Inclusion

At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics.

Compliance

If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone.

Related Jobs

View all jobs

Solutions Architect (Retail/CPG)

Databricks London, United Kingdom

Senior Data Platform Architect

Snowflake London, United Kingdom
Permanent

Data Lead (Fabric)

Hays Technology London, United Kingdom
£450 – £500 pd

D365 CE, Copilot & Microsoft Fabric Specialist - - Partner

Opus Recruitment Solutions London, United Kingdom
£500 – £650 pa Contract

Data Architect - Bristol Opportunity

Hays Technology Bristol, Bristol (county), United Kingdom
£80,000 – £88,000 pa Hybrid

Fabric / BI Developer

IT Talent Solutions Guildford, Surrey, United Kingdom

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Where to Advertise Data Engineering Jobs in the UK (2026 Guide)

Advertising data engineering jobs in the UK requires a different approach to most technical hiring. Data engineers occupy a distinct discipline that sits between software engineering, data science and cloud infrastructure — and the strongest candidates identify firmly with the data engineering community rather than with adjacent roles. General job boards consistently conflate data engineering with data analysis, data science and BI development, producing high application volumes but low candidate quality for specialist pipeline and platform roles. This guide, published by DataEngineeringJobs.co.uk, covers where to advertise data engineering roles in the UK in 2026, how the main platforms compare, what employers should expect to pay, and what the data says about hiring across different role types.

New Data Engineering Employers to Watch in 2026: UK and Global Companies Driving the Data Revolution

Data engineering is at the heart of the digital economy, transforming raw data into actionable insights, powering analytics, AI systems, and cloud infrastructure. As the UK and global markets continue to invest heavily in data platforms, pipelines, and real-time analytics, demand for skilled data engineers is growing rapidly. For professionals exploring opportunities on www.DataEngineeringJobs.co.uk , the critical question is: which companies are expanding, hiring, and shaping the future of data-driven business? This article highlights new data engineering employers to watch in 2026, including UK startups, scale-ups, and international firms expanding in the UK.

How Many Data Engineering Tools Do You Need to Know to Get a Data Engineering Job?

If you’re aiming for a career in data engineering, it can feel like you’re staring at a never-ending list of tools and technologies — SQL, Python, Spark, Kafka, Airflow, dbt, Snowflake, Redshift, Terraform, Kubernetes, and the list goes on. Scroll job boards and LinkedIn, and it’s easy to conclude that unless you have experience with every modern tool in the data stack, you won’t even get a callback. Here’s the honest truth most data engineering hiring managers will quietly agree with: 👉 They don’t hire you because you know every tool — they hire you because you can solve real data problems with the tools you know. Tools matter. But only in service of outcomes. Jobs are won by candidates who know why a technology is used, when to use it, and how to explain their decisions. So how many data engineering tools do you actually need to know to get a job? For most job seekers, the answer is far fewer than you think — but you do need them in the right combination and order. This article breaks down what employers really expect, which tools are core, which are role-specific, and how to focus your learning so you look capable and employable rather than overwhelmed.