Data Engineer

ATG (Auction Technology Group)
City of London
3 months ago
Applications closed

Related Jobs

View all jobs

Data Engineer - AI Analytics and EdTech Developments

Data Engineer

Data Engineer

Data Engineer

Data Engineer

Data Engineer

You have a passion for building scalable, reliable data systems that enable data scientists, ML engineers, and analysts to do their best work. You understand that great data products require more than just moving data; they need robust pipelines, data quality assurance, and thoughtful architecture. Not only do you put reliability and scalability at the heart of everything you do, but you are adept at enabling data-driven decisions through proper data modeling and pipeline design. You will be comfortable working cross-functionally with Product, Engineering, Data Science, Analytics, and MLOps teams to develop our products and improve the end-user experience. You should have a strong track record of successful prioritization, meeting critical deadlines, and enthusiastically tackling challenges with an eye toward problem solving.


Key Responsibilities

  • Data Pipeline Development & Management
  • Design, build, and maintain robust ETL/ELT pipelines that support analytics, ML models, and business intelligence
  • Develop scalable batch and streaming data pipelines to process millions of auction events, user interactions, and transactions daily
  • Implement workflow orchestration using Airflow, Dagster, or similar tools to manage complex data dependencies
  • Build data validation and quality monitoring frameworks to ensure data accuracy and reliability
  • ML & Analytics Infrastructure
  • Build feature engineering pipelines to support ML models for search, recommendations, and personalization
  • Integrate with feature stores to enable consistent feature computation across training and inference
  • Create datasets for model training, validation, and testing with proper versioning
  • Data Quality & Monitoring
  • Implement comprehensive data quality checks, anomaly detection, and alerting systems
  • Monitor pipeline health, data freshness, and SLA compliance
  • Create dashboards and reporting tools for data pipeline observability
  • Debug and resolve data quality issues and pipeline failures
  • Collaboration & Best Practices
  • Work closely with Data Scientists and ML Engineers to understand data requirements and deliver reliable datasets
  • Partner with Software Engineers to integrate data pipelines with application systems
  • Establish and document data engineering best practices, coding standards, and design patterns
  • Mentor junior engineers on data engineering principles and best practices

Key Requirements

  • Required Qualifications: BSc or MSc in Computer Science, Data Engineering, Software Engineering, or a related field, or equivalent practical experience
  • 5+ years of experience building and maintaining data pipelines and infrastructure in production environments
  • Strong programming skills in Python, with experience in data processing libraries (Pandas, PySpark)
  • Expert-level SQL skills with experience in query optimization and performance tuning
  • Proven experience with workflow orchestration tools (Airflow, Dagster, Prefect, or similar)
  • Hands‑on experience with cloud platforms (AWS preferred) including S3, Redshift, EMR, Glue, Lambda
  • Experience with data warehousing solutions (Redshift, Snowflake, BigQuery, or similar)
  • Experience with version control systems (Git) and CI/CD practices for data pipelines

Technical Skills

  • Experience with distributed computing frameworks (Apache Spark, Dask, or similar)
  • Knowledge of both batch and streaming data processing (Kafka, Kinesis, or similar)
  • Familiarity with data formats (Parquet, ORC, Avro, JSON) and their trade-offs
  • Understanding of data quality frameworks and testing strategies
  • Previous work with vector databases (Pinecone, Milvus, etc)
  • Experience with monitoring and observability tools (Prometheus, Grafana, CloudWatch)
  • Knowledge of infrastructure-as-code tools (Terraform, CloudFormation)
  • Understanding of containerization (Docker) and orchestration (Kubernetes) is a plus

Nice-to-Have

  • Familiarity with dbt (data build tool) for data transformation workflows
  • Knowledge of Elasticsearch or similar search technologies
  • Experience in eCommerce, marketplace, or auction platforms
  • Understanding of GDPR, data privacy, and compliance requirements
  • Experience with real-time analytics and event-driven architectures (Flink, Materialize)


#J-18808-Ljbffr

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Engineering Tools Do You Need to Know to Get a Data Engineering Job?

If you’re aiming for a career in data engineering, it can feel like you’re staring at a never-ending list of tools and technologies — SQL, Python, Spark, Kafka, Airflow, dbt, Snowflake, Redshift, Terraform, Kubernetes, and the list goes on. Scroll job boards and LinkedIn, and it’s easy to conclude that unless you have experience with every modern tool in the data stack, you won’t even get a callback. Here’s the honest truth most data engineering hiring managers will quietly agree with: 👉 They don’t hire you because you know every tool — they hire you because you can solve real data problems with the tools you know. Tools matter. But only in service of outcomes. Jobs are won by candidates who know why a technology is used, when to use it, and how to explain their decisions. So how many data engineering tools do you actually need to know to get a job? For most job seekers, the answer is far fewer than you think — but you do need them in the right combination and order. This article breaks down what employers really expect, which tools are core, which are role-specific, and how to focus your learning so you look capable and employable rather than overwhelmed.

What Hiring Managers Look for First in Data Engineering Job Applications (UK Guide)

If you’re applying for data engineering jobs in the UK, the first thing to understand is this: Hiring managers don’t read every word of your CV. They scan it. They look for signals of relevance, credibility, delivery and collaboration — and if they don’t see the right signals quickly, your application may never get a second look. In data engineering, hiring managers are especially focused on whether you can build and operate reliable, scalable data systems, handle real-world data challenges and work effectively with analytics, BI, data science and engineering teams. This guide breaks down exactly what they look at first in your application — and how to shape your CV, portfolio and cover letter so you stand out.

The Skills Gap in Data Engineering Jobs: What Universities Aren’t Teaching

Data engineering has quietly become one of the most critical roles in the modern technology stack. While data science and AI often receive the spotlight, data engineers are the professionals who design, build and maintain the systems that make data usable at scale. Across the UK, demand for data engineers continues to rise. Organisations in finance, retail, healthcare, government, media and technology all report difficulty hiring candidates with the right skills. Salaries remain strong, and experienced professionals are in short supply. Yet despite this demand, many graduates with degrees in computer science, data science or related disciplines struggle to secure data engineering roles. The reason is not academic ability. It is a persistent skills gap between university education and real-world data engineering work. This article explores that gap in depth: what universities teach well, what they consistently miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data engineering.