Data Engineer
Magentic
- London
- £110,000-120,000 per year
- Permanent
- Full-time
- Design and operate performant, scalable ingestion pipelines processing high-volume data from global supply chain and procurement systems.
- Define, evolve, and manage data schemas and catalogues—from raw staging to high-quality analytics and feature stores—ensuring consistency and discoverability.
- Build end-to-end monitoring and observability for your pipelines: owning data quality, latency, completeness, and lineage at every stage.
- Champion secure, governed data practices: access controls, secrets management, encrypted data-in-transit/at-rest, and compliance with frameworks like GDPR.
- Collaborate closely with AI, Platform, and Product teams, provisioning data sets, feature tables, and contracts for analytics and machine learning at scale.
- Continuously improve efficiency and reliability via testing, CI/CD automation, cost/performance tuning, and incident/root-cause reviews.
- Experience working at startups, scaleups or at companies with a big focus on data quality, DataOps and data management at scale.
- Expertise in Cloud-Native Data Engineering: 5+ years building and running data warehouses and pipelines in AWS or Azure, including managed data services (e.g., Kinesis, EMR/Databricks, Redshift, Glue, Azure Data Lake).
- Programming Mastery: Advanced skills in Python or another major language; writing clean, testable, production-grade ETL code at scale.
- Modern Data Pipelines: Experience with batch and streaming frameworks (e.g., Apache Spark, Flink, Kafka Streams, Beam), including orchestration via Airflow, Prefect or Dagster.
- Data Modeling & Schema Management: Demonstrated expertise in designing, evolving, and documenting schemas (OLAP/OLTP, dimensional, star/snowflake, CDC), data contracts, and data cataloguing.
- API & Integration Fluency: Building data ingestion from REST/gRPC APIs, file drops, message queues (SQS, Kafka), and 3rd party SaaS integrations, with idempotency and error handling.
- Storage & Query Engines: Strong with RDBMS (PostgreSQL, MySQL), NoSQL (DynamoDB, Cassandra), data lakes (Parquet, ORC), and warehouse paradigms.
- Observability & Quality: Deep familiarity with metrics, logging, tracing, and data quality tools (e.g., Great Expectations, Monte Carlo, custom validation/test suites).
- Security & Governance: Data encryption, secrets management, RBAC/ABAC, and compliance awareness (GDPR, CCPA).
- CI/CD for Data Systems: Comfort with automation, infrastructure as code (Terraform), version control, and release workflows.
- Collaborative Spirit: Experience working closely with platform, ML, and analytics teams in a fast-paced, mission-driven environment.
- Competitive Equity: play a real part in Magentic’s upside.
- A salary of £110,000-£120,000
- Visa sponsorship available; (note we are only accepting candidates who are currently based in the UK.)
- Hybrid London HQ (3-4 days in the office)
- Annual team retreat—a fully-funded off-site to recharge, bond, and build.