Data engineer

  • Data engineer
  • Permanent
  • London, UK

Who are Metapack?

We are a tech company that works with a lot of the world’s biggest ecommerce players to integrate them with over 450 carriers around the world to make delivery easy. We are a multi-tenant SaaS platform. We give them the platform to help consumers decide their delivery preference and track the parcel’s progress whilst also providing the retailer with intelligent smart decisions about how to send the parcel – all underpinned with lots of data. We work with well-known global retailers and major brands such as ASOS, Adidas, Burberry, John Lewis, Boohoo, eBay, and Zalando. In fact, we work with so many retailers and carriers it’s highly likely that you’ve interacted with us at some point when ordering goods online!

In August 2018, we were acquired by Fortune 100’s 2 nd fastest growing company, We have super ambitious and exciting plans all centred around our tech. Metapack will play a role in shipping around 600 million parcels in 2018 and with the wider family the number rises to 2.5bn parcels. Metapack has been growing at 40% year on year over the last 5 years and continues to grow at a rapid rate.

Why would I want to be a Data Engineer at Metapack?

Data is key to the Metapack’s strategy. We work at scale, pace and with the latest architecture patterns and tech. We process thousands of events per second and our massive dataset keeps growing at a staggering pace. We keep improving our data platform and data engineering stack to accommodate growth, enable novel solutions and provide the best service to our customers.

We have a flat and open engineering culture where data, & evidence beats opinion and hierarchy, backed by honest and frank discussions. We passionately believe in forming autonomous, cross functional teams who are empowered to deliver our ambitious strategy. With ownership comes the ability to operate largely independently away from Board meetings and old world thinking but with the financial support of a high performing tech company. Energy and passion for our business and customers is a part of the Metapack culture – and we love working with like-minded people.

What would I be doing?

  • Contributing to the design, build and operational management of our data lake and analytics solution on top of proven AWS data technologies like S3, Athena, Lambda, Kinesis, Glue
  • Using state of the art technologies like Airflow and Spark to process data and get our dataset just right
  • Developing frameworks and solutions that enable us to acquire, process, monitor and extract value from our massive dataset
  • Supporting the Data Analysts and Data Scientists with automation, tooling, data pipelines and data engineering expertise
  • Delivering highly reliable software and data pipelines using Software Engineering best practices like automation, version control, continuous integration/continuous delivery, testing, security, etc.
  • Define, implement and enforce automated data security and data governance best practices within the solutions designed
  • Mentoring more junior colleagues and being mentored by more senior colleagues

What key skills and experience do I need?

  • A Software Engineering background
  • Experience developing and supporting robust, automated and reliable data pipelines in Python and SQL
  • Experience with data processing frameworks like Pandas or Spark
  • Experience with streaming data processing
  • AWS, Azure or Google Cloud experience
  • Continuous integration/delivery environment experience with a passion for automation
  • Knowledge of a Data Orchestration solutions like Airflow, Oozie, Luigi or Talend
  • Knowledge of both relational and non-relational database design and internals
  • Knowledge of how to design distributed systems and the trade-offs involved
  • Experience with working with software engineering best practices for development, including source control systems, automated deployment pipelines like Jenkins and devops tools like Terraform

It would be great if you also could bring

  • Practical understanding of GDPR and other considerations regarding data security
  • Knowledge and direct experience of using business intelligence and analytics tools (Tableau, Looker, Power BI, etc.)
  • Production experience working with very large datasets
  • Experience with big data cloud technologies like EMR, Athena, Glue, Big Query, Dataproc, Dataflow.
  • Data Science/Machine Learning know-how
  • A desire to constantly challenge the norm
  • Willing to attend conferences, webinars and meet-ups and share the learning

What are the perks?

  • 25 days holiday, 10% bonus (paid quarterly), pension, enhanced maternity and paternity leave, group life insurance scheme, private medical healthcare
  • Discounted gym membership, cycle to work scheme, interest free season ticket loan
  • Breakfast, dinner, fresh fruit, snacks and drinks
  • Dynamic, open culture with lots of social activities