Data Engineer (Multiple Roles) - AI SaaS

Vortexa
City of London
1 month ago
Create job alert

Vortexa is a fast‑growing international technology business founded to solve the immense information gap that exists in the energy industry. By using massive amounts of new satellite data and pioneering work in artificial intelligence, Vortexa creates an unprecedented view on the global seaborne energy flows in real‑time, bringing transparency and efficiency to the energy markets and society as a whole.


Role

Processing thousands of rich data points per second from many and vastly different external sources, moving terabytes of data while processing it in real‑time, running complex prediction and forecasting AI models while coupling their output into a hybrid human‑machine data refinement process and presenting the result through a nimble low‑latency SaaS solution used by customers around the globe is no small feat of science and engineering. This processing requires models that can survive the scrutiny of industry experts, data analysts and traders, with the performance, stability, latency and agility a fast‑moving startup influencing multi‑$m transactions requires.


The Data Production Team is responsible for all of Vortexa's data. It ranges from mixing raw satellite data from 600,000 vessels with rich but incomplete text data, to generating high‑value forecasts such as the vessel destination, cargo onboard, ship‑to‑ship transfer detection, dark vessels, congestion, future prices, etc.


The team has built a variety of procedural, statistical and machine learning models that enabled us to provide the most accurate and comprehensive view of energy flows. We take pride in applying cutting‑edge research to real‑world problems in a robust, long‑lasting and maintainable way. The quality of our data is continuously benchmarked and assessed by experienced in‑house market and data analysts to ensure the quality of our predictions.


You'll be instrumental in designing and building infrastructure and applications to propel the design, deployment, and benchmarking of existing and new pipelines and ML models. Working with software and data engineers, data scientists and market analysts, you'll help bridge the gap between scientific experiments and commercial products by ensuring 100% uptime and bullet‑proof fault‑tolerance of every component of the team's data pipelines.


Qualifications

  • Experienced in building and deploying distributed scalable backend data processing pipelines that can go through terabytes of data daily using AWS, K8s, and Airflow.
  • With solid software engineering fundamentals, fluent in both Java and Python (with Rust good to have).
  • Knowledgeable about data lake systems like Athena, and big data storage formats like Parquet, HDF5, ORC, with a focus on data ingestion.
  • Driven by working in an intellectually engaging environment with the top minds in the industry, where constructive and friendly challenges and debates are encouraged, not avoided.
  • Excited about working in a start‑up environment: not afraid of challenges, excited to bring new ideas to production, and a positive can‑do will‑do person, not afraid to push the boundaries of your job role.
  • Passionate about coaching developers, helping them improve their skills and grow their careers.
  • Deep experience of the full software development life cycle (SDLC), including technical design, coding standards, code review, source control, build, test, deploy, and operations.

Preferred Skills

  • Have experience with Apache Kafka and streaming frameworks, e.g., Flink.
  • Familiar with observability principles such as logging, monitoring, and tracing.
  • Have experience with web scraping technologies and information extraction.
  • A vibrant, diverse company pushing ourselves and the technology to deliver beyond the cutting edge.
  • A team of motivated characters and top minds striving to be the best at what we do at all times.
  • Constantly learning and exploring new tools and technologies.
  • Acting as company owners (all Vortexa staff have equity options) – in a business‑savvy and responsible way.
  • Motivated by being collaborative, working and achieving together.
  • A flexible working policy – accommodating both remote & home working, with regular staff events.
  • Private Health Insurance offered via Vitality to help you look after your physical health.
  • Global Volunteering Policy to help you 'do good' and feel better.


#J-18808-Ljbffr

Related Jobs

View all jobs

Data Engineer

Data Engineer

Data Engineer

Data Engineer

Data Engineer

Data Engineer

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.