AI Sr. Data Engineer

Lenovo
Edinburgh
1 month ago
Applications closed

Related Jobs

View all jobs

AI Data Analyst

Junior Data scientist

Senior Data Scientist

Junior Data scientist

Head of Data Engineering (AI)

Head of Data Engineering (AI)

The Lenovo AI Technology Center (LATC)—Lenovo’s global AI Center of Excellence—is driving our transformation into an AI‑first organization. We are assembling a world‑class team of researchers, engineers, and innovators to position Lenovo and its customers at the forefront of the generational shift toward AI. Lenovo is one of the world’s leading computing companies, delivering products across the entire technology spectrum, spanning wearables, smartphones (Motorola), laptops (ThinkPad, Yoga), PCs, workstations, servers, and services/solutions.

This unmatched breadth gives us a unique canvas for AI innovation, including the ability to rapidly deploy cutting‑edge foundation models and to enable flexible, hybrid‑cloud, and agentic computing across our full product portfolio. To this end, we are building the next wave of AI core technologies and platforms that leverage and evolve with the fast‑moving AI ecosystem, including novel model and agentic orchestration & collaboration across mobile, edge, and cloud resources.

Responsibilities
  • Data Creation & Annotation: Design, build, and implement processes for creating task‑specific training datasets. This may include data labeling, annotation, and data augmentation techniques.
  • Data Pipeline Development: Leverage tools and technologies to accelerate dataset creation and improvement. This includes scripting, automation, and potentially working with data labeling platforms.
  • Data Quality & Evaluation: Perform thorough data analysis to assess data quality, identify anomalies, and ensure data integrity. Utilize machine learning tools and techniques to evaluate dataset performance and identify areas for improvement.
  • Big Data Technologies: Utilize database systems (SQL and NoSQL) and big data tools (e.g., Spark, Hadoop, cloud‑based data warehouses like Snowflake/Redshift/BigQuery) to process, transform, and store large datasets.
  • Data Governance & Lineage: Implement and maintain data governance best practices, including data source tracking, data lineage documentation, and license management. Ensure compliance with data privacy regulations.
  • Collaboration with Model Developers: Work closely with machine learning engineers and data scientists to understand their data requirements, provide clean and well‑documented datasets, and iterate on data solutions based on model performance feedback.
  • Documentation: Create and maintain clear and concise documentation for data pipelines, data quality checks, and data governance procedures.
  • Stay Current: Keep up‑to‑date with the latest advancements in data engineering, machine learning, and data governance.
Qualifications
  • Education: Bachelor's or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, Statistics, Mathematics, or a related field.
  • Experience: 15+ years of experience in a data engineering or data science role.
  • Programming Skills: Mastery in Python and SQL. Experience with other languages (e.g., Java, Scala) is a plus.
  • Database Skills: Strong experience with relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., MongoDB, Cassandra).
  • Big Data Tools: Experience with big data technologies such as Spark, Hadoop or cloud‑based data warehousing solutions (Snowflake, Redshift, BigQuery).
  • Data Manipulation: Proficiency in data manipulation and cleaning techniques using tools like Pandas, NumPy, and other data processing libraries.
  • ML Fundamentals: Solid understanding of machine learning concepts and techniques, including data preprocessing, feature engineering, and model evaluation.
  • Data Governance: Understanding of data governance principles and practices, including data lineage, data quality, and data security.
  • Communication Skills: Excellent written and verbal communication skills, with the ability to explain complex technical concepts to both technical and non‑technical audiences.
  • Problem Solving: Strong analytical and problem‑solving skills.
Bonus Points
  • Experience with data labeling platforms (e.g., Labelbox, Scale AI, Amazon SageMaker Ground Truth).
  • Experience with MLOps practices and tools (e.g., Kubeflow, MLflow).
  • Experience with cloud platforms (e.g., AWS, Azure, GCP).
  • Experience with data visualization tools (e.g., Tableau, Power BI).
  • Experience with building and maintaining data pipelines using orchestration tools (e.g. Airflow, Prefect)
What we offer
  • Opportunities for career advancement and personal development
  • Access to a diverse range of training programs
  • Performance‑based rewards that celebrate your achievements
  • Flexibility with a hybrid work model (3:2) that blends home and office life
  • Electric car salary sacrifice scheme
  • Life insurance

This role is open for the Edinburgh, Scotland location only. Candidates must be based there, as the position requires working from the office at least three days per week (3:2 hybrid policy).

Seniority level: Mid‑Senior level

Employment type: Full‑time

Job function: Information Technology and Research

Industries: IT Services and IT Consulting


#J-18808-Ljbffr

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.