Sr. Data Engineer

Lenovo
Edinburgh
2 days ago
Create job alert

The Lenovo AI Technology Center (LATC)—Lenovo’s global AI Center of Excellence—is driving our transformation into an AI‑first organization. We are assembling a world‑class team of researchers, engineers, and innovators to position Lenovo and its customers at the forefront of the generational shift toward AI. Lenovo is one of the world’s leading computing companies, delivering products across the entire technology spectrum, spanning wearables, smartphones (Motorola), laptops (ThinkPad, Yoga), PCs, workstations, servers, and services/solutions.

This unmatched breadth gives us a unique canvas for AI innovation, including the ability to rapidly deploy cutting‑edge foundation models and to enable flexible, hybrid‑cloud, and agentic computing across our full product portfolio. To this end, we are building the next wave of AI core technologies and platforms that leverage and evolve with the fast‑moving AI ecosystem, including novel model and agentic orchestration & collaboration across mobile, edge, and cloud resources.

Responsibilities
  • Data Creation & Annotation: Design, build, and implement processes for creating task‑specific training datasets. This may include data labeling, annotation, and data augmentation techniques.
  • Data Pipeline Development: Leverage tools and technologies to accelerate dataset creation and improvement. This includes scripting, automation, and potentially working with data labeling platforms.
  • Data Quality & Evaluation: Perform thorough data analysis to assess data quality, identify anomalies, and ensure data integrity. Utilize machine learning tools and techniques to evaluate dataset performance and identify areas for improvement.
  • Big Data Technologies: Utilize database systems (SQL and NoSQL) and big data tools (e.g., Spark, Hadoop, cloud‑based data warehouses like Snowflake/Redshift/BigQuery) to process, transform, and store large datasets.
  • Data Governance & Lineage: Implement and maintain data governance best practices, including data source tracking, data lineage documentation, and license management. Ensure compliance with data privacy regulations.
  • Collaboration with Model Developers: Work closely with machine learning engineers and data scientists to understand their data requirements, provide clean and well‑documented datasets, and iterate on data solutions based on model performance feedback.
  • Documentation: Create and maintain clear and concise documentation for data pipelines, data quality checks, and data governance procedures.
  • Stay Current: Keep up‑to‑date with the latest advancements in data engineering, machine learning, and data governance.
Qualifications
  • Education: Bachelor's or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, Statistics, Mathematics, or a related field.
  • Experience: 15+ years of experience in a data engineering or data science role.
  • Programming Skills: Mastery in Python and SQL. Experience with other languages (e.g., Java, Scala) is a plus.
  • Database Skills: Strong experience with relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., MongoDB, Cassandra).
  • Big Data Tools: Experience with big data technologies such as Spark, Hadoop or cloud‑based data warehousing solutions (Snowflake, Redshift, BigQuery).
  • Data Manipulation: Proficiency in data manipulation and cleaning techniques using tools like Pandas, NumPy, and other data processing libraries.
  • ML Fundamentals: Solid understanding of machine learning concepts and techniques, including data preprocessing, feature engineering, and model evaluation.
  • Data Governance: Understanding of data governance principles and practices, including data lineage, data quality, and data security.
  • Communication Skills: Excellent written and verbal communication skills, with the ability to explain complex technical concepts to both technical and non‑technical audiences.
  • Problem Solving: Strong analytical and problem‑solving skills.
Bonus Points
  • Experience with data labeling platforms (e.g., Labelbox, Scale AI, Amazon SageMaker Ground Truth).
  • Experience with MLOps practices and tools (e.g., Kubeflow, MLflow).
  • Experience with cloud platforms (e.g., AWS, Azure, GCP).
  • Experience with data visualization tools (e.g., Tableau, Power BI).
  • Experience with building and maintaining data pipelines using orchestration tools (e.g. Airflow, Prefect)
What we offer
  • Opportunities for career advancement and personal development
  • Access to a diverse range of training programs
  • Performance‑based rewards that celebrate your achievements
  • Flexibility with a hybrid work model (3:2) that blends home and office life
  • Electric car salary sacrifice scheme
  • Life insurance

This role is open for the Edinburgh, Scotland location only. Candidates must be based there, as the position requires working from the office at least three days per week (3:2 hybrid policy).

Seniority level: Mid‑Senior level

Employment type: Full‑time

Job function: Information Technology and Research

Industries: IT Services and IT Consulting


#J-18808-Ljbffr

Related Jobs

View all jobs

Sr. Software/Data Engineer, Autonomy (AWS/Databricks)

Systematic Trading - Python Quant Data Engineer - Vice President

Data Scientist

Data Scientist

Data Scientist

Sr. Manager Data Analytics - iCasino

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Data Science Jobs for Career Switchers in Their 30s, 40s & 50s (UK Reality Check)

Thinking about switching into data science in your 30s, 40s or 50s? You’re far from alone. Across the UK, businesses are investing in data science talent to turn data into insight, support better decisions and unlock competitive advantage. But with all the hype about machine learning, Python, AI and data unicorns, it can be hard to separate real opportunities from noise. This article gives you a practical, UK-focused reality check on data science careers for mid-life career switchers — what roles really exist, what skills employers really hire for, how long retraining typically takes, what UK recruiters actually look for and how to craft a compelling career pivot story. Whether you come from finance, marketing, operations, research, project management or another field entirely, there are meaningful pathways into data science — and age itself is not the barrier many people fear.

How to Write a Data Science Job Ad That Attracts the Right People

Data science plays a critical role in how organisations across the UK make decisions, build products and gain competitive advantage. From forecasting and personalisation to risk modelling and experimentation, data scientists help translate data into insight and action. Yet many employers struggle to attract the right data science candidates. Job adverts often generate high volumes of applications, but few applicants have the mix of analytical skill, business understanding and communication ability the role actually requires. At the same time, experienced data scientists skip over adverts that feel vague, inflated or misaligned with real data science work. In most cases, the issue is not a lack of talent — it is the quality and clarity of the job advert. Data scientists are analytical, sceptical of hype and highly selective. A poorly written job ad signals unclear expectations and immature data practices. A well-written one signals credibility, focus and serious intent. This guide explains how to write a data science job ad that attracts the right people, improves applicant quality and positions your organisation as a strong data employer.

Maths for Data Science Jobs: The Only Topics You Actually Need (& How to Learn Them)

If you are applying for data science jobs in the UK, the maths can feel like a moving target. Job descriptions say “strong statistical knowledge” or “solid ML fundamentals” but they rarely tell you which topics you will actually use day to day. Here’s the truth: most UK data science roles do not require advanced pure maths. What they do require is confidence with a tight set of practical topics that come up repeatedly in modelling, experimentation, forecasting, evaluation, stakeholder comms & decision-making. This guide focuses on the only maths most data scientists keep using: Statistics for decision making (confidence intervals, hypothesis tests, power, uncertainty) Probability for real-world data (base rates, noise, sampling, Bayesian intuition) Linear algebra essentials (vectors, matrices, projections, PCA intuition) Calculus & gradients (enough to understand optimisation & backprop) Optimisation & model evaluation (loss functions, cross-validation, metrics, thresholds) You’ll also get a 6-week plan, portfolio projects & a resources section you can follow without getting pulled into unnecessary theory.