AI Engineer - Distributed Inference Specialist
Do you want to be a spectator or a player as the world races to develop AGI?
Are you ready to be a pioneer of AI?
Why join us?
AtDanucore, we are on the hunt forBRILLIANT MINDSto join a team of visionaries and innovators dedicated to building distributedsupercomputersandAI systemswhich are:
Faster⚡️ –from building and deploying AI datacentres at speed to optimising the AI workloads that run on them we want to be the fastest
Cheaper–AI should be accessible to all. We lower the costs of AI deployment with careful hardware deployments and software systems to ensure efficient resource utilisation.
Kinder–Our systems are designed to benefit humanity. We do not allow our systems to participate in military, gambling or pornography applications
Greener–We optimise energy consumption with an integrated hardware and software solution to leverage renewable energy, optimise heat recovery - all running under energy aware orchestration systems to optimise workloads
Cleverer–We develop agentic AI systems and to make our systems intelligent and constantly improving
Help us build systems to ensure the power of frontier AI remainsaccessibleand give userssovereigntyover their AI systems
Join us in ensuring that the most transformative technology in human history remains in the hands of humanity itself. Let's make AI development transparent, accessible, and aligned with the interests of humanity, not just the profits of a few. ⚡
About the Role
This role is for those obsessed with pushing the boundaries of AI model performance.
We're looking for someone who gets excited about shaving milliseconds off inference time, every percentage point of GPU utilization gained and how many Watts were consumed to achieve it. ⚡️
You'll work directly with cutting-edge models — from LLMs to multimodal systems — and large GPU clusters, finding innovative ways to make them run faster, more efficiently, and more accessibly on diverse hardware setups. ️
What We're Looking For
In team members:
- Passion for AI: A strong desire to influence the future of technology and its societal impact.
- Willingness to Learn: we're looking for future experts with curious minds and a growth mindset.
- Open-Mindedness: Ready to challenge the norm and think outside the box?
and for the role:
- Evidence of deploying and optimising AI models in multi gpu and multi node systems ️ ️
- Good working knowledge of leading AI runtimes: PyTorch, vLLM, TensorRT, ONNX Runtime, Llama.cpp ♂️➡️⏱️o
- Experience with distributed inference engines: Ray Serve, Triton Inference Server, vLLM, SLURM
- Knowledge of AI compilers: OpenXLA, torch.compile, OpenAI's triton, MLIR, Mojo, TVM, MLC-LLM ⚙️
- Good working knowledge of inter-process communication: message queues, MPI, NCCL, gRPC
- Good working knowledge of high performance networking: RDMA, RoCE, Infiniband, NVIDIA GPUDirect, NVLink, NVIDIA DOCA, MagnumIO, dpdk, spdk
- Experience with model quantisation, pruning, and sparsity techniques for performance optimisation.
And bonus points if you have:
- a homelab, blog, or a collection of git repos showcasing your talents and interests
- made contributions to open-source projects or publications in the field of AI/ML systems optimisation
Let us know which of the above you have worked with / are relevant in your cover letter! ✨
Key Responsibilities
- Design and implement high-performance distributed inference systems for running large language models and multimodal AI models at scale
- Optimise model serving infrastructure for maximum throughput, minimal latency, and optimal power efficiency ⚡
- Develop and maintain deployment pipelines for efficient model serving, and monitoring in production
- Research and implement cutting-edge techniques in model optimisation, including pruning, quantisation, and sparsity methods
- Design, build and configure experimental hardware setups for model serving and optimisation ️
- Design and implement robust testing frameworks to ensure reliable model serving ✅
- Collaborate with the team to build and improve our distributed inference platform, making it more accessible and efficient for users
- Monitor, optimise and document system performance metrics, including latency, throughput, power consumption and benchmark scores
How Can We Tempt You?
Exceptional Financial Package: Enjoy a competitive compensation structure, including an enticing EMI scheme that rewards your brilliance.
Envious Compute Power: Gain access to a vast array of cutting-edge computing resources to bring your ideas to life!
Support for Your Vision: We believe that the brightest minds often have their own innovative projects. Let's collaborate! Share your ideas, and work with our team and support network to make them happen!
Make an Impact: Join a passionate team dedicated to creating positive change in the world. The future is ours to shape, and together we can ensure it's for the better.
Dynamic Start-Up Culture: Dive in from day one! Experience the thrill of a start-up environment where you can roll up your sleeves and make a real difference right away.
How to Apply
Email your cover letter and CV to with subject "AI Engineer - Distributed Inference"
In your cover letter, please include details of:
- what parts or technologies mentioned in this job advert you have experience with and can add value with
- links to any public work e.g. github profile, blogs or papers