General Information
Req #
WD00085739
Career area:
Research/Development
Country/Region:
China
State:
Beijing
City:
北京(Beijing)
Date:
Friday, August 22, 2025
Working time:
Full-time
Additional Locations:
- China - Beijing - 北京(Beijing)
Why Work at Lenovo
We are Lenovo. We do what we say. We own what we do. We WOW our customers.
Lenovo is a US$69 billion revenue global technology powerhouse, ranked #196 in the Fortune Global 500, and serving millions of customers every day in 180 markets. Focused on a bold vision to deliver Smarter Technology for All, Lenovo has built on its success as the world’s largest PC company with a full-stack portfolio of AI-enabled, AI-ready, and AI-optimized devices (PCs, workstations, smartphones, tablets), infrastructure (server, storage, edge, high performance computing and software defined infrastructure), software, solutions, and services. Lenovo’s continued investment in world-changing innovation is building a more equitable, trustworthy, and smarter future for everyone, everywhere. Lenovo is listed on the Hong Kong stock exchange under Lenovo Group Limited (HKSE: 992) (ADR: LNVGY).
This transformation together with Lenovo’s world-changing innovation is building a more inclusive, trustworthy, and smarter future for everyone, everywhere. To find out more visit www.lenovo.com, and read about the latest news via our StoryHub.
Description and Requirements
Advisory Engineer, Large Model On-Device Inference (C++)
Job Description:
- Responsible for the architecture design, development, maintenance, optimization, and innovative exploration of AI on-device inference engines.
- Responsible for verifying and analyzing product stability, performance, and accuracy.
Requirements:
- Master’s degree or higher in Computer Science, Networking, Communications, or related fields;
- Familiar with operating system principles, with extensive programming experience on Windows/Linux systems and product development experience.
- Proficient in C/C++ programming, including modern standards such as C++11; familiar with STL; knowledgeable in common scripting languages such as Shell and Python.
- Solid understanding of network communication principles; proficient in developing applications using TCP/UDP/HTTP; experienced with mainstream network programming models.
- Familiar with at least one inference engine in terms of application, underlying principles, and customized development, including inference engine architecture, operators, and model conversion tools.
- Hands-on development experience with llama.cpp is a plus
- Prior experience in parallel computing frameworks such as CUDA or SYCL is a plus
Additional Locations:
- China - Beijing - 北京(Beijing)
- China
- China - Beijing
- China - Beijing - 北京(Beijing)
举报职位