Skip to content

[Feature]: Request for NPU Support on New Surface Devices to Enhance AI Model Execution #1184

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
behroozbc opened this issue May 10, 2025 · 3 comments
Labels
Upstream Tracking an issue in llama.cpp

Comments

@behroozbc
Copy link

Background & Description

LLamaSharp currently supports efficient execution of AI models on CPU and GPU. With the introduction of Neural Processing Units (NPUs) in new Surface devices (e.g., Surface Pro 11 and Surface Laptop 7 powered by Snapdragon X Elite), adding NPU support could significantly improve the performance and speed of AI model execution.

Objective:

Leverage NPU hardware acceleration for faster AI model inference.
Enhance user experience for Surface device owners with NPU capabilities.
Technical Details:

Target Platform: New Surface devices with NPU (e.g., Qualcomm Snapdragon X Elite).
Benefits: NPUs are optimized for matrix operations and deep learning tasks, offloading compute from CPU/GPU.

API & Usage

No response

How to implement

No response

@martindevans
Copy link
Member

Is this something llama.cpp already has support for?

@behroozbc
Copy link
Author

I found out this issue related to it,
ggml-org/llama.cpp#7772

@martindevans martindevans added the Upstream Tracking an issue in llama.cpp label May 11, 2025
@martindevans
Copy link
Member

It looks like there's a lot of work left to do in llama.cpp before we can support anything on our end.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Upstream Tracking an issue in llama.cpp
Projects
None yet
Development

No branches or pull requests

2 participants