Qualcomm® AI HubAI Hub
HomeAppsLLM Chat Windows

LLM Chat Windows

Deploy and run LLMs via CLI

A sample CLI app demonstrating how to integrate an on‑device LLM with Qualcomm® Gen AI Inference Extensions (Genie) APIs, optimized for Snapdragon® NPU acceleration.

1. Choose a compatible model
2. Get source code and dependencies
3. Build and run app

Compatible Models

App Information

Operating System

Windows 11+

Language

C++

Runtime

Qualcomm® Gen AI Inference Extensions (Genie)

Use Case

Text Generation

Join us on Slack

Join our growing community of 2000+ developers in the AI Hub Slack Channel building the future of AI
Join Now