-
-
Notifications
You must be signed in to change notification settings - Fork 159
Open
Labels
bugSomething isn't workingSomething isn't working
Description
Describe the bug
Feature Request
Description
Currently, LLM Unity only supports CPU inference on Android (LLMLib.cs line 671-674 only adds "android" architecture regardless of GPU settings). This results in very slow inference speeds (2-3 tokens/sec) on mobile devices.
I would like to request Android GPU acceleration support using Vulkan backend.
Current Behavior
numGPULayers setting is visible in Inspector but has no effect on Android PossibleArchitectures() in LLMLib.cs only returns CPU architecture for Android:
csharp
else if (Application.platform == RuntimePlatform.Android)
{
architectures.Add("android"); // Only CPU, no GPU option
}
Steps to reproduce
No response
LLMUnity version
v3.0.0
Operating System
None
Leuconoe
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working