Skip to content

Is the QNN backend support the model of Llama 3.2 3B instead of XNNPACK? #9311

Open
@tombang

Description

@tombang

🐛 Describe the bug

I have run the XNNPACK tutorial and the pte file can run normally. but when I run the tutorial of Llama 3 8B and change the model to Llama 3.2 3B, the model load fail on the Android device of Qualcomm 8Gen 2, the fail code is 1.

QNN:2.26
SDK:r27b

Versions

QNN:2.26
SDK:r27b

cc @cccclai @winskuo-quic @shewu-quic @cbilgin

Metadata

Metadata

Assignees

No one assigned

    Labels

    module: qnnIssues related to Qualcomm's QNN delegate and code under backends/qualcomm/need-user-inputThe issue needs more information from the reporter before moving forwardpartner: qualcommFor backend delegation, kernels, demo, etc. from the 3rd-party partner, Qualcomm

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions