Skip to content

AttributeError: module 'intel_extension_for_pytorch.quantization' has no attribute 'WoqActQuantMode' #456

Open
@azhuvath

Description

@azhuvath

Describe the issue

Getting the error while trying to run the below command.

Step 2: Generate quantized model with INT4 weights

Provide checkpoint file name by --low-precision-checkpoint

python single_instance/run_llama_quantization.py --ipex-weight-only-quantization --output-dir "saved_results" --int8-bf16-mixed -m meta-llama/Llama-2-7b-chat-hf --low-precision-checkpoint "saved_results/gptq_checkpoint.pt"

Do I need to source install IPEX as opposed to pip install?

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions