Skip to content

How to recover stage quantization from finetuning stage after an error #957

Closed
@jiangjiadi

Description

@jiangjiadi

After executing examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py, I obtained three models at different stages: stage_sparsity, stage_finetuning, and stage_quantization. My question is, if an error occurs while running the stage_quantization phase, how can I resume the process from the stage_finetuning model? Given that both the stage_sparsity and stage_finetuning phases are resource-intensive, restarting from the beginning would be inefficient and time-consuming.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions