Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bug Report: RuntimeError in Google Colab Due to Missing llama-quantize Files #2007

Open
AbhigyanSrivastav opened this issue Mar 13, 2025 · 0 comments

Comments

@AbhigyanSrivastav
Copy link

Description

A recent commit introduced checks for different file extensions (.exe and non-.exe) to support both Linux and Windows environments. However, this change results in a RuntimeError when the required files (llama-quantize, llama-quantize.exe, quantize, or quantize.exe) are not found in the expected directories, causing the llama.cpp installation to fail in Google Colab.

Commit Details

  • Commit: 222b1e7
  • Author: Gennadii Manzhos
  • Date: 2025-02-16
  • Changes:
    • Fixed llama-quantize error on Windows WSL by modifying save.py (fixes GGUF saving issue).
    • Added extension checks for .exe and non-.exe files for Linux and Windows.

Pull Request Details

  • PR: Gemma 3 Bug Fixes
  • Author: Daniel Han
  • Merged by: Daniel Han
  • Files Changed: 7
  • Additions: 224
  • Deletions: 115
  • State: Merged

Error Message

RuntimeError: Unsloth: The file ('llama.cpp/llama-quantize' or 'llama.cpp/llama-quantize.exe' if you are on Windows WSL) or 'llama.cpp/quantize' does not exist. But we expect this file to exist! Maybe the llama.cpp developers changed the name or check extension of the llama-quantize file.

Steps to Reproduce

  1. Install llama.cpp in a Google Colab environment.
  2. Ensure that llama-quantize or quantize files are missing or renamed.
  3. The installation process fails with a RuntimeError.

Expected Behavior

The installation process should complete successfully when the required files are correctly identified and present.

Actual Behavior

The installation fails with a RuntimeError due to missing or misidentified llama-quantize or quantize files in Google Colab.

Additional Information

For reference, please see:

Suggested Fix

Consider modifying the file detection logic to accommodate different execution environments, particularly for Google Colab. A fallback mechanism or a clear warning message with resolution steps would improve usability.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant