Skip to content

Allow truncation when embedding #14493

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 4 commits into
base: master
Choose a base branch
from

Conversation

huydt84
Copy link
Collaborator

@huydt84 huydt84 commented Jul 2, 2025

Sometimes it frustrates me because llama-server automatically stops when slot.n_ctx < input token length in embedding task. I want it to be able to truncate the input token, as an option.

@huydt84 huydt84 requested a review from ngxson as a code owner July 2, 2025 04:31
@huydt84
Copy link
Collaborator Author

huydt84 commented Jul 4, 2025

@ngxson Please check

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants