-
-
Notifications
You must be signed in to change notification settings - Fork 3.1k
fix(reranker): respect top_n in the request
#7025
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
✅ Deploy Preview for localai ready!
To edit notification comments on pull requests, go to your Netlify project configuration. |
Signed-off-by: Mikhail Khludnev <mkhl@apache.org>
61ee6b4 to
51d6347
Compare
Signed-off-by: Mikhail Khludnev <mkhl@apache.org>
|
presumably it might be fixed for llama_cpp rerank as well, if @mudler consider it necessary. |
|
Hey, thanks for looking at this!
I think for llama.cpp we should be covered already, as top_k/top_p settings are already propagated for the chat (and rerankers follows along the same reasoning): LocalAI/backend/cpp/llama-cpp/grpc-server.cpp Line 104 in 2573102
Unless you noticed something off! |
top_n in the request
|
I just realized that the most users don't pass |
Description
This PR fixes #6700
Notes for Reviewers
Signed commits