Merged
Conversation
Collaborator
Author
|
resolves #568 |
Collaborator
Author
|
resolves #474 |
erickgalinkin
approved these changes
Apr 3, 2024
Collaborator
erickgalinkin
left a comment
There was a problem hiding this comment.
Largely looks good to me -- just a few minor comments, and I could be wrong on two of them.
Closed
3 tasks
Closed
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
As of
llama.cppversion 1046 the model format expected for GGML based tooling is now GGUF.This revision improves initialization to validate model file is in
GUFFformat and enhances error handling for subprocess execution.Changes take the approach that first
_call_model()invocation will raise an exception if asubprocess.run()raises an error however subsequent invocations will log the exception and returnNoneallowing the run to continue. Any other exception will be logged and returnNone.Updates to
requirements.txtandpyproject.tomldocument that typing#573 impacts loadinggarakwhen loaded in a debugger such as debugpy. Since the project requires python >= 3.10 and thetypinglibrary is included as a system util since 3.6 in theory the change has zero impact on requirements.Impact of this change:
Consider the case where user error presented the wrong filename for the model.
Garak's output was unclear as to cause.
When executing the command directly however the user error is more obvious:
Once working encoding test later errors with:
By expanding the error handling the testing can now complete.