Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ BINARY_NAME=local-ai
DETECT_LIBS?=true

# llama.cpp versions
CPPLLAMA_VERSION?=5bbe6a9fe9a8796a9389c85accec89dbc4d91e39
CPPLLAMA_VERSION?=5e43f104cca1a14874e980326a506b44fde022b8

# whisper.cpp version
WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp
Expand Down
1 change: 0 additions & 1 deletion backend/backend.proto
Original file line number Diff line number Diff line change
Expand Up @@ -165,7 +165,6 @@ message Reply {

message GrammarTrigger {
string word = 1;
bool at_start = 2;
}

message ModelOptions {
Expand Down
14 changes: 7 additions & 7 deletions backend/cpp/llama/grpc-server.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -469,7 +469,7 @@ struct llama_server_context
bool has_eos_token = true;

bool grammar_lazy = false;
std::vector<common_grammar_trigger> grammar_trigger_words;
std::vector<common_grammar_trigger> grammar_triggers;

int32_t n_ctx; // total context for all clients / slots

Expand Down Expand Up @@ -709,7 +709,7 @@ struct llama_server_context
slot->sparams.grammar = json_value(data, "grammar", default_sparams.grammar);
slot->sparams.n_probs = json_value(data, "n_probs", default_sparams.n_probs);
slot->sparams.min_keep = json_value(data, "min_keep", default_sparams.min_keep);
slot->sparams.grammar_trigger_words = grammar_trigger_words;
slot->sparams.grammar_triggers = grammar_triggers;
slot->sparams.grammar_lazy = grammar_lazy;

if (slot->n_predict > 0 && slot->params.n_predict > slot->n_predict) {
Expand Down Expand Up @@ -2393,12 +2393,12 @@ static void params_parse(const backend::ModelOptions* request,
llama.grammar_lazy = true;
for (int i = 0; i < request->grammartriggers_size(); i++) {
common_grammar_trigger trigger;
trigger.word = request->grammartriggers(i).word();
trigger.at_start = request->grammartriggers(i).at_start();
llama.grammar_trigger_words.push_back(trigger);
trigger.type = COMMON_GRAMMAR_TRIGGER_TYPE_WORD;
trigger.value = request->grammartriggers(i).word();
// trigger.at_start = request->grammartriggers(i).at_start();
llama.grammar_triggers.push_back(trigger);
LOG_INFO("grammar trigger", {
{ "word", trigger.word },
{ "at_start", trigger.at_start }
{ "word", trigger.value },
});
}
}
Expand Down
1 change: 0 additions & 1 deletion core/backend/options.go
Original file line number Diff line number Diff line change
Expand Up @@ -122,7 +122,6 @@ func grpcModelOpts(c config.BackendConfig) *pb.ModelOptions {
for _, t := range c.FunctionsConfig.GrammarConfig.GrammarTriggers {
triggers = append(triggers, &pb.GrammarTrigger{
Word: t.Word,
AtStart: t.AtStart,
})

}
Expand Down
1 change: 0 additions & 1 deletion pkg/functions/parse.go
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,6 @@ type GrammarConfig struct {
type GrammarTrigger struct {
// Trigger is the string that triggers the grammar
Word string `yaml:"word"`
AtStart bool `yaml:"at_start"`
}

// FunctionsConfig is the configuration for the tool/function call.
Expand Down
Loading