This repository was archived by the owner on Jul 22, 2025. It is now read-only.
File tree Expand file tree Collapse file tree 2 files changed +4
-3
lines changed Expand file tree Collapse file tree 2 files changed +4
-3
lines changed Original file line number Diff line number Diff line change @@ -99,7 +99,8 @@ def tokenizer
9999 end
100100
101101 it "limits the system message to 60% of available tokens" do
102- prompt = DiscourseAi ::Completions ::Prompt . new ( "I'm a system message consisting of 10 tokens" )
102+ prompt =
103+ DiscourseAi ::Completions ::Prompt . new ( "I'm a system message consisting of 10 tokens okay" )
103104 prompt . push ( type : :user , content : five_token_msg )
104105
105106 dialect = TestDialect . new ( prompt , llm_model )
@@ -109,7 +110,7 @@ def tokenizer
109110
110111 expect ( trimmed ) . to eq (
111112 [
112- { type : :system , content : "I'm a system message consisting of 10" } ,
113+ { type : :system , content : "I'm a system message consisting of 10 tokens " } ,
113114 { type : :user , content : five_token_msg } ,
114115 ] ,
115116 )
Original file line number Diff line number Diff line change @@ -18,7 +18,7 @@ def response(content, tool_call: false)
1818 model : "gpt-3.5-turbo-0301" ,
1919 usage : {
2020 prompt_tokens : 8 ,
21- completion_tokens : 13 ,
21+ completion_tokens : 12 ,
2222 total_tokens : 499 ,
2323 } ,
2424 choices : [
You can’t perform that action at this time.
0 commit comments