Skip to content
This repository was archived by the owner on Jul 22, 2025. It is now read-only.

Commit 7357280

Browse files
authored
FEATURE: Add old OpenAI tokenizer to embeddings (#1487)
1 parent 6f8960e commit 7357280

File tree

2 files changed

+2
-1
lines changed

2 files changed

+2
-1
lines changed

app/models/embedding_definition.rb

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ def tokenizer_names
2323
DiscourseAi::Tokenizer::GeminiTokenizer,
2424
DiscourseAi::Tokenizer::MultilingualE5LargeTokenizer,
2525
DiscourseAi::Tokenizer::OpenAiTokenizer,
26+
DiscourseAi::Tokenizer::OpenAiCl100kTokenizer,
2627
DiscourseAi::Tokenizer::MistralTokenizer,
2728
DiscourseAi::Tokenizer::QwenTokenizer,
2829
].map(&:name)

plugin.rb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010

1111
gem "tokenizers", "0.5.4"
1212
gem "tiktoken_ruby", "0.0.11.1"
13-
gem "discourse_ai-tokenizers", "0.2.0", require_name: "discourse_ai/tokenizers"
13+
gem "discourse_ai-tokenizers", "0.3.1", require_name: "discourse_ai/tokenizers"
1414
gem "ed25519", "1.2.4" #TODO remove this as existing ssl gem should handle this
1515

1616
# we probably want to move all dependencies directly in to the Discourse Gemfile, this

0 commit comments

Comments
 (0)