Skip to content

Commit

Permalink
squash! llama : std::move llm_bigram_bpe from work_queue
Browse files Browse the repository at this point in the history
Rename lama_priority_queue -> llama_priority_queue.
  • Loading branch information
danbev committed Aug 20, 2024
1 parent 29c5129 commit a9fdcfd
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions src/llama-vocab.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -322,7 +322,7 @@ struct llm_tokenizer_spm {
// TODO: there are a lot of common parts between spm and bpe tokenizers, should be refactored and reused

template<typename T, typename Container = std::vector<T>, typename Compare = std::less<typename Container::value_type>>
class lama_priority_queue : public std::priority_queue<T, Container, Compare> {
class llama_priority_queue : public std::priority_queue<T, Container, Compare> {
public:
using std::priority_queue<T, Container, Compare>::priority_queue;

Expand All @@ -344,7 +344,7 @@ struct llm_bigram_bpe {
};

using queue_storage = std::vector<llm_bigram_bpe>;
using queue = lama_priority_queue<llm_bigram_bpe, queue_storage, comparator>;
using queue = llama_priority_queue<llm_bigram_bpe, queue_storage, comparator>;
llm_symbol::index left;
llm_symbol::index right;
std::string text;
Expand Down

0 comments on commit a9fdcfd

Please sign in to comment.