|
16 | 16 | #include "model.h" |
17 | 17 | #include "stable-diffusion.h" |
18 | 18 | #include "util.h" |
19 | | -#include "vocab.hpp" |
20 | | -#include "vocab_mistral.hpp" |
21 | | -#include "vocab_qwen.hpp" |
22 | | -#include "vocab_umt5.hpp" |
23 | 19 |
|
24 | 20 | #include "ggml-alloc.h" |
25 | 21 | #include "ggml-backend.h" |
@@ -1340,36 +1336,6 @@ void ModelLoader::set_wtype_override(ggml_type wtype, std::string tensor_type_ru |
1340 | 1336 | } |
1341 | 1337 | } |
1342 | 1338 |
|
1343 | | -std::string ModelLoader::load_merges() { |
1344 | | - std::string merges_utf8_str(reinterpret_cast<const char*>(merges_utf8_c_str), sizeof(merges_utf8_c_str)); |
1345 | | - return merges_utf8_str; |
1346 | | -} |
1347 | | - |
1348 | | -std::string ModelLoader::load_qwen2_merges() { |
1349 | | - std::string merges_utf8_str(reinterpret_cast<const char*>(qwen2_merges_utf8_c_str), sizeof(qwen2_merges_utf8_c_str)); |
1350 | | - return merges_utf8_str; |
1351 | | -} |
1352 | | - |
1353 | | -std::string ModelLoader::load_mistral_merges() { |
1354 | | - std::string merges_utf8_str(reinterpret_cast<const char*>(mistral_merges_utf8_c_str), sizeof(mistral_merges_utf8_c_str)); |
1355 | | - return merges_utf8_str; |
1356 | | -} |
1357 | | - |
1358 | | -std::string ModelLoader::load_mistral_vocab_json() { |
1359 | | - std::string json_str(reinterpret_cast<const char*>(mistral_vocab_json_utf8_c_str), sizeof(mistral_vocab_json_utf8_c_str)); |
1360 | | - return json_str; |
1361 | | -} |
1362 | | - |
1363 | | -std::string ModelLoader::load_t5_tokenizer_json() { |
1364 | | - std::string json_str(reinterpret_cast<const char*>(t5_tokenizer_json_str), sizeof(t5_tokenizer_json_str)); |
1365 | | - return json_str; |
1366 | | -} |
1367 | | - |
1368 | | -std::string ModelLoader::load_umt5_tokenizer_json() { |
1369 | | - std::string json_str(reinterpret_cast<const char*>(umt5_tokenizer_json_str), sizeof(umt5_tokenizer_json_str)); |
1370 | | - return json_str; |
1371 | | -} |
1372 | | - |
1373 | 1339 | bool ModelLoader::load_tensors(on_new_tensor_cb_t on_new_tensor_cb, int n_threads_p, bool enable_mmap) { |
1374 | 1340 | int64_t process_time_ms = 0; |
1375 | 1341 | std::atomic<int64_t> read_time_ms(0); |
|
0 commit comments