@@ -610,7 +610,7 @@ enum llm_tensor {
610
610
LLM_TENSOR_CLS_OUT,
611
611
};
612
612
613
- static const std::map<llm_arch, std::map<llm_tensor, std::string >> LLM_TENSOR_NAMES = {
613
+ static const std::map<llm_arch, std::map<llm_tensor, const char * >> LLM_TENSOR_NAMES = {
614
614
{
615
615
LLM_ARCH_LLAMA,
616
616
{
@@ -1566,32 +1566,32 @@ struct LLM_TN {
1566
1566
return LLM_TENSOR_NAMES.at(arch).at(tensor);
1567
1567
}
1568
1568
1569
- std::string operator()(llm_tensor tensor, const std::string & suffix) const {
1569
+ std::string operator()(llm_tensor tensor, const char * suffix) const {
1570
1570
if (LLM_TENSOR_NAMES.at(arch).find(tensor) == LLM_TENSOR_NAMES.at(arch).end()) {
1571
1571
return "__missing__";
1572
1572
}
1573
- return LLM_TENSOR_NAMES.at(arch).at(tensor) + "." + suffix;
1573
+ return std::string( LLM_TENSOR_NAMES.at(arch).at(tensor) ) + "." + suffix;
1574
1574
}
1575
1575
1576
1576
std::string operator()(llm_tensor tensor, int bid) const {
1577
1577
if (LLM_TENSOR_NAMES.at(arch).find(tensor) == LLM_TENSOR_NAMES.at(arch).end()) {
1578
1578
return "__missing__";
1579
1579
}
1580
- return ::format(LLM_TENSOR_NAMES.at(arch).at(tensor).c_str() , bid);
1580
+ return ::format(LLM_TENSOR_NAMES.at(arch).at(tensor), bid);
1581
1581
}
1582
1582
1583
- std::string operator()(llm_tensor tensor, const std::string & suffix, int bid) const {
1583
+ std::string operator()(llm_tensor tensor, const char * suffix, int bid) const {
1584
1584
if (LLM_TENSOR_NAMES.at(arch).find(tensor) == LLM_TENSOR_NAMES.at(arch).end()) {
1585
1585
return "__missing__";
1586
1586
}
1587
- return ::format(LLM_TENSOR_NAMES.at(arch).at(tensor).c_str() , bid) + "." + suffix;
1587
+ return ::format(LLM_TENSOR_NAMES.at(arch).at(tensor), bid) + "." + suffix;
1588
1588
}
1589
1589
1590
- std::string operator()(llm_tensor tensor, const std::string & suffix, int bid, int xid) const {
1590
+ std::string operator()(llm_tensor tensor, const char * suffix, int bid, int xid) const {
1591
1591
if (LLM_TENSOR_NAMES.at(arch).find(tensor) == LLM_TENSOR_NAMES.at(arch).end()) {
1592
1592
return "__missing__";
1593
1593
}
1594
- return ::format(LLM_TENSOR_NAMES.at(arch).at(tensor).c_str() , bid, xid) + "." + suffix;
1594
+ return ::format(LLM_TENSOR_NAMES.at(arch).at(tensor), bid, xid) + "." + suffix;
1595
1595
}
1596
1596
};
1597
1597
@@ -4916,7 +4916,7 @@ struct llama_model_loader {
4916
4916
static const int TENSOR_NOT_REQUIRED = 1;
4917
4917
static const int TENSOR_DUPLICATED = 2;
4918
4918
4919
- struct ggml_tensor * create_tensor(struct ggml_context * ctx, const std::string & name, const std::vector <int64_t> & ne, int flags = 0) {
4919
+ struct ggml_tensor * create_tensor(struct ggml_context * ctx, const std::string & name, const std::initializer_list <int64_t> & ne, int flags = 0) {
4920
4920
const struct ggml_tensor * cur = check_tensor_dims(name, ne, !(flags & TENSOR_NOT_REQUIRED));
4921
4921
4922
4922
if (cur == NULL) {
@@ -4926,7 +4926,7 @@ struct llama_model_loader {
4926
4926
return create_tensor_for(ctx, cur, flags & TENSOR_DUPLICATED);
4927
4927
}
4928
4928
4929
- struct ggml_tensor * create_tensor_as_view(struct ggml_context * ctx, struct ggml_tensor * base, const std::string & name, const std::vector <int64_t> & ne, size_t offset, bool required = true) {
4929
+ struct ggml_tensor * create_tensor_as_view(struct ggml_context * ctx, struct ggml_tensor * base, const std::string & name, const std::initializer_list <int64_t> & ne, size_t offset, bool required = true) {
4930
4930
const struct ggml_tensor * cur = check_tensor_dims(name, ne, required);
4931
4931
4932
4932
if (cur == NULL) {
@@ -4939,7 +4939,7 @@ struct llama_model_loader {
4939
4939
4940
4940
std::array<int64_t, GGML_MAX_DIMS> dims;
4941
4941
for (size_t i = 0; i < GGML_MAX_DIMS; ++i) {
4942
- dims[i] = i < ne.size() ? ne[i] : 1;
4942
+ dims[i] = i < ne.size() ? ne.begin() [i] : 1;
4943
4943
}
4944
4944
4945
4945
struct ggml_tensor * tensor = ggml_view_4d(ctx, base,
0 commit comments