Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
18 changes: 9 additions & 9 deletions src/llama-quant.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ static void zeros(std::ofstream & file, size_t n) {
}
}

static std::string remap_layer(const std::string & orig_name, const std::vector<int>& prune, std::map<int, std::string>& mapped, int& next_id) {
static std::string remap_layer(const std::string & orig_name, const std::vector<int> & prune, std::map<int, std::string> & mapped, int & next_id) {
if (prune.empty()) {
return orig_name;
}
Expand All @@ -39,7 +39,7 @@ static std::string remap_layer(const std::string & orig_name, const std::vector<
if (mapped.count(blk)) {
// Already mapped, do nothing
} else if (std::find(prune.begin(), prune.end(), blk) != prune.end()) {
mapped[blk] = "X";
mapped[blk] = "";
} else if (blk < prune.front()) {
mapped[blk] = std::to_string(blk);
next_id = blk + 1;
Expand All @@ -48,13 +48,13 @@ static std::string remap_layer(const std::string & orig_name, const std::vector<
++next_id;
}

return mapped[blk] == "X" ? mapped[blk] : new_name.replace(match.position(1), match.length(1), mapped[blk]);
return mapped[blk].empty() ? mapped[blk] : new_name.replace(match.position(1), match.length(1), mapped[blk]);
}

return orig_name;
}

static std::string remap_imatrix (const std::string & orig_name, const std::map<int, std::string>& mapped) {
static std::string remap_imatrix (const std::string & orig_name, const std::map<int, std::string> & mapped) {
if (mapped.empty()) {
return orig_name;
}
Expand Down Expand Up @@ -628,7 +628,6 @@ static void llama_model_quantize_impl(const std::string & fname_inp, const std::
gguf_set_val_u32(ctx_out.get(), "general.file_type", ftype); // TODO: use LLM_KV

if (!prune_list.empty()) {
gguf_set_val_bool(ctx_out.get(), "general.pruned", true);
uint32_t block_count = 0;
ml.get_key(LLM_KV_BLOCK_COUNT, block_count);
gguf_set_val_u32(ctx_out.get(), ml.llm_kv(LLM_KV_BLOCK_COUNT).c_str(), block_count - prune_list.size());
Expand Down Expand Up @@ -667,10 +666,11 @@ static void llama_model_quantize_impl(const std::string & fname_inp, const std::
tensors.reserve(ml.weights_map.size());
for (const auto & it : ml.weights_map) {
const std::string remapped_name(remap_layer(it.first, prune_list, mapped, next_blk_id));
if (remapped_name == "X") {
if (it.first.find("attn_v.weight") != std::string::npos ||
it.first.find("attn_qkv.weight") != std::string::npos ||
it.first.find("attn_kv_b.weight")!= std::string::npos) {
if (remapped_name.empty()) {
if (false
|| it.first.find("attn_v.weight") != std::string::npos
|| it.first.find("attn_qkv.weight") != std::string::npos
|| it.first.find("attn_kv_b.weight")!= std::string::npos) {
pruned_attention_w++;
}
LLAMA_LOG_DEBUG("%s: prunning tensor %s\n", __func__, it.first.c_str());
Expand Down
14 changes: 6 additions & 8 deletions tools/quantize/quantize.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -293,24 +293,22 @@ static bool parse_layer_prune(const char * data, std::vector<int> & prune_layers
}

const auto block_ids = string_split<std::string>(data, ',');

for ( const auto & block_id : block_ids) {

for (const auto & block_id : block_ids) {
int id;
try {
std::stoi(block_id);
id = std::stoi(block_id);
} catch (...) {
printf("%s: invalid layer id '%s'\n\n", __func__, block_id.c_str());
return false;
id = -1;
}

int id = std::stoi(block_id);
if (id < 0) {
printf("\n%s: invalid layer id '%s'\n\n", __func__, block_id.c_str());
return false;
}
prune_layers.emplace_back(id);
}

sort(prune_layers.begin(), prune_layers.end());
prune_layers.erase(std::unique(prune_layers.begin(), prune_layers.end()), prune_layers.end());
return true;
}

Expand Down
Loading