lora: make sure model keep track of associated adapters (#18490)
* lora: make sure model keep track of associated adapters * deprecate llama_adapter_lora_free * minor : std::unordered_set over std::set --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
This commit is contained in:
parent
2a13180100
commit
a7e6ddb8bd
7 changed files with 24 additions and 22 deletions
|
|
@ -468,7 +468,11 @@ llama_model::llama_model(const llama_model_params & params) : params(params), pi
|
|||
pimpl->has_tensor_overrides = params.tensor_buft_overrides && params.tensor_buft_overrides[0].pattern;
|
||||
}
|
||||
|
||||
llama_model::~llama_model() = default;
|
||||
llama_model::~llama_model() {
|
||||
for (auto * lora : loras) {
|
||||
delete lora;
|
||||
}
|
||||
}
|
||||
|
||||
void llama_model::load_stats(llama_model_loader & ml) {
|
||||
pimpl->n_elements = ml.n_elements;
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue