Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

v0.15.0

Compare
Choose a tag to compare
Loading
@githubnemogithubnemo released this 19 Mar 15:05
· 102 commits to main since this release
b34d8a2
This commit was created on GitHub.com and signed with GitHub’sverified signature.
GPG key ID:B5690EEEBB952194
Verified
Learn about vigilant mode.

Highlights

peft-v0 15 0

New Methods

CorDA: Context-Oriented Decomposition Adaptation

@iboing and@5eqn contributedCorDA: Context-Oriented Decomposition Adaptation of Large Language Models for Task-Aware Parameter-Efficient Fine-tuning . This task-driven initialization method hastwo modes, knowledge-preservation and instruction-preservation, both using external data to select ranks intelligently. The former can be used to select those ranks that correspond to weights not affiliated with knowledge from, say, a QA dataset. The latter can be used to select those ranks that correspond most to the task at hand (e.g., a classification task). (#2231)

Trainable Tokens: Selective token update

The newTrainable Tokens tuner allows for selective training of tokens without re-training the full embedding matrix, e.g. when adding support for reasoning / thinking tokens. This is a lot more memory efficient and the saved checkpoint is much smaller. It can be used standalone orin conjunction with LoRA adapters by passingtrainable_token_indices toLoraConfig. (#2376)

Enhancements

LoRA now supports targeting multihead attention modules (but for now only those with_qkv_same_embed_dim=True). These modules were tricky as they may expose linear submodules but won't use their forward methods, therefore needing explicit support. (#1324)

Hotswapping now allows different alpha scalings and ranks without recompilation of the model when the model is prepared using a call toprepare_model_for_compiled_hotswap() before compiling the model. (#2177)

GPTQModel support was added in#2247 as a replacement for AutoGPTQ which is not maintained anymore.

Changes

  • It's now possible to useall-linear astarget_modules for custom (non-transformers) models (#2267). With this change comes a bugfix where it was possible that non-linear layers were selected when they shared the same name with a linear layer (e.g.,bar.foo andbaz.foo).
  • The internal tuner API was refactored to make method registration easier. With this change the number of changes to numerous files is reduced to a singleregister_peft_method() call. (#2282)
  • PEFT_TYPE_TO_MODEL_MAPPING is now deprecated and should not be relied upon. UsePEFT_TYPE_TO_TUNER_MAPPING instead. (#2282)
  • Mixed adapter batches can now be used in conjunction with beam search. (#2287)
  • It was possible thatmodules_to_save keys wrongly matched parts of the state dict if the key was a substring of another key (e.g.,classifier andclassifier2). (#2334)
  • Auto-casting of the input dtype to the LoRA adapter dtype can now be disabled viadisable_input_dtype_casting=True. (#2353)
  • The config parametersrank_pattern andalpha_pattern used by many adapters now supports matching full paths as well by specifying the pattern with a caret in front, for example:^foo to targetmodel.foo but notmodel.bar.foo. (#2419)
  • AutoPeftModels do not reduce the embedding size anymore if the tokenizer size differs from the embedding size. Only if there are more tokens in the tokenizer than in the embedding matrix, the matrix will be resized. This is to prevent resizing of embedding matrices in models that have 'spare' tokens built-in. (#2427)

What's Changed

New Contributors

Full Changelog:v0.14.0...v0.15.0

Contributors

  • @githubnemo
  • @Qubitium
  • @thedebugger
  • @makelinux
  • @bluenote10
  • @BenjaminBossan
  • @pzdkn
  • @innerlee
  • @bingwork
  • @faaany
  • @iboing
  • @CCLDArjun
  • @henryzhengr
  • @d-kleine
  • @NilBiescas
  • @jiqing-feng
  • @5eqn
  • @gslama12
githubnemo, Qubitium, and 16 other contributors
Assets2
Loading

[8]ページ先頭

©2009-2025 Movatter.jp