Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

adaption for moe models #2101

Open
wants to merge 23 commits into
base: main
Choose a base branch
from
Open
Changes from 1 commit
Commits
Show all changes
23 commits
Select commit Hold shift + click to select a range
b75c001
adaption for moe models
Sep 26, 2024
c29810b
FIX: Change check if past_key_values is empty (#2106)
BenjaminBossan Sep 27, 2024
aa3bd8f
DOC Update source install instruction (#2110)
Salehbigdeli Sep 30, 2024
2a80735
FIX Refactor OFT, small changes to BOFT (#1996)
Zeju1997 Oct 1, 2024
ae297f0
ENH: Improved attribute access for modules_to_save (#2117)
BenjaminBossan Oct 2, 2024
ca8462b
FIX low_cpu_mem_usage consolidates devices (#2113)
BenjaminBossan Oct 2, 2024
534d361
TST Mark flaky X-LoRA test as xfail (#2114)
BenjaminBossan Oct 2, 2024
d9d3059
ENH: Warn when from_pretrained misses PEFT keys (#2118)
BenjaminBossan Oct 2, 2024
8d9ecbe
FEAT: Adding exclude modules param(#2044) (#2102)
JINO-ROHIT Oct 3, 2024
e6f927b
FIX BC breaking change to boft conv2d scaling variable (#2127)
Zeju1997 Oct 7, 2024
859fd88
FEAT: VeRA quantization using bitsandbytes (#2070) (#2076)
ZiadHelal Oct 7, 2024
5e91b54
Bump version to 0.13.2.dev0 (#2137)
BenjaminBossan Oct 8, 2024
9918977
FEAT: Support torchao (#2062)
BenjaminBossan Oct 8, 2024
a724834
FIX: PiSSA now works with Conv1D layers (#2103) (#2104)
suyang160 Oct 8, 2024
3b314cc
FIX Type annoations in vera/bnb.py (#2139)
BenjaminBossan Oct 9, 2024
85e3202
ENH Make PEFT configs forward compatible (#2038)
BenjaminBossan Oct 9, 2024
8efa0cb
FIX Raise mixed adapter infer with missing adapter (#2090)
BenjaminBossan Oct 9, 2024
1eab9bd
FIX Prompt learning with latest transformers error (#2140)
BenjaminBossan Oct 9, 2024
5758a7e
ENH LoRA notebook for NER task (#2126)
JINO-ROHIT Oct 10, 2024
0aa7e3a
FIX TST NaN issue with HQQ GPU test (#2143)
BenjaminBossan Oct 10, 2024
c925d0a
FIX Bug in target module optimization if suffix (#2144)
BenjaminBossan Oct 10, 2024
749b924
Bump version to 0.13.2.dev0 (#2145)
BenjaminBossan Oct 11, 2024
669ce90
Merge branch 'dhr_moe'
Oct 12, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions src/peft/tuners/lora/tp_layer.py
Original file line number Diff line number Diff line change
Expand Up @@ -62,6 +62,7 @@ def __init__(
self.is_parallel_a = isinstance(base_layer, backend.RowParallelLinear)
self.fan_in_fan_out = fan_in_fan_out
self._active_adapter = adapter_name
self.is_expert = base_layer.is_expert

megatron_config = kwargs["megatron_config"]
parallel_linear_kwargs = {"megatron_config": megatron_config}
Expand Down Expand Up @@ -131,6 +132,7 @@ def update_layer(
skip_bias_add=True,
init_method=init_method,
config=megatron_config,
is_expert=self.is_expert,
)
lora_b = nn.Linear(in_features=r, out_features=self.out_features, bias=False, dtype=torch.float32)
else:
Expand All @@ -142,6 +144,7 @@ def update_layer(
gather_output=gather_output,
init_method=init_method,
config=megatron_config,
is_expert=self.is_expert,
)
self.lora_A[adapter_name] = lora_a
self.lora_B[adapter_name] = lora_b
Expand Down
Loading