Commit 068e672 1 parent f2cac91 commit 068e672 Copy full SHA for 068e672
File tree 1 file changed +2
-2
lines changed
vllm/attention/backends/mla
1 file changed +2
-2
lines changed Original file line number Diff line number Diff line change @@ -231,15 +231,15 @@ def process_weights_after_loading(self):
231
231
.flatten (start_dim = 0 , end_dim = 1 ).contiguous ()
232
232
233
233
tp_size = get_tensor_model_parallel_world_size ()
234
- self .o_proj_absored = RowParallelLinear (
234
+ self .o_proj_absorbed = RowParallelLinear (
235
235
self .W_UV_O .shape [0 ] * tp_size ,
236
236
self .W_UV_O .shape [1 ],
237
237
bias = False ,
238
238
# TODO(lucas) figure out how to properly forward quant_method
239
239
#quant_config=self.o_proj.quant_method,
240
240
)
241
241
242
- self .o_proj_absored .weight = torch .nn .Parameter (self .W_UV_O .T )
242
+ self .o_proj_absorbed .weight = torch .nn .Parameter (self .W_UV_O .T )
243
243
else :
244
244
self .W_UV = W_UV
245
245
self .W_UK = W_UK
You can’t perform that action at this time.
0 commit comments