Skip to content

Commit

Permalink
merge conflict
Browse files Browse the repository at this point in the history
  • Loading branch information
mvpatel2000 committed Oct 10, 2023
1 parent f2fef59 commit 6640ebd
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions megablocks/layers/mlp.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
from megablocks.layers import weight_parallel as wp
from megablocks.layers.arguments import Arguments, InitFn
from megablocks import turbo_util as turbo
from megablocks import grouped_gemm_util as grouped_gemm
from megablocks import grouped_gemm_util as gg
import stk
import torch
import torch.nn.functional as F
Expand Down Expand Up @@ -522,6 +522,6 @@ def forward(self, x, tokens_per_expert):
self.args.quantize_rematerialize_num_bits)

# Compute the MLP.
x = grouped_gemm.gmm(x, w1, batch_sizes, trans_b=True)
x = gg.ops.gmm(x, w1, batch_sizes, trans_b=True)
x = F.gelu(x, approximate="tanh")
return grouped_gemm.gmm(x, w2, batch_sizes)
return gg.ops.gmm(x, w2, batch_sizes)

0 comments on commit 6640ebd

Please sign in to comment.