japanese-mpt-7b / custom_embedding.py
higuhigu-lb's picture
0847334f7c1baf70652c5d27a20b940a54fc28bd82501ee07212e50019f9eeb9
7a4bc3f
raw
history blame
305 Bytes
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch import Tensor
class SharedEmbedding(nn.Embedding):
def forward(self, input: Tensor, unembed: bool=False) -> Tensor:
if unembed:
return F.linear(input, self.weight)
return super().forward(input)