X-Git-Url: https://fleuret.org/cgi-bin/gitweb/gitweb.cgi?a=blobdiff_plain;ds=sidebyside;f=mygpt.py;h=d0fda7e4182878043e74a260f0676654fc12193f;hb=eaed6307836d88abe7c0f4be733a38364ba20e2f;hp=7119c7a4ef12d49d0f1e164a0839437496d955af;hpb=bfcef9a8c82ed45528601e85725166241bbee916;p=culture.git diff --git a/mygpt.py b/mygpt.py index 7119c7a..d0fda7e 100755 --- a/mygpt.py +++ b/mygpt.py @@ -201,6 +201,26 @@ class QKVAttention(nn.Module): ############################## +class NoiseInjector(nn.Module): + def __init__(self): + super().__init__() + self.noise_std = 0.0 + + def forward(self, x): + if self.noise_std > 0: + x = x + torch.randn(x.size(), device=x.device) * self.noise_std + return x + + +def set_noise_injection(model, noise_std): + for m in model.modules(): + if isinstance(m, NoiseInjector): + m.noise_std = noise_std + + +############################## + + class MyGPT(nn.Module): def __init__( self, @@ -228,7 +248,10 @@ class MyGPT(nn.Module): for b in range(nb_blocks): trunk_blocks += [ WithResidual( - CacheWrapper(nn.LayerNorm((dim_model,))), + CacheWrapper( + nn.LayerNorm((dim_model,)), + NoiseInjector(), + ), QKVAttention( dim_in=dim_model, dim_qk=dim_keys, @@ -241,6 +264,7 @@ class MyGPT(nn.Module): WithResidual( CacheWrapper( nn.LayerNorm((dim_model,)), + NoiseInjector(), nn.Linear(in_features=dim_model, out_features=dim_hidden), nn.ReLU(), nn.Linear(in_features=dim_hidden, out_features=dim_model),