X-Git-Url: https://fleuret.org/cgi-bin/gitweb/gitweb.cgi?a=blobdiff_plain;f=mygpt.py;h=5ea927e09b7b1835bc1222a6ddf5329868d26da9;hb=3c5ce93138700c33a055f83ac1a46efb2975e28a;hp=24ba34591bb6833e9f0a550a4e104074486cec4a;hpb=43fbfaac1850098f5b1a9470c8e6ca3d5ab479fe;p=mygptrnn.git diff --git a/mygpt.py b/mygpt.py index 24ba345..5ea927e 100755 --- a/mygpt.py +++ b/mygpt.py @@ -656,16 +656,13 @@ class Caterpillar(nn.Module): self.rec_K[:, :, t0:t1] = next_K.flatten(2, 3) if self.training and self.proba_flashback: - # insert_flash_back( - # self.rec_V, - # V, - # self.rec_K, - # K, - # t0, - # t1, - # CL, - # proba=self.proba_flashback / CL, - # ) + # insert_flash_back(self.rec_V,V,self.rec_K,K,t0,t1,CL,proba=self.proba_flashback / CL,) + + # This piece of code makes the assumption that there is + # nothing informative before t0, otherwise we'd have to + # implement a cache for V and K too. This should not be + # too much of a problem since this is used only during + # train, where full sequence are available n = torch.arange(N, device=X.device)[:, None, None, None] t = torch.arange(t0, t1, device=X.device)[None, None, :, None]