X-Git-Url: https://fleuret.org/cgi-bin/gitweb/gitweb.cgi?a=blobdiff_plain;ds=sidebyside;f=sky.py;h=d2a4568919c31f097966979988914a529652880a;hb=3b41e2797fc340fd11cb35015b57c3cae1e8447b;hp=6ba3882f681fb81e99b617385bbb430c310562f8;hpb=29679cb42710602037fee650a5672f01a3292077;p=culture.git diff --git a/sky.py b/sky.py index 6ba3882..d2a4568 100755 --- a/sky.py +++ b/sky.py @@ -37,13 +37,14 @@ class Sky(problem.Problem): token_background = 0 first_bird_token = 1 nb_bird_tokens = colors.size(0) - 1 - token_forward = first_bird_token + nb_bird_tokens - token_backward = token_forward + 1 token2char = ( "_" + "".join([chr(ord("A") + n) for n in range(len(colors) - 1)]) + "><" ) + def nb_token_values(self): + return len(self.colors) + def __init__( self, height=6, @@ -60,9 +61,6 @@ class Sky(problem.Problem): self.nb_iterations = nb_iterations self.avoid_collision = avoid_collision - def direction_tokens(self): - return self.token_forward, self.token_backward - def generate_frame_sequences(self, nb): frame_sequences = [] @@ -118,8 +116,14 @@ class Sky(problem.Problem): dtype=torch.int64, ) + fine = torch.empty(self.nb_iterations * self.speed) + + t_to_keep = ( + torch.arange(self.nb_iterations, device=result.device) * self.speed + ) + for l in range(self.nb_iterations * self.speed): - fine = collision_okay() + fine[l] = collision_okay() for n in range(self.nb_birds): c = col[n] result[l, i[n], j[n]] = c @@ -139,40 +143,24 @@ class Sky(problem.Problem): i[n] += vi[n] j[n] += vj[n] - if fine: + result = result[t_to_keep] + fine = fine[t_to_keep] + + if fine[-1]: break - frame_sequences.append( - result[ - torch.arange(self.nb_iterations, device=result.device) * self.speed - ] - ) + frame_sequences.append(result) return frame_sequences ###################################################################### - def generate_token_sequences(self, nb): + def generate_prompts_and_answers(self, nb): frame_sequences = self.generate_frame_sequences(nb) - - result = [] - - for frame_sequence in frame_sequences: - a = [] - if torch.rand(1) < 0.5: - for frame in frame_sequence: - if len(a) > 0: - a.append(torch.tensor([self.token_forward])) - a.append(frame.flatten()) - else: - for frame in reversed(frame_sequence): - if len(a) > 0: - a.append(torch.tensor([self.token_backward])) - a.append(frame.flatten()) - - result.append(torch.cat(a, dim=0)[None, :]) - - return torch.cat(result, dim=0) + frame_sequences = torch.cat([x[None] for x in frame_sequences], dim=0) + prompts = frame_sequences[:, : frame_sequences.size(1) // 2].flatten(1) + answers = frame_sequences[:, frame_sequences.size(1) // 2 :].flatten(1) + return prompts, answers ###################################################################### @@ -203,92 +191,82 @@ class Sky(problem.Problem): return x - def seq2img(self, seq, scale=15): - all = [ - self.frame2img( - seq[:, : self.height * self.width].reshape(-1, self.height, self.width), - scale, + def seq2str(self, seq): + result = [] + for s in seq: + result.append("".join([self.token2char[v] for v in s])) + return result + + def save_image( + self, + result_dir, + filename, + prompts, + answers, + predicted_prompts=None, + predicted_answers=None, + ): + if predicted_prompts is None: + predicted_prompts = 255 + + if predicted_answers is None: + predicted_answers = 255 + + def add_frame(x, c, margin): + y = x.new_full( + (x.size(0), x.size(1), x.size(2) + 2 * margin, x.size(3) + 2 * margin), + 0, ) - ] + if type(c) is int: + y[...] = c + else: + c = c.long()[:, None] + c = c * torch.tensor([192, 192, 192], device=c.device) + ( + 1 - c + ) * torch.tensor([255, 255, 255], device=c.device) + y[...] = c[:, :, None, None] + y[:, :, margin:-margin, margin:-margin] = x + return y - separator = torch.full((seq.size(0), 3, self.height * scale - 1, 1), 0) + margin = 4 - t = self.height * self.width + img_prompts = add_frame(self.frame2img(prompts.to("cpu")), 0, 1) + img_answers = add_frame(self.frame2img(answers.to("cpu")), 0, 1) - while t < seq.size(1): - direction_tokens = seq[:, t] - t += 1 + # img_prompts = add_frame(img_prompts, 255, margin) + # img_answers = add_frame(img_answers, 255, margin) - direction_images = self.colors[ - torch.full( - (direction_tokens.size(0), self.height * scale - 1, scale), 0 - ) - ].permute(0, 3, 1, 2) - - for n in range(direction_tokens.size(0)): - if direction_tokens[n] == self.token_forward: - for k in range(scale): - for l in [0, 1]: - direction_images[ - n, - :, - (self.height * scale) // 2 - scale // 2 + k - l, - 3 + scale // 2 - abs(k - scale // 2), - ] = 0 - elif direction_tokens[n] == self.token_backward: - for k in range(scale): - for l in [0, 1]: - direction_images[ - n, - :, - (self.height * scale) // 2 - scale // 2 + k - l, - 3 + abs(k - scale // 2), - ] = 0 - else: - for k in range(2, scale - 2): - for l in [0, 1]: - direction_images[ - n, - :, - (self.height * scale) // 2 - scale // 2 + k - l, - k, - ] = 0 - direction_images[ - n, - :, - (self.height * scale) // 2 - scale // 2 + k - l, - scale - 1 - k, - ] = 0 - - all += [ - separator, - direction_images, - separator, - self.frame2img( - seq[:, t : t + self.height * self.width].reshape( - -1, self.height, self.width - ), - scale, - ), - ] - - t += self.height * self.width - - return torch.cat(all, dim=3) + img_prompts = add_frame(img_prompts, predicted_prompts, margin) + img_answers = add_frame(img_answers, predicted_answers, margin) - def seq2str(self, seq): - result = [] - for s in seq: - result.append("".join([self.token2char[v] for v in s])) - return result + separator = img_prompts.new_full( + (img_prompts.size(0), img_prompts.size(1), img_prompts.size(2), margin), 255 + ) + + img = torch.cat([img_prompts, img_answers], dim=3) - def save_image(self, input, result_dir, filename): - img = self.seq2img(input.to("cpu")) image_name = os.path.join(result_dir, filename) - torchvision.utils.save_image(img.float() / 255.0, image_name, nrow=6, padding=4) + torchvision.utils.save_image( + img.float() / 255.0, image_name, nrow=6, padding=margin * 2, pad_value=1.0 + ) - def save_quizzes(self, input, result_dir, filename_prefix): - self.save_image(input, result_dir, filename_prefix + ".png") + def save_quizzes( + self, + result_dir, + filename_prefix, + prompts, + answers, + predicted_prompts=None, + predicted_answers=None, + ): + self.save_image( + result_dir, + filename_prefix + ".png", + prompts, + answers, + predicted_prompts, + predicted_answers, + ) ###################################################################### @@ -296,12 +274,21 @@ class Sky(problem.Problem): if __name__ == "__main__": import time - sky = Sky(height=6, width=8, speed=2, nb_iterations=2) + sky = Sky(height=6, width=8, speed=4, nb_iterations=2) + + prompts, answers = sky.generate_prompts_and_answers(4) + + predicted_prompts = torch.rand(prompts.size(0)) < 0.5 + predicted_answers = torch.rand(answers.size(0)) < 0.5 - start_time = time.perf_counter() - token_sequences = sky.generate_token_sequences(nb=64) - delay = time.perf_counter() - start_time - print(f"{token_sequences.size(0)/delay:02f} seq/s") + sky.save_quizzes( + "/tmp", "test", prompts, answers, predicted_prompts, predicted_answers + ) + + # start_time = time.perf_counter() + # token_sequences = sky.generate_token_sequences(nb=64) + # delay = time.perf_counter() - start_time + # print(f"{token_sequences.size(0)/delay:02f} seq/s") # print(sky.seq2str(seq[:4])) @@ -319,9 +306,9 @@ if __name__ == "__main__": # seq = (1 - m) * seq + m * 23 # print(seq.size()) - img = sky.seq2img(token_sequences) + # img = sky.seq2img(token_sequences) # print(img.size()) - torchvision.utils.save_image( - img.float() / 255.0, "/tmp/world.png", nrow=6, padding=6, pad_value=0 - ) + # torchvision.utils.save_image( + # img.float() / 255.0, "/tmp/world.png", nrow=6, padding=6, pad_value=0 + # )