Update.
[culture.git] / sky.py
diff --git a/sky.py b/sky.py
index 1e6ed4d..040ec67 100755 (executable)
--- a/sky.py
+++ b/sky.py
@@ -37,91 +37,101 @@ class Sky(problem.Problem):
     token_background = 0
     first_bird_token = 1
     nb_bird_tokens = colors.size(0) - 1
-    token_forward = first_bird_token + nb_bird_tokens
-    token_backward = token_forward + 1
 
     token2char = (
         "_" + "".join([chr(ord("A") + n) for n in range(len(colors) - 1)]) + "><"
     )
 
-    def __init__(self, height=6, width=8, nb_birds=3, nb_iterations=2):
+    def __init__(
+        self,
+        height=6,
+        width=8,
+        nb_birds=3,
+        speed=2,
+        nb_iterations=2,
+        avoid_collision=True,
+    ):
         self.height = height
         self.width = width
         self.nb_birds = nb_birds
+        self.speed = speed
         self.nb_iterations = nb_iterations
+        self.avoid_collision = avoid_collision
 
-    def direction_tokens(self):
-        return self.token_forward, self.token_backward
-
-    def generate_seq(self, nb, return_iterations=False):
-        pairs = []
-        kept_iterations = []
+    def generate_frame_sequences(self, nb):
+        frame_sequences = []
 
         for _ in tqdm.tqdm(range(nb), dynamic_ncols=True, desc="world generation"):
-            while True:
-                iterations = []
-
-                f_start = torch.zeros(self.height, self.width, dtype=torch.int64)
+            i, j, vi, vj = (
+                torch.empty(self.nb_birds, dtype=torch.int64),
+                torch.empty(self.nb_birds, dtype=torch.int64),
+                torch.empty(self.nb_birds, dtype=torch.int64),
+                torch.empty(self.nb_birds, dtype=torch.int64),
+            )
 
-                i, j, vi, vj = (
-                    torch.empty(self.nb_birds, dtype=torch.int64),
-                    torch.empty(self.nb_birds, dtype=torch.int64),
-                    torch.empty(self.nb_birds, dtype=torch.int64),
-                    torch.empty(self.nb_birds, dtype=torch.int64),
-                )
+            def collision_okay():
+                if not self.avoid_collision:
+                    return True
 
-                col = (
-                    torch.randperm(self.colors.size(0) - 1)[: self.nb_birds]
-                    .sort()
-                    .values
-                    + 1
-                )
+                count = torch.zeros(self.height, self.width, dtype=torch.int64)
 
                 for n in range(self.nb_birds):
-                    c = col[n]
-
-                    while True:
-                        i[n], j[n] = (
-                            torch.randint(self.height, (1,))[0],
-                            torch.randint(self.width, (1,))[0],
-                        )
-                        vm = torch.randint(4, (1,))[0]
-                        vi[n], vj[n] = (vm % 2) * 2 - 1, (vm // 2) * 2 - 1
-                        if (
-                            i[n] - vi[n] >= 0
-                            and i[n] - vi[n] < self.height
-                            and j[n] - vj[n] >= 0
-                            and j[n] - vj[n] < self.width
-                            and f_start[i[n], j[n]] == 0
-                            and f_start[i[n] - vi[n], j[n]] == 0
-                            and f_start[i[n], j[n] - vj[n]] == 0
-                        ):
-                            break
+                    count[i[n], j[n]] += 1
+                    count[i[n] - vi[n], j[n]] += 1
+                    count[i[n], j[n] - vj[n]] += 1
 
-                    f_start[i[n], j[n]] = c
-                    f_start[i[n] - vi[n], j[n]] = c
-                    f_start[i[n], j[n] - vj[n]] = c
+                return count.max() <= 1
 
-                f_end = f_start.clone()
+            col = (
+                torch.randperm(self.colors.size(0) - 1)[: self.nb_birds].sort().values
+                + 1
+            )
 
-                for l in range(self.nb_iterations):
-                    iterations.append(f_end.clone())
-                    f_end[...] = 0
-                    nb_collisions = 0
+            while True:
+                while True:
                     for n in range(self.nb_birds):
-                        c = col[n]
+                        while True:
+                            i[n] = torch.randint(self.height, (1,))
+                            j[n] = torch.randint(self.width, (1,))
+                            vm = torch.randint(4, (1,))
+                            vi[n], vj[n] = (vm % 2) * 2 - 1, (vm // 2) * 2 - 1
+                            if (
+                                i[n] - vi[n] >= 0
+                                and i[n] - vi[n] < self.height
+                                and j[n] - vj[n] >= 0
+                                and j[n] - vj[n] < self.width
+                            ):
+                                break
+
+                    if collision_okay():
+                        break
+
+                result = torch.zeros(
+                    self.nb_iterations * self.speed,
+                    self.height,
+                    self.width,
+                    dtype=torch.int64,
+                )
+
+                fine = torch.empty(self.nb_iterations * self.speed)
 
-                        pi, pj, pvi, pvj = (
-                            i[n].item(),
-                            j[n].item(),
-                            vi[n].item(),
-                            vj[n].item(),
-                        )
+                t_to_keep = (
+                    torch.arange(self.nb_iterations, device=result.device) * self.speed
+                )
+
+                for l in range(self.nb_iterations * self.speed):
+                    fine[l] = collision_okay()
+                    for n in range(self.nb_birds):
+                        c = col[n]
+                        result[l, i[n], j[n]] = c
+                        result[l, i[n] - vi[n], j[n]] = c
+                        result[l, i[n], j[n] - vj[n]] = c
 
                         if (i[n] == 0 and vi[n] == -1) or (
                             i[n] == self.height - 1 and vi[n] == 1
                         ):
                             vi[n] = -vi[n]
+
                         if (j[n] == 0 and vj[n] == -1) or (
                             j[n] == self.width - 1 and vj[n] == 1
                         ):
@@ -130,226 +140,133 @@ class Sky(problem.Problem):
                         i[n] += vi[n]
                         j[n] += vj[n]
 
-                        if not (
-                            f_end[i[n], j[n]] == 0
-                            and f_end[i[n] - vi[n], j[n]] == 0
-                            and f_end[i[n], j[n] - vj[n]] == 0
-                        ):
-                            nb_collisions += 1
-
-                        f_end[i[n], j[n]] = c
-                        f_end[i[n] - vi[n], j[n]] = c
-                        f_end[i[n], j[n] - vj[n]] = c
+                result = result[t_to_keep]
+                fine = fine[t_to_keep]
 
-                iterations.append(f_end.clone())
-
-                if nb_collisions == 0:
+                if fine[-1]:
                     break
 
-            kept_iterations.append(iterations)
-            pairs.append((f_start, f_end))
+            frame_sequences.append(result)
 
-        result = []
-        for p in pairs:
-            if torch.rand(1) < 0.5:
-                result.append(
-                    torch.cat(
-                        [
-                            p[0].flatten(),
-                            torch.tensor([self.token_forward]),
-                            p[1].flatten(),
-                        ],
-                        dim=0,
-                    )[None, :]
-                )
-            else:
-                result.append(
-                    torch.cat(
-                        [
-                            p[1].flatten(),
-                            torch.tensor([self.token_backward]),
-                            p[0].flatten(),
-                        ],
-                        dim=0,
-                    )[None, :]
-                )
-
-        if return_iterations:
-            # iterations = torch.cat([ torch.cat([ x[None, None] for x in l], dim = 1) for l in kept_iterations ], dim=0)
-            return torch.cat(result, dim=0), kept_iterations
-        else:
-            return torch.cat(result, dim=0)
+        return frame_sequences
 
     ######################################################################
 
-    def generate_seq_old(
-        self,
-        nb,
-    ):
-        pairs = []
-
-        for n in tqdm.tqdm(range(nb), dynamic_ncols=True, desc="world generation"):
-            f_start = torch.zeros(self.height, self.width, dtype=torch.int64)
-            f_end = torch.zeros(self.height, self.width, dtype=torch.int64)
-            n = torch.arange(f_start.size(0))
-
-            for c in (
-                (torch.randperm(self.nb_bird_tokens) + self.first_bird_token)[
-                    : self.nb_birds
-                ]
-                .sort()
-                .values
-            ):
-                i, j = (
-                    torch.randint(self.height - 2, (1,))[0] + 1,
-                    torch.randint(self.width - 2, (1,))[0] + 1,
-                )
-                vm = torch.randint(4, (1,))[0]
-                vi, vj = (vm // 2) * (2 * (vm % 2) - 1), (1 - vm // 2) * (
-                    2 * (vm % 2) - 1
-                )
-
-                f_start[i, j] = c
-                f_start[i - vi, j - vj] = c
-                f_start[i + vj, j - vi] = c
-                f_start[i - vj, j + vi] = c
-
-                for l in range(self.nb_iterations):
-                    i += vi
-                    j += vj
-                    if i < 0 or i >= self.height or j < 0 or j >= self.width:
-                        i -= vi
-                        j -= vj
-                        vi, vj = -vi, -vj
-                        i += vi
-                        j += vj
-
-                f_end[i, j] = c
-                f_end[i - vi, j - vj] = c
-                f_end[i + vj, j - vi] = c
-                f_end[i - vj, j + vi] = c
-
-            pairs.append((f_start, f_end))
-
-        result = []
-        for p in pairs:
-            if torch.rand(1) < 0.5:
-                result.append(
-                    torch.cat(
-                        [
-                            p[0].flatten(),
-                            torch.tensor([self.token_forward]),
-                            p[1].flatten(),
-                        ],
-                        dim=0,
-                    )[None, :]
-                )
-            else:
-                result.append(
-                    torch.cat(
-                        [
-                            p[1].flatten(),
-                            torch.tensor([self.token_backward]),
-                            p[0].flatten(),
-                        ],
-                        dim=0,
-                    )[None, :]
-                )
-
-        return torch.cat(result, dim=0)
-
-    def frame2img(self, x, upscale=15):
-        x = x.reshape(-1, self.height, self.width)
+    def frame2img(self, x, scale=15):
+        x = x.reshape(x.size(0), self.height, -1)
         m = torch.logical_and(
             x >= 0, x < self.first_bird_token + self.nb_bird_tokens
         ).long()
         x = self.colors[x * m].permute(0, 3, 1, 2)
         s = x.shape
-        x = x[:, :, :, None, :, None].expand(-1, -1, -1, upscale, -1, upscale)
-        x = x.reshape(s[0], s[1], s[2] * upscale, s[3] * upscale)
+        x = x[:, :, :, None, :, None].expand(-1, -1, -1, scale, -1, scale)
+        x = x.reshape(s[0], s[1], s[2] * scale, s[3] * scale)
 
-        x[:, :, :, torch.arange(0, x.size(3), upscale)] = 0
-        x[:, :, torch.arange(0, x.size(2), upscale), :] = 0
+        x[:, :, :, torch.arange(0, x.size(3), scale)] = 0
+        x[:, :, torch.arange(0, x.size(2), scale), :] = 0
         x = x[:, :, 1:, 1:]
 
         for n in range(m.size(0)):
             for i in range(m.size(1)):
                 for j in range(m.size(2)):
                     if m[n, i, j] == 0:
-                        for k in range(2, upscale - 2):
-                            x[n, :, i * upscale + k, j * upscale + k] = 0
-                            x[n, :, i * upscale + upscale - 1 - k, j * upscale + k] = 0
+                        for k in range(2, scale - 2):
+                            for l in [0, 1]:
+                                x[n, :, i * scale + k, j * scale + k - l] = 0
+                                x[
+                                    n, :, i * scale + scale - 1 - k, j * scale + k - l
+                                ] = 0
 
         return x
 
-    def seq2img(self, seq, upscale=15):
-        f_first = seq[:, : self.height * self.width].reshape(
-            -1, self.height, self.width
-        )
-        f_second = seq[:, self.height * self.width + 1 :].reshape(
-            -1, self.height, self.width
-        )
-        direction = seq[:, self.height * self.width]
-
-        direction_symbol = torch.full(
-            (direction.size(0), self.height * upscale - 1, upscale), 0
-        )
-        direction_symbol = self.colors[direction_symbol].permute(0, 3, 1, 2)
-        separator = torch.full((direction.size(0), 3, self.height * upscale - 1, 1), 0)
-
-        for n in range(direction_symbol.size(0)):
-            if direction[n] == self.token_forward:
-                for k in range(upscale):
-                    direction_symbol[
-                        n,
-                        :,
-                        (self.height * upscale) // 2 - upscale // 2 + k,
-                        3 + upscale // 2 - abs(k - upscale // 2),
-                    ] = 0
-            elif direction[n] == self.token_backward:
-                for k in range(upscale):
-                    direction_symbol[
-                        n,
-                        :,
-                        (self.height * upscale) // 2 - upscale // 2 + k,
-                        3 + abs(k - upscale // 2),
-                    ] = 0
-            else:
-                for k in range(2, upscale - 2):
-                    direction_symbol[
-                        n, :, (self.height * upscale) // 2 - upscale // 2 + k, k
-                    ] = 0
-                    direction_symbol[
-                        n,
-                        :,
-                        (self.height * upscale) // 2 - upscale // 2 + k,
-                        upscale - 1 - k,
-                    ] = 0
-
-        return torch.cat(
-            [
-                self.frame2img(f_first, upscale),
-                separator,
-                direction_symbol,
-                separator,
-                self.frame2img(f_second, upscale),
-            ],
-            dim=3,
-        )
-
     def seq2str(self, seq):
         result = []
         for s in seq:
             result.append("".join([self.token2char[v] for v in s]))
         return result
 
-    def save_image(self, input, result_dir, filename):
-        img = self.seq2img(input.to("cpu"))
+    def save_image(
+        self,
+        result_dir,
+        filename,
+        prompts,
+        answers,
+        predicted_prompts=None,
+        predicted_answers=None,
+    ):
+        if predicted_prompts is None:
+            predicted_prompts = 255
+
+        if predicted_answers is None:
+            predicted_answers = 255
+
+        def add_frame(x, c, margin):
+            y = x.new_full(
+                (x.size(0), x.size(1), x.size(2) + 2 * margin, x.size(3) + 2 * margin),
+                0,
+            )
+            if type(c) is int:
+                y[...] = c
+            else:
+                c = c.long()[:, None]
+                c = c * torch.tensor([192, 192, 192], device=c.device) + (
+                    1 - c
+                ) * torch.tensor([255, 255, 255], device=c.device)
+                y[...] = c[:, :, None, None]
+            y[:, :, margin:-margin, margin:-margin] = x
+            return y
+
+        margin = 4
+
+        img_prompts = add_frame(self.frame2img(prompts.to("cpu")), 0, 1)
+        img_answers = add_frame(self.frame2img(answers.to("cpu")), 0, 1)
+
+        # img_prompts = add_frame(img_prompts, 255, margin)
+        # img_answers = add_frame(img_answers, 255, margin)
+
+        img_prompts = add_frame(img_prompts, predicted_prompts, margin)
+        img_answers = add_frame(img_answers, predicted_answers, margin)
+
+        separator = img_prompts.new_full(
+            (img_prompts.size(0), img_prompts.size(1), img_prompts.size(2), margin), 255
+        )
+
+        img = torch.cat([img_prompts, img_answers], dim=3)
+
         image_name = os.path.join(result_dir, filename)
-        torchvision.utils.save_image(img.float() / 255.0, image_name, nrow=6, padding=4)
+        torchvision.utils.save_image(
+            img.float() / 255.0, image_name, nrow=6, padding=margin * 2, pad_value=1.0
+        )
 
-    def save_quizzes(self, input, result_dir, filename_prefix):
-        self.save_image(input, result_dir, filename_prefix + ".png")
+    ######################################################################
+
+    def nb_token_values(self):
+        return len(self.colors)
+
+    def generate_prompts_and_answers(self, nb):
+        frame_sequences = self.generate_frame_sequences(nb)
+        frame_sequences = torch.cat([x[None] for x in frame_sequences], dim=0)
+        prompts = frame_sequences[:, : frame_sequences.size(1) // 2].flatten(1)
+        answers = frame_sequences[:, frame_sequences.size(1) // 2 :].flatten(1)
+        return prompts, answers
+
+    def save_quizzes(
+        self,
+        result_dir,
+        filename_prefix,
+        prompts,
+        answers,
+        predicted_prompts=None,
+        predicted_answers=None,
+    ):
+        self.save_image(
+            result_dir,
+            filename_prefix + ".png",
+            prompts,
+            answers,
+            predicted_prompts,
+            predicted_answers,
+        )
 
 
 ######################################################################
@@ -357,31 +274,41 @@ class Sky(problem.Problem):
 if __name__ == "__main__":
     import time
 
-    sky = Sky(height=6, width=8, nb_iterations=100)
+    sky = Sky(height=6, width=8, speed=1, nb_iterations=4)
 
-    start_time = time.perf_counter()
-    seq, it = sky.generate_seq(nb=64, return_iterations=True)
-    delay = time.perf_counter() - start_time
-    print(f"{seq.size(0)/delay:02f} samples/s")
+    prompts, answers = sky.generate_prompts_and_answers(4)
 
-    print(sky.seq2str(seq[:4]))
+    predicted_prompts = torch.rand(prompts.size(0)) < 0.5
+    predicted_answers = torch.rand(answers.size(0)) < 0.5
 
-    for t in range(len(it[0])):
-        img = torch.cat([sky.frame2img(f[t]) for f in it], dim=0)
-        torchvision.utils.save_image(
-            img.float() / 255.0,
-            f"/tmp/frame_{t:03d}.png",
-            nrow=8,
-            padding=6,
-            pad_value=0,
-        )
+    sky.save_quizzes(
+        "/tmp", "test", prompts, answers, predicted_prompts, predicted_answers
+    )
+
+    # start_time = time.perf_counter()
+    # token_sequences = sky.generate_token_sequences(nb=64)
+    # delay = time.perf_counter() - start_time
+    # print(f"{token_sequences.size(0)/delay:02f} seq/s")
+
+    # print(sky.seq2str(seq[:4]))
+
+    # for t in range(len(it[0])):
+    # img = torch.cat([sky.frame2img(f[t]) for f in it], dim=0)
+    # torchvision.utils.save_image(
+    # img.float() / 255.0,
+    # f"/tmp/frame_{t:03d}.png",
+    # nrow=8,
+    # padding=6,
+    # pad_value=0,
+    # )
 
     # m = (torch.rand(seq.size()) < 0.05).long()
     # seq = (1 - m) * seq + m * 23
 
-    img = sky.seq2img(seq)
-    print(img.size())
+    # print(seq.size())
+    # img = sky.seq2img(token_sequences)
+    # print(img.size())
 
-    torchvision.utils.save_image(
-        img.float() / 255.0, "/tmp/world.png", nrow=6, padding=6, pad_value=0
-    )
+    torchvision.utils.save_image(
+    # img.float() / 255.0, "/tmp/world.png", nrow=6, padding=6, pad_value=0
+    )