+
+def run_tests(model, quiz_machine, deterministic_synthesis, local_device=main_device):
+ with torch.autograd.no_grad():
+ model.eval().to(local_device)
+
+ nb_test_samples, acc_test_loss = 0, 0.0
+ nb_samples_accumulated = 0
+
+ for input in quiz_machine.batches(model, split="test"):
+ input = input.to(local_device)
+
+ bs = model(mygpt.BracketedSequence(input))
+ output = bs.x
+
+ loss = F.cross_entropy(output.transpose(1, 2), input)
+
+ acc_test_loss += loss.item() * input.size(0)
+
+ nb_test_samples += input.size(0)
+
+ test_perplexity = math.exp(min(100, acc_test_loss / nb_test_samples))
+
+ log_string(f"test_perplexity {n_epoch} model {model.id} {test_perplexity}")
+
+ model.main_test_accuracy = quiz_machine.produce_results(
+ n_epoch=n_epoch,
+ model=model,
+ result_dir=args.result_dir,
+ deterministic_synthesis=deterministic_synthesis,
+ )
+
+
+def one_epoch(model, quiz_machine, local_device=main_device):
+ model.to(local_device).train()
+
+ optimizer = torch.optim.Adam(model.parameters(), lr=args.learning_rate)
+
+ nb_train_samples, acc_train_loss = 0, 0.0
+
+ for input in quiz_machine.batches(model, split="train"):
+ input = input.to(local_device)
+
+ if nb_train_samples % args.batch_size == 0:
+ optimizer.zero_grad()
+
+ output = model(mygpt.BracketedSequence(input)).x
+ loss = F.cross_entropy(output.transpose(1, 2), input)
+ acc_train_loss += loss.item() * input.size(0)
+
+ nb_train_samples += input.size(0)
+
+ loss.backward()
+
+ if nb_train_samples % args.batch_size == 0:
+ optimizer.step()
+
+ train_perplexity = math.exp(min(100, acc_train_loss / nb_train_samples))
+
+ log_string(f"train_perplexity {n_epoch} model {model.id} {train_perplexity}")
+
+ run_tests(model, quiz_machine, deterministic_synthesis=False)
+
+ model.to(main_device)
+
+
+######################################################################
+
+# This is the key routine that decides what generated quizzes to keep
+
+
+def compute_valid_quizzes(token_logprobas):
+ warnings.warn("validation with uniform constraints", RuntimeWarning)
+ l = token_logprobas.min(dim=-1).values.sort(dim=-1).values
+ return (l[:, 0] < math.log(0.1)) & (l[:, 1] > math.log(0.5))
+
+
+def compute_valid_quizzes_(token_logprobas):
+ l = token_logprobas.sum(dim=-1).sort(dim=-1).values
+ return (l[:, 0] < math.log(args.proba_not_understands)) & (
+ l[:, 1] > math.log(args.proba_understands)
+ )
+
+
+def extract_valid_quizzes_and_logprobas(recorded):
+ validated_quizzes, validated_logprobas = [], []
+ for quizzes, token_logprobas in recorded:
+ validated_indices = compute_valid_quizzes(token_logprobas)
+ validated_quizzes.append(quizzes[validated_indices])
+ validated_logprobas.append(token_logprobas[validated_indices])
+
+ if len(validated_quizzes) > 0:
+ return torch.cat(validated_quizzes, dim=0), torch.cat(
+ validated_logprobas, dim=0
+ )
+ else:
+ return None, None
+
+
+######################################################################
+
+
+def create_c_quizzes(models, quiz_machine, nb_for_train=1000, nb_for_test=100):
+ nb_to_create = nb_for_train + nb_for_test
+
+ recorded_quizzes_logprobas = []
+
+ nb_validated = 0
+
+ while nb_validated < nb_to_create:
+ model_for_generation = models[torch.randint(len(models), (1,))]
+
+ c_quizzes = quiz_machine.generate_quizzes(
+ nb_to_create,
+ model_for_generation=model_for_generation,
+ temperature=args.generation_temperature,
+ )
+
+ c_quizzes = c_quizzes[quiz_machine.non_trivial(c_quizzes)]
+
+ if c_quizzes.size(0) > 0:
+ token_logproba = quiz_machine.solution_token_logprobas(models, c_quizzes)
+ recorded_quizzes_logprobas.append((c_quizzes, token_logproba))
+
+ (
+ validated_quizzes,
+ validated_logprobas,
+ ) = extract_valid_quizzes_and_logprobas(recorded_quizzes_logprobas)
+
+ if validated_quizzes is not None:
+ nb_validated = validated_quizzes.size(0)
+
+ log_string(
+ f"keep c_quizzes model {model_for_generation.id} nb_accumulated {nb_validated} / {nb_to_create}"
+ )
+
+ # store the new c_quizzes which have been validated
+
+ quiz_machine.reverse_random_half_in_place(validated_quizzes)
+ quiz_machine.store_c_quizzes(validated_quizzes[:nb_for_train], for_train=True)
+ quiz_machine.store_c_quizzes(
+ validated_quizzes[nb_for_train:nb_to_create], for_train=False
+ )
+
+ ######################################################################
+ # save images with their logprobas
+
+ vq = validated_quizzes[:72]
+ vl = validated_logprobas[:72]
+
+ if vq.size(0) > 0:
+ prefix = f"culture_c_quiz_{n_epoch:04d}"
+ filename = os.path.join(args.result_dir, prefix + "_logp.pth")
+ torch.save(vl, filename)
+ # with open(file_name, "w") as logp_file:
+ # for l in vl:
+ # s = " ".join([str(x.item()) for x in l])
+ # logp_file.write(s + "\n")
+
+ quiz_machine.save_quiz_illustrations(args.result_dir, prefix, vq)
+
+
+######################################################################
+