- # Re-generate the lookahead_reward pessimistically in the
- # previous iterations
- ar_mask = (t < u).long() * (t % it_len == index_lookahead_reward).long()
- ar(result, ar_mask, logit_biases=-optimistic_bias)
- snapshots.append(result[:10].detach().clone())
-
- # Generate the state
- ar_mask = (t >= u).long() * (t < u + state_len).long()
+ # Generate the next state but keep the initial one, the
+ # lookahead_reward of previous iterations are set to
+ # UNKNOWN
+ if u > 0:
+ result[
+ :, u + self.world.index_lookahead_reward
+ ] = self.world.lookahead_reward2code(greed.REWARD_UNKNOWN)
+ ar_mask = (t >= u + self.world.index_states).long() * (
+ t < u + self.world.index_states + self.world.state_len
+ ).long()
+ ar(result, ar_mask)
+
+ # Generate the action and reward with lookahead_reward to +1
+ result[
+ :, u + self.world.index_lookahead_reward
+ ] = self.world.lookahead_reward2code(greed.REWARD_PLUS)
+ ar_mask = (t >= u + self.world.index_reward).long() * (
+ t <= u + self.world.index_action
+ ).long()