An overview of LEAF on the Sawyer Push and Reach task. The initial and goal images are encoded by the encoder of the VAE into latent states. Random states are sampled from the currently learned manifold of the VAE latent states, and are used to infer the current frontier. The currently learned deterministic policy is used to reach a state in the frontier from the initial latent state. After that, the currently learned stochastic policy is used to reach the latent goal state. A reconstruction of what the latent state in the frontier decodes to is shown. For clarity, a rendered view of the actual state reached by the agent is shown alongside the reconstruction.