Exact Bayesian inference is not possible for nonlinear models. Instead, one must use approximate inference frameworks, and Variational Inference is one such approach. It factorises the posterior density and optimises the parameters of the factors so as to minimise the KL divergence between the true and approximate posterior. It also provides a lower bound on the the model evidence. This bound is the "negative variational free energy".
The original idea behind variational inference was proposed in the context of neural networks, whereby the suggestion was to replace the standard cost function (e.g. mean square error or "energy") with the "variational free energy", F, so as to improve generalisation performance.