Nick Bostrom's Simulation Hypothesis

My question is not about the validity of his premises - because I actually don't believe they logically lead to the proposed conclusion but about the test itself.

"Currently, supercomputers using a technique called lattice quantum chromodynamics and starting from the fundamental physical laws that govern the universe can simulate only a very small portion of the universe, on the scale of one 100-trillionth of a meter, a little larger than the nucleus of an atom, said Martin Savage, a UW physics professor."

My understanding of physics is moderate however not in-depth. This is the general statement on QCD

Lattice QCD is a well-established non-perturbative approach to solving the quantum chromodynamics (QCD) theory of quarks and gluons. It is a lattice gauge theory formulated on a grid or lattice of points in space and time. When the size of the lattice is taken infinitely large and its sites infinitesimally close to each other, the continuum QCD is recovered.[1]

Could someone perhaps explain to me in laymans terms why QCD is being supposed for the simulation engine - is this necessary or contingent?

I find this idea fascinating and I understand the heuristic approach I'm simply trying to fill in the gaps in my understanding. Thanks in advance

Staff: Mentor

There is nothing stating that QCD would be the basis. QCD is used as an example of how we do computer simulations today to extract information from complex theories. In QCD simulations, the theory is put on a lattice and the article claims that we may be able to see the effects of a theory being put on a lattice if we are part of a simulation.

There is nothing stating that QCD would be the basis. QCD is used as an example of how we do computer simulations today to extract information from complex theories. In QCD simulations, the theory is put on a lattice and the article claims that we may be able to see the effects of a theory being put on a lattice if we are part of a simulation.