Microscopic background?

Is the geometry of space a macroscopic manifestation of an underlying microscopic statistical structure? Is geometrodynamics - the theory of gravity - derivable from general principles for processing information, asks Ariel Caticha 2005 in “The Information Geometry of Space and Time”? A model of geometrodynamics based on the statistical concepts of entropy, information metric, and entropic dynamics. The model shows remarkable similarities with the 3+1 formulation of general relativity. There is, in addition, a gauge symmetry under scale transformations. The general theory of relativity (GR) can be derived from an underlying “statistical geometrodynamics” in much the same way that thermodynamics can be explained from an underlying statistical mechanics?

The geometry of space is the geometry of all the distances between test particles and this geometry is of statistical origin. We deal with probability distributions and not with “structureless” points. (The statistical geometrodynamics developed here is a model for empty vacuum; it does not include matter.) The information geometry we introduce does not define the full Riemannian geometry of space but only its conformal geometry. This appears at first to be a threat to the whole program but it turns to be just what we need in a theory of gravity. GR can be obtained in some appropriate limit. But with scale invariance. The scale factor σ(x) needed to assign a Riemannian geometry to space is arbitrary chosen. The essence of the dynamics of GR lies in the embeddability of space in spacetime. The σ(x) can be chosen so that the evolving geometry of space sweeps out a four-dimensional spacetime.

Perhaps the most interesting are the revision it requires of the notion of distance, the statistical structure of both time and space, and the recognition that spacetime is not a fundamental notion. The statistical nature of geometry could provide mechanisms that would eliminate the infinities pervading quantum field theories, either through decoherence or through the finite number of distinguishable points within a finite volume. Furthermore, it would make the Lorentz and CPT symmetries have only statistical validity and it might bear on the subject of CP violation and matter-antimatter asymmetry. On the other hand, the scale invariance might be relevant to cosmological issues such as the early inflation and the late accelerated expansion of the universe.

This is what I shall discuss. The notion is very much like TGD. Notice there is a need for two sheets of space, space and spacetime, matter and antimatter/dark matter/statistics and entropy.

The background space, a non-linearly interacting scalar field, the Higgs field or Skyrme field, is usually assumed to be non-local. There is a field of energy sometimes called the vacuum energy, zero energy ontology, the Higgs field etc. These field potentials (homogenous and isotropic 'vacuum') are suggested to explain the expanding space, measured as instance from gravitational lensings coming from dark matter (DM). This DM is making a dimming effect when we look back in spacetime. But DM is not understood. Usually it is claimed to be WIMPs (weakly interacting particulate matter). The result is a cosmological constant interpreted as DE, about 75% of all energy. A very big unexplained 'constant', indeed. It is linked to gravity usually. "Dark energy is just possibly the most important problem in all of physics", says Michael Turner in “Dark Energy and the New Cosmology”.

Dark energy has the following defining properties:

- it emits no light;

- it has large, negative pressure, pX is about-pX; and

- it is approximately homogeneous (more precisely, does not cluster significantly with matter on scales at least as large as clusters of galaxies).

Because its pressure is comparable in magnitude to its energy density, it is more \energy-like" than \matter-like" (matter being characterized by p >>p). Dark energy is qualitatively very different from dark matter. Sufficiently large negative pressure leads to repulsive gravity (the second Friedmann equation).

Frank Wilczek talks of the need for a ‘grid’ to explain the vacuum.

It is about interpretation, says Kea. Look at the work of Starkman et al on extra galactic plane subtractions from the WMAP data that suggest the zero angular correlation beyond an angle of roughly 60 degrees, in agreement with Riofrio's cosmology.

D.L. Bennett 1996, thesis. Multiple Point Criticality, Nonlocality and Finetuning in Fundamental Physics: Nature - in for example a field theory - seeks out values of action parameters that are located at the junction of a maximum number of phases in a phase diagram of a system that undergoes phase transitions. This is multiple point criticality, or non-locality. This latter often takes the form of a lattice, and the consistency of any field theory in the ultraviolet limit requires an ontological fundamental scale regulator. A gauge group is taken to be a Planck scale predecessor with no assumed symmetry.

The AGUT gauge group SMG3 has one SMG factor for each family of quarks and leptons. Ambiguities that arise under mappings of the gauge group SGM3 onto itself result in the Planck scale breakdown to the diagonal subgroup of SMG3. The diagonal subgroup is isomorphic to the usual standard model group. In the context of a Yang-Mills lattice gauge theory, the principle of multiple point criticality states that Nature seeks out the point in the phase diagram at which a maximum number of phases convene. This is the multiple point. The physical values of the three SMG gauge couplings at the Planck scale are predicted to be equal to the diagonal subgroup couplings corresponding to the multiple point action parameters of the Yang Mills lattice gauge theory having as the gauge group the AGUT gauge group SMG3.

This gives a geometric trinity that can be interpreted as a categorical group theory?

Graham Dungworth: The tetrahedral form would provide the lowest energy state, more stable than a polyneutron. It's a 15 quark assemblage. Fifteen quarks as proton quarks for lepton conservation would require 5 electrons. This 15 quark assemblage conserves lepton number as the overall charge for the trimer is +1. Energy wise is a four fold 4 *0.511 MeV saving which fortuitiuously is similar to that for the H2 formation ca. 2 ev below H states.

Matti Pitkänen, TGD: The 'holy' trinity is one basement.

On the interaction and massivation.

Richard M. Weiner “The mysteries of fermions”. It is conjectured that all known fermions are topological solitons. No bosonic leptons or baryons have yet been found, although predicted by SUSY. Another specific property of fermions is the Pauli exclusion principle, which states that, unlike in the case of bosons, a given quantum state cannot be occupied by more than one fermion. Maybe these both has a common ground in physics, through the Skyrme mechanism, analog to the Higgs mechanism?

Gerardus THooft, Nobel lecture: The mass generation mechanism should not be seen as spontaneous symmetry breaking, since in these theories the vacuum doesn't break the gauge symmetry. 'Hidden symmetry' is a better phrase. Or the Higgs mechanism.

But Kea/Marni says there is no fairy fields, and no whackalinos (WIMPs), no DM, no Higgs boson, no constant speed of light. Her background is composed of algebra and logics. The braid pictures do not directly describe a physical space, but rather categorical objects, non-locality is very much a feature of the categorical picture, starting with entanglement for ordinary quantum mechanics. Action at a distance requires a deeper discussion. The idea of a changing c is supposed to be an observer dependent one ... us being the observers. So in that sense there is no absolute time and distance. An old fashioned space is best thought of as an infinite dimensional category, built from paths, and paths between paths, and paths between paths between paths, and so on. That is, classical geometry is complicated precisely because it is 'unnatural'.

WAU, what a statement! She says:

My opinion happens to be different. Since I do not believe that Quantum Gravity is based on classical symmetry principles, I have no reason to take the Higgs boson seriously. Moreover, I am well aware of its effective role in a local theory, but that theory just simply does not apply to quantized masses ... just like Newton's laws break down if you push them too far. A cosmological constant is not an explanation of anything. It is an ad hoc constant of integration. The static universe was a mistake then, and the Dark Force is a mistake now. There is no inflation. Louise's infinite c condition explains the horizon problem perfectly well, and gives PREDICTIONS. Riofrio has unearthed much evidence that c has varied. Of course it is a conversion factor ... and hbar must also vary. Also the fine constant alpha? And you are beginning to see that this c variation is a statement about mass. Compton lengths has to be considered. In fact, there is a conformal boundary so that we can say M=t in Planck units, where t is a cosmic time parameter which one sets to zero in the Big Bang epoch. Alternatively, Riofrio's law is a form of Kepler's law, the general validity of which you probably do not doubt.

It is a radically new wiev, different from every ‘New Cosmology’ model. Graham Dungworth tries to make the new approach compatible with the SM on galaxyzoo forum.

Louise usually discusses her theory in the context of General Relativity, with which I suspect you are not arguing. In my opinion, once we start talking about a varying hbar (keeping alpha constant) we are outside General Relativity, as a mathematical formalism. However, even Einstein believed that 19th century Riemannian geometry was inadequate to capture the underlying principles of relativity. Here we have two basic options: (i) Louise's varying c rule or (ii) the favoured Dark Force within conventional geometry, fixing c arbitrarily. The claim is that (i) is far, far simpler and more natural.

Matti Pitkänen, TGD, uses 2-surfaces and Fermi surface analogs (the’ Big Book’) for the job. Leakage between surfaces do the braidings. “p-Adic length scale hypothesis and CP breaking provide a simple explanaton once again. There is already old evidence that neutrinos can appear in several p-adic length scales. Suppose that the p-adic mass squared scales of neutrinos and and antineutrinos differ by 2 (p =about 2^k by p-adic length scale hypothesis). This means k(nubar)= k(nu)+1.” The argument that hbar and c are absolute constants is correct only in GRT framework and in standard quantum mechanics.

The Fermi analog gives a coupling to condensed matter physics. Matti has also an own ‘Higgs mechanism’ for the massivation. The concept of mass naturally arises in p-adic models as inverse transition probability with a dimensional constant of proportionality.

Inflationary scenario is a possible explanation for flat spacetime but plagued by extra-ordinary clumsy Higgs potentials and loss of predictivity. TGD explanation is based on quantum criticality. At quantum criticality there are no scales and this translates to the vanishing of 3-D curvature tensor and flatness of 3-space during the phase transition which corresponds to an increase of Planck constant scaling up quantum scales. Cosmological expansion would indeed occur as rapid quantum phase transitions in TGD Universe rather than as smooth continuous expansion. Bubbles within bubbles. TGD cosmology is fractal containing cosmologies within cosmologies in all scales and the scaled version of the same universal critical cosmology allows also to model what happens at RHIC. It would be important to test this cosmology since it is universal unlike the inflationary scenarios.

The negative pressure expansion is not due to negative kinetic energy as in quintessence scenario nor due to the addition of cosmological constant term to Einstein's equations but simply due to the constraint force taking care that space-time surface is surface in M^4xCP_2. One cannot of course exclude phenomenological modeling in terms of cosmological constant.

As I explained the measured c, I call it c# has c only as upper bound in sub-manifold gravitation. Argument is childishly simple and explains the apparent variation of Earth-Moon distance, which is something GRT people should take with utmost seriousness.

In the case of Planck constant TGD predicts that the Planck constant for single sheet for the covering is n-fold multiple of the ordinary one. If one treats the whole bundle of sheets as single object it is the ordinary one but now one has to find a manner to treat the system non-perturbatively. This is by no means prevented by conversion factor identification. Also the more radical option in which hierarchy of Planck constant does not follow from basic TGD is completely consistent with conversion factor interpretation.

Speed of light.

Luoise Riofrio, okt 2010, The c Change Continues . speed of light Dr. Daniel Gezari interests include tests of Lorentz invariance and searching for evidence of a preferred reference frame for light. Lunar Laser Ranging Test of the Invariance of c . The speed of light c as 8 m/sec less than its formerly accepted value. Perhaps c has slowed since the canonical measurement? Even more surprising, Gezari claims that c varies by as much as 200 m/sec depending on the observer's direction. This would undermine Lorentz Invariance, which says that c is the same regardless of direction. 'the speed of light seems to depend on the motion of the observer' (blue and redshift) undermine the basis of Special Relativity.

Matti: It does not make sense in GRT framework. “In general relativity one can only say that light moves along light-like geodesics. Riofrio's formula also leads to non-sensical results since it leads to time variation of elementary particle masses. In TGD framework the sub-manifold geometry allows to speak about varying c, call this velocity c#. These times are in general different since the motion along light-like geodesic of the space-time surface is in general longer than along the light-like geodesics of imbedding space and depends on how wiggly the space-time sheet is. The value of c# equals for Robertson-Walker cosmology to c#=sqrt(g_aa) and increases since space-time approaches flatness (cosmological t=0?). (OBS, the negative pressure was calculated by Friedmanns cosmology).

Kea: I don't think Louise's picture is so different to yours really. Remember that we are always mixing up several notions of time here. Your picture of increasing c could be viewed as a forward time evolution in Louise's picture. Since I take stringy dualities seriously, this actually makes sense.

Witten: For massive particles a vector current could create either vector or scalar particles, for massless particles though, Lorenz invariance forbid creation of vector particles, only scalar particles are allowed.

Graham D: Wherever we might be in the universe we are confronted by an absolute frame of reference from the neutrino temperatures; of mass energy continually created and destroyed which currently is blue shifted towards us, wherever we are of incoming neutrinos at -1315 km/second contra to the Hubble flow, where the size factor for compression =1 +z , which we use to determine matter DM outflow. This is reminiscent of part of the steady state model that attempted to create new particle mass , eg. the H atom from space to balnce the eventual disppearance of the outflow of galaxies to beyond the horizon. Today, the alternative explantion for big Delta is an accelerated expansion from supernova standard candles to hyperbolic geometry. Irrespective of absolute neutrino massess this low cumulative mass, even should it change will question Lorentz invariance. Yet if complete the cycle of creation and eventual destruction and rebirth, all will find comfort.

Lunar Laser-Ranging Detection of Light-Speed Anisotropy and Gravitational Waves Reginald Cahill LLRE shows a variation in c. He also claims a correlation with spacecraft flyby anomalies. LLRE has reported the Moon's semimajor axis increasing by 3.82 cm/yr, anomalously high. If the Moon were gaining angular momentum at this rate, it would have coincided with Earth about 1.5 billion years ago. Apollo lunar samples (which this scientist has had the honour to touch) show that the Moon is nearly old as the Solar System, over 4.5 billion years. Measurements using tidal sediments and eclipse records show that the Moon is receding at about 2.9 cm/yr. LLRE disagrees with independent measurements by 0ver 0.9 cm/yr, a huge anomaly.

If the speed of light is slowing according to GM=tc^3, the time for light to return will increase, making the Moon appear to recede faster as measured by LLRE. Predicted change is 0.935 cm/yr, precisely accounting for a 10 sigma anomaly acceleration caused by "dark energy." The explanation for DE and the apparent acceleration of redshifts is something a child could understand. Since redshifts are related to the speed of light, it is not the Universe accelerating but the speed of light slowing down.

Ed Witten 2010 “A new look at the path integral of quantum mechanics”, abstract:

The Feynman path integral of ordinary quantum mechanics is complexified and it is shown that possible integration cycles for this complexified integral are associated with branes in a two-dimensional A-model. This provides a fairly direct explanation of the relationship of the A-model to quantum mechanics; such a relationship has been explored from several points of view in the last few years. These phenomena have an analog for Chern-Simons gauge theory in three dimensions: integration cycles in the path integral of this theory can be derived from N=4 super Yang-Mills theory in four dimensions. Hence, under certain conditions, a Chern-Simons path integral in three dimensions is equivalent to an N=4 path integral in four dimensions.

And in The Problem Of Gauge Theory 2008 he says: Based on real experiments and computer simulations, quantum gauge theory in four dimensions is believed to have a mass gap. This is one of the most fundamental facts that makes the Universe the way it is. The mass gap means simply that the mass m of any state (orthogonal to the vacuum) is bounded strictly above zero. There is no mass gap in electromagnetism; the photon is massless, so electromagnetic waves can have any positive frequency. That is why we can experience light waves in everyday life.

By contrast, the mass gap in strong interactions means that the minimum frequency needed to probe the world of SU(3) gauge theory (which describes the strong interactions) is mc2/~, where m is the smallest mass.

Taking from experiment the value of the smallest mass, this frequency is of order 10^24 sec−1, which is high enough (with room to spare) that this world is way outside of our ordinary experience. While it is very large compared to our ordinary experience, the mass gap is in one sense very small: it is zero in the asymptotic expansion. As a result, we do not have a really good way to calculate it, though we know it is there from real experiments and computer simulations.

So in short, this mass gap is one of the most basic things that makes the Universe the way it is, with electromagnetism obvious in everyday life and other forces3 only accessible to study with modern technology. The mass gap is the reason, if you will, that we do not see classical nonlinear Yang-Mills waves. They are a good approximation only under inaccessible conditions.

DM –fuzz?

Kea do not believe in the Dark Force, and by that she means the DE-approach in SM. But it doesn’t rule out the DM-problem. This is because the DM-problem involves also antimatter scenarios and anionic Yang-Mill massgap phases. The wavefunction does indeed change as you move from point to point by the phase angle, not an observable but crucial to Yang and Mills.

Graham D: Unlike the Abelian U(1), SU(2) is difficult in that there's the rotations in both complex planes of an inner space with those of weak isospin space – non-Abelian and non commutative. For every particle there is an antiparticle and these have opposite attributes or truths. The attributes or truths of the electron are mass charge, electric U1 charge and weak isospin charge, handled the gauge symmetry SU(2). There's no colour charge, that's covered in SU(3) symmetries. The initial Minos results appear to contradict the standard model and these truths. Dark matter is not just mirror matter. The standard model is recovered by those four neutrino masses, but they are two pairs, not a threesome and a oneness. Nor does the mirror or former mirror properties, before the Minos data, preserve parity. Those neutrinos aren't born equal, some are more massive and less numerous than others. Additionally, they cannot exist in isolation; they are in equilibrium with photon energies at all temperatures.

There is no Higgs but DM. The conventional interpretation is that the inertial force or residual particle mass (energy) is given by the Higgs fields and scalar potentials ( 2 of them). There are actually 4 since two of these are spatially complex dimensions. Of the 4 fields there is one field that gives rise to the so called Higgs boson, the lowest energy potential. It's a top down theory where the empirically discovered mass of the top quark was made possible by the theoretical treatment of the Higgs field components at energies of ca 100GeV (cf, the proton mass ca. 1GeV is 95% approximated by the gluon fields).

One ‘allo’ neutrino (Grahams def of gen 1), only one of which is associated with postulated dark matter. What happens to the other two? They belong to the other phase that is the vacuum, tied into it. Ordinary matter and dark matter consort together, of opposite parity, mirror forms but their mass charges are different and consequently their abundances are different in the vacuum vapour. Because their abundances are different ca. 8 fold more DM than OM, the vacuum as a consequence is enriched in left handed chirality, because the DM removes right handed species. Hence you can't treat mirror chirality soley with real particles in Nature without addressing the pulling or pushing power of the vacuum.

Kea and Graham talk of 'false vacuum'.

Unless all charges are matched particle antiparticle pairs don't annihilate. A u quark annihilates with a u bar antiquark but not with say a dbar antiquark since they carry different mass charges as well as all other conjugate charges. The standard model doesn't predict everything of course eg as to numerosity of generations nor why there are these pairings between leptons and hadrons and you need two measured masses to predict others nor the magnitude of charges. Neutrinos don't carry electric charge or colour charge. There's no pattern there nor why they all carry the gravitational charge. This latter force or interaction as mediated by a graviton boson, spin2! although ignored in the standard model is a truth all particles carry.

This gives a very interesting situation. Einstein, Lorenz invariance, Pauli exclusion and Heisenberg’s uncertainty are reinterpreted, maybe also Feynman? No wonder the theoretical physics has made no progress lately. Is this probable? No, say most. But something is seriously wrong, that almost all agree upon. So let’s look further on this issue.

References.

D.L. Bennett 1996, thesis. Multiple Point Criticality, Nonlocality and Finetuning in Fundamental Physics . http://www.arxiv.org/abs/hep-ph/9607341v1 the anti-GUT approach.

Carl Brannen. Spin Path Integrals and Generations. 2010 http://arxiv.org/abs/1006.3114v1 and Found. Phys. (2010), dOI: 10.1007/s10701-010-9465-8.

C. A. Brannen (2008), http://www.brannenworks.com/koidehadrons.pdf .

Reginald Cahill, Lunar Laser-Ranging Detection of Light-Speed Anisotropy and Gravitational Waves http://arxiv.org/pdf/1001.2358

Ariel Caticha, The Information Geometry of Space and Time. http://arxiv.org/abs/gr-qc/0508108v2

Jerry Decker, comment on Riofrios blog.

Graham Dungworth, http://www.galaxyzooforum.org/index.php?topic=277933.0. Dark matter tread.+ Minos neutrinos tread.

Daniel Gezari Lunar Laser Ranging Test of the Invariance of c http://arxiv.org/abs/0912.3934

Gerardus ’t Hooft “Determinism and Dissipation in Quantum Gravity”, 2000. hep-th/0003005

Yoshio Koide. A fermion-boson composite model of quarks and leptons. Phys Lett. B, 120:161–165, 1983 [abstract]. http://linkinghub.elsevier.com/retrieve/pii/0370269383906445

Matti Pitkänen, The phases are divided according to primes in TGD. p-Adic Mass Calculations: New Physics. Higgs and massivation in TGD framework, oct. 2010 http://matpitka.blogspot.com/ and comments on Keas blog.

http://tgd.wippiespace.com/public_html/genememe/genememe.html

Louise Riofrio, The c Change Continues, Oct 17, 2010, http://riofriospacetime.blogspot.com

Marni D. Sheppeard, Neutral particle gravity with nonassociative braids, ViXra.org oct. 2010 1010.0029v1.pdf

M. D. Sheppeard, http://pseudomonad.blogspot.com/

Minos, http://physicsworld.com/cws/article/news/42957

SPACE.com -- Dark Energy and Dark Matter Might Not Exist, Scientists Allege

Starkman et al on extra galactic plane subtractions from the WMAP data

Michael Turner, “Dark Energy and the New Cosmology”. http://arxiv.org/abs/astro-ph/0108103

Eric Verlinde, http://en.wikipedia.org/wiki/Erik_Verlinde , http://en.wikipedia.org/wiki/Gravity_as_an_entropic_force

Richard M. Weiner “The mysteries of fermions”. http://arxiv.org/abs/0901.3816

Frank Wilczek 342_Origin_of_Mass.pdf http://frankwilczek.com/Wilczek_Easy_Pieces/342_Origin_of_Mass.pdf

Ed Witten, 2010, “A new look at the path integral of quantum mechanics”, http://arxiv.org/abs/1009.6032

E.Witten, Baryons In The 1/N Expansion. Nucl. Phys. B160, 57 (1979). On the Skyrme model.

Sidney Coleman ,Edward Witten, CHIRAL SYMMETRYB REAKDOWNIN LARGE-N CHROMODYNAMICS. Phys.Rev.Lett. 1980. http://www.slac.stanford.edu/cgi-wrap/getdoc/slac-pub-2493.pdf