alexa Distance and Mass Correction on Logarithmic Metrics for High Z's, Due to the Non-Cero Angular Momentum in an Isotropic Rotational Universe Hypothesis

ISSN: 2090-0902

Journal of Physical Mathematics

  • Research Article   
  • J Phys Math 2017. 8: 234, Vol 8(2)
  • DOI: 10.4172/2090-0902.1000234

Distance and Mass Correction on Logarithmic Metrics for High Z's, Due to the Non-Cero Angular Momentum in an Isotropic Rotational Universe Hypothesis

Pons-Rullán B*
PhD, Independent Researcher, Mallorca, Spain
*Corresponding Author: Pons-Rullán B PhD, Independent Researcher, Mallorca, Spain, Tel: +34629647564, Email: [email protected]

Received Date: Jun 10, 2017 / Accepted Date: Jun 26, 2017 / Published Date: Jun 28, 2017


Dark may not be on the same sentence of a Scientific Hypothesis. Let’s surmise here again an alternative model in which the time turn around the space with non-zero angular momentum. Then, there should be a center and we may observe anisotropies, and indeed although they are not explicit in space, but are observed over time, and we identify a Beginning and an arrow-of-time. If so, to preserve the metrics invariance through scale, estimated distances to deep galaxies may be corrected by replacing the FLRW metrics by another logarithmic based, according to now a day observations in supernovae Ia range up to 0.2<Z<2, predicting how much further over brightness adjustment might be. Universe may seem to accelerate its expansion because the assumptions, not because the Ghost Energy, and as corollaries either G, c, h, α or Ho, seem to be from our point of view as observers, logarithmic variable in look-back-time. That means that together with the isotropy in the metric expansion of space-time, it requires a x5,75 measurement of the mass in CMB, as observed, the near“keplerian” velocity distribution in very ancient galaxies, and a non-dependent on the radius distribution of the velocities of stars and galaxies in our locality, with enough time and appropriate normalization. Also as corollaries it may explain the weakness of gravity relative to the other forces; the Horizon Problem; simultaneity; baryonic asymmetry; the homogeneity and granularity of the CMB; the extra large size of the galactic black holes; the very fast configuration and the narrow range of dimension of the mass of galaxies and stars; the absence of small G, K, M, L, T, orange and red stars from Population III; the distribution of ages of quasars; the highest density than expected of stars and heavy metals in the very deep space; the lowest density than expected of brown and black dwarfs; the absence of slow neutron stars; the decreasing change of the critical mass for black holes and its progressive deactivation; that the Universe is younger and older than estimated -according to the observer-; and alternative descriptions to Hyperinflation, Big Bang -it was a long boring and no forceful process-;...GR and QM are deterministic and needs an upgrade to irreversibility for high Z’s.

Facts: entropy is not isotropic; Universe is expanding; a Bang means linear momentum, that may have been evolved some way to angular momentum, while everything orbits and rotates; there is much more matter than antimatter; time and space coordinates are not ergodic; and reality is mostly not linear. Not too much restrictions to start. The following will be structured in two sub papers. Further details justifying this Bartolo's Conjecture, circumstantial evidences of Lateralization, and development of the conclusions can be found in “Eppur si Muove”: www.bartolo. and a contextualization of the questions in a wider frame, in “Todología”:


We will develop the following hypothesis from the reinterpretation of the Accelerated Expansion of the Universe, based on the measurements of supernovae Ia, with the apriorism of a linear and isotropic spacetime metric; from the alternative assumption of Decelerated Expansion -so that matter and energy will add 0 from the point of view of an observer outside our 4D manifold-, and compare it with several non-linear metrics that may fit the measured values.

With this alternative approach, the result of a rotational Universe is a logarithmic scale as a consequence, not as an ansatz, which can also afford temporary covariance -and though the energy conservation-, but requires anisotropy that is not observed in space... unless an additional space-time dimension is conceived. Then, if we assume a metric that preserve the classic first principle of thermodynamics, the logarithmic metric with isotropic space, implies a multiple rotational time distance and means somehow non linearity in cdt.

If the Big Bang was as a “Bang”, how did the radial linear momentum shift into angular momentum? Why, and from when, the Universe conserves it? Here it is claimed to be because a non linear time dimension rotating, instead than a Potential Dark Energy supposed to be emerging non-linearly on constant time first in the hyperinflation, after stopping, then awaking again...

Neither GR nor QM are compatible while they are based on incompatible ansatzs, but they both share to be deterministic and ergodic, and this is enough to ensure that at some scale, they are both only good approximations, though reality forgets and not always time and space derivatives can be exchanged. If cdt changes by dt/ln(t), as we will find as a consequence, then GR evolves in confluence with the nonlinearity and irreversibility paradigm, while a second time dimension allow the causality principle in a single decision trajectory from the past to the future, but a way back through other infinite alternative paths from the future to the past.

More than that, in the KAM context, an algorithmic spiral relates two points of time by a constant angle and maintain the invariance with its expansion, proposing naturally the Gold Ratio as the most stable relation from a perturbative forcing between all “orbits”. As FLRW metric is linear, it does not have this property on expansion frame.

Modern stochastic-chaotic-dissipative paradigm contemplates linearity as an exception in complex systems, and the GR or QM have not been adapted to upgrade them to norm. Birkhoff summarized it: only in linear relations, space and time comply with ergodic property.

Linearity in space related to time is ergodicity and related to scale fractality, but both are useful simplifications to make reality understandable. By Lyapounov, space and time do not have pair evolution; while by Poincaré, the recurrence may repeat the storyboard again and again in conservative systems. To suppose that a “divine observer” of a 4D manifold would understand the temporal dimensions of different nature as spatial dimensions by their properties of symmetry, is a hypothesis as valid as the opposite, which has the disadvantage of not being able to play the dice.

Adding spatial dimensionality does not change the symmetry group or the flatness, but adding a single temporal dimension changes the properties of curvature -and though of gravity- in the very beginning and in the future, as changes the path of the proper time (even the constancy of Λ). Space and time are both coordinates, and while they have different sign near each event, GR doesn't know anything about the irreversibility of the arrow: GR needs to play dice. They are more different indeed: we ride inside space but on top of time. Space is the landscape and time the horse. The hypothesis brings a way to solve the Cauchy problem in a 4D deterministic space-time [1]. Odds are apparently against: [2].

“Divine observer”: Campbell’s Theorem states that any analytical solution of the Einstein field equations in N dimensions can be locally embedded in a Ricci-flat manifold in (N+1) dimensions. A circumference of two-spatial and one-time dimensions (2+1), spins around a point. A sphere turn around a one-dimensional axis. A 5-dimensional hypersphere rotates over a symmetric three-dimensional space, presented as a dual time spherical manifold or in a space-time of 6 dimensions. If that hypersphere expands and spins, dimensions might grow in logarithm rate to preserve the invariance to scale and to describe the different nature of the axis and hypersphere, we'll call the non-isotropic dimensions different because their symmetry properties: time.

The number of arbitrary constants sizes the distance from a model to the fundamental reality. If the universe has a conservative non-zero momentum from an outsider divine observer, a model considering this Λ from the inner observer as constant, would not be fundamental. If time is another coordinate like space, expands as any other spatial dimension and spins, there might be more non-isotropic dimensions than a single one and nor the speed of light and the gravitational constant,... and hardly no constant, are arbitrary, but only their relations in units changing in each moment of time. G and c remain constant in every observation, for every observer, in every time, in all space -“soft cosmological principle”-, but an observer measuring them in other epoch will use his clock, not the clock he would use at our time. The isotropy will be in space and the anisotropy in time, but maybe some lateralization clues, beyond the arrow of time, would be identified [3-8].

GR equations do not show any preference for any particular shape, scale or metrics; but we need symmetries to find solutions. A universe in which everything, including elementary particles, is twice as large (or small) as in our universe and in which the duration of a second is twice (or half) as long, would be completely equivalent to our universe only if time relation with space coordinates do not changes. If the expansion is linear or logarithmic, the metrics would be different, but laws of physics would be the same. [9] If the Universe twists around a 3D-space axis (with the remaining inertia Λ -conserves the momentum-, after the particle generation, ɱ), the three-dimensional space axis expands isotropically with an also expanding revolution time-surface related both to the spin, where the radius represents one dimension of time -tr-: Figure 1.


Figure 1: Path on a dimensional time dimensionality.

For an outsider observer any distance and any time grows on a logarithmic scale and dimensions do not have the same real or imaginary nature than for a human insider observer. Maybe a new perspective about Bergson and Einstein 1922 time conception could fit here (time as itself or proper, as a path on a dimensional time dimensionality) [10].

Lanczos 1924 [11], Gamow 1946 [12] or Gödel 1949 [13,14], ventured a Universe with non-zero momentum, and from time to time, some body tries to rescue the hypothesis [15-24]. If we could measure the tangential speed of the temporal spiral -ta'- from our inertial reference, it would be as a stone at the end of a "temporary sling" growing 3D space in proportional volume to the cube of the radial time (tr) increase. Expansion would be related for a divine observer to the speed of radial time, tr', apparently against our experience which feels time as constant and measures an accelerating growth. An outsider divine observer may "see" it that way, but the human sizing would be determined by our limitation in the embedding dimensional perception, that makes the clock seems to have constant speed.

For Abbot's Flatland citizens embedded in our reality, if their time could be represented as our third space dimension, we would see all their time as a dimension and we should maybe seem to be for them Chiral Gods. They would describe us as our own projections on their space. A drawing of a right hand can be switched for the left hand -reflection- only with in a lower group of transformations -rotation- if we access to extra dimensional spaces. So, one extradimension repairs one symmetry, and their asymmetric time would be for us their destiny seen as a hole. Any broken symmetry can be restored from the point of view of a “growing block” observer living in more dimensions. With no esoteric purposes, but didactical a Chiral God would improve invariance from a complete Hilbert space-time with one or more dimensions, and tensor-geometrical description in a manifold is only a transport of the references from the ortoND to the (N-1)D hypersurface with N-1 free degrees. So asymmetries that configures our reality can be restored to the Nothing, considering more dimensions, which may not be folded up in our reality, but projected from a single asymmetry in more dimensionality: the rotation.

Recalling J.L. Borges poetry -“God moves the player and the player plays the chess,/ which God beyond which God began the game?”-, we'll qualify this hypothetic outsiders as chiral and ergodic divinities and ourselves as mortal enantiomers.

GR is preserved with a soft principle of constant c: while c remains as a constant relation, not necessarily as an absolute value over all times. Divine observer would measure the variability of speed of causality and information, and mortal observer would size the expansion, both exegeses because their own references, but they would not agree about which one is constant. For a divine observer all 6 would be orthogonal dimensions, but from the human point of view inside the manifold shape of the space-time, the perspective is limited in representation 3 perpendicular contravariants referring to a fourth frame -time- taken as constant, distinguished from the others because it only happens in a non-commutative sense, but it's a summary of 3 time covectors (time is an obvious anisotropy of space-time).

For an outsider observer, in this simulation, classical mechanics applies analogous to what we could call “surface of temporal dimensions”, deemed for description purposes as spatial euclidean dimensions, some of them symmetric and some anti-symmetric, (as a 3D simplification we model a cylinder rotating and expanding, in which its section is built with 2 time coordinates). As we measure identical expansion in all dimensions of the axis -distance/time-, we do not need more than one space dimension to emulate mathematically the “divine observer” perspective in his conceptual reference to our space. A single spacedual time diagram model of a single real dimension -isotropic spindleand two imaginary dimensions -temporal surface- all transformed into real, and all referred to a time frame.

equation (1)

In other words, to simplify the simulacrum 2 symmetric-spatial dimensions will be replaced by 2 anti-symmetric-temporal dimensions, with no need of the complexity of far beyond algebras (3,3) or (4,4) if we consider also scale and Λ, as dimensions [25,26]. This closed and axisymmetrical model (isotropic for a resident in 1D spindle-space: living in the surface of the sphere, simplified for our purposes in a cylinder), describes a universe that emulates ours, as a resident observer would assume if he lives in 6D.

In this spacetime diagram simulation, time perceived by the mortal observer as constant, is the spiral path of time: the vector addition of temporary increases,

equation (2)

To the mortal enantiomeric observer resident in that single spatial dimension, t' -the rate of progression of time- would be constant, but not for the outsider. In the beginning, the difference was important, but it happens that hundreds of billions of years later, sinω long ago is negligible in relation to cosω, and for both observers t'= tr'. Different masses, like different ta', generates different never-crossing spiral paths, and no mass will have as its path just the radius itself. As the interpretation for rotation has allways been spatial, Hawking 1969: “These models could well be a reasonable description of the universe that we observe, however observational data are compatible only with a very low rate of rotation” [27]. But in this hypothesis time spins, not the space, and it is indeed at a very low rate now a days.

The incorporation of Hyperinflation and Dark Energy to the models to fit the observations results a Hubble “roller coaster” function, -Hubble Flux, Ho(t)-, with bizarre loops, accelerations, decelerations, to avoid resignation of the Conservation of Energy Law -inflatons?-. If from the point of view of an outsider observer (GR does not conceive a Λ conservation not dependent of the observer), the Universe as a whole in its rotation preserves the angular momentum (Λ, Cosmological Constant), it would be a very fundamental configuration parameter of reality. Taking the classic descriptive rotational deterministic model, the areal velocity may be kept:

equation (3)

and the tangential speed of time would be

equation (4)

To preserve momentum, the angular time speed of growth decreases faster than the growth of tr; and the ratio between the derivatives of both time dimensions is not constant. Therefore, at least a divine observer, in any way won't conceptualize this rotating speed, -perceived time-, as constant. In the Beginning it would be very obvious for any observer, but now even Him/Her should be very accurate in assessing the decline, and we can compromise that t'=tr' in current universal scale,... but not at the first moments of time, near the Big Bang. From a divine outsider observer measuring the expansion in a n-dimensional euclidean geometry and angular time would decrease faster than the radial expansion. As the ice dancer extends his arms (Figure 2), from the audience perspective, rotates more slowly than he "expands" the length of his arms. Assuming Λ & ɱ as constants, the angular velocity


Figure 2 : Calculation of length and the angular velocity of the dancer (Gα1/tr!).

equation (5)

and the total amount of gravity would be time-dependent and may be constant only in the space dimensions:

equation (6)

From an unimaginable divine existence in 6 or 8 dimensions (quaternions), all of them are perpendicular, but some are symmetrical-axis- and other -rotating- are not, depending if they are spinning or not. From another limited perspective, if we conceptualize this model as a single dimension space which the finger imaginatively draws in the air -curvature Ωk=1/tr-, it'll expands, but more slowly. The growth of the one-dimensional simulated space, with a single degree of freedom: up or down-, would be the maximum for any movement of an impossible inhabitant living in this "needle on the vinyl record" (analogy to explain that asymmetrical time dimensions twist rigidly in a path depending of the mass) because in case of being overtaken the ant could jump forward to the expansion itself. In a dynamic system paths can't cross out the attraction of “fixed point”, and that limitation holds the causality maximum speed: c.

The ant in the dancer finger (resident in time), which also gets weight while the dancer extends his arm (perceiving the time and the strength from a one-dimensional space), can’t go further than the finger and can't spin faster, but has its legs inevitably attached to the end of the nail and only investing energy he could break against the centrifugal acceleration and walk to the center, drawing a different spiral path

equation (7)

which is obviously the same as saying that

equation (8)

For a divine observer either space and time expands alike in its three coordinates, allowing mathematically replace and integrate in reference to radial time, and for the divine observer the speed of expansion will be α1/tr2. In our role of divinity for the ant at the end of the finger of a dancer, Expansion has proportional growth to the radial space-time

equation (9)

locally, but for an observer resident in the one-dimensional space. For us a unit of time is always the same unit of time, i.e., the dancer turns spinning down for the public sitting in the stands of Olympus. The ant in his nail rides on tr, and imposes its reference measuring constant Ho, to what divine observer consider only as relations between spatial dimensions units relative to units of temporal dimensions, c. For the divine observer, the angular momentum may be conservative, and though for the inner observer, c will decrease with time. This makes mortal observer perspective "see" movement with the assumption that time is linear (as watching the landscape moving from inside a train) α1/tr, and clears out from the equation

equation (10)

The relation between time of the observers keeps a transformation and c must decrease α1/tr, from the point of view of an observer that has a different clock. But, from the outsider observer point of view, if c is constant, Ho α1/tr!

In this model any particle moves over time in his own timedimensional spiral characterized by its mass: since its creation has a past and though is not a point but a temporary spiral path. So analogous a photon would be a perpendicular dimensional particle in space, not in time: simultaneous in each synchronized manifold, with no past and no future. Two space-like points in the hypersurface of the present are each outside of the light-cone of the other, so each one has access only to the past of the other: to preserve causality, a simultaneous photon broadcast may not be seen by the other. If we “see” a photon is because our cone intersects with another cone in a hole trazoid manifold and cross to the time like zone (Figure 3). Not near of a huge mass, cones grow parallel and as time goes on, the points in space-time tends to be not linear but hyperbolically related in causality between them. So, the causality relation draws a null hypersurface between simultaneous coordinates.


Figure 3: Relation between null hyper surface and simultaneous coordinates.

A photon dilutes linearly with Expansion, and the speed of causality -c- would be the expansion from our point of view (the speed of the landscape from the perspective of the traveler). The speed of light for a divine observer, would be only the expansion ratio between space and time units, that is, adjusting the system measures a matter of scale (1sg always equivalent to 4,775 109 m to the outsider observer). Why there should be a "speed of light" if it's only a convention of units of measurement? It's just a way to express our perception of expansion from the perspective of an observer measuring it from inside the manifold with a clock that measures a non-constant time.

G, T, P, c, α, H α1/tr!

Time is a length derivative for a photon and its emission would be simultaneous to observation, but both paths, photon and massobserver, have not the same invariance axis. The point of emission of a photon follows a temporal spiral path increasing both temporal and angular time, but the space where the photon expands follows only radial time. In the next instant other spinning paths -stories- will intersect with expanding paths -movements-, but would never meet again. The divine observer "sees" a spatial dimension that expands homogeneously and only conceives the speed of time expansion. For the mortal enantiomeric observer who does not "see" that space and time are expanding together with him inside, the speed of Causality is the increase of radial distance in time, rather than the expansion rate.

Acceleration could be as well represented as a temporary vector. If we measure the space in terms of our clock, it would be as if the ant moves to a temporary lower pitch spiral, where finds a past field and share the reality with other ants which remained at rest to us some time ago (being reality the full path since the beginning of time to now, and therefore radially remembered). Our apparent c, inversely proportional to radial time, mathematically re-integrating, results a model with logarithmic scale factor:

equation (11)

That is, as it was said by Teller 1949 [28], from the equations of a rotating system, the space-time expands at logarithmic scale.

If we change the metric considering the time as a variable according to a(t), so that the speed of causality is constant from the perspective of a divine observer, our measure of c can’t be constant, and we would measure the distance to an object in light years variables regarding this scale factor (each megaparsec -cosmological correction-, but also every second would be smaller in look-back-time, retaining the ratio, while older). We measure distances calibrated on the brightness according to the parameters of Cepheid in six galaxies, applying to larger scales what it works in our galactic environment, assuming the metric in which the time -and thus the speed of light with "mortal" references- remains constant. If the Universe preserves a non-zero angular momentum, to be spatial-isotropic but temporal-anisotropic, while angular time remains negligible to radial time, a logarithmic metric would be finest than a Friedmann metric.

In the absence of a reference step that affirms or denies, beyond the local super cluster of galaxies, a distance expressed in current light-years admit the vagary of increase with the average speed of light ¢, corrected by the scaling factor, as we would have considered the cosmological present time tro=1, taking it as a unit:

equation (12)

Dc, proper distance corrected.

DL, proper distance estimated under the assumption of constant time.

¢, average of c over the log scale

Standard candles supernovae Ia distances estimation, may include other variables like metallicity -more luminosity with less heavy particles, that may be more common the more further they are-; and needs a better knowledge of the processes. According metrics normally in use the Z=1, is given in t ro/2, but according to a metric in which the time is also affected by the scale factor, radial time would be transported to an axis as

equation (13)

and would also apply the above Dc correction, plus a correction of the time scale transporting an axis with a linear scale -cdt- by

equation (14)

According to this metrics the speed of light at every moment of the past seems inversely proportional to the time, because it's referenced to constant time, but still remains constant for a divine observer, as the relation between tr and Dc: 2π, is constant.

If the assumptions were correct and model fits with the choice of what is relevant, regardless of the values of Λ, ɱ, c and Ho, we should be measuring a redshift corresponding to the difference between Dc in each metric with its cosmological correction of its proper distance dependent on the above parameters. The question is: regarding what model applies the correction? Linear, de Sitter, ΛCDM, Benchmark,...?

It happens that when calculating the distance correction

equation (15)

it is identical result in the model of linear expansion, (ɱ=0, Λ=0, with c constant and a(t)=t/t0 and the logarithmic simulation; i.e., in both cases proportional to ln(1+Z). So if we compare them both, Table 1, we can save mathematical rhyme and verse and ignore at purposes of apportioning the cosmological correction. By the way, we can save relativistic corrections because the comparison of measurement criteria has nothing to do with distance themselves in absolute terms, or the use recessional velocities.

OverD (%)= ((ln(1/Z+1)/(1-(1/(Z+1)))

equation (16)

The overestimation of the distance we would be measuring with the brightness may be the gap between the excess in the distance that we would obtain if the speed of light were inversely proportional to time, taking it as in constant expansion, with respect to the excess if we took it with time in a logarithmic brake Table 1.

Z 0.5 1 2 3 5 10
t lineal 0.67 0.5 0.33 0.25 0.17 0.09
DL 1.21 1.39 1.65 1.85 2.13 2.64
t log 0.71 0.61 0.51 0.47 0.43 0.4
Dc 1.18 1.27 1.37 1.42 1.48 1.53
% over 3 12 28 43 71 112

Table 1: Over Redshift of recessional velocities.

To reach those conclusions, we have not needed any parameters or any universal constant! Below Z =0.2, in the few billions of light-years closer, but beyond the reach of the methods of direct estimation of distances, the difference is less than 1%, so the metric models FLWR would gain an application limit, being optimistic about the ability of future predictable astronomical techniques, Z<0.2.

“%overD” means the extra-redshift, i.e. we'll overestimate in 12% the distance at Z=1 vs a linear no-matter model. Any other Z may have to be correct with a precise ratio.

Up to here, we have assumed that Λ is a fundamental constant for an outsider god, but variable for inner observers, and precisely that invariance fixes a limit in the application of the logarithmic scale, and may not happen when ω begin unless negligible compared to tr. Then, for the very beginning of Expansion, we could speculate a reformulation of a more complex metric in the same time dimensional basis, wherein the reference time considers ω.

Minkowski space-time will increase dimensions and split dt in dtr&dta, then c would not be constant for all observers, even for divine, and other constants are treated as what they are: variables. The implications of non-conservative magnitude would be deep: Noether asymmetry. From Sato 1990 [29] to Kochanek 1996 [30] and Belinchón 2014 [31], some Λ(t) variable anisotropic models has been proposed as maths jokes, meaning also α, c o G. In any case, the logarithmic scale has not much observational sense above Z=11.1 [32], unless maybe could apply until Z> 1,089, although to guess and follow a dark path until another better model.

Following the example of Z=1, what we believe according to Friedmann-Lemâitre metrics it happened in 0,5tr (without cosmological correction), it was less time ago according to the way we measure it now: 0,61tr (also without correction of Dc). When we measure the brightness of supernovae Ia 0,5tr, they seem to be further than we expect because in fact they are further away than where they might be. It can’t balance because the metric used as implicit assumption applies a bias. In 0,5tr, we assume from the equations above the distance is 1.27 times estimated when it is 1.39 times, up 12%. In both cases it was considered c constant from our perspective: elongating the wavelength, but not the time in which it happens; and having no direct references we can’t know but in relative terms, how distant supernovae are in each Z, or its youth, but about what we assume on them Figure 4.


Figure 4: Illustration of Amanullah's 2010 Compilation.

Merged from Supernovae Cosmology Project & 42 Gamma-Ray Burst [33-36]

While W-Virginis were not identified, the Universe was much younger than the Earth itself, and it is a good starting point to think about the supernovae Ia critical mass is well known even at high Z's, but maybe that means digging in our heels. Up to Z = 2 (max. Z = 1.914), this simulation is consistent with the measurements from Perlmutter, Riess, Schmidt, and much more harvested in the SCP, and even with other estimations based on gamma ray bursts, up to Z = 7 (1.44 to 6.6), which has been used to predict an accelerated expansion, as the metric and luminosity constancy have been taken for granted -Figure 4-. The corrected curve returns the distances to the environment of a Universe with density of matter and without acceleration. "Quod erant demostrandum”.

The current mainstream explanation of the paradox of the bizarre behavior of the expansion brings us to a bigger and darker problem than the one that is trying to solve: the “fifth-column” Constant Λ, or the Quintessence scalar field Λ(t). Perhaps the convergence of the results is a fluke, perhaps a Confirmation Bias, but can also be considered an argument for upgrading the Conjecture of an spinning universe to Hypotheses -since has been deduced from an isotropic Universe with non-0 Angular Momentum-, and maybe in the future, with better tools like DES, it will refine and measuring older Z's, eventually confirm or reject this drill.

It means not only that the expansion is decelerating, but also that Transparency corresponding to Z=1089 would be in a younger age of the universe expressed in current seconds although a larger distance, which means nothing because the units of space and time are conditioned by our observation criteria. From the perspective of an outsider observer with metric in which the time jointly expands as the space, look back time tends to infinity

equation (17)

and time of the first phase of the expansion was immeasurably large (although sized in very tiny units). It has probably taken longer to reach from the Beginning to Decoupling, -in this diminishing space-time metric referring to our time units- that from that Era to us (in a clock running from there to now). Then h is also variable as the scale factor:

equation (18)

From our metric we interpret as explosive a process that a divine observer or an observer who lived in those first seconds and years of life of the Universe, would understand as a very peaceful and progressive development, and would not have any sense to set out a Problem of Isotropy or Hyperinflation, because there is no Problem to fix. Calling the process Big Bang is just a bias in the criteria of the observer: it seems explosive to us because we compress in our clock that huge time... as we show an entire dramatic series concentrated in a split of a second... we would not understand the history and all would seem dizzying to us.

The speed of time has been breaking logarithmically since the Beginning, and in this sense the concept of time is analogous to the concept of “the distance between changes”. In the very first few seconds there were many changes, yes, but only because the speed of our clock -one second per second- transported to a "position" of the time axis, such that the speed of time was much, much greater. The universe would have according to our patterns of measurement of time an age of 37% less and a diameter 58% higher, that is to say the age that we are today: 13,700 and billions of years of age (with Guth's conjecture, the diameter is Triplicate). Nothing exploded and we are not only decelerating, but almost stopping!

According to the patterns if time rotates around space, the Universe would be 37% younger and 58% larger, taking as reference the age that we date today: 13,720 billion years (the radius with the A. Guth 1981 Hypothesis would be x3 [37]). When photons pass trough the mass, the Universe had 1/3 of its age even from our point of view, which is a very long time. Nothing burst and we are not only slowing down the expansion, but almost stopping!

So, while the very long decoupling era -+/- 115.000 our-years-, an observer from then had time to extend the poly-tropic star model to the whole Universe itself. He could model an adiabatic expansion of a ionized and degenerated plasma with high specific heat in thousand degrees magnitude, as it might be just beyond CMB. We would expect to identify some kind of pattern similar to the surface of the Sun: convective bubbles in the transition phase from convection to conduction in transparency and explain the slight inhomogeneity and its rotational nature. We know the size of central galactic singularities can't be explained by the same process of massive stars collapse, because the Eddington limit, unless then that limit was then 3 to 5 times larger... as it's obvious from this hypothesis perspective. Then the galaxies birth from the inhomogeneity of a convection surface would be easy and natural, with stars not hundreds but millions times more massive. The same reason would explain the why the IMF modeling doesn't apply to the early youth and we are not able to find any orange, red or brown Population III small stars with clean spectrum.

Exotic Matter

The overlap between predictions of the Bariogenesis and calculations by the resonant analysis of the cosmic background radiation, estimations by virial theorem, by rotational speeds in super clusters and galaxies, by X-rays pressure gradients in hydrostatic equilibrium of hot gas, by spectrography of rotational gas in superclusters, by computer simulation of the rate of formation of structures (Sachs & Wolfe), soft&strong gravitational lensing, by Siuyaez & Zeldovich Effect in the CMB, Boomerang project... provides consistency to the hypothesis of dark matter, but can a hypothesis be called theory containing the word "dark"? Maybe we may think otherwise: change the paradigm, as we bury the geocentrism, heliocentrism, the estate of rest, the linear motion, determinism,... and also bury anthropocentric, mathematical totalitarian, fine adjustment, conservation information principles,...

The best argument for darkness is do not find a better explanation.

Dark Matter includes cold and relativistic baryon components (intergalactic gas, gaseous filaments, neutrinos, planets, planetoids and belts, brown dwarfs, black dwarfs, neutron stars, "strange" stars, black holes, halo, hydrogen, water,...) which existence is observable and extensible by the Cosmological Principle, where we do not observe them (OGLE & MACHO surveys). Although is difficult to "weigh" and if so, with a significant margin of error, even in the best estimation, baryonic matter is not enough to explain the measurements obtained. Mainstream solves the “dark matter problem” with a more complex hypothesis that the question: implicit assumption that if the measurement is correct, there must be exotic matter that interacts only gravitationally. Another alternative option would be that the implicit Friedmann Metric of constant time, distorts measurements, and will be the approach here...

We do not know how much mass contains the black holes of Population III, we did not even know the existence of dwarf galaxies like Segue1, with apparently 3400 times more invisible matter than visible [38]. What part of dark matter is unknown matter for the resolution of our devices? What incredible things will we find when we watch the Universe with neutrinos or gravitational waves telescopes? Extrapolating the models to the early Universe, there should be more massive and far-flung protoplanetary belts and halos than we assume in our solar system, more black dwarfs, many more brown dwarfs (IMF), much many more slow neutron stars, and who knows if much many, many, many more black holes (because the massive gap between the center of the galaxies and binary systems and because the unnova silent decay modeling).

No margin for exotic matter remain in Nucleogenesys. The question is not whether or not dark matter exists, but what's the nature of the exotic matter, whether it is cold or relativistic, and which is its distribution and how decays on time. On one side, baryonic matter may spread out (neutrinos, gas, ...), but in any case also with higher density as close it is to the center (black holes, brown and black dwarfs, orphans exoplanets, neutral hydrogen,...). Dark matter hypotheses leads to bizarre distribution of exotic matter: "profile NPW" [39] or other nonsingular isotherm distribution. La Silla Observatory, has mapped the orbits of more than 400 stars in a volume four times greater up to now in more than 13000 light years from the Sun. "The amount of mass fits well with what we see -stars, dust and gas-. This leaves a narrow room for extra matter -dark matter- that we expected find. Our calculations show that should have been clearly seen in our measurements, but simply, it wasn't there!", says the team leader Bidin et al. 2014 [40].

There are clear evidences of galactic halos. Observing Bullet clusters collision occurred 150 million years ago [41], dark matter would be associated with normal matter and not with gas, and measuring the deflection by gravitational lensing, halo is left behind after rubbing, which limits to the inner visible galaxy the presence of dark matter. Other measurements such as MACS J0025.4-1222 reconfirm the effect [42], although estimates its gravitational collapse limit their mass, and it's not enough to explain gravity effects. If massive halos would be really associated with faster rotational discs, therefore with brighter galaxies, we may expect a correlation between the speed of a satellite galaxy to the main galaxy and its disk rotation speed.

The dwarf galaxies Fornax and Sculptor, describes a uniform distribution of dark matter [43]. Nor gaseous filaments justify the peculiar distribution needed to explain dark matter (in 2008, AEE, XMM-Newton telescope) [44]. With our sizing hardware most part of Dark Matter seems to be exotic and bizarre... but more than this: strange. In every new approach we find new hidden mass: the very old Segue1 does not have 5 or 6 times more dark than visible matter, but x3.400 [38]. It seems that the most of the stars are older out of the main sequence, the more dark matter referred in older galaxies.

We start from the prejudice of seeing galaxies as gravitationally consistent systems, not aware about intergalactic expansion. The problem of rotational or virial speeds ceases to exist as such, accepting stars and galaxies are in fact escaping from their gravitational systems at lower speeds than recessional (Figure 4), following a hyperbolic, logarithmic, golden or other runaway spiral path pattern, still remaining tens, hundreds, or thousands of galactic orbits, before dispersion is evident in some tail as a comet.

According to constant-time metrics, since the formation of the very first galaxies the Sun has completed about 50 orbits to the Milky Way -70 or 80 in logarithmic metric-, so maybe it's not that the rotational speeds require dark matter for justify why stars remain in galaxies, but are simply escaping in elliptical orbits. Something holds but not kidnap them, and there is not enough time to scatter through intergalactic space, which expands faster together than dispersed locally. This would implies that the recessional velocity measures apparent expansion speed, which is unlike what intergalactic space-time relative to intragalactic expands, resulting in a younger universe than estimated, which may be consistent with the previous argument that estimates 1/3 less age Figure 5.


Figure 5: Azcorra proposed model.

I. Azcorra proposes a model in which tangential velocity of rotation adds to Ho expansion. Accepting that the galaxy is fading, he estimates that the Sun will remain "only" 20,000 million years of being in the Milky Way, which far exceeds their life expectancy and is more than the "official" age of the universe up to now. [45].

Mainstream means that galaxies are gravitationally bounded systems: do not change their volume with expansion. Is a cluster a gravitationally bounded system? and so is a super cluster?... even wider filamentary structures and walls, with no gradualicity? The theoretical effect of expansion in the Sun-Earth system is 44 orders of magnitude smaller than the internal system gravitational forces. Applied to the solar orbit around the galactic center, the effect is 11 orders of magnitude less than the acceleration due to own gravitational effects of ligation. Even at giant galaxy cluster scales, the effect of the expansion is 7 orders of magnitude smaller than the acceleration due to internal self-gravitation of the cluster itself ... but this is today! If G depends on the inverse of time, the effect of modification of the gravitational ligation should be noted in high Z's,... and also in the fine structure.

The “problem of the big numbers” -strange coincidences in the power of 10 in the macro and micro constants- led the terse Dirac 1937 to speculate on the variability of G as the inverse of the time [46] and proposed the LNH, describing its incompatibility with FLRW. He was criticized by Zwiky 1939 [47] and also supported by Chandrasekhar 1939 [48]. Sciama 1953 [49], and later Brans & Dicke 1961 with an scalar-tensor VSL theory [50], developed the conjecture including the Mach Principle. Gamow 1946 first derided, and later call the hypothesis smart [51], and proposed to verify Sommerfeld's constant [52]. The alternative chosen by the Mainstream has been the “teleological principle” or “petitio principii” as usual: rescuing the Boltzmann rectification (which may be delayed physics some decades), updated with Guth 1981 [37] and reformulating hypothesis following the geocentralism and denying de Natural Selection or the Coase Theorem about scarcity: Anthropic Principle. We want to feel special.

Why gravity would be related to inverse to the square of the distance –conceptually, a surface- and not to distance -proportional- or volume -cubed-? Newton did not know the value of G, but GxM, and according to this hypothesis, G outstanding decreases linearly, but may only be measurable from Z> 0.2. Inevitable but consciously influenced by the narrative confirmation bias, considering the metric logarithmic time, gravitational density has been diluted

equation (19)

“pari passu” to the photon density (αT3).

Considering these factors, is not in super symmetric particles, right-handed antineutrinos, axions or WIMP´s, where we find the explanation about exotic matter. In look-back-time, the apparent gravity would be

equation (20)

higher than its projection to our time, that means 5,746 times more from the Detachment to today, coinciding with the proportion of exotic matter calculated with the resonant CDM model.

If so, galaxies should be larger and increasingly less dense in relation >(Z+1)1/3, because intragalactic expansion is maybe less but anyway expanding like intergalactic space-time itself, and because in fact they are in runaway spirals. On a sample in Z~4, Ferguson 2003 notes that galaxies were smaller, irregular and more massive than our days. In the same way, comparing large galaxies in SDSS, between Z=0.2 and Z=1, those with more than 1.5 kpc radius multiplied by 500 [53]. Sizing galaxies with dispersion speeds -σ-, Osiris have found 6 times higher densities than current average (slightly less than expectable (Z+1)3), in four elliptical galaxies in Z~1, half of size ("a bit larger" than expected Z+1) and only an increase of 1.8σ (slightly less than Z+1 expectable, if the space intragalactic expands similar to intergalactic) [54]. GN-Z11 is 25 times smaller with a 1% of the stellar mass of our Milky Way.

Brightness is more or less on the fourth power of maximum speed (Tully-Fisher) or stellar velocity dispersion (Faber- Jackson): "grosso modo" its squared mass. With the I band specter, we can estimate distance (we have already analyzed its bias) and weight. The mass calculated by the virial theorem depends on the radius of the cluster, the square of the velocity dispersion and inversely to G, but we have seen that both G and radius are linearly dependent to radial time. As their dependence is canceled one against the other, it would be proportional to Δσ2, which for the above example Z=1, means a mass of 3.24 times over-estimated. In any case virial weighing of galaxies should include a further decrease (Z+1)1/3 of baryon/exotic ratio, which reaches to 4:1.

Distribution of stellar rotational speed in the Milky Way, is observed as expected p2/r3 by the keplerian laws up to 10,000 light-years -on the same order of magnitude as its thickness-, Figure 6, while mass distributed further than 20% of its radius is "running away", followed for an increase according to the expected for a spiral orbit escape, but it still remains 2/3 radius with approximately constant rotational speeds, when according to the classical model should be decreasing ... if we did not consider the Expansion! In a deep galaxy far away, rotational speed -vt- won't be proportional to 1/√r but

equation (21)

which makes them dependent on the distance of the galaxy -furtherand the reference radius for measuring -smaller-. Four articles in 2017 based on the European telescope in Chile -ESO- point to fit in more "Keplerian" velocity curves in very distant galaxies, [55-58], just as it is expected to be following the present hypothesis, and it is waiting to expand data with the VLT and JWST telescopes (Figure 6).


Figure 6: Graphical Illustration of VLT and JWST

A unit of length taken by an observer with straightedge billions of years ago, was shorter if we measure according to the straightedge of our clock, but it was "the same" for the rules of physics then, if we transport then our space-time metric. Although most of the mass of the galaxy is in spiral orbit not enough to be retained by gravity, while the expansion exceeds “the runaway spiral", the distance of the peripheral mass grows but less, than the growth of the length-unit taken as a standard at every moment of the life of the Universe.

If so, the rotational speeds should be kept by expanding space-time as they were billions years ago, being the distribution according to Kepler's third law a brake, but may not be the model. We assume centripetal force Fg= GMm/r2 must be balanced with the centrifugal “force”

equation (22)

and then we define the Kepler's rotational speeds expected. What if the measure did not apply constant time metrics? r could not be canceled on both sides of the equation at different times in history, rc ≠ rg. "Dark matter had less influence on the early universe. Observations of distant galaxies carried out with the VLT suggest that they were dominated by ordinary matter”. (ESO 1709).

Rotational speeds distribution would be a fossil record of the maximum speeds, corrected by logarithmic metric

equation (23)

which provides graduality and fits with the near “keplerian” velocities distributions in very far away galaxies -0.6<Z

We could say the same for the virial speed in super clusters. The outline of a galaxy is fuzzy and to adjust rc to ln(tr) on the same axis and system of units, we need a common reference -normalization-: a galactic radius R such as holds horizontal line

equation (24)

If intragalactic behaves metrically the same as intergalactic spacetime, gravity can’t be a brake dependent of the distance between mass, which is only possible if G decreases with distance (in fact decreases with time all over the space the same), closing the argumentative circle. As an example, in 0,61tr, when the space-time units were half of our time, the velocity distribution curve of the Milky Way was "keplerian" to 6 kpc (twice than in our time). In other words: an outer star in our galaxy has the same vt than when it was at 13,000 million light years away, and covers the same amount of space units per unit of time, but both have grown proportionately and in fact it escapes following a spiral path.

If G decreases as so, the Chandrasekhar limit may increase in lookback- time arrow, and with enough time single white dwarfs would decay beyond the electronic exclusion pressure even without feeding, and maybe before their evolution to “diamond” black dwarf. In high Z's, that adds another correction to the distance estimated by supernovae, because then the critical mass was higher (i.e. Z=0.5, x 2.7; Z=1, x4.5; or Z=2 x7.4); and so was the luminosity, but not its spectrum.

From September 2014 several observations GW has been made in LIGO in the order of hundreds of Mpc: on the more optimistic models of density of black holes coalescence events, and more than this, of bigger masses than models expected [59,60]. If times rotates over space coordinates, while bigger masses are easier to fix at low frequencies, and beyond that it is indeed a bias for the more massive events, we can advance the masses that will be found in the next 5 years on (Figure 7),(which they claim to achieve increase distances to thousands of Mpc). Those masses are and will be bigger than the more massive known stellar black hole, maybe up to 250 Mo at very high Z’s.


Figure 7: Mean of masses of stars in terms of actual mean mass.

The constant 8πG/c4, -which relates the second derivative of the metric tensor Gμυ, to the relativistic energy-momentum tensor, Τμυ-, would be dependent on the volume of the Universe: as the Universe Expands, less mass is needed for the same effect in curvature. Galaxies configuration around a black hole were built in a size according to the change in the order of magnitude of the Eddington limit: central galaxy mass versus stellar black holes mass may have grown per millions due to the proportionality of the critical mass to the volume, -αtr3-, and its multiple stellar origin collision and migration, is not a reasonable hypothesis.

History left us its footsteps on distribution: peaks during hunger activity periods of its central black hole (the more outward the more frequent should be, we must appreciate even a tendency to increase vt with the radius); the more pronounced valleys, the more closer near the center (decelerating expansion α1/tr2, gravity may force the curve to the classical α1/√r, escaping the rest of stars as the tale of a comet); and for the same reason above, these "fossil peaks" will be lower (it will not split on the radius, but as the movement of a whip), being consistent to those observed without the need of non-baryonic Dark Matter. So, G is comover, not proper.

Parallel sizes slight approach in G's, between -0,1% y el +0,7%, [60-62]. G'(tr) may decrease if Expansion is decreasing. There has not been evidences with frequencies of atomic clocks or decreases in the rotation speeds of the bodies (although and beyond its “relativity” gravitational interpretation) or orbital decay (LLR, with mirrors on the Moon); but also alternative scenarios (if the brightness depends on G7 and radius of 1/G, temperatures may boil the ocean few hundred years ago). Jordan 1971 even proposed Relativity modifications [63], but had not impact either.

Kaluza & Klein related for any extra-dimension the color of light with G:

equation (25)

So, being the Sommerfeld constant α1/εhc, as hα1/t and cα1/t, we may either size very small direct variations (depending if h variation is or is not “pari passu” with c), or try to interpret indirect observations. Since then observational astrophysics has tried to verify the hypothesis of the constancy of G through the constancy of fine structure constant, α, -DEEP2 confirms the color of oxygen up to Z=0.7-0.9 [64]. Color modification depends on the electron's charge, e, on the permittivity ε, on the Planck constant, h, and on the speed of causality, c, -α e/εhc-; if ii would not be so, either Cosmological Principle is wrong at large scale, or charges, or h or c, changes with time and expansion. In those hypotheses, h changes with the scale factor ln(tr)=1/(Z+1), and c with the inverse of radial time, so

equation (26)

The effects of G' in the gravitational dynamics, in stellar behavior, in the light of stars, in the rotation of binary pulsars, or the Hubble constant itself, may be less than 1% variability on color spectrum in the last 1.5 billion years, Z~0,35; and in 0,5<Z<1, so we may be able to observe 7%<|G'/G|<18%, which means less than 10-11 or 10-12 per year. [65]. Analyzing the Oklo natural reaction, Shlyakhter 1982 [66] supports a change even less than 10?17 per year, over the last two billion years. Lamoreaux 1994 found a 4,5 parts in 108, decreasing of α, [67]. J.Webb et al, from 1997 to 2011 edit several papers [68] analyzing the changes of the spectrum of light from distant quasars, trough metal clouds, with a so slight effect that it would only be clear to Z's high -Δ1/100.000-, in a consistent range with in this Conjecture and with differences depending on the direction, which may indicate laterality, [69].

Einstein 1911: “The principle of the constancy of the speed of light can be kept only when one restricts oneself to space-time regions of constant gravitational potential”. All this sounds familiar to the MOND [70] and its relativistic expansion TeVeS, [71], but much more fits with the MOG or STVG models [72,73]: assuming G/c as a constant; not by its possibilism, but because both depend analogously on 1/tr,

equation (27)

Interaction between mass is not given by its proper distance, which depends on c and a(t), and therefore excluding expansion. Gravitational force at large scale changes to comoving criteria. Relationships between variables that depend alike proportional to time would be constant, T/G, P/T, c/G, c/T,... Λ/ɱ(?)... because their values are only adjustment of units. If G is inversely proportional to the radial time, rewinding the film looking back in time, gravity was in the same order of magnitude as the other interactions almost a trillionth of a trillionth of what, in terms of our measurement prejudices, was the first second.

This alternative interpretation to the exotic Dark Matter leads us to reconsider the acting distances of "Newtonian" forces as "comovers" and the geometrical measuring depending of c, as “proper” length. From our point of view the gravitational constant weakens with the Expansion, but from the "comoving" point of view, it doesn't. It doesn't matter how much energy will increase the hadron colliders: they will simulate a very partial view about what happened on these temperature ranges in terms of GTU theories, because particles behavior would not be only a matter of temperature, but also depends on the space-time density forces range and the angular speed.

If Dark Energy were the answer, it would be split to the different questions: Initial Hyperinflation, Acceleration of constant or variable Expansion with respect to volume or to time, in its vacuum or scalar versions. Λ'(t). How the universe conserves energy? Is the amount of dark energy related to the amount of dark matter? If Dark Matter grows with time, does Dark Energy grows in look back time? Granularity of the CMB draw typical boiling emerging patterns of auto-similar structures of 1º and frequency around 200, compatible with flat curvature and null critical density (as a condition for the energy-matter balance to be null).

From Statistical Mechanics, Dark Matter comes from the incompatibility with the Virial Theorem. From Classical Mechanics, from it comes from the incompatibility with Kepler's Laws. From General Relativity, it is the focus of gravitational lens, but with tremendous variance. From Electromagnetism, it is an interpretation of the harmonic analysis of inhomogeneities in the CMB. From Quantum Mechanics, it is an alternative explanation of the missed mass on CP asymmetry balance. From the Numerical Calculus, it is the output of the tautological adjustment of the galactic genesis. From Special Relativity, the Dark Energy (Cosmological), comes from the incompatibility with the limit of the rate of causality, c. From Thermodynamics, different versions deals with the compatibility of Increasing Expansion with Energy Conservation Law (Figure 8).


Figure 8: Future estimations of very deep luminosity distances will find extra-red-shift beyond the extra-red-shift found until now, and it will not fit with the upper curve today used as concordant, and estimated distances with FLRW metrics will grow even more than its own correlation m-M=43+5log((Z+1)lg(z+1)) or m-M=44+2.5lg(Z) or m-M=44x0.06 [76].

If the Dark Matter is the answer, it would not be neither the same “Dark Matter”, with different properties to the different questions: each drill needs its own “matter”. To explain the rotational speed in galaxies we need a profile distribution inverse to the baryonic matter, but in a cluster it seems to be associated to baryonic matter. There are different proportions according to the method of weighing used and the object analyzed (they measure three orders of magnitude between x3.5 as a minimum measurement by gravitational lens and x3400 of Segue1) [38]. The oldest galaxies have velocity distributions "keplerians" alike, and means that exotic itself would be of baryonic origin, so we more task: find exotic matter, find the why and how migrates to the halo, while the halo has to decrease its radius to increase its effect, and how it decay the one in the other; in another roller coaster “deus ex machina” [74]. If so, the decay may release energy, the fragmentation modeling with 4/5 of dark matter does not fit, and the CMB would not contain that much of darkness from Zeldovich harmonic analysis.

It is not the same Dark Matter to justify masses of galactic holes, stellar mass distributions or B-modes. The distribution, proportion and decay are not compatible to justify masses of galactic holes, distributions of stellar masses and simulations of galactic genesis. We need two or three types of dark energy and even some more dark matter forms, we need non-gravitational migration dynamics to the halo and decays from baryonic to non-baryonic, which are also incompatible with each one arbitrary evolution without an identifiable pattern. They are so dark that darkness does not allow to see anything, like ether, at all. It might be one of the worse theory ever announced. Science never proves the certainty of a hypothesis, but its falsity, and the "darkness" is as "false" and empty as the ether.

We don't need exotic and bizarre distributed matter to balance equations, no ghost forces, not even constants, not anthropocentric principles, beyond the fact that has been so many years without appearing in the detectors, even with restrictions of interacting only with weak WZ particles. Although the rotational speeds are apparently running away, Expansion will always be faster and galaxies will evolve while they get away one from another, decreasing its density and increasing its size until fuzzy star clouds, increasingly distances one from each other, switching off, leaving a tail of stars and dust in its track. Sad and dull destination... or maybe there is another way...


Reality do not mind about us or about our perspective. We use axioms (isotropic, homogeneous, flatness) that fits with specific useful measurements, but GR itself does not prefer a particular metric; and more than this, because its formal, we use linear space-time dimensionality relation (cdt) and symmetries (spherical, axial,...) on a plain Minkowski manifold (-1) to afford the equations. So, to estimate distances in deep space we use Friedmann metrics as it is consistent with isometry, homogeneity, space-time symmetry, invariance of c, linearity,... and we size accelerated expansion with this implicit axiom, because it has been a very good approximation in "large but local spacetime scale", (Euclidean metrics is also accurate for its own range). Maybe at larger scales there are better metrics for modeling than FLRW, and if so, only with another ruler and/or another clock, data do not mean accelerating expansion, and though do not mean Dark Energy (which it is easier to understand).

GR is deterministic and assumes a linear relation between time and space. Time is another dimension like the other 3-spacial, but time has a preferred arrow and space don't. Time can be modelled dynamically as any other dimension considering this circumstance. If time rotates around space, to preserve c as constant, metrics has to be logarithmic, we change a squared graph paper to size distances by a gold log graph paper, to be conservative, isotropic and invarianced, but also not-linear and not-deterministic. Then, main constants as G, c, H,... must be variable from our perspective, but not for an observer in each moment, holding a Soft Cosmological Principle. See also [75-79].

Extra-red-shift data means maybe that the Universe is accelerating and there is a Dark Energy hidden, or alternatively maybe that time rotates over space. If so, without any change in Relativity or Big Bang Theory, Dark Energy is not that Dark, but angular momentum, Dark Matter is not exotic, Universe is not accelerating, black holes will be smaller with time, Big Bang was a very boring and slow process, so will be Big Crunch, and all fits with observations and can be verified with our day technologies.


Citation: Pons-Rullán B (2017) Distance and Mass Correction on Logarithmic Metrics for High Z's, Due to the Non-Cero Angular Momentum in an Isotropic Rotational Universe Hypothesis. J Phys Math 8: 234. Doi: 10.4172/2090-0902.1000234

Copyright: © 2017 Pons-Rullán B. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Post your comment

captcha   Reload  Can't read the image? click here to refresh
Leave Your Message 24x7