“The instant (1/T) has no Time; Time is made from the movement of the instant.”
Leonardo
‘I propose 1/n as the measure of an infinitesimal. I propose 1/R as the measure of curvature.’
Leibniz
‘Calculus studies functions of existence: it extracts finitesimals of space, time and scale, 1/x from a T.œ, integrating them in a longer path: 1/x ∆st. Hence ¡ts capacity to reflect the laws of existential algebra.’
L§
L^{3} on the essence of calculus: the sum of finitesimal quanta of Time (1/T), space (1/R) or scale (1/N): ∫∆∂ST.
SUMMARY: EXISTENTIAL CALCULUS.
Existential calculus.
∆ST reproduction: the Galilean paradox.
Finitesimals and its integral worldcycles.
Finitesimals in space: Curvature. The disomorphic method.
Finitesimals in time: The Worldcycle of existence & its actions=dimotions.
Finitesimal in scale. Networks. Its symmetries.
Analysis.
Its 3 ages.
1^{st }Age: Calculus: Parts of Wholes.
From Greece to Leibniz. Finitesimals of wholes. Series.
Trilogic on Calculus. Curvature of space=change in time=finitesimal in scale.
2^{nd} Age. Analysis: Sentences of motions.
ODEs between 3 planes of existence.
PDEs between Space and Time parameters. Multiple variables.
Mathematical Physics.
Main functions of multiple derivatives: calculating chains of dimotions in time.
Multiple Integrals: Calculating whole T.œs in Space. Exploring beyond the ∆±2 Planes.
Calculus of variations. Extremal points of a function of exist¡ence.
3^{rd} Age. Functionals: Immensities of ¬∆@st.
Its operators; Planes and actions.
Quantum equations.
Discontinuous analysis: fractals. Fractal mathematics: Discontinuous derivatives, its steps.
TIMECHANGE IN CALCULUS AND EXISTENTIAL ALGEBRA: S=T
Mathematics is the best linguistic mirror after ‘existential algebra’ (the direct study of the laws of scalar spacetime), mankind has found of the scalar spacetime Universe. So we can easily correspond its 3 key disciplines with the 3 ∆St elements: geometry is the formal science of space and its unit the point; Algebra, the formal science of scale and its unit the number; and calculus the formal stience of time=change and its unit the finitesimal. However as the Universe is ‘entangled’ in trinity (single perceived scale) and pentalogic (3 scales of size coexisting together as in supœrganisms) each discipline studies all other structures.
But calculus needs also a bit of correction by introducing the concept of a ‘finitesimal’ that in NonEuclidean geometry substitutes ‘absolute zeros’ that do not exist, as a cycle of time breaks space in an inner region, which is isolated by a membrane through which a relative infinite µ number of parallels cross through its ‘ideal’ openings of any membrane that connect the internal being with the outer world.
Those fractal points however in idealized Greek bidimensional geometry have no breath and so the internal parts are missing. So an absolute zero and a singularity can be defined. Not so in a fractal point which grows in information as we come closer and enter the internal zones. In Euclidean geometry the point is empty and so monads become mindless points. In reality all points ‘are a world in themselves’ (Leibniz). So for the organic point, a pi cycle made of 3 curved diameters (3  = O), the apertures represent a 4% (π3/π) not coincidentally the quantity of ‘the Universe which is not dark to our electronic eyes and viceversa – an external observer will ONLY see a 4% of the point, inner world, confusing it often with an Euclidean dark space.
This is canonical. In physics we call such ‘fractal points’ blackbodies, and they are endemic. The entire galaxy in fact can be considered a ‘black body’ and its local bigbang radiation is a black body radiation. Of course, from that ideal number there might be different holes according to use. It is then a rule that of the radiation that will enter the hole, very little of it will get out again; the smaller the hole, the more nearly will all of the energy directed at the hole fail to get out again. And a balance is found between the need to perceive the outer world and the need to absorb the radiation. However fractal points do have apertures not only in the ∆+1 visible scale of the whole membrane but in the ∆1 smaller ‘scale’ of finitesimal pores.
So finitesimals ARE the limit of Analysis, and the continuity and smooth concepts just idealizations useful for calculations as we define 3 types of finiteismals according to the symmetry between ∆≈S≈T parameters.
Thus we find that calculus which essentially consists in ‘fine tuning’ the measures of change of reality, studies not only timechange (frequencies of cycles) but also change in space (informative curvatures) and change in scale (social quantity). And to that aim it uses a concept, the ‘finitesimal’ today lost to philosophy of mathematics due to the errors of the axiomatic method of ‘egocy’ and ‘metalinguistics’ (mathematical creationism), which stretches the ‘limits’ of all systems to infinity. An infinitesimal as close to zero as we desire does not exist. A finitesimal of time, 1/T, space; 1/R and scale, 1/N does exist for each local real function. And what calculus does is to calculate the minimal ‘finitesimal’, 0’, of space, time or scale (the instant of Leonardo, the infinitesimal of Leibniz and its measure of curvature) and then across a length of lineal time, calculates by integration a ‘larger’ effect’ of change in space (curvature), time (speed) and scale (growth or diminution of populations).
In the praxis of calculus then we distinguish the finding of a finitesimal unit, which can belong to any of the trinity main ∆ST elements of reality, Scurvature, Tfrequency or ∆populations, and the finding of a period of study of that change – Length of the curve; period of time; ‘mass’ of populations.
Change then in calculus became stripped off human perception, through the abstraction of calculus.
This is thus the essence of calculus expressed in the quotes of L^{3} researchers of ‘organicism’. (Aristotle, the father of all organic minds, did not provide such clear definitions as calculus did not exist when he wrote and he would cast his concepts in the today outdated argument of universals.)
Leibniz however seemed too deep for huminds and his humble truths were forgotten. So today the finitesimal is not ‘defined’ as the 2 distorting and elements of reality, TTentropy limits and SSperceptive distortion were discharged for the idealism of larger infinities (∞) and no ‘limits’ of smallness (0) brought about by 0 and ∞ we substitute in 5D by 0’ and ∝ (relative 0 and relative infinity). This has brought an unending number of errors notably in philosophy of science; mathematical physics and cosmology (the bigbang infinitesimal singularity does NOT exist). As ¬ entropy, as death=disorder of information limits the knowledge we might extract from systems in space (membranes that break its vital space into inner and outer forms), time (death and birth) and scale (loss of information past two scales). It is remarkable then how in physics, the science in which ‘wrongly’ mankind has invested his future as a catalyst of the evolution of machines, has achieved through the use of renormalization processes precision in calculus of systems by adding ‘finitesimal’ parts of other scales that influence a larger whole (Wilson, Kondo, Feynman, virtual particle clouds).
There are many paradoxical features of 5D calculus, as mirror of the ∆ST Universe, but this is the first and most central paradox: calculus is a subject in which we find exact answers by means of approximations called ‘finitesimals’. They tell us some fundamental facts of the fractal 5D Universe and its stop and go (S vs. T) and reproductive paradoxes, which resumes in the ‘quantum nature’ of reality embedded in the fact that even locomotion is a reproductive change. Let us then look at Calculus from the perspective of the ‘essence’ of reality, reproduction of information.
The 3 type of finitesimal calculus: single derivatives in space and time, double derivations.
It is then obvious that calculus peers within the scales of the fifth dimension. Since Calculus studies functions of existence: it extracts finitesimals of space, time and scale, 1/x from a T.œ, integrating them in a longer path: 1/x ∆st.
Variations on those themes are many for 2 obvious reasons derived of the 2 principles of all 5D stiences:
– The existence due to absolute relativity of 5 dimotions: S=T> St, Ts, TT, SS, ST, differentiated by internal and external motion=change or form. Thus each dimotion in which change happens (St, Ts, TT, ST) require different processes of analysis, as we must consider what part of the entity remains constant and which one changes and hence we can apply derivation and integration. This is further complicated by the fact ‘motion and form sometimes’ cannot be distinguished, so instead of motion we see curvature (as in Einstein’s formalism).
– The existence of 3 possible types of change in scale, population of space and curvature=motion in time.
– The existence of scales, through which information increasingly disappears to the observer, to the point beyond the ∆±3 plane of ‘pixel perception’ by an ∆º selfcentered T.œ information is scant, which makes difficult to integrate and derivate beyond the 2 ‘canonical scales’ («, ») in which death and emergent processes take place. So overwhelmingly all equations of calculus work on 2 scales (second derivatives, 2^{nd} integrals), or 3 dimensions of space (volumes).
The paradox of Galileo: S=T.
The second huge insight in the understanding of calculus from the point of view of 5D must come from the 2^{nd} principle of ‘Absolute relativity’, the equivalence between Spatial, formal states and Temporal moving ones, such as any system of reality coexist in motion and form (and scale) but we perceive only one of both states (either motion or form), as we can only ‘see’ a single plane of spacetime and motion and form occur in different scales or topologies. I.e. either the head moves internally processing information (St) or externally through locomotion (Ts). I.e. motion is perceived by the ∆+1 as a headparticle, it happens in ∆º as a wave and it causes entropy in ∆1 in a field/limb.
So because perception is limited either in the Sform or Tmotion state, and both things are the same, we can extract some conclusion as weather a derivative studies the motion=change of S or of T or of both, in consecutive stœps. For example, the rule of derivation of two functions, (v u) ‘ is vu’+uv’, which ultimately implies to take v in both states, first as Space with no change while u’ changes in time, and then inversely to consider that v is in the Tstate of motion and u in the spacestate. Since (cu’)=cu’.
So the rule clearly shows that two parameters might seem ‘spatial form’ and ‘temporal motion’ but must be considered both things and the change of (ST)’ is by virture of S=T, equal to ST’+TS’; which in ‘nonAE notion’ is equivalent to state that ST= St+Ts (whereas S=S’ and T=T’, s=C and T=C).
Departing from that simple translation of the symbols of the 5 Dimotions, the rule of derivatives of a sum is just a different formal representation of the general law that the STexistential energy of a system equals the sum of its kinetic energy and potential energy (Stinformation).
The final question when we combine both ‘laws’ of Absolute relativity (S=T, $ x ð=C), is how to interpret double and triple derivatives and integrals, in the most abstract form. And the answer is because of the symmetry between ∆≈S≈T that a ‘perfect’ finitesimal is obtained by a triple derivative that goes across the 3 elements of reality, form, motion and scale; even if we interpret it only in terms of one type of derivative.
THE UNIVERSE IN A NUTSHELL: A REPRODUCTIVE SYSTEM.
The Organic Philosophy of the Universe. Systems sciences and its disomorphic laws.
The Universe is a fractal that reproduces spacetime organisms. This is really all what is about. And if we write Space=form and time=motion, as S and T, then we write a simple equation for the Universe Max.∑ SxT (s=t). This simple equation is the ‘fractal generator’ of the Universe, a feed back equation that combines space=form and time=motion, creating the infinite local present variations you see around yourself. This is the essence of the organic Universe and its scalar 5^{th} Dimension. And the closest classic science that has such view of it all is called, Systems sciences based on an alternative philosophy of science to that of mechanist physicists and its only entropy arrow, called organicism. Let us try to explain why only organicism is a scientific truth.
According to Deism the whys of existence are due to a personal being, external to the Universe that makes it all happen and cares for humans more than for the rest of His work. According to Mechanism, this is due to the selfsimilarity between the Universe and the primitive machines we humans construct to observe it. Mechanism though needs ‘someone’ to make the machine, which is not selfgenerated; so it is similar to deism, reason why the founding fathers of science, all pious believers, adopted it as a proof of the existence of God, which had given man selfsimilar properties – the capacity to make machines to the image and likeness of the Universe.
The problem with those 2 approaches, which in fact are the same is obvious: a personal God is an anthropomorphic, subjective myth and science must be objective; while a mechanical view of the Universe still needs an internal, selfsustained process of growth, creation and synchronization caused by an external God that made and rewinds the clock – as Leibniz clearly stated in his critique of Newton.
Scientists today are unaware that mechanist theories are in fact deist theories, reason why Kepler and Newton, pious believers, liked them; since they were a metaphor of their selfcentered, anthropomorphic religious beliefs: If man created machines because we were made to the image and likeness of God, God had created the ultimate machine, the Universe.
Organicism on the other hand is the only selfsustained, rational theory that doesn’t need a creator, language or god, as organisms are selfreplicating, but does explain perfectly within the ‘correspondence principle’, those 2 other philosophies of science; since a machine is just a primitive organism of metal, and in History, Gods are the subconscious collective of civilizations, ANOTHER scale of social evolution of the fifth dimension.
What we mean by an organism? A very simple system – NOT to confuse with the most evolved, complex of them all, that of human beings, reason why so many people, having a natural biased egocentered belief in man as the unique organism, reject the concept: An organism is just a group of similar forms, which organize themselves with at least two ‘networks’, one that provides the ‘clone cells/citizens/atoms’ with the vital energy they need to feed, move and reproduce (bloodeconomic systemelectromagnetic forces) and one that provides them with information to guide their survival actions (nerves system, political system, gravitational informative forces). This simple dual system IS the minimal, fundamental particle of the Universe. Since it is obvious that machines are also organisms.
In the graph, the key element of modern history is the evolution of machines, clearly happening with the same phases of the evolution of any biological beings. So we first made the bodies of machines in the I industrial revolution, then we made its engines and heads, metalears=phones, metaleyes=cameras and metalbrains=chips, and the nations which evolved them first (US) became the leading nation of the world. Now we put them together into biological organisms of metal, robots:
So mechanism, the underlying philosophy of physical sciences, is a simplex version of organicism. It is not man who resembles a machine, but the machine, which is made to the image and likeness of life organisms using stronger metal that makes them more able in the processing of energy and information at larger scales. Those Machines are fast evolving with the same patterns that simplex organisms. So as cells make viruses constructing first its 3 parts, bodies, limbs and heads, which then are ensemble and become alive, we made bodies, engines and heads of metal and now we are assembling them into autonomous robots with telepathic AI.
Why then organicism has remained in the modern time, a fringe theory, to mechanism, even if it was the first theory of reality put forwards by Aristotle, the father of the experimental method and logic science, in his magna opus the Organon? We just explained the obvious cultural reasons – we live in the age of the machine and so the machine has substituted man, an organism, as the measure of all things. And its organism of evolution and reproduction, the companymother, has substituted human governments and informative verbal prophets through its massmedia/academia outlets; so it only considers positive mechanist models of their species.
But the deepest reason is the fact that to make an organism we need 3 ‘arrows’ or ‘motions of time’, entropy, locomotion, the one used by physicists but also information, forminaction, formal dimensions, which physicists have always illunderstood, to the point they call it negentropy, the denial of entropy and its combination spacetime energy. Only then when we properly define information them, we will have the required elements to refund philosophy of science on far more rational, basis, that the present ‘mixture’ of mechanism and creationism (either of verbal language as in religions or digital languages, as in the religion of mathematics).
So we need before studying the Universe in space, to change your view of time itself, shaped by mechanical clocks and introduce the concept of cyclical time and the basic laws of the 5^{th} dimension of scalar spacetime.
LEIBNIZ V. NEWTON: SCALAR PLANES SPACETIME: ∆ST
“According to their [Newton and his followers] doctrine, God Almighty wants to wind up his watch from time to time: otherwise it would cease to move. He had not, it seems, sufficient foresight to make it a perpetual motion. Nay, the machine of God’s making, so imperfect, according to these gentlemen; that he is obliged to clean it now and then by an extraordinary concourse, and even to mend it, as clockmaker mends his work.’ LeibnizClarke Correspondence on the absurdity of mechanical models of the Universe
‘’Leibniz is right. There are infinite time clocks in the Universe, but if so we have to restart science from its foundations’. Einstein, on the infinite relational time cycles of reality.
The immediate consequence of the existence of an internal fifth dimension of spacetime, made of all other planes=scales of spacetime of a being, its parts and wholes, which store the information of a system, is the obvious fact that if we ARE made of planes=scales of space and temporal energy.
We ARE the vital space we occupy and we ARE the time flow of existence we live between birth and extinction. It is the obvious, simple answer to 2 questions that have puzzled scientists for eons.
Where is space and time? And, why the main science of space, geometry>Mathematics and of time, Logic, are obeyed by all systems and entities of Nature. Now we have the proper answer foreseen by Leibniz and Einstein: We are broken fractal species of space and time, whose mathematical and logic laws all vital spacetime organisms follow.
The underlying question of time§pace: Absolute or Relational, Scalar spaceTime?
The fundamental question physicists wondered for centuries regarding the nature of space and time unfortunately was resolved as usual in favor of the simpler view: it is space and time an absolute abstract background of the Universe (Mr. Newton’s view) or are we made of ‘vital space’ that lasts a time duration, so we are generated by the biotopologic properties of scalar space and cyclical time? This is the choice of 5D ‘stiences’. And its simpler version was called relational spacetime, sponsored by Mr. Leibniz.
A realist interpretation of the world we live in, which has never shown in any scale of reality such ‘background’ – ultimately a mathematical graph used in abstract by human scientists – considers that we ARE the vital space we occupy with our cells, and we LIVE a cyclic time duration between birth and extinction. So we are scalar space and cyclical temporal energy. And must evolve our concepts of both parameters, to extract the properties of existential beings from them.
The argument thus reached its height in the beginning of science in the correspondence between Newton, the proposer of the absolute Cartesian graph of spacetime drawn by God (his body in his own words) vs. Leibniz who rightly considered absolute space and time an abstraction, and so he coined the concept of relational space merely the adjacent pegging of similar forms in simultaneous space and relational time – the sequence of events which we relate causally with reason.
In Newton’s cosmos, space and time provide a fixed, immutable and eternal background, through which particles move. Space and time are the stage of intersecting lines sketched in the illustration. Fact is this ‘mathematical artifact’ made with pen and paper by earlier physicists, called the Cartesian graph, useful to measure ‘translation in space’ is no where to be seen in reality. Unfortunately as time went by the graph became somehow ‘real’ as scientists’ felt the ‘mathematical language’ created reality.
It meant also the invention of an absolute ‘continuous space’ and a single ‘lineal time’ that extends to infinity contradicting the obvious fact that all ‘spaces’ are broken, divided by membranes, and all beings have a finite time duration. Further on, as we kept exploring smaller scales of reality, we never found the drawings of God, not even a solid still substance, but always ‘motions’ tracing closed timespace cycles; since even particles turned to be also ‘vortices of timespace motions’.
So the true, sound experimental and logic theory was Leibniz’s who rightly considered absolute space and time an abstraction, and so he coined the concept of relational space merely the adjacent pegging of similar forms in simultaneous space and relational time – the sequence of events which we relate causally with reason origin of the ‘Scalar spacetime’ model of 5D in which are the space we occupy and the time we last – as in the graph where there is no longer abstract background lines.
This realist concept was NOT adopted by physicists despite its sheer evidence. Unfortunately Physicists sided with Newton not with Leibniz on the question of what is space and time – an abstract background put by God or the substance of which we are all made; and so the conceptual jump would not happen.
But if space is what objects occupy that distance between the red square’s vital spade and the yellow ‘circle’ must have something. Horror vacuum comes then into place: indeed the Universe must be scalar. There must be very small parts between them, which we do not see. And that is what we have proved with microscopes – as we probe smaller distances forms with motion, spaces with timemotions appear and there it seems no limit to the fractal scales of the Universe. It is the fifth dimension of spacetime, which as the sum of all those ‘planes of reality’ includes within it all other dimensions.
Next, to explain all this properly came Einstein. One of the fundamental discoveries of Einstein is that in our universe, there is indeed no fixed spacetime background. In Einstein’s theory of general relativity, which replaces Newton’s theory of mechanics and the gravitational force, the geometry of spacetime is not fixed. Instead it is an evolving, dynamical quantity – a topology; and it is the substance of which reality is. So we are topological beings, geometries of space with motions of temporal energy.
What Newton called absolute spacetime is not: Space is the sum of all discontinuous vital spaces occupied by different beings: ∑s=S Lineal time, T the sum of all the finite lifedeath cycles of all beings T=∑t.
Since space and time do exist and so if they are not in the background we ‘are’ vital space and cyclic temporal energy. The simple idea behind the structure of the fractal Universe is then to consider time=change=motion and Topologic, formal space= extension the 2 elements of which all beings are made.
Wheeler said ‘Spacetime tells matter how to move; matter tells spacetime how to curve’. More precisely Spacetime is geometry in motion. Time is change, the perception of change moves time; time is motion; space is its opposite, stillness, form, the information of temporal energy. And so it is all about 2 parameters: time=motion and space=form.
Look around you, all what you see are ‘spaceforms’ with ‘timemotion’. We are all spacetime, forms in motion, ‘informmotion’, ‘information’, forms in action, play with the words of what you are.
Because both are always messed up in practical terms is often easier to measure, ‘forms with a little motion’, we shall call ‘information’, sT; and motions with a little form, we shall call energy, Ts, and talk of beings made of spatial information and temporal energy, as there is no ‘yin (information)’ without a little yang (energy). And call the pure absolute motion without form, TT, entropy; and the absolute form without motion, SS, language. So we can establish a gradation of combined spacetime dualities, SS<St<ST<sT<TT, with a symbol < for an increase of motion over form, or ‘arrow of energetic entropy’, which we shall call ‘past arrow’ or arrow of ‘death’ as it erases information, devolving a system to simpler forms; and vice versa, an inverse arrow, TT>sT>ST>St>SS, with the symbol > of an increase of form over motion, or ‘arrow of information’, which we shall call the ‘future arrow’ or ‘arrow of life’ as it increases information. It then becomes evident that the intermediate state, ST, with a balanced quantity of Spatial information and temporal energy, S=T, is the ‘state of present’, that doesn’t seem to change as it is a balance of form and motion. And it is the preferred state for any system of spacetime in the Universe, akin to the concept of ‘beauty’ (balance between cyclical forms and lineal motions), of reproduction and creative communication (as it brings together the two poles of reality merging and combining them). This state is seek by all systems. So in physics we find it akin to the state of ‘minimal energy’ hence ‘more form’, in which most particles like to remain; in biology is akin to the age between adolescence of maximal energy and the 3^{rd} age of maximal information, or age of reproduction, in which most people like to live and so on.
So we have 5 ‘states of spacetime’ of which we are all made that have topological, formal, geometric and temporal, moving properties. Because those states are messing ‘dimensions of formal space’ and ‘motions of time’, we have coined a new word, ‘Dimotion’ (ab. Ð, which is the capital of ð, similar to D of Dimension and T of time). The main property of a Dimotion is to be holographic, having both spatial dimension and time motion; and we shall soon find that Dimotions are closely related to the 3 classic dimensions of space, to which we must add a function=motion in time, as their study becomes a new discipline of knowledge called topological evolution.
In the graphs above if space is what large objects occupy, what they are, the distance between the red square’s vital space and the yellow ‘circle’s vital space must have something; even if it is tinier. Horror vacuum comes then into place: indeed the Universe must be scalar. There must be very small parts between them, which we do not see. And that is what we have proved with microscopes – as we probe smaller distances forms with motion, spaces with timemotions appear
It is the fifth dimension of spacetime, which Is the sum of all those ‘planes of reality’ coexisting in organic scales, whose form and scalar connection through networks we shall study now constructing reality from its ultimate substances space=form and time=motion, which geometry calls a ‘topologic bidimensional variety of spacetime’.
TOPOLOGICAL SUPŒRGANISMS: THE MAIN PARTICLE OF TIME0SPACE.
“All what is possible, demands to exist” Leibniz, on the chain of beings.
This said the devil is in the details. So the next question is How many variations can come out of just 1+1=3 ‘elements/substances’, cyclical time motions, scalar spatial forms and its spacetime energy combinations?
And the answer provided by topology which studies geometric forms with spatial dimension and time motions is only 3; which perform the 3 organic functions of all systems of vital spacetime of which we are all made: A 4 or 5D Universe has only 3 ‘topological varieties’ each one best suited to perform one the 3 organic vital functions of any physic or biologic system –gauging information (1D spheres, the topology that stores more information in lesser space, hence used in all particles and heads in the height dimension), lineal or cylindrical forms that move the system (2D, the shortest distance between two points, hence used as fields or limbs in the length dimension) and hyperbolic bodywaves, a mixture of the other two topologies that reproduces the system and stores its energy in the width dimension (3D); which are similar to the 3 ‘conserved quantities’ of physics, angular, lineal momentum and energy, but not quite… as it is impossible to translate the ‘game of existence’, into the limited understanding and terminology humans use to describe it, plagued with conceptual errors that limits our use of the correspondence principle. For that reason, after much time wasted in attempting translations we start from scratch with the concepts of TTTsStSS and ST dimotions and its topological, qualitative and organic translation. So we define the ‘Fractal Generator of Vital spacetime topologies for all systems of Nature:
G: $T(limbsfields)<ØSi=TeBodywaves>O§ð (particleheads).
Where @mental Oheads/minds/particles are nonE points crossed by ∞ parallels of information
Topological organisms like those of the graph are the true meaning of a ‘fractal Universe’, and its description with the common laws of time space, the goal of Systems Sciences, realized through the 5D models of those papers: every structure of reality follows certain basic laws, derived of the fact all of them are made of spatial information and temporal motion, combined into infinite energy species. So ‘motion’=Time; Information, form=Space combine together into infinite energy bodies and waves. And this is so obvious that already the old Taoists just looking with simple, naïve eyes to reality said, ‘from yin=information and yang=motion, come all the ∞ qi=energy beings of reality’.
We call the organic properties of scalar space and cyclical time that structure all organisms of reality, ‘Disomorphic laws’ (isomorphic laws – that is, similar laws based in the same dimensional motion of spacetime or ‘dimotions’).
Even languages have a trilogic structure, red(entropy)<Green(energy)>Blue(information); F(x)=G(y), length=entropymotion x height=information = width=interaction, studied latter in the analysis of the Universal grammar, including that of words: Subject (information) <Verb: STaction> Name (entropy of the subject.
But the 3 topological varieties of a 4 or 5 Dimensional spacetime are in reality ‘networks’ of adjacent points. Thus we can instead of drawing the 3 canonical forms of lineal limbs/fields of motions, cyclical, spherical heads/particles and hyperbolic iterative bodywaves, systems made of 3 networks, which correspond to those forms, and that is what we find in nature: supœrganisms in which 3 networks with that form deliver information, energy and motion to its parts:
An organism IS any kind of spacetime systems, in which parts are gathered by networks, forming a new scale, which in praxis are 3, a ‘feeding network’, an informative network or common language, which forms therefore a ‘lanwave’ of similar beings and a combined space+time=energy network that reproduces both, So those are the same 3 dimensional motions that topology describes for a good reason. Modern topology defines its 3 varieties of form as networks of points. So we need to add 2 more dimensional motions to reality, the ∆1 parts that make those networks, and the whole (the 4^{th} and 5^{th} dimensions of scale).
In the graph we show its main species in the Universe, the one we shall study in most papers. The beauty and simplicity of reality the graphs shows contrasts heavily with the pedantic description of it using only the mathematical language, a theme studied in depth when considering the inflationary nature of information.
5D universe. The metric of organic scales: faster, smaller parts code larger energetic wholes.
We said that forms evolve through topological networks into larger wholes. So we need a final dual dimensional motion of parts that become wholes (4^{th} Dimotion of social evolution) and wholes that dissolve into parts (5^{th} entropic dimotion) to complete the organic outlook of the Universe; as shown in the graph for different species.
In the graph the Universe is a fractal that reproduces ‘forms with motion’, informations and then organizes them in networks and systems that evolve into larger organic systems creating the scalar structure of reality.
We call the sum of all those coexisting scales of parts and wholes the fifth dimension.
Thus reality has a final key feature overlooked for too long: the coexistence of all those systems of space and time in several scales of relative size from the smallest atom to the largest galaxy that put together create a dual scalar ‘4^{th} and 5^{th} Dimension of parts and wholes, which we shall call the ‘social dimension of evolution’ and the ‘entropic dimension of dissolution’.
This function, as all spacetime metric functions, is simple. So we write using ð for cyclic time instead of t:
$ (Lineal Size/ Space Volume) x ð (cyclic speed of its time clocks) = C¡: Constant Plane of timespace (ab.∆¡)
But how we travel in ‘size’ in space and ‘speed of our time cycles. Here is where the biggest discovery of 5D comes into play: We travel through the worldcycle of life and death, as we are born in a smaller seed with faster time cycles, evolve as an organism coming out in the ∆ºscale within a larger world of slower ‘Deep time cycles’, to die back dissolving our information again into cellular space.
It is the same process in all 5D journeys of all species that live and die travelling through 3 planes of 5D spacetime; from the smallest black hole that is born with an enormous ‘metabolic temperature’, to the new species, routinely born as small individuals (first mammal rat, first robots with small chips; first human likely the Homo Floresiensis, who had the same morphology and used technology and likely spoke, etc.) Then a reproductive radiation multiplies the seed into a larger herd of clones, joined by emergent physiological networks whose slower ‘entropic, informative and reproductive networks, create an ∆º supœrganism that lives tages and dissolves back into ∆1.
So 5D adds to the 4D formalism of worldlines, a dimension of growth, shaping the worldcycles of life and death. Reason why we call 5D metric the function of existence, because its multiple ‘solutions’ are the origin of all the varieties of Space and Time beings, there are – a whole family of functions.
As we keep exploring in depth, 5D metrics and its associated concepts of Space=form and Time=motion in all its varieties, we shall see it is the origin of multiple ‘solutions’, a whole family of function, from where we shall derive most of the logic relationships and particular equations of each science.
In the complex models of existential illogic, we derive all the particular equations of each science from it.
So according to those metrics, smaller systems in space have faster time clocks. And as information is stored in the frequency and form of those cycles, smaller systems have more information, coding larger ones: genes code cells, memes societies and particles’ quantum numbers code atoms and molecules.
We shall use the metrics of the 5^{th} scalar dimension to explain the fractal, nested Universe and its scales, shown in the graph. As 5D metrics balances the survival and symbiotic existence of all parts of the Universe, and all parts of a super organism, and defines ‘what codes information’ – the small being, and what codes energy the larger whole, establishing the ‘harmony’ of all the scales of the Universe, and explaining all its fundamental constants which are ratios between spatial volumes and informative clocks of temporal energy.
It follows from a nested structure and the search for creative, organic balances, a symbiotic relationship between the ∆¡ smaller parts that have more speed of time clocks, which carry its information in the form and frequency of its cycles, coding larger systems. And the larger, ∆+¡ larger envelopes, membranes (static, dimensional view) or angular momentums (dynamic view as time=motions) which have more spatial energy and enclose and control in synchronicity its faster smaller parts, creating the coexisting scales and symbiotic cycles of super organisms in any system of the Universe.
For example, chips become smaller as they evolve into faster brains. Every 2 years a chip doubles its capacity to think, as it dwindles in size. Such process follows a generic law of evolution I call the ‘Black hole Law’, which computer scientists know as the ‘Chip paradox’ or ‘Moore Law’: maximal informative capacity= minimal spatial extension. The reason is obvious: to think, to calculate you have to communicate information, forms between elements of any informative system. The smaller the brain, the faster the communication that takes place within that brain and the faster you can calculate and process information in a logic manner. And vice versa: larger wholes accumulate more energy and are stronger than parts, so they can protect and feed them. So wholes and parts coexist in several scales forming super organisms. Since organic reality arises of the synchronicities of those parts and whole made symbiotic thanks to the simple metric of the 5^{th} dimension and its homology as spacetime beings. So the addition of topology and the simple metric laws of 5D that make smaller system run faster time cycles are all the elements we need to construct reality.
Latter we shall study those regions, evident in physical equations as there are asymptotic barriers – Lorentz transformations in the cbarrier; negative temperature in the 0 barrier; etc. What then the Universe conserves is easy to see: the total volume of spacetime of each scale; that is its energy. We will also elaborate latter on those concepts. To mention now that the symbol ∆ is both a visual reminder of the two different arrows of ‘growth in space, inverse to the loss of information’, and a tribute to one of the few predecessors of this work, in the formal arena – Wilson’s renormalization mathematical apparatus, which finally realized of those discontinuities using a symbol Λ, for the energy scale under which a measurement of a physical parameter is performed. According to Wilson every scale of the Universe and the fields of spacetime that define them have its energy cutoff Λ, i.e. the theory is no longer valid at energies higher than Λ, and all degrees of freedom above the scale Λ are to be omitted. But Λ is related to a size of space. For example, the cutoff could be the inverse of the atomic spacing in a condensed matter system, and in elementary particle physics it could be associated with the fundamental “graininess” of spacetime caused by quantum fluctuations in gravity. The failure to remove the cutoff Λ from calculations in such a theory merely indicates that new physical phenomena appear at scales above Λ, where a new theory is necessary. As today only with the use of Wilson’s renormalization, which simply eliminates absolute zeros and infinities outside the scaling of spacetime of a given plane, quantum physics makes sense
Recap. Besides S and T we need ∆, the scalar property of reality, by which ‘spacetime beings’ are made of smaller ones and form part of larger wholes. We talk of multiple planes of spacetime because each scale has different parameters, as parts become wholes, unit of a next scale. We will explain this as the product of network formation, emergence and other disciplines of General Systems Sciences, of which 5D is a formal view. In mathematics ‘fractal points will evolve’ through networklines into topological planes of spacetime. This discontinuity between scales is real. There are transition regions between planes, which can only be crossed with loss of information, therefore only by energy, by entropic motion, by ‘death’ of a system.
So we unify the Universe, as a fractal whole whose parts are made to its image and likeness: entities of timespace, with a vital body and a linguistic mind/brain/particle of information surrounded by a membrane that separates them from other similar parts that put together create the puzzle of the Universe. This is the biologic concept behind all realities: The supœrganism that combines particles/heads of space=information and limbs/fields of time=motion into iterative cycles of spacetime energy, giving birth to the causal trinity logic of reality, to which we add the logic of smaller parts and larger wholes to form the pentalogic of 5D existence; the next step in the complex comprehension of reality, which will allow us to connect those supœrganisms of scalar spacetime and its networks with the ‘details’ of each science.
Yet before we do so, because the Universe of spacetime is entangled, we need to consider if briefly the quality of time missed in most scientific analysis. The fact that time motion can be both lineal cyclical, and a combination of both, as space forms can be also of 3 varieties; but scientists only consider lineal time and lineal space, deforming enormously their comprehension of the Universe. Since it is of the interaction of those 2 forms of space and motions of time, from where the proper formulation of the principles of conservation of energy, its combination happens.
TIME CYCLES
The causal repetitive laws of ‘stiences’
A Universe of ∞ time clocks of different size & speed differs from lineal time described with a single mechanical clock which equalize all time clocks of the universe, elongated into a lineal ‘secondminutehourdayyear’ system of equalized time clocks (of light waves, mechanical clocks, earth’s astronomical clocks). As Galilean physics, born of ballistics, simplified the nature of cycles of timespace into lineal durations, to measure best the locomotions of cannonballs: Time is cyclical, all clocks of time and laws of science are based in the cyclical patterns of nature. But physicists developed ballistics and denied the truth that we can know the future because it will repeat the causality of the past, and we can change it by changing that causality, in History by repressing the lethal memes of the tree of metal and enhance the welfare memes that make us survive.
Lineal and cyclical time render the same functions as one is the inverse of the other, measured by frequency, T=1/ƒ, but the philosophical implications of cyclical time, are enormous, as we regain the information provided by those cycles, origin of the laws of science, which would not exist if there were not cyclical patterns; including the cycles of history and economics. The most important of them being, the fact that a time cycle breaks reality (1^{st} knot theorem) in an outer and inner region, creating a membrane that encloses a vital space, the ‘substance of which we are all made’.
Reality is a fractal system made of topological organisms of coexisting scales of space and cyclical time which close its ‘internal vital point content’ with the entropic limit of those time cycles, in its vital territorial bodywaves, synchronized symbiotically by 5D metrics. As we are all ultimately ¬∆@St; dust of spacetime.
Why there are 2 forms of time, the long lineal Time and the ‘short’ frequency steps we integrate into the larger whole? Because there are 2 ±¡ scales of 5D reality whose metric, SxT=∆±¡ defines larger space systems as having slower time cycles. So we consider an ∆¡ ‘quanta of time frequency or ‘finitesimal derivative’ of the larger whole represented with the concept of lineal time; as in the classic formula,V(st)=ƒ(t) l(s). We can measure Space e, Vt=S with lineal time as a single unit, or as a sum of frequency steps, with more detail.
Those 2 forms of motion are lineal motion with a bit of form, Ts, or locomotion and cyclical form with a bit of motion, Stinformation, stored in the frequency and form of its time cycles that come together into S=T, energy. So we express the main law of science, the principle of conservation of energy in terms of the conservation of time=motion and space=form as two varieties that approach and transform each other ad eternal:
‘All what exists are time motions that transform between lineal open and cyclic closed forms ad eternal: SióTe’
Reality is a constant game of transformation of ‘cyclical spatial form’ and lineal timemotion. As Taoism said ‘tao=reality’, is composed of yin=cyclical form and yang=lineal entropy, whose generator, Si≤=≥Te we call the function of existence, as To exi=st is to combine S&T, pure form and motion, TTentropy and SSeeds of form, into St & Ts, information and kinetic energy, exi, till finally they become one. The knowledge of that game, expressed in infinite variations, both of language and species, is the mind of the Universe, what is all about.
.
The many mirrors calculus puts on 5D.
It is then evident that the Universe we have described is amenable to the methods of calculus, which fit like a glove the concepts analyzed above. Let us see how starting by the essential element of reality, reproduction.
REPRODUCTION OF FORM IN 5D AND ITS MATHEMATICAL MIRROR: CALCULUS.
The Universe is a ∆fractal of spatial information and temporal energy, (ab. ∆ST) that reproduces information, formsinaction, forms of space with motions in time.
But reproduction has an essential feature: it happens in a lower plane as a seed that reproduces, integrates its parts and evolves into a whole – the exact method we use in the mathematical discipline of calculus.
As all this is what actually calculus calculates: It finds a finitesimal part of reality and then integrates it as a sum, through a path that might be a motion in space, a growth in ∆scale or a repetition in ‘time; whereas the function of existence of the form displaces and reproduces its orthogonal parameters of form and motion. So physical forms are constantly reproducing, ‘calculating’ and the equivalence between the tools of calculus as mirror of the process of reproductive locomotion become crystal clear.
Let us then from the general texts on 5D bring here only the specific analysis of relativity of motion and form, with the 3 features that are essential to the process of calculus:
– The symmetry between spaceform (height dimension) and timemotion (lengthdimension), or S=T mimicked in calculus by the orthogonal smooth form of ∂x/∂y, and its differentials.
– The finitesimal nature of change=motion in time, as it happens from minimal parts in a lower scale of the fifth dimension.
– The reproductive nature of all motion, as the reproduction of information in an ∆1 wave state then integrated, as it collapses in its particle form.
Reproduction as the origin of the 5 Dimotions of existence.
How many types of reproduction there are, is a complex subject that would require a whole treatise in existential algebra published elsewhere on those texts.
Let us then define the most important for calculus, according to the general method of existential algebra that distinguish always 3±¡ possibilities:
∆ Social reproduction: A reproduction might be persistent in time, creating a process of scalar social wholes; when the reproduction ‘lasts memorially in time’, larger social wholes, ∆1 ∑∆1 > ∆^{0} (using symbols of existential algebra).
Ts: Locomotion & Lineal Inertia. If the reproduction fades away at the same rate it happens, we observe then a locomotion, through spacetime.
St: Angular momentum: If the reproduction doesn’t move in space, we observe a cycle of spacetime, which can be equal to the previous cycle.
SSvortex motion: In the reproduction shrinks in space size and increases in cyclical time speed (according to 5D metrics: S x T=C), we talk of growth of informative frequency, travelling to a smaller scale of the fifth dimension.
TT: if the reproduction grow in size tracing a +π cycle, we talk of an entropic reproduction that slows down the motion of the system.
Those are the 5 essential forms of reproduction in ∆ST, which correspond to 5 Dimotions of spacetime and are studied by the functions of calculus in physical systems. All of them however will be reproductions that happen through small Si=Te steps of motion and form. So we have next to understand how reproductive motions happen through Stops and motions, particles and wave states, finitesimal after finitesimal from the perspective of the whole – as a minimal 1/n angular curvature in curved ‘space’ motions, as a minimal l ƒrequency=1/T step in lineal time wave motions as a 1/n minimal cell in scalar motions. They are the 3 ∆ST finitesimal steps in all calculus which as we saw have always the same 1/n formula, either representing curvature of spacetime, frequency of time, or population in space. How the 3 ‘concepts’ are actually symmetric, is due to the…
Galilean Paradox: SóT: Relativity of space Dimensions=Forms=Motion in time: 5 Universal Dimotions
Galileo’s time and space Principle of Relativity is the fundamental conceptual thought behind the relationship between time=motion and space=form and how one can be converted into another: All what exists is made of space=form and time=motion. And yet physicists know that we cannot distinguish motion from form. That any being in motion from its point of view seems to be still and all other things moving around it. This is the principle of Relativity of motion.
Physicists then without much thought about that fascinating duality, went on to use mathematics to calculate the relative motion of each entity of reality respect to other system, which seems static from both points of view. This is called Galilean relativity, latter refined by Einstein’s relativity, and essentially is concerned with the mathematical calculus of what we shall call the 2nd Dimotion of time=change, locomotion. Fine, but we are more interested on the duality of space=form and motion=time and its entangled relationships –the reasons why we do NOT see together motion and form, even if all systems have both.
The conclusion is then rather obvious: one of the two parameters of reality is ‘hidden’ to perception; we either see motion or form, ‘waves or particles’ (quantum complementarity), distances and lines or points in motion (as in the night when fast cars in a picture appear as lines). So physicists calculate only one when in fact we must assess the existence of 2; and since we cannot distinguish them, logically we must equal them. ‘Form=motionfunction; space=time; Si=Te’.
Relativity then becomes a duality, Si=Te, which is at the heart of every law of the Universe. Whereas the primary element, the ultimate substance is time=motion. As space is a Maya of the senses – a slice of time motion. Form is what a ‘still mind’, makes of that motion to ‘perceive’, information, formsinaction.
Since we see Earth still and flat but it is round and moving. Galileo’s profession was ballistics – the study of cannonballs motion. So he chose ONLY motion and lost the chance to start physics with a complex philosophical understanding of its Si=Te dual Principle of relativity, which Poincare defined latter clearly when he said that ‘we cannot distinguish motion from stillness’. An example is quantum/relativity duality. In detail quantum space has ‘dark energy’ because it has expansive motion that extends into a plane of space, but when seen at larger scales without detail its entropic motion seems static space – a dual area of scattering length and width. So in the galaxy we see either dark energy motion or expanding space: T=S. A motion of time is equivalent to a dimension of space: Distance and motion cannot be distinguished so they must be taken as two side of the same being, a space=time Ðimotion (ab. Dimensional Motion):
S= T; DimensionDistance = Timemotion = ST Ðimotion
Earth moves in time, but we see it as a still form in space because reality is a constant game of ∞ motions, but the mind focus those motions and measures them at still distances. For huminds, motion is relative to our systems of measure and perception, which are lightbased; hence a fixed crod speed/ distance. Reason why Einstein’s relativity postulates a maximal T:cspeed, measured as if observer and observable were still to each other (Constant S); which at our scale we correct with Lorentz Transformations.
As it happens the identity between spatial states of ‘form’ and temporal states of ‘motion’, which become stops and steps of all reproductive motions become the fundamental ‘present state’ of the Universe, and the essential tool of calculus to ‘solve’ its differential equation (D’alambert’s method of separation of variables); and his unending philosophical and logicomathematical consequences will appear in many parts of those texts.
But physicists just substitutes the Earth’s still distances for motions, and it took another 300 years for Einstein to realize the relativity of motion and its measure made essentially time and space, motion and form two sides of the same coin. Still this realization was not explored philosophically and so it gave birth to a series of illunderstood dualities between ‘states of measure and form’ (particles, head gauging form, information) and ‘states of motion’ (wave states).
It is then essential to grasp that motion and form coexist as 2 different states depending on 5D scale and detail: Motions are perceived by minds that stop motion into form, into information, as distances. So if we see slow motion in the night, a car’s headlight seems a long distance line ‘still’ picture. But this means also that the 3 ‘Euclidean still dimensions’ must have motion; they are ‘bidimensional STholographic, topologic dimotions’. So we have 3 Space + 1 Time + 1 5^{th} dimension of scales = 5 Dimensional motions. None of them is a Dimension of pure spatial form or a pure time motion but a combination of both. Even if mentally we tend to reduce motion and focus on forms, all has motion=time, and form =space: this is the meaning of ‘spacetime’, the messing of both into 5 dimotions, the fundamental element of all realities.
Relativity states ‘we cannot distinguish motion=time from position=space’. So all what exists is a composite of both, undistinguishable Si=Te, 5 ‘Dimensional motions’ (Ab. Dimotions), broken in infinite fractal, vital time space organisms, composed of topological Dimotions: height=information; length=locomotion; width=reproduction; form=social evolution of parts into wholes & entropy=dissolution of a whole into its parts in a lower scale of the fifth dimension (term we keep for the whole range of scales of the Universe); whose study is both mathematical, the main science that studies how those 5 Dimotions entangle in simultaneous Space, connected to each other topological adjacent parts, which create superorganism, and Logic; the main stience of time that observes how those pentalogic, entangled superorganisms move and evolve, change in sequential relational time, living a worldcycle of life and death.
As all is time&space, the 2 experimental primary mirrorstiences of time&space become the most important to extract the Disomorphic=equal laws of those 5 Dimotions that all systems have in common. Since while those Dimotions are broken, in vital organisms, separated by cyclical time membranes, they are the same.
In the graph Galilean relativity was ill understood, as the true question about timechange is why ‘the mind sees space as a still, when in detail is made of smaller selfsimilar quanta, in motion. The paradox defines mental spaces as still simplified views of the more complex whole.
The 3 ¡logic paradoxes of space topology (closed informative curvedO vs. open, free entropic lineal forms), timemotion (stillness vs. motion) and ∆scale, (continuous whole vs. discrete forms; single scale vs. multiple one)s, are essential to the perception of a simplified ‘spatial mind universe’ in a single flat still plane vs. the full, more detailed complex picture in time, of a curved, discrete and moving Universe. Those paradoxes resume the 5 elements of reality, Space=form, time=motion, scales and the mind that measures them, within its own entropic limits.
They are also essential to all the elements of calculus and mathematics at large and its methods of solutions; specially the inversion between finitesimal lineal steps (as a step between two points is NEVER curved) and the cyclical form of longer ‘integral paths’. So lineal approximations are the essential tool of calculus and mathematics to resolve many equations.
What neither mathematicians nor physicists fully understand (though some inroads in abstract were made through the Noether’s concepts of symmetry) is that each stœp of a method of solution is not ‘gratuitous’; but must be grounded in a real property of the 5D ∆ST symmetries and conservation laws of the Universe, which are not so many – hence the repetition of methods. Specifically, the aforementioned 3 paradoxes between ∆+1 curved closed worldcycles, sum of lineal steps, which gives birth to the most used method of lineal approximations; the equivalence between Space and time, in all Stœps of dimotions, which gives birth to the method of separation of variables on differential equations and more broadly allows to move around relative space and time parameters in equations joined by an operand of ‘equivalence’ (≈ not =). And the 2 conservation laws of the Universe, conservation of those ‘beats’ of existence, S=T in relative present, eternal balance, justifying the equivalence operands. And conservation of the ‘volume of spacetime’ of each plane of the Universe, by virtue of the 5D metric equation SxT=C, which justifies all the procedures regarding scales – solution of differential equations by separations of scales, renormalization procedures (Wilson), and harmonizes those scales allowing constant but balanced transfers of energy and information, St=Ts.
5d metrics expresses the conservation of time.
The paradoxes of Relativity, discontinuity, parts and wholes, scales are all related to the reductionist nature of minds that bias reality. Minds reduce dimensions to the relevant ones, eliminating all dark spaces: continuity is the result. Of all formal languages that map out reality 2 are paramount, Time ¡logic & mathematics of Scalar Spatial information.
A 5D Metric function, S(0Mind) x T(∞universe)=constant world is the function of all mind languages who only perceive from its selfcentered point its language mirror confused with the whole Universe (Ego paradox, basis of psychology). Ænthropic huminds reduce the multiple clocks of time and vital spaces of reality to the single human clock and spatial scale, rejecting the organic properties of other Universal systems.
The main laws of 5D are the metric functions of the scalar Universe, which relate the spatial size and speed of temporal clocks of all scales of Nature. Both parameters are inverted: when systems grow in size the speed of its clocks, its ‘time cycles’, diminish proportionally, both in biological and physical systems. And vice versa. Smaller clocks tick faster and information processing carried by the frequency of those cycles accelerates, as it happens in chips, particles or life metabolism. So we write: S x T= C.
The mind thus starts it all with its linguistic ‘still mapping’ stopping its world in a locked ‘crystal image’, measure of its self. But even perception is social, linguistic. The Universe can only be explained if ‘perception’ exists within the language, as when you think words, you sense words, when your eye sees light and maps into an electronic mapping you are seeing. And when an atom maps a geometric image in its ‘locked’ ‘stopped’ spin, it must perceive that geometry as information.
Physicists made the Galileo’s paradox, the cornerstone of their theory of measure, but they failed to study the deep implications it has for every aspect of the structure of the Universe, from the duality between spatial mental, linguistic forms and physical motions; to the balances achieved by the similarity of both space and time, which becomes the fundamental ‘function of present’ Si=Te, and hence with the metric function of scales, $ x ð = K, the two essential functions to formalize single planes Si=Te, and multiple scales of spacetime. Yet as Si=Te maximizes SxT=K (5×5>6×4). We unify both in 1 function:
Max. S x T = C, which defines for each fractal vital spacetime organism its Function of existence, as all species will try to maximize its motionentropytime for its fieldlimbs, its informationspatial states for its particleheads, whose product will give us its vital reproductive energy. Moreover the function has an immediate biologic meaning, because as we are made topologically of ‘fieldslimbs’ of lineal space with motion provided by the energy we absorb to also reproduce our bodieswaves, and the information we need to linguistically guide our motions with particleheads, the very essence of survival is to increase our S=position, mental forms of space and T=entropic motions of time (whereas time=motion & space=form are the two limiting Dimotions with ‘energy=reproduction, s=t, locomotion, sT and information, St, are the intermediate 3 dimotions).
The fifth dimension is made of the ‘different coexisting scales’, which from the simplest forces through particles, atoms, molecules, matter, organisms, super organisms, planetary systems and galaxies, create an ‘organic network structure’, which amazing enough since it was discovered at the beginning of science with telescope and microscopes, was not formalized till I introduce its metric function in the milieu of systems sciences, as a single lineal time motion is a dogma physicists don’t dare to challenge. Yet science cannot advance in its fundamental principles unless the formalism of the fifth dimension is accepted and used to fully understand the cyclical, repetitive patterns=laws of science of each discipline that studies a scale of the fifth dimension and its species.
Reproduction of form in 5D and its essential mathematical tool: calculus.
The Universe is a fractal that reproduces information, formsinaction, forms of space with motions in time. This is the essence of it all. But space is a maya of the senses, the synchronous view of a series of cycles of time motions, knotted in the simultaneous perception of an observer; what physicists call a ‘frame of reference’.
Thus time=change is the fundamental element of reality, and this makes Algebra of timechange, specifically calculus perhaps the most important experimental science of time, besides logic, which we have upgraded to existential algebra, which explores the vital, organic whys of those changes.
It is the Galilean Paradox: S=T. We cannot distinguish time from form. In as much as each frame of reference or mind locks in a knotmirror of the motions of the Universe from its point of view. So each point of space is a perceiver relative field of motions, which from its perspective knot as forces ‘attracted’ by its frame of reference. Yet if we cannot distinguish motion from form each point is entangled to those motions and is made of motion and form, of the particle and wave states.
Locomotion as reproduction of form solves the Paradoxes of Zeno and the meaning of discontinuity. As motion is reproduction of information, of form, since particles are knots of perception of form, fractal points, monads, that move by reproducing in a lower 5D plane, as ∆1 waves, its information, as formsinaction.
So all forms of change can be reduced to the ultimate function of existence, reproduction, a back and forth travel through 2 scales of the fifth dimension, as a form becomes a seed that reproduces, evolves socially and forms its whole again. The extraordinary capacity of Calculus, which extracts at ∆1 level a ‘finitesimal’ (Leibniz’s 1/n definition of infonitesimal as a minimal part of a whole and ALSO, by virtue of S=T, a minimal ‘curvature’ of a time cycle, which is then integrated for a time duration of the event, either locomotion, or volume of population in space or S=T continuous=smooth change in time happens precisely because CALCULUS perfectly mimics the process of change and reproduction of form between ∆º and ∆1 scale which is the basis of all timechange also in physics. Change thus is change reproduced in a lower plane as a seed that evolves into a whole.
It is then not so much in physics but in calculus where we find the strongest model of the laws of 5D and locomotion as a reproductive process of form, even if the experimental proofs are scattered all over physics. Indeed, the entire world of quantum physics can only make sense if we consider that particles MOVE AS WAVES and gauge information as stop particles. Because waves can be transparent to each other but particles collide. A simple proof: the atomic nucleus is so small compared to its particles that if they wouldn’t move as waves, transparent to each other, they would be always colliding and the nucleus would never remain stable. In fact, when we get pictures of those particles outside its shells, (electrons) they move in zig zag as they stop and change motion constantly. As usual physicists just make an axiomatic rule and subvert the law of causality converting the mathematical mirror derived of the fact in the cause of the fact – in this case they say this is due to the Pauli exclusion principle without providing the mechanism for particles to avoid collision if moving.
It follows that beings with more information, reproduce slowly and we can hardly see them moving. The limit of it being complex life superorganims on Earth, whose reproduction takes 9 months. It happens ‘inside’ the reproductive mother, and it reproduces in the adjacent space after ‘tearing’ the topological knot of the umbilical chord. A similar very slow process of reproduction happens in physics with the weak interaction that reproduces a form with even more information evolving the mass of particles, so the range of the force is minimal and the new particle appears adjacent to the one that disappears, dying for the new hatched ‘baby’ to be born.
This is the essence of it all. Motion is reproduction of information, of form. Since particles are knots of perception of form, fractal points, monads, which move by reproducing through a lower plane of the 5^{th} dimension, as ∆1 waves, its information, as formsinaction; all forms of change can be reduced to the ultimate function of existence, reproduction, a back and forth travel through 2 scales of the fifth dimension, as a form becomes a seed that reproduces, evolves socially and forms its whole again.
Space is a Maya of the senses, the synchronous view of a series of cycles of time motions, knotted in the simultaneous perception of an observer; what physicists call a ‘frame of reference’.
Thus time=change is the fundamental element of reality, and this makes Algebra of timechange, specifically calculus perhaps the most important experimental science of time, besides logic, which we have upgraded to existential algebra, which explores the vital, organic whys of those changes.
It is the Galilean Paradox: S=T. We cannot distinguish time from form. In as much as each frame of reference or mind locks in a knotmirror of the motions of the Universe from its point of view. So each point of space is a perceiver relative field of motions, which from its perspective knot as forces ‘attracted’ by its frame of reference. Yet if we cannot distinguish motion from form each point is entangled to those motions and is made of motion and form, of the particle and wave states.
Thus systems reproduce its form, travelling across scales of the fifth dimension: they reproduce a finitesimal form creating a reproductive wave, which integrated as a population of space give us back a whole.
Such discontinuous locomotion solves Zeno Paradoxes as the finitesimal is the limit of one ‘step’.
RECAP. Calculus study functions of existence: it extracts finitesimals and integrate them as a reproductive wave.
hence the enormous value of calculus to reflect mathematically the laws of existential algebra.
The fundamental paradoxes of relativity (S=T) become then the ‘backbone’ that justifies the methods of calculus and differentials, as each ‘slice’ of integral calculus is ultimately an S=T stop and go process. Reality thus is the exhaustion method of calculus down to a finitesimal – creationism and the inflationary properties of languages unfortunately became the ‘standard’ justification as finitesimal became infinitesimals, discontinuous stop and go processes of reproduction of locomotion, a continuous ‘flow’ and so 0’ became 0, which does NOT exist. 0 is indeed the definition of no existence; 0’, the minimal existential quanta of any system of spacetime.
TRILOGIC ON CALCULUS. CURVATURE OF SPACE; CHANGE IN TIME; FINITESIMASL IN SCALE.
The enormous advantage of algebraic dimotions and calculus over all other forms of study of motion, including physics which can be considered basically the application of the mathematics of change to the study of nature is the fact that it can study all the elements and dimotions of the Universe from the ‘mind’s perspective’ – that is to study timemotions, spacechange (volumes, lines measures), scalar change and the entropic limits of reality. Let us briefly introduce those 3∆¡ problems, which in fact gave origin to calculus.
Time=Motion=Speed.
XVII C. science was concerned with problems of motion. Copernicus introduced the concept of Earth rotating on its axis, revolving around the sun. The earlier theory of planetary motion, which presupposed an earth absolutely fixed in space in the center of the universe, was discarded. The theory involving an earth in motion invalidated the laws and explanations of motion that had been accepted since Greek times. New insights were needed to the question of why objects stay with the moving Earth seemed called for. All of these motions—those of objects near the surface of the earth and those of the heavenly bodies—take place with variable velocity, and many involve variable acceleration. But the branches of mathematics that existed before calculus was created were not adequate to treat them. So a method was required and that started up calculus.
Space= form=curvature.
The 2^{nd} major problem of XVII C. mathematical physics was the determination of tangents to various curves. Its deeper significance is that the tangent to a curve at a point represents the direction of the curve at the point, as small steps are lineal, open free, but the long term motion closes into itself. This key element of ‘scalar time’, which makes easier to predict longer lifedeath cycle and curved trajectories is the key to the interplay between small scale lineal tangent points and large scale.
Its practical use was to find out the best angle for the motion of a projectile shot from a cannon. Since, if a projectile moves along a curve, the direction in which the projectile is headed at any point on its path is the direction of the tangent at that point. The invention of the telescope and microscope also stimulated great interest in the action of lenses. To determine the course of a light ray after it strikes the surface of a lens, we must know the angle that the light ray makes with the lens, that is, the angle between the light ray and the tangent to the lens. So the study of the behavior of light was, next to the study of motion, the most active scientific field in that century, the question of finding the tangent to a curve was a major one.
¬ Entropic Limits and STreproductive maximal
A 3^{rd} class of problems besetting XVII C. scientists was maxima and minima. The motion of cannon balls obsessed Galileo, the weapons master of the Venetian arsenal, seeking the determination of the maximum range. As the angle of elevation of a cannon is varied, the range—that is, the horizontal distance from the cannon to the point at which the projectile again reaches the ground—also varies. The question is, at what angle of elevation is the range a maximum? Another maximum and minimum problem arises in planetary motion. As a planet moves about the sun, its distance from the sun varies. What are then the maximum and minimum distances?
0’finitesimal scales: All those problems required calculus, which was based in the finding of a finitesimal quanta of Time (1/T), space (1/R) or scale (1/N). Then considering the system suffered a ‘deterministic’ rate of change, defined by a mathematical function (that introduced the specific type of rate often associated to each of the 5 Dimotions of the Universe, St, Ts, ST, SS and TT), the scientist could ‘extend’ that rate of change into a ‘sum of spacetime stœps’, which gave us a ‘lineal period in time, or continuous surface in space’, over which the longer time period was calculated. The ‘magic outcome’ (as this simple foundations of calculus were messed by creationsit egocy) thus solved the essential problems of mathematics.
THE FINITESIMAL: Þ
∆1: Lebiniz’s definition of S=T Finitesimals: 1/n: minimal curvature. ∆1 unit.
The key concept of 5D calculus is a finitesimal. A finitesimal in lineal spacetime is a frequency step or wavelength. A finitesimal in curved spacetime is a minimal curvature of a clock cycle. A finitesimal in scale is a minimal unit of population.
But we use other term for any finitesimal 0’; that is, a bit more than 0’, which can be either curvature, ¡1 unit of population or frequency motion. 0’ is then the mental finitesimal – the minimal quantity in existence of a being, which still remains the being; the seed, the mind of the species, the mothercell that must therefore exist for any being, as the template which will develop the immanent program of exi»st¡ence, which does NOT need to be stored within the finItesimal 0’.
And inversely as the reciprocal of SSminds with no motion are TTentropy with maximal motion; and the reciprocal of zero is infinity, and almost 0’, immensity, which we write with the symbol ∝. The finitesimal of entropy is the largest domain of the being, beyond which the being dies. It is a real definition of the borders of the mind, as in its equation, 0’mind x ∝ spacetime cycles of the Universe = K World; 0’ x ∝ = K.
We thus talk of immensity as the entropic limit in which the being no longer is. And both obviously act in calculus as the limits of a definite integral. Thus we can define calculus with the 5 Dimotions of existence. Since the finitesimals of ∆ST will be integrated between its two limits of SS an TT, to give us its whole ‘worldcycle’ in time, or ‘closed circle’ in space, sum of its ‘stœps’ or ‘curvatures’, or its ‘wholeness’ in scale, sum of its finitesimal parts, showing the deep entanglement and symmetry between ∆, S and T:
∆st.
We shall use as windows does not let me put ∝, œ as the best symbol for immensity, which is the symbol of the whole superorganism, that is ultimately an alternative symbol for ∝, as the whole tends to be the limit of existence of the being, more exactly its world, O’ x ∝ = œ.
The subtle difference being that ∝ is external to the being that perceives it, but has pure entropy; that is potentially feeds the creation of multiple Kaleidoscopic monadworlds, and œ is specific, ∝ has been ordered to become a ‘whole:ab. œ’.
But let’s not fancy too much ourselves with existential algebra, its profound paradoxes and symbols, returning to classic calculus.
I propose then 3 alternative symbols, L, for the classic limit as the finitesimal of space, ƒ, for Finitesimal in time, as frequency, and þ as the symbol of a palingenetic cycle, for finitesimal of scale, and will be using ƒ for commodity, with þ as the less confusing symbol for all the cases. And write as the general formula:
Þ:1/¡, whereas ¡, might be in classic mathematics, N, the whole population, T, the period or R the radius:
Þ(∆)=1/n; þ(S)=1/R=K; þ(T)=1/T=ƒ; Þ(@)=0’; Þ(¬)=∝
It is fascinating to observe that the 3 finitesimals of scale, space and time have the same equation in classic mathematics, 2 of them discovered by Leibniz; the guy who unlike Newton always ‘hit a target nobody sees’ (:
This of course is only the beginning; and as usual we shall pounder more the philosophical aspects of 5D calculus, leaving for ‘pros’ with imagination, a humble realization that new beginnings are simple but always found by amateurs without the burden of knowledge that an entire new world of calculus of which I have just swimmed on the surface with unfocused diving glasses, awaits to the brave.
What is the 1 in the equations of finitesimals, a whole or a stœp
Thus the infinitesimal does not exist – being space quantic, there will be always a limit, a microcycle of time or quanta of population in space, to signify the finitesimal point, as Leibniz rightly understood and defined it with a simple powerful form: 1/n.
Indeed in the Universe finitesimals tend to be structured as in a russian doll, such as the biggest wholes, n> have the smallest finitesimals, 1/n>0’. But and this is the incredible magic insight of Leibniz’s ¡n:finitesimal, 1/n is also the formula for a curvature in space. And as S=T, for the minimal motion of a clock of time. So we do have a concept in ∆ST¡1 that we shall then ‘integrate’ through a relative path with the finite limit of a worldcycle, where the function is meaningful (that is has a value, ƒ(x,y)≥1/n), and relates ∆ and ∆1 through its ‘stœps’ of change. We does connect in this manner calculus to 5D reality, no longer base in human invented, axiomatic concepts of absolute zeros and infinities, limits and the paradoxes enclosed within them.
The 0’ size is thus the finitesimal. In praxis, we humans only observe a finitesimal from our mind perspective, whose minimal form is an h’ quantum of the planck scale, and accordingly we see a Universe of inverse relative size, being humans in the ∆º middle view (at cellular level) as physicists wonder without realizing this is NOT a coincidence, but a natural law of the scalar, fractal organic structure of the Universe:
So we accept Leibniz’s concept of a finitesimal, as ALL organic systems have a minimal cellular quanta and a maximal enclosure, which in mathematics can be represented in the o1 finitesimal circle, closed above, as it becomes the 1 element in ∆1 of the ∆º whole, which is represented by the 1 equivalent graph, which is opened above into the wholeness of a larger Universe (but will have also a limit normally in the decametric logarithmic scale of the ∆º whole world embedded in the ∆+1 truly infinite Universe).
What the 3±¡ finitesimals of existence have is the 1 on ‘top’? As we can consider 0’ 1/∝ and ∝, 1/0.
And it can mean two things, the whole as 1=∝ in the 01 palingenetic Universe or the 0’ as the finitesimal in the 1∝ Cartesian domain. So we find that immensity can be a finitesimal.
If 1 is taken as the finitesimal, it becomes then a step in a curvature, which tends to be ‘lineal’, as all steps are discontinuous motions between two points, which can always be closed with a straight line.
So by definition the minimal curvature step is always a line of infinite curvature (: And we need two steps to find a ‘real curvature’, whose maximal value, for a step back and forth will be 360º.
If 1 is the finitesimal of time, frequency, it will be the minimal event, hence a closed time cycle. And as such it will be the sum of finitesimal curvature steps.
If 1 is the finitesimal of a population, then it will be its minimal meaningful part, often a ‘seed’ or mindsingularity, and n its whole population, and the smaller the finitesimal 1 is the larger the npopulation will be in the nested Universe, when we measure ∆2 finitesimal ‘bites’ of energyfeeding.
But for a finitesimal to persist as a unit of population it must not be erased after a single ‘stœp’; so its cycle must be repeated in time.
So we realize there is a chain relationship between the 3 finitesimals such as its reciprocal, a bit of space =form, a beat of time=cyclical motion and a bite (a piece) of population are nested parts of larger wholes:
∑∑s = ∑T= ∆1.
A deep result in both calculus and existential algebra, which can be said as follows:
Space is a slice of time which is a slice of ∆planes, and so we grow in dimensional motions and wholeness when we move from space/curvature steps to fulfill a whole time cycle, which however is just a frequency of memoriless form, that only when persists by repeating its cycles in the same region of spacetime becomes the unit of population.
This growth of reality is essential to grasp the complex nature of calculus when we move beyond the first pages dedicated to the analysis of its dimotions and operands, to ODEs and PDEs of physical systems and beyond; which are also nested systems of complex dimotions in which finitesimals of scale, time and space are considered all together.
Unfortunately humind’s unaware of those symmetries, and even the simplest concepts of linearity, cyclicality and scale just ‘calculate’ as if they were performing some magical trick; so our purpose will be to enlighten philosophically its calculus extracting general laws of reality from calculus, as calculus is by far the closest formal language humans have learned that mimics the laws of the Universe.
Indeed, essentially what a calculus operation does in its essential form, an ODE or PDE is to find the closest thing to a finitesimal, which is a differential (as it is a o’ piece or ‘lineal step’ the closest possible as a piece of space to the curvature piece around the derivative=tangent); and then it integrates it in an interval of time; between its ‘original seed’ and ‘entropic limit’.
This is the fundamental use of calculus today, because it was born on Physics, which studies locomotion.
But calculus also works very often on ∆1 finitesimals, which are integrated in a longer time, or whole in space, or worldcycle in time with the same limits.
And then we realize such operations are exactly the inverse of the previous one. The finitesimal of scale is now the smaller part, which we integrate in a volume of space, or through a motion of time. Why is that possible? Because in each scale a new game of existence does happen. Reason why we wrote the equation as a double feedback equation:
∑∑∑Só∑∑Tó∑∆¡ó∆º= S
If the Universe had only a scale of spacetime, the first equation will be truth. As it is made of planes of spacetime, where each new whole becomes a quanta of a smaller scale; starting again the game, in operations where we differentiate and/or integrate twice, we are emerging and descending through planes of existence. And this is what makes calculus so magic. As the second and third derivative has also a full meaning. It is the rate of rate of change of space into time into acceleration into jerk. It is the growth of growth of a point into a line and a plane and a volume. It is a line that curves, and closes a cycle and becomes a spiral of infinite curvature.
It might then be to ignore all together those symmetries and iterations of ∆ST across planes, which convert one into each other, but that is the deeper structure of reality even if most humind’s are onedimensional and get a headache thinking paradoxically.
And finally we have the limits; once we have scaled up and down as many times as required, we will still just be part of a whole, so limits exist and prevent us from doing what physicists do, ‘from here to infinity’, getting then in all kind of troubles, singularities, infinities that they eliminate and renormalize. All that gets them to real results but if they had the proper understanding of finitesimals and immensities, they would use cutoff earlier on, for ‘singularities’ of bigbangs; charges and masses, understand the wormholes that on those singularities just transfer energy and information between planes, and so on.
To show them will be therefore the second task of this paper as it grows and we enter into ODEs, PDEs and mathematical physics, sometime in the fall of 2019… But for the impatient one, a sample…
THE CURVATURE OF SPACE.
A Disomorphic example on how to understand homology of stiences.
The differences between classic calculus and 5D calculus are thus small, mainly conceptual, but on the ‘fringes’ it will have real consequences for understanding paradoxes of physics; and the philosophical foundations of mathematics. On techniques and what mathematicians like most, crunching numbers and equations, very little new at this stage is expected. And this is comforting cause as I say is summerday and I don’t want to write too much. But do not dismiss the paper. Because this is real existential calculus. Consider the finitesimal of space, curvature. 1//R is the simplest one; the curvature of a circle. It follows that the straight line has a finitesimal curvature. But also that there must be, an immense curvature – as now ‘absolute infinite does not exist’ because a limited infinite is the reciprocal of 0’.
But how curvature can be immense, close to infinity? Mathematicians define curvature for a curve as df/ds where ϕ is the inclination of the sensed tangent and s is the arc length measured from some fixed point.
This limits curvature; but physicists have the not resolved paradox that curvature in Relativity can be infinite, or rather immense. The solution? 5D (:
It is also ‘embedded’ in the complicated formulae and principles of relativity. In relativity the principle of equivalence between acceleration and an attractive vortex of mass; and the principle of Newton that gives a change of motion, hence an acceleration to the curvature of a motion that is not lineal. Thus curvature and acceleration are similar concepts and the more curved a ‘curve’ is, the faster is growing its speed.
Those are you might say trivial results; after all in physics to maintain in orbit a satellite we need an acceleration, a=v^{2}/R, hence a higher curvature requires a higher acceleration; but that is precisely the beauty of 5D, to reduce to the synoptic laws of 5D metric, S=T, SxT=C, and the symmetries of ∆=S=T, an astounding array of phenomena.
Thus we extend Einstein’s Principle of equivalence between force and acceleration, to both charge and mass, and to the curvature of space=time to both the ∆1 charge and ∆+1 mass, physical scales. It is then the implicit concept in Newton’s equivalent formulae, as G, or in Coulomb’s k factor.
This is also embedded in the solution to that differential equation:
Which the reader will notice has a second derivative above, that is usually the symbol of acceleration. So we observe as in many other cases the ∆≈S≈T symmetries between functions in space, scale and time.
But in 5D we can define curvature by the S=T symmetry also as a measure of acceleration, this is possible because as we diminish in size, according to the metrics of 5D, $xð=K (whereas $ is a symbol for lineal space and ð for cyclical time; so we could write in physical terms, Lxƒ=K, but as usual 5D existential algebra symbols are more general and new to use them in any science).
So when we carry curvature in space to acceleration in time, it becomes ‘frequency’, and curvature can have in space any angle above 2π=360º; or in terms of length.
But those forces are conserved in physics. So To understand what physics conserves let us consider a 5D metric equivalent – a 2 D vortex equation, VxRo=K. As the vortex diminish in size it turns faster. In cyclical time, ð cycles of perception that happen when the point returns to the memorial ‘singularity’ happen more often, as, its unit is the closing of a cycle.
The increase of curvature therefore implies an increase on the acceleration of the system, and both are indeed equal concepts: 1/R, ðT/ðS. But now for a given Space perimeter, its higher curvature=acceleration implies a shorter ‘unit of time perception’. So for the same Spatial distance travelled, even at the same lineal speed, more time units have been consumed in a smaller time cycle; as we go down in size scales of the fifth dimension. Both angular speed and existential cycles accelerate; due to the vortex 5D metric: V(ð) x R($) = K. So a slow large turning galaxy might shrink to the size of an atom; which lives a tiny fraction in a tiny size, but in fact its spatial distance traversed is roughly maintained. And indeed latter we will see how in the 3Dimensional spacetime of the ∆±3 scales of the galatom, a beta decay is in 5D metrics equivalent in time duration to a quasar 15 billion bigbang cycle, and the 5D metric of the proton equivalent to the Schwarzschild event horizon of the black hole.
5D metrics conserves 2 things: The ‘energy’ volume of spacetime of all scales and its beats of existence:
 The worldline distance the being travelled – which is in fact a worldcycle distance as it is the sum of all the perimeters, travelled slowly in the large spacetime, faster in the smaller spacetime. So the spacetime volume of the different ∆scales is the same. And because energy is the only parameter used by huminds in all scales, the conservation of the total volume of spacetime of the Universe is equivalent to the conservation of Energy.
 But to conserve the same ‘length of internal perception in existential beats’, the smaller being much live far less time. And indeed, the neutron cycle in beta decay is 15 minutes.
The galaxy cycle in a quasar bigbang cycle is 15 billion years.
They are two different curvatures, because the curvatureaccelerationforce of the charge is immensity compared to the almost 0’ curvature of the galaxy.
But alas! Both are the same. So we can unify both forces, with the simple concept of curvature=acceleration=attractive force, as a faster more curved sink will attract, like a hurricane stronger.
Einstein’s derived his formalism from Poisson. It is more detailed because it is according to the Galilean paradox (Si=Te) the spatial still perspective of those vortices as a series of simultaneous derivative measures.
So it reduces the temporal continuous Newtonian view of a spacetime vortex into an ∞ number of infinitesimal detailed pictures, focusing not on the speed but on the curvature of the vortex (which is the spatial definition of a moving cyclical speed – the faster it turns, the more curvature it has in ‘still mathematics’).
Let us do the maths in the simpler Newton’s formalism, whereas by the paradox of Galileo S (Curvature) = T (accelerated motion). So the Universal Constants (G, k), define the curvature of 2 spacetime vortices at the ∆1 quantum charge and ∆+1 cosmic mass scales (∆ is the symbol for the different ±¡ scales of the fifth dimension within a given organic system). Its formalism of a vortex of time space is then Newton’s Unification Function: M,Q= ω^{2} r^{3} /U.C.(G,k)
It applies to all vortices of timespace from particles to planets to galaxies. For example if we substitute for the Earthsun system we obtain G, (1st ever theoretical deduction) and if we substitute for the Bohr Radius and Proton Mass, we obtain k with a 10^{39} higher curvature value, the exact difference between both forces that solves its hierarchy problem. As curvature in space is symmetric to rotational speed in time, so it is symmetric to the attractive force of any vortex. It works marvels when we translate electromagnetic jargon to Newtonian jargon. For example it shows the ‘isomorphism’ (systemic jargon for an equal ‘form’ between scales) between atoms and galaxies, which Hatoms of the cosmic scale.
Since when we translate electromagnetic function into gravitational mass vortices, the proton radius becomes the Schwarzschild radius of a black hole and its electronic orbitals its star clouds, a result foreseen by Relativity that modeled galaxies as Hydrogen atoms in the EinsteinWalker Metric of the Cosmos.
Let us put some easy numbers by substituting the parameters in that Unification function for the values of the sun (mass) minus earth (rotational speed and radius) to get G, which any high school student can do:
Sun mass = 2 × 10^{30} kg; Earth’s angular velocity 2 × 10^{7} rad. per sec. Earth’s orbit = 150 million kms. Result: G=6.67 × 10 ^{11} kg1 m^{ˆ3} rad. sec.^{2}
This is standard gravitational theory. What has never been done, because the fractal systemic view of the fifth dimension was not known till recently, is to substitute in the same function of gravitational cosmological masses the mass radius and speed of the spacetime vortex by the values of the fundamental quantum spacetime vortex, a hydrogen atom/charge.
If the thesis of a fractal universe made of hierarchical scales is truth, then those values should give us the value of the universal constant of charges, the Coulomb constant.
Indeed, if we substitute for the proton (mass) and the Bohr electronic orbital (speed and radius)
4 × 10^{16} rad. sec. 1 = w (electron); 5.3 × 10^{11} m. (Bohr radius); proton mass = 1.6 × 10^{ 27 }kg.
Then we get a G, which is 2×10^{39} stronger than the gravitational radius; thus, the hydrogen atom behaves as a selfsimilar fractal scale in the quantum world to a solar system.
And then you can get also the electron radius expressed in the jargon of a quantum gravitational world using the translated ‘Gravitational Coulomb constant’: G(k)M/c^{ˆ2}.
Since in that expression M is the mass of a proton, G(k), the electromagnetic constant is a gravitational constant, and c, light speed, that expression is exactly the Schwarzschild radius of a quantum black hole.
Thus, the electron Bohr radius, which is the final radius of minimal size and energy in electrons, is isomorphic to the event horizon of a black hole in the quantum gravitational world.
Those results (more than a decade old), are a first theoretical deduction of Ke departing from G and the enormous simplification of the parameters of the electron radius till arriving to the same expression that a black hole radius cannot be by chance. They are mathematical deductions, one of the three standard forms of proof in science.
Yet a theoretical calculus of those values cannot be exact ‘by chance’, unless the theoretical model behind it – the fractal selfsimilar structure as $T (Space population) x ð§ (Temporal frequency) entities of all physical systems is right. Thus, the previous calculus is a clear proof that both, charges and masses, are unified as values of the same type of spacetime vortices in the 2 different scales of spacetime of the Universe. And they are geometrically unified from the p.o.v. of geometrical relativity not from quantum theory, as Einstein wanted it.
Galaxies, (Galaxies≈Atoms) thus resolve the philosophical question on how many 5D scales exist; as we find enough selfsimilarity to ‘run again’ another game of fractal scales (not identical but selfsimilar as in a Mandelbrot fractal) both by quantitative and qualitative methods between the atom and the galaxy. A question that might be extended to the ST dualities of open, ‘entropic strings’ and closed ‘cyclical informative strings’, in a possible larger and smaller scale of microscopic strings and superstrings:
Ouroboros the Universal Snake, bites its tail on the string quantum and cosmological selfsimilar scales, as perceived from the human ∆o mind. Philosophy of stience would then argue that those scales are real, but part of its selfsimilarity is mental: that is, the loss of information in the perception of scales make humans extract the same information from the upper and lower 10^{±30} scales.
Alas! In this showcase of multiple meanings, jumping from mathematics, to physics, to metaphysics, solving questions seek for centuries in classic science we show the essential nature of 5D – not so much crunching numbers but ‘seeing’ what nobody sees.
As I improve the papers, we will focus better the equations already resolved but vastly more profound that people think.
WORLDCYCLE OF EXISTENCE.
In the next graph repeated ad nauseam in those papers we see the essence of the process of a worldcycle of existence: the creation of a finitesimal form which will reproduce and then collapse into a superorganism.
All what exists is a supœrganism of vital space tracing a 0sum worldcycle of time through 3 scales of the 5^{th} dimension: Born as a seed of fast time cycles in a lower 5D scale (∆1:Max. T x Min. S), emerging as an organism in ∆o, living 3 ages of increasing information, as its time clocks slow down in its ∆+1 world to die in a time quanta back to ∆1. Yet the maximal point Si=Te where reproduction happens defines the classic age, maturity, beauty, balance, survival of the system, all disomorphic jargons.
The 3 ages of life emerge in human social superorganisms as the 3 ages of cultures and its 3 artistic styles: Min.S x Max. T (infantile epic, lineal art, as in treccento, Greek kuroi; Si=Te; balanced beauty, when form and size are in balance, the classic mature age; and Max. S x Min. T: baroque, 3rd age of a civilisation, whose subconscious mind is the art of its ‘neuronal artists’, the age of maximal form and a ∆st for a no future, which is the age of war and death of cultures).
We talk of 3 ∆±1 scales of worldcycles as the being live in a placenta, then emerges as organism in a world:
þ: 01: its palingenetic o1 social evolution in the accelerated time sphere of existence, till becoming 1 (01 bounded unit circle in ¡logic mathematics; quantum probability sphere of particles in physical systems; palingenetic fetal age in biologic systems; 09 memetic learning childhood in social systems). It is the highly ordered world cycle as a ‘placental motherenergy world’ is nurturing as memorial cyclical spacetime has erased errors of previous generations.
– c: The outer 1∞ world, in which it will deploy its 2nd world cycle of existence in an environment which is open, entropic (1∞ hyperbolic unbounded Cartesian plane in ¡logic mathematics; thermodynamic entropic statistical molecular populations in physics; Darwinian struggle between populations in biology; idologic dogeatdog capitalist, nationalist competitive eco(nomic)systems in the super organisms of history. In this 1∞ existence the world cycle is not ensured to continue, as the entropy of the world system can cut it off.
ω: The existential life cycle, though is part of a larger world of hierarchical social scales (§ D¡), where it performs 5 survival actions through ∆±4 Planes selfcentered in its mind, beyond which it cannot longer perceive, to become if successful a new superorganism of the infinite planes of God, the game of existence.
In graph, physical, biologic & social worldcycles show to which extent 5D laws enlighten our understanding of reality. Matter States are physical time ages, from left pure solid, crystal, §top state, to an even more solid ∆+1 boson condensate, etc. We see that systems either move a step at a time within a plane of existence (gas, liquid, solid) or they can jump « two states at once, (as in the case sublimation) within that plane, or most often between two planes, as in « scattering & entropic death), to become a different Dimotional state. We can then see how the fundamental elements of 5D time appear on the graph: the worldcycle is local and complete. There are 2 inverse arrows from an entropic past (plasma), in a lower plane (ion particles) to the 3 ages of the matter states with increasing form (gas to solid), to end in a higher plane of existence as a bosonEinstein condensate. Do those worldcycles happen for the whole Universe? (cyclic bigbang). Unlkely…
It is then clear that calculus is the closest mathematical mirror of the commonest process of timechange: the creation of finitesimals that reproduce in clonic waves forming ‘spatial organic systems’, in the most complex worldcycle, or mere herds, or locomotions imprinting information in a lower field of entropic space – you name it. As we study mathematical physics with calculus, we shall be commenting precisely in the unity of all process of calculus – a process of finding finitesimals to integrate through time locomotions or spatial populations, mimicking what is the essence of time=change, the reproduction of finitesimal parts into wholes.
The next question is then how to write the worldcycle of existence in calculus; which is selfevident: ∆st = 0
Whereas the whole trajectory of a T.œ=∆ST in space and time will finally become a zero sum. But a function of existence does vary its ‘rate of change’ as the system goes through 3 ages changing its parameters of spatial form and temporal motion; represented in calculus by the S and T parameters.
Its usefulness becomes then more clear when we consider the standing points and draw the function=worldcycle of existence in terms of the SxT existential momentum of its 3 ages.
In praxis often human calculus will deal with the ∆scalar main parameter that defines the system for each scale (mass, temperature, momentum)… embedded in an outer spacetime world. And so calculus as humans practice it is not so much about the function of existence but a ‘partial’ analysis of a ‘dimotion of existence’ performed by an ∆0 being in a larger ˙∆+1 world: ∆º ∂st+1 = 0’ That is, we take a T.œ, defined in scale, space and time and study its minimal finitesimal change, a relative zero (in physics using Lagrangians), and then integrate it through a ‘lineal sumperiod’.
Calculus on the function of existence.
Since I haven’t told you this one thousand one nightmare times (: the function of exist¡ence is all. You are a repetitive fractal of spacetime and your purpose is to exist, to conserve your time, but your time is just the form of information, your vital space, reproduced in all the scales that rise from the bottom line of your gravitational and light spacetime, going upwards into scales. Reproduction is the game. But the worldcycle makes errors in the reproduction of your ilogon and those errors that are statistically seen in space as a normal distribution, in time as a repetitive sequence of actions and events slowly wear you down, and as errors of copying information repeat and accumulate your function of existence looses freshness and you age.
So because each stœp of your existence you repeat your sequential actions each derivative is one of such stœps a zig zag up information right motion, up information right motion, whose tangent is the derivative of each quanta of your time. All this said then we can study the worldcycle with calculus. In fact is the best way to study the worldcycle with calculus, in the orthogonal graph of information and motion, information and motion, stop and step, particle and wave state, up and right up and right, as you age, first rising fast young and bold, reaching higher accelerations in your second derivative, as space is time, the curve represents in its form of space, its motion of time, and that is your first derivative, seeking a standing point of constant speed but that is not possible because speed is reproduction and you reproduce your form, with lesser skill past the prime time of your standing point the maximal and minimum no far before, no long ahead:
Let us remember the general laws for any possible function of existence:
If we draw the ‘existential momentum’, SxT of the system in the left side, and the lineal time of the system, T in the bottom side.
So sinusoidal bell curve functions represent a worldcycle, though the symmetry is broken in the moment of entropic death when the collapse is extreme in a ‘falling line’ as death happens in a single moment of time:
4D»∆1(seed)∑∆:$T(limbfield)<ØS≈T (iterative bodywave)> O§ð (particlehead)«5D∆1(death)
A key theme of vital mathematics is the representation of a worldcycle in lineal time, with ± exponentials & its inverse, logarithmic curve around the key points of change of phase… as growth of ‘entropymotion’ diminishes. So we move from ‘adolescence’ of max. growth of both parameters (sT energy and sT information) to the y”=0 point of youth, where the logarithmic part grows slower. Together they form, one half of the total graph of a cycle of existence, till reaching the y’=0 point of Max. (S≥≤T), which then becomes negative, happening a decay of the whole system in two negative curves.
The conservation of time in its 5 y’Ù y” =0, standing points that define the 5 SS, Ts, ST, St & TT moments of generation, youth, maturity, 3^{rd} age and entropic death thus become the essential points (maximal and minimal) of the equations of calculus, the sinusoidal function of existence and all its derived elements.
Let us suppose that on a certain interval a≤t≤b we are given a function S = f(t) which is not only continuous but also has a derivative at every point. Our ability to calculate the derivative enables us to form a clear picture of the graph of the function. On an interval on which the derivative is always positive the tangent to the graph will be directed upward. On such an interval the function will increase; that is, to a greater value of t will correspond a greater value of f(t). On the other hand, on an interval where the derivative is always negative, the function will decrease; the graph will run downward.
We have drawn the graph of an ∆st function of the general form, S (any dimension of a whole world cycle or T.Œ) = f(T) – Any time motion or action.
It is defined on the interval between a minimal quanta in space or time (t1) and its limit as a function (d).
And it can represent any S=T duality, or more complex 5Ds=5Dt forms or simpler ones. We can also change the s and t coordinates according to the Galilean paradox, etc. Hence the ginormous numbers of applications, but essentially it will define a process of change in spacetime between the emergence of the phenomena at ST1 AND ITS DEATH mostly by scattering and entropic dissolution of form at d.
And in most cases will have a bell curved from of fast growth after emergence in its first age of maximal motion (youth, 1D) till a maximal point where it often will reproduce into a discontinuous parallel form (not shown in the graph at Max. S x Max. T; which will provoke its loss of energy and start its diminution till its extinction at point d.
Thus the best way to express quantitatively in terms of ST parameters (mostly information and energy), for any world cycle of any timespace super organism is a curve where we can find those key standing points in which a change of age, state or motion happens.
Of a special interest thus are the points of this graph whose abcissas are t_{1,2,3,4,5}.
At the point t0 the function f(t) is said to have a local maximum; by this we mean that at this point f(t) is greater than at neighboring points; more precisely for every t in a certain interval around the point x0.
A local minimum is defined analogously. For our function a local maximum occurs at the points t0 and t3, and a local minimum at the point t1.
At every maximum or minimum point, if it is inside the interval [a, b], i.e., if it does not coincide with one of the end points a or b, the derivative must be equal to zeroth (0’).
This last statement, a very important one, follows immediately from the definition of the derivative as the limit of the ratio ΔS/ΔT. In fact, if we move a short distance from the maximum point, then ∆S≤0.
Thus for positive ΔT the ratio ΔS/ΔT is nonpositive, and for negative ΔT the ratio ΔS/ΔT is nonnegative. The limit of this ratio, which exists by hypothesis, can therefore be neither positive nor negative and there remains only the possibility that it is zeroth.
By inspection of the diagram it is seen that this means that at maximum or minimum points (it is customary to leave out the word “local,” although it is understood) the tangent to the graph is horizontal.
At the points t2, and t4 also the tangent is horizontal, just as it is at the points t1, t3, although at these points the function has neither maximum nor minimum. In general, there may be more points at which the derivative of the function is equal to zeroth (stationary points) than there are maximum or minimum points.
One of the simplest and most important applications of the derivative in that sense is in the theory of maxima and minima.
Criteria for maxima and minima; study of the graphs of curves.
If throughout the whole interval over which x varies the curve is convex upward and if at a certain point x0 of this interval the derivative is equal to zeroth, then at this point the function necessarily attains its maximum; and its minimum in the case of convexity downward. This simple consideration often allows us, after finding a point at which the derivative is equal to zeroth, to decide thereupon whether at this point the function has a local maximum or minimum.
Now, the apparently equal nature on a first derivative of the minimal and maximal points of a being, have also deep philosophical implications, as it makes at ‘first sight’ indistinguishable often the processes of ‘reproductive expansion’ towards a maximal and explosive decay into death, the ‘two reversal’ points of the 5D (maximal) and 4D (minimal) states of a cycle of existence, for which we have to make a second assessment (second derivative) to know if we are in the point of maximal life (5D) or maximal death (4D) of a world cycle. And to know if the cycle will cease in a continuous flat encephalogram or will restart a new upwards trend.
Or in other words is any scalar, e>cc>m bigbang both the death and the birth of matter?
Finitesimal Quanta, as the limit of populations in space and the minimal action in time.
So there is behind the duality between the concept of limits and differentials (Newton’s vs. Leibniz’s approach), the concept of a minimal quanta in space or in time, which has been hardly explored by classic mathematics in its experimental meaning but will be the key to understand ‘Planckton’ (Hplanck constants) and its role in the vital physics of atomic Planes.
It is then essential to the workings of the Universe to fully grasp the relationship between Planes and analysis. Both in the down direction of derivatives and the up dimension of integrals; in its parallelism with polynomials, which rise dimensional Planes of a system in a different ‘more lineal social inter planar way’.
So polynomials and limits are what ¬Algebra is to calculus; space to time and lineal ¬Algebra to curved geometries.
The vital interpretation though of that amazing growth of polynomials is far scarier.
Power laws by the very fact of ‘being lineal’, and maximise the growth of a function ARE NOT REAL in the positive sense of infinite growth, a fantasy only taken seriously by our economists of greed and infinite usury debt interest… where the eª exponential function first appeared.
The fact is that in reality such exponentials only portrait the decay destruction of a mass of cellular/atomic beings already created by the much smaller processes of ‘re=production’ which is the second dimension mostly operated with multiplication (of scalars or anti commutative cross vectors).
So the third dimension of operands is a backwards motion – a lineal motion into death, because it only reverses the growth of sums and multiplications polynomials makes sense of its properties.
Let us then see how the operations mimic the five dimensions, beyond the simplest ST, SS and TT steps, namely reproductive and 4D5D inverted arrows.
We can establish as the main parameter of the singularity, its time frequency, which will be synchronised to the rotary motion or angular momentum of the cyclical membrane. They will appear as the initial conditions and boundary conditions of a derivative/integral function, which often will be able to define the values of the vital energy within, as the law of superposition should work between the 3 elements, such as:
Determination of the greatest and least values of a function.
In numerous technical questions it is necessary to find the point t at which a given function f(t) attains its greatest or its least value on a given interval.
In case we are interested in the greatest value, we must find x0 on the interval [a, b] for which among all x on [a, b] the inequality ƒ(to)≥ƒ(t) is fulfilled.
But now the fundamental question arises, whether in general there exists such a point. By the methods of modern analysis it is possible to prove the following existence theorem:
If the function f(t) is continuous on a finite interval, then there exists at least one point on the interval for which the function attains its maximum (minimum) value on the interval [a, b].
From what has been said already, it follows that these maximum or minimum points must be sought among the “stationary” points. This fact is the basis for the following wellknown method for finding maxima and minima.
First we find the derivative of, f(t) and then solve the equation obtained by setting it equal to zeroth.
If t_{1}, t_{2}, ···, t_{n}, are the roots of this equation, we then compare the numbers f(t_{1}, f(t_{2}), ···, f(t_{n}) with one another. Of course, it is necessary to take into account that the maximum or minimum of the function may be found not within the interval but at the end (as is the case with the minimum in figure) or at a point where the function has no derivative.
Thus to the points t1, t2, ···, tn, we must add the ends a and b of the interval and also those points, if they exist, at which there is no derivative. It only remains to compare the values of the function at all these points and to choose among them the greatest or the least.
With respect to the stated existence theorem, it is important to add that this theorem ceases, in general, to hold in the case that the function f(t) is continuous only on the interval (a, b); that is, on the set of points x satisfying the inequalities a <t < b.
It is then necessary to consider an initial time point and a final time point, birth and death, emergence and extinction to have a determined solution.
Derivatives of higher orders.
We have just seen how, for closer study of the graph of a function, we must examine the changes in its derivative f′(x). This derivative is a function of x, so that we may in turn find its derivative.
The derivative of the derivative is called the second derivative and is denoted by y”=ƒ”(x)
Analogously, we may calculate the 3^{rd} derivative y”‘=ƒ”‘(x) or, the derivative of nth order. But as there are not more than 3 ‘similar derivatives, with meaning’ in time (speed, acceleration, jerk) or space (distance, area and volume), beyond the 3^{rd} derivative the use of derivatives is only as an approximation to polynomial equations, whose solvability itself is not possible by radicals beyond the 3^{rd} power.
So it must be kept in mind that, for a certain value of x (or even for all values of x) this sequence may break off at the derivative of some order, say the kth; it may happen that f(k)(x) exists but not f(k + 1)(x). Derivatives of arbitrary order are therefore connected to the symmetry between power laws and ∫∂ operations in the 4th and inverse 5th Dimension, through the Taylor formula. For the moment we confine ourselves to the second and third derivatives for ‘real parameters’ of the 3 space volumes and time accelerations.
The second derivative has then as we have seen a simple significance in mechanics. Let s = f(t) be a law of motion along a straight line; then s′ is the velocity and s″ is the “velocity of the change in the velocity” or more simply the “acceleration” of the point at time t. For example, for a falling body under the force of gravity: That is, the acceleration of falling bodies is constant.
Significance of the second derivative; convexity and concavity.
The second derivative also has a simple geometric meaning. Just as the sign of the first derivative determines whether the function is increasing or decreasing, so the sign of the second derivative determines the side toward which the graph of the function will be curved; but in terms of time represents the second derivative of the curve of existence. That no longer accelerates its growth, hence the end of youth, and viceversa, the moment in which it does accelerate its decay, thus the beginning of the third age.
So we can consider the same concept in the ‘discreet’ baguas of life cycles as it is NOT a mere ideal curve but one that do happens in all forms of life. This simple law with deep cases because it is essential to the worldcycle:
Suppose, for example, that on a given interval the second derivative is everywhere positive. Then the first derivative increases and therefore f′(x) = tan α increases and the angle of inclination of the tangent line itself increases. Thus as we move along the curve it keeps turning constantly to the same side, namely upward, and is thus, as they say, “convex downward.” On the other hand, in a part of a curve where the second derivative is constantly negative the graph of the function is convex upward.
Because it is the clear proof of what is all about: reproduction in space of frequencies of time.
The function is more than its equation – A path of existence through the whole plane.
The function of existence is the whole plane divided by the line that must be grown by the nonE method of rising points into curves of motion, which divide an energy information plane in an act for creation with a path in S=T, the path of present through squares in which information and energy are orthogonal.
We have found thus the simplest spacetime curve, the S=T, curve of existence between an integral 1/3^{rd} of the plane in path with a 2/3^{rd} Lébesgue integral so to speak of the external/internal path of the curve.
The curve is thus a point in motion, equivalent to a line of distance, equivalent to a ratio between 2 parts, 2/3rds to the left and 1/3^{rd} to the right. But the beauty of it is that we take from the curve square points.
In the graph, a classic, Taoist representation of the 3 ages of life and its inverse parameters of youth (max. energy) and old age (max Information) represented by the triads of the I Ching, and a modern graph of duality showing those parameters as a semicycle, which in certain simple beings like light are in fact both the ages of time of a physical wave and its form in space, as light quanta, h=exi, is indeed both our basic cycle of time and surface of energetic space of which all are made.
One of the oldest graphs from 92.c. ‘The error of Einstein’, pioneer book on 5D physics is the understanding of a Galaxy as a representation of the game of existence, and its deep metaphysical implications in terms of mental spaces, which summarizes in a huge metaphysical thought regarding the way a mind perceives a mental space:
Minds diminish the information they observe from reality as reality becomes further away in ∆ST distances (scale, form or motion), to a point in which they only perceive the purest forms of mental space, which are the waves of existence in cyclical form (our perception of the galaxy) or lineal form (our perception of the light of the quantum scale, or in scalar form (fractal perception of networks in hyperbolic space).
In the graphs above and below, the minimal reality is a 3D² form seen in a single plane, with a singularity @mind a membrane and a vital energy within. When we make a holographic broken image of this reality the simplest way to do it is in four cartesian regions, TT, ST, ts, and ss, which correspond to the +1 +1, +1 1, 1 +1 and 1 1 quadrants of the plane.
It is then when the Lebesgue inverse function matters to integrate the Y perspective of the S=T symmetry that the function know taken as a topological partition of a vital motion of a wave of similar particles that will collapse t the end of its journey through a plane of the fifth dimension, takes place.
The least path action implies thought that the end of the path taken in the Cartesian but also the imaginary plane collapses in the same point, regardless of how many oaths have been taken in the ‘compressed’ iplane where the coexistence of paths in particle space has sunk the plane to a √ root value for the dense line that then can be even further reduced to a point that will potentially trace a full s=t, valuing in present time the spacetime dilation of the integrated,
The wave form as an integral expression of the function of existence.
How many possible forms might acquire the function of existence? The answer that might surprise the reader is depending on the number of parameters, duration in time of our analysis and type of dimotion studied, from smaller steps of a single dimotion to the whole worldcycle an all the sequential dimotions of a T.œ. there are ‘infinite solutions’ – as all equations are ultimately ‘partial equations’ of the fractal generator, SóT.
Consider the commonest form of the Universe, a wave. If we consider that y measures NOT the value of ST of the system as a constant ‘volume’ of existential momentum, but the value of its ‘degree of increase or decreste’ at each moment of time, hence y’ over x(t), we obtain the exact form of a wave, with a first half wave in which the growth from youth to maturity constantly diminish but is still positive, till the middle point of maturity at y’=0, where the growth starts to be negative, followed by a fast decline as we age, till a maximal point of ‘degeneration’, where we normally die by sudden sickness; but if we overcome that point somewhere around the 70 years age, we will have a slow down of our aging, towards a point of no ‘change at all’ – the point of death; when we simply disappear from this plane of spacetime existence.
Thus when we perceive a wave of light, we are in fact, perceiving time=change, and creating a mental space of the lifedeath cycle of a single photon as space is just the memorial tail of our slow time perception.
RECAP. The function of existence in its fractal variations and cx. Pentalogic HAS infinite paths=forms but all end in a 0’ sum.
Time is cyclical as all clocks of time return to its point of origin, so all time cycles including those of life of its vital spacetime beings are finite. Further on those time cycles break ‘space’ into inner and outer parts, so vital space is broken by the membranes and angular momentums of those time cycles that make spacetime beings also finite in spatial information. And an obvious experimental facts about timespace: cycles of time, vital spaces and the species made of them, coexist in several scales of relative size from particles to galaxies, each one with clocks of time of different speeds. So spacetime is fractal broken in scales that added create a new 5^{th} dimension of spacetime.
The dual functions of 5D Absolute Relativity, the function of 5D scales, SxT=C & the function of equality between form and motion, SI=TE, develops in 3 ages with 3 standing points, a max. point of existence, Si=Te or mature age, a young age of Max. T=motion, and an old age of Max.S=information; between birth in ∆1 Form & Tentropic death. The search for spacetime, Energy=information balances in a classic reproductive age of conserved time is thus the goal of all exist¡ences, but only the whole achieves the immortality of timespace, as we shall see egocy errors of fractal mindpoints of space trying to stop the flow of time from a single selfish point of view, accelerates the imbalance that brings death equations. We are richer in our still property at that 0Tmoment, when all is quiet so for time to keep moving, a reversal of entropy takes place.
The connection between existential algebra and calculus: Dimotions as actions. Reproduction as change.
We said often that time=motion is all; and space just the Maya of the senses, the mind’s mapping of the fractal points ‘that hold a world in themselves’. But the ultimate arrow of time is that of scalar growth between planes of the fifth dimension, as parts must become before wholes; the upwards arrow matters more than the down arrow. And so of the 3 parameters that define objectively between ¬ limits, and vitalized by a mind’s program, ∆ST, any being, ∆scale matters more – numbers of algebra in mathematics. Then it comes time perceived in one given plane, T, and finally Space, the most evident but shallow part of the whole. For that reason Algebra matters more and includes calculus, the temporal view of mathematics that tries to capture all modalities of change with a simple scalar process of adding ‘¡n≈finitesimals’ of scalar change to analyze the larger processes of change in the whole scale.
This is done in calculus with the simple methods of ‘finding the parts=derivatives’ and adding them together =integrate them either over scale, spatial volume or temporal frequencies. In this manner something so simple as a finitesimal change becomes the seed of all possible variations of change (dimotions each studied by an operand) across scale=size, spatial population or temporal frequency of events.
The study of the 5 Dimotions of the Universe is carried out in spatial geometry by calculus; in NonAristotelian Logic by Existential algebra. Thus both languages have many deep common structures worth to compare, even if calculus was born on the praxis of analysis of one single dimotion, locomotion, in the milieu of physical sciences and only slowly extended to the understanding of the other dimotions of the Universe.
Thus we shall bring in this second paper on algebra both sciences together.
Even if Existential algebra is much wider and ultimately a logic stience, as it is also the underlying structure of mathematical algebras, including those of group theory that deal with an ‘extensive catalog’ of the dimotions and evolutions of the Universe, and reticular Boolean algebras that deal with the @mind mirrors of logic and numbers. In the original plan I had envisioned a much larger output of papers for academia, taken from my 30 years notebooks, so Existential Algebra would have deserved one of his own. But time is running out…
Existential algebra and calculus study time change. How can then unify all time changes? The answer comes from existential algebra and its finding that all forms of change can be reduced to reproductive change. Which itself can be considered a travel down and up two scales of the fifth dimension. Thus changes happens on finitesimal parts that emerge and affect larger wholes.
The function of existence is a function of reproduction in scale (as a 5D journey) in time (as a conjunction of the 5 Dimensional motions of existence), and space, as a simultaneous growth of clone information; formalized in the fractal generator of 5D metrics, Max. ∑Te x S¡ (s=t) = c; as reproduction happens in a ‘present s=t state’, of balance when the relative past of lesser informed flows of entropic time, Te, becomes Imprinted by Spatial information: Past TTentropy x Future SSform = Reproductive ST Present
Change happens informatively through increase of finitesimal parts, entropically when you loose those scalar parts. Reproduction of form or its annihilation at the finitesimal scale in calculus is mirrored by a simple function, F(x+h)/F(x), that calculates ratios of change, for different operands that mirror the 5 Dimensional motions of existence, which can potentially change.
Thus calculus uses a unit of change, h, to mirror different changes In the 5 dimotions of existence. Since change once it happens in small units, in small scales, in small instants of time; differentiate in 5 type of dimensional motions = actions: TT, feeding, entropic and moving, Ts, changes, informative & perceptive St, SS changes, reproductive changes, ST, proper.
That diversification is studied better with different algebraic operands; but all can be derived into its finitesimal units of change and integrated, for different scalar groups, social functions and paths of dimotional change.
Thus what both disciplines, calculus and existential algebra have in common is the object of its linguistic mirrors: Times=changes, all kind of them.
That apparently they seem so different wears witness to the ultimate nature of mindmonads, ‘infinity mirrors’ that reflect always different points of view on reality and its imagination to slightly bend that reality to the point of view of the mind.
Still it is more remarkable in its common elements than in its differences.
Calculus has its emphasis in numbers hence in the scalar analysis of huge social groups in motion; while existential algebra has its emphasis in discrete dimotions, hence on the study of individual T.œs experiencing a transformation.
The very essence of calculus is to study in synchronous spatial dimotion huge amounts of numbers, which will erase its ‘discrete’ form to appear as a continuous susceptible to be studied at the ∆+1 scale of the whole.
The emphasis of Existential algebra is the study of that whole as an individual subject to sequential dimotions.
But in both cases the dynamic process of study are the 5 Dimotions of time=change of the universe.
Finally logic systems and Boolean algebras become the syntax of verbal and computer minds that describe with its sentences the dynamic dimotions of reality. So its language is closer to that of Existential Algebra, reason why we include it in this paper, instead of the more advanced models of existential algebra termed, monologic, duality, trinity, pentalogic and dodecalogic.
To fully grasp that essential connection between ∆st and calculus mirrors, we must first understand how species on one hand, and equations on the other, probe in the Planes of reality to obtain its quanta of spacetime converted either in motion steps or information pixels, to build up reality.
The connection between existential algebra and calculus is qualitative: both study initially the finitesimal action of existence, which become the finitesimal quanta of spacetime, whose repetitive accumulation causes the phenomena of timechange. Existential algebra though studies the qualitatively in terms of sequences between the 5 Dimotions, and calculus quantitatively focusing in one single dimotion spread in a group of scalar numbers.
This is the case because the actions of beings happen through finitesimals extracted from other ∆plane scales.
In all Planes, the simpler actions of any being are extractions of motion, energy, entropy=motion and form from lower ∆i Planes:
A T.œ perceives only the ∆±3 planes from where it extracts energy or information. As its actions and dimotions are architectonically performed through planes of 5D where each main action relates to an interval of scales:
∆43: The system extracts indistinguishable boosts of entropic of motion (man from gravitation).
∆32: The system extracts bits of information (Light in man)
∆21: The system extracts bites of energy (amino acids in man)
∆1 0: The system seeds its minimal seed of reproduction.
∆0+1: The system connects socially with other systems to evolve into a whole.
So simpler Actions start at finitesimal level, gathering in sequential patterns in existential algebra, as ‘time flows’ and in population and spatial patterns – in integral herds of numbers in calculus.
We and all other beings perceive from ∆3 quanta (light in our case), feed on amino acids, (∆2 quanta for any ∆º system), seed with seminal ∆1 cellular quanta (electrons also, with ∆1 photon quanta).
For each action of spacetime we shall find a whole, ∆º T.œ, which will enter in contact with another world, ∆±i, from where it will extract finitesimals of space or time, energy or information, entropy or motion, and this will be the finitesimal ∂ ƒ(x), which will be absorbed and used by the species to obtain a certain action, å.
Analysis allow us to extract actions from wholes, reason why there are not really use beyond the third derivative of a being, as super organisms coexist in 3 only Scalar Planes. It also works in terms of a volume, as its derivative is a plane, then its unitcell or point… So to speak, if you derivate a world, you get its organism, and if you derivate it again you get its cell and then its molecular parts. And then if you do that in time, you get its speed and then its acceleration and then its jerk.
The magic of derivation
Because of the symmetry between ∆≈S≈T, to extract finitesimals of smaller scales the process is the same. We derive the whole, which diminishes its ‘dimensions= power’ as the system looses its larger whole, but increases its number of ∆¡ visible particles, whereas the difference of value between both, shows the ratio and structure of its entropy, energy and information networks, sum of its components that form the whole. As certain functions define more specialized T.œs than others. So the parts of a whole vary according to topological structure.
PART II. CALCULUS OF DIMOTIONS OF EXISTENCE.
THE 3 AGES OF ANALYSIS.
The underlying order of all structures of the entangled Universe between its S, T and ∆ Components once more shows in the 3 ages of calculus, which we can terms as the scalar age, when the main question was that between parts and wholes (from Greece through Newton), the temporal age when its main focus was the description of the 5 Dimotions of physical systems (from Leibniz to Heaviside) and finally the spatial view, when its main focus is, besides the completion of the previous ages, its use to the description of mental spaces (from Gauss through Riemann and Hilbert to Einstein and quantum spaces).
Thus to put some order in such a vast subject, we shall do as usual a diachronic analysis of its informative growth in complexity in 3 ages, barely touching the essential elements of each of them; from its:
I Age: Scalar view, from the Greeks to Newton ns Leibniz. The beginning of calculus was verbal, logic, in the Greek age, with the discussion of finitesimals (5D infinitesimals with a minimal size), and Universals. This philosophical analysis was retaken by Leibniz. Whereas the duality of derivatives as limits vs. differentials – tangents of change (Newton’s vs. Leibniz’s approach), represents the duality of a minimal quanta in spacetime (Leibniz’s infinitesimal) or in scale (Newton’s limit), hardly explored in philosophy of mathematics, but a key concept in 5D Planes, Universal Constants and quantum physics.
Newton on the other hand, a practical English man with little interest for the whys came to the concept through the study of limits, of power series – the scalar view, without much interest on what they meant. They whys were covered by Yahweh and his biblical studies to prove that God had sent him comets to teach him gravitation as the ‘chosen one’ after Kepler, who knew that ‘Him had waited 5000 years to find an intelligence like his, me Kepler, to show him his clock work’. After so much evident truth, who were those humble believers to contest God’s wise decisions? Leibniz though was more interested in meaning and so he did find the true finitesimal, 1/x. To the question of who copied who the answer is obvious, and the fact is not yet resolved merely shows that mathematicians still do NOT understand the foundations of calculus in its trinity useful for ∆, S and T, the 3 components of reality. Because they came through different methods, Newton found the ∆scalar power series of finitesimal changes that grow internally in ‘speed of change’ as they accumulate larger power series factors; so each summand of the power series can be taken as a scalar ever larger change per unit of time; while Leibniz found the ST geometric analysis of external change, mostly useful for locomotion vs. the higher interest of power series understood as a summand at a time, for internal change and growth.
Both are completely different approaches that serve an essential duality between internal evolutionary ‘biological’ growth vs. external physical motion, which instead of opening a proper philosophy of calculus based in the whole range of changes in time (best served by derivatives), space populations (best served by integrals) and scalar growth (best served by sums of series), just brought the quintessential monologic ego centered, ænthropic man, Mr. Newton, to argue, as he had done also with Boyle, on the primacy of its ‘ceteris paribus’ discovery of a vast region of mental mathematical spaces, suit to study ALL forms of change=time, which truly made ¬Algebra, the queen of all experimental sciences, of which physics, given the reductionism of its practitioners is just a subdiscipline.
So as huminds still ignore that all is about the trinity of ∆, S and T (power series, integrals and derivatives) and their egocy is the only ∞ truth (Einstein), they have not yet understood what they found discoverying calculus.
II Age, Motion view: Needless to say because power series are yet the less understood, and internal growth and biological scalar series ignored, the Newtonian approach had less obvious uses than the approach of calculating locomotions in time sequences and spatial external evident growth in intergral forms. So from Leibniz to Heaviside its methods became the fundamental applications to physics of locomotion, and its two essential dimotions, Ts, and TT (locomotion and entropy), which became the magic of calculus. While the level of complexity of ∆∫∂ studies is maintained in strict realist basis, as physicists try to correspond those finitesimals and wholes with experimentally sound observations of the real world at the close range of Planes in which humans perceive. While the formalism of its functions is built from Leibniz’s finitesimal 1/n analysis to the work of Heaviside with vectors and ∇ functions. Partial derivatives are kept then at the ‘holographic level’ of 2 dimensions (second derivatives on ∆±2).
∆ will be thus the general symbol of the 5^{th} dimension of mental wholes or social dimension and ∫∂ the symbol of the 4th dimension of aggregate finitesimals or entropic dimension.
III Age, spatial view: from Riemann and Einstein to the present. The extension of analysis happens to infinite dimensions with the help of the work of Riemann and Hilbert, applied by Einstein and quantum physicists to the study of Planes of reality beyond our direct perception (∆≥3).
This implies that physicists according to 5D metrics, P$t x Tƒ=K must describe much larger structures in space extension and time duration (astrophysics) and vice versa, much faster populous groups of T.œs in the quantum realm; so ‘functionals’ – functions of functions – ad new dimensions of time, and Hilbert quasiinfinite spaces and statistical methods of collecting quasiinfinite populations are required in the relentless pursuit of huminds for an allcomprehensive ‘mental metric’ of a block of timespace, where all the potential histories and worldcycles of all the entities they study can be ‘mapped’.
The impressive results obtained with those exhaustive mappings bare witness of the modern civilisation based in the manipulation wholesale of electronic particles, but the extreme ‘compression’ of so huge populations in time and space blurs its ‘comprehension’ in ‘realist’ terms, and so the age of ‘idealist science’, spearheaded by Hilbert’s imagination of points lines and congruences detaches mathematical physics and by extension analysis from reality.
±¡: The digital and existential era, is the last age of humind mathematics, where Computers will carry this confusing from the conceptual perspective, detailed from the manipulative point of view, Analysis to its quantitative exhaustion. But for ethic reasons, as a ‘vital humind’, we shall not comment or advance the evolution of the future species that is making us obsolete.
Instead we consider a different version of calculus of change – existential algebra.
The generator equation of Analysis’ ages.
A Generator equation of Analysis in time and scale resumes the 3±∆ fields of the scalar Universe through mathematical mirrors:
Γ Analysis: ∆i: Fractal Mathematics (discontinuous analysis of finitesimals) < Analysis – Integrals and differential equations (∆º±1: continuous=organic space): youth: ODEs<∑ØPDEs<∑∑ Functionals ≈ < ∆+i: Polynomials (diminishing information on wholes).
The 3±∆ approaches of mathematical mirrors to observe the Planes of reality is thus clear: Fractal maths focuses on the point of view of the finitesimals, and its growing quantity of information, enlarging the perspective of the @observer as we probe, enlarging smaller Planes of smaller finitesimals; and in the opposite range polynomials observer larger Planes with restriction of solutions, as basically the wholes we observe are symmetric within its internal equations, and the easiest solutions are those of a perfect holographic bidimesional structure (where even polynomials can be reduced to products of 2manifolds).
Now within analysis proper, we find that the complexity or rather ‘range’ of phenomena studied by each age of analysis increases, from single variables (ODEs) to multiple variables (PDEs) to functions of functions (Functionals).
So the most balanced, extended field is that of differential equations focused on the ∆±1 organic (hence neither lineal not vortex like but balanced S=T), PLANES of the being, where we focus on finding the precise finitesimal that we can then integrate properly guided by the function of growth of the system. And we distinguish then ODE, where we probe a single ST symmetry or PDE obviously the best mirror, as we extend our analysis to multiple S and T dimensions and multiple STST variations of those STep motions; given the fact that a ‘chain of dimensions’ do not fair well beyond the 3 ‘sss’, distanceareavolume dimensions of space and tttt deceleration lineal motioncyclical motion acceleration related time motions that can ‘change’ a given event of spacetime.
So further ODE derivatives are only significant to observe the differences between the differential and/or fractal and polynomial approaches – this last comparison, well established as an essential method of mathematics, worth to mention in this intro.
A space of formal ¬Algebra thus is a function of space, which can be displayed as a continuous sum of infinitesimals across a plane of spacetime of a higher dimension.
In such a geography of Disomorphic spacetime the number of dimension matters to obtain different operations but we are just gliding on the simpler notions of the duality ¬Algebra=polynomials vs. Analysis: integrals of infinitesimals.
Yet soon the enormous extension of ‘events’ that happen between the 3 ∆±1 planes of T.œs as forms of entropic devolution or informative evolution across ∆±i, converted analysis in a bulky stience much larger than the study of an STsingle plane of geometry, the 2 planes of topology and the polynomials of ¬Algebra – which roughly speaking are an approximation to the more subtle methods of finding dimensional change proper of analysis – even if huminds found first the unfocused polynomials and so we call today Taylor’s formulae of multiple derivatives, approximations to Polynomials.
Since Derivatives & integrals often transcend planes relating wholes and parts, studying change of complex organic structures through its internal changes in ages and form.
Polynomials are better suited for simpler systems, Planes of social herds and dimensional volumes of space, with a ‘lineal’ social structure of simple growth.
So in principle Analysis was a subdiscipline of ¬Algebra. But as always happens, time increases the informative complexity of systems and refines closer to a better linguistic focus with finer details the first steps of the mind. So ¬Algebra became with Analysis more precise, measuring dimensional polynomials and its finite steps.
In any case such huge size of ∆nalysis is a clear proof that in mathematics and physics the ∆ST elements of reality are also its underlying structure.
As such since ∆Planes are the less evident components of the Universe, Analysis took long to appear, till humans did not discovered microscopes to see those planes but while maths has dealt with the relativism of human individual planes of existence, philosophy has yet to understand Leibniz’s dictum upon discovery ‘finitesimals’, 1/n, mirror reflections of the (in)finite whole, n: ‘every pointmonad is a world in itself’.
Analysis was already embedded in the Greek Philosophical age, in the disquisition about Universals and Individuals. Thus a brief account of Analysis in its 3±1 ages, through its timegenerator:
Ps (youth: Greek age) < St: Maturity (calculus) > T (informative age: Analysis) >∆+1:emergence: Functionals (Hilbert Spaces)<∆1: Humind death: Digital Chip thought…
Whereas its 3 ‘Planes’ are: ∆1: Derivatives > ∫∆: integrals > ∆+1 differential equations.
Thus Analysis also studies the scales of the 5th dimension and its evolution of parts into wholes.
Derivative vs. integral in time and space.
Derivatives & integrals calculate ratios of change in the 3 elements of reality, space, time=change and its Planes, ∆ST The main error of ‘axiomatic’ analysis is to force continuity, infinitesimals and infinities, without considering the discontinuous limits of derivatives and integrals of parts and wholes between planes of existence.
Leibniz vs. Newton already argued what is an infinitesimal part, the unit of derivatives. The answer in a Universe, which is a fractal scalar system of stœps of spacetime, S<T>S, or S=S=S, T=T=T, is a minimal quanta in SCALE, TIME or SPACE, which by virtue of 5D metric, SxT=K, will be a ‘minimal unit’ on that equation (Min. S x Max. T), for the quanta of space, Min. T x Max. S for the quanta of frequency, or minimal cyclical bit of information.
So for example the inverse, ƒ=1/T of a long duration in time, will be its short quanta. The inverse of a population, taken as a whole, 1, 1/n will be a scalar space quanta. So all systems have an infinitesimal, which is a cutoff limit, NOT really an ‘infinitely small’ (an error of the continuous dogma of the axiomatic method), but a ‘finitesimal’. So we can obtain through a derivative a finitesimal unit of time, space or scale, a minimal action, a minimal pointvolume or a minimal cellular quanta.
And for that reason we can approach finitesimals with ‘differentials’, which become the minimal ‘lineal steps’ of a long curve. And for that reason we can use ‘affine functions in space’ and lineal approximations, as by definition the shortest path between two points is a line, and so in a discontinuous Universe of Stœps, quanta are minimal lineal units; minimal fractal points of a population or a whole.
It is a complete overhauling of the dogmatic attempts to prove the ‘hypothesis of the continuum’, but it is not my fault that humans are so offtrack with the reality of the Universe as it is, not as their ego tries to impose.
The minimal quanta of ∆1 space and time. Chains of Dimotions expressed as chains of equations.
Derivatives are the essential quantitative minimal action absorbed by any Tœ, (ab. spacetime organism).
Integrals then sum a minimal derivative quanta in space or a minimal action in time for any being in existence.
They are best for spatial growth of information as the 3 stages or states of the being through its world cycle of existence, have discontinuities or changes of phase that cannot be integrated. Hence time sequences are better studied with existential algebra. Further on sequences can be come more complex, if we consider tridimensional actions as combinations of S and T states, stt, tst, tss, sss which is the origin among other things of the 3!=6 variations of species according to hierarchy on its physiological networks studied in trinity.
Further on, and this will constantly be the limit of mathematical analysis of reality, we should stress once more than the larger more complex actions of gender reproduction and social evolution are qualitative, taking place both in longer time spans and longer spatial surfaces – hence better described with qualitative logic languages.
So existential algebra studies in depth the qualitative connection of the a,e,i,o,u actions between Planes. That qualitative analysis at the larger scale and for sequences that imply changes of state cannot be overlooked and require the approach of existential algebra.
And calculus is its mathematical, analytic development (quantitative understanding of 1st second and 3rd derivatives – extracting ‘1,2,3 Dimotions’ from the invisible gravitational, light spacetime or feeding Planes.
The limit of calculus however is bridged by the fact that the other families of social operands (±, x÷, x^{a} can reflect better the ‘social reproductive actions’.
So another duality way to differentiate algebraic operands is to consider that classic polynomial operands mirror social complex actions, and calculus operands reflect better also in ‘trinity scales’ (3 first derivatives or ternary integrals) the simplex actions. Each Operand specializes in one Dimotion (angular sine/cosine in Perception, ± in back and forth locomotions, x ÷ in complementary and social evolution, log xª in reproduction) and OVER all of them a new Plane of existence is accessed by analysis. So operands guide the mathematical equations through a vital process of stœps (stops and steps) and will allow us to ‘vitalize’ equations, as we have done with points with ‘numerical parts’ as the essence of a mathematical T.œ
¬Æ thus sets a limited number of logic propositions that can happen when a system or group of T.œs interact through its 5 Ðimotions, as an point of view, can potentially change its state between those 5 Dimotions, and the limits of its function of existence, such as the being can only exists without permanent disruption of its ‘vital constants’, (conserved energy, angular and lineal momenta – energy and membrain). All systems can exist with the infinite cutoff limits of space (membrain) and time (death), which are set as part of the fractal Universe. Only the whole if potential or real in existence can be talked off as a function of infinity but not perceived.
So a point will start any of the 5 Ðimotions and we need formal symbols to address the Ðimotion of any being in existence, and the states of switch between Ðimotions.
Does the being stop before switching Ðimotion? If so it would simple to establish then for each sequential steps of a being:
∆¡ Ð1,3,2,4,3…. and so on as a simple 5 letter process of the actions of a being (whereas 4Ð entropy refers to feedings not dean, only in it final state being ‘that entropy’… So we know all sequential of a being ends in 4Ð.
Can then we run a sequential for any species through its life as a complete deterministic sequence?
There is there the sequence of all sequences, the perfect worldcycle=life sequence?
Questions those for advanced existential algebra.
RECAP. Actions in timespace are the main finitesimal part of reality, its quantity of time or space. In pentalogic operands mirror as actions the 5D vowels (a,e,i,o,u) that define the five dimotions of existence. In calculus they first extract the minimal timespace quanta of the actions of the being, integrating them across a population of space or a length of time. Thus actions vitalize the operands of calculus, relating them to existential algebra.
The fantasy of the continuum substitutes the reality of discontinuous sum.
The age of calculus represents a great advance over simpler polynomial operands and statistical≈Tprobabilistic methods of studying, the parts and wholes of a system; and its S=T interaction between its spatial form and temporal motions; as it differentiated by studying the change of each of the previous 5 Dimotional operands of algebra, all modes of change=time natural to any system of the Universe – as the best mathematical language that mirrors them.
This duality of ∆scales and S=T dimotions represented in algebra by numerical systems and operands was studied in a bulk manner with polynomials and Statistics=Tprobabilistic methods that appeared first, using inverse scalar operands on the ladder of 3 scales of numbers, the sum, the product and different exponentials and inverse roots and logarithms – which reduced the range of exponentials to the 3 scales that matter to the Universe, the Log_{10} or social evolution of triads into 3×3+I new decametric scales, the exponential Ln of maximal growth or death=decay and the log_{2} or ‘power set’ of all parts of a whole.
Probability showed in time (frequency events) or space (statistical populations) that events were not perfect, but accumulated errors in its reproduction, shaping a normal distribution often converted into an entropic population of disconnected individuals forming herds or events that withered away due to those errors of reproduction, when trying to reproduce the perfect mean; to form an identical statistical ‘boson’ species, a perfect form that transcended scale – a feat of becoming a whole ∆+1 that was only achieved in Dirac’s distributions by immortal pointparticles or resonances that amplified the perfect information of ∑ finitesimal 0’s to become the whole ∆+1.
But what about all the range of variations between the entropic Gauss curve and the Dirac perfection? All the different specific actions of species acting in holographic Ts, ST, St dimotions that were neither perfect resonances of evolving form, or disaggregated herds of entropic aleatory motions?
This required to calculate the finitesimal form of change=time of each specific dimotion of the Universe mirrored by each specific operand, and add a given number of ‘stœps’ that measured the repetition of such minimal dimotions of a t.œ in a given length of time or volume of space to get the outcome of an existential dimotion.
Thus by calculating each finitesimal and then integrate it through the entire interval in space, time or scale in which it was performed, calculus gets a more accurate depiction of the whole event, no longer an entropic aleatory change, but a purposeful action. And because both derivatives and integrals could be done in space, time or scale, all the range of possible variations, could be studied, regardless of complexity, from changes of time, to variations in space represented as 0’ actions =, d>0’; happening in minimal time (Lagrangians)…
The detail and specification reached by the finitesimal of an specific dimotion=operand of time=change; either angular perception – sin/cos; social evolution, ±; re=production or locomotion as reproduction of information, x÷; entropic decay and growth, e^{x}, made calculus the queen of all operands, almost a magic tool, as its foundations were ignored.
Mathematicians unable to understand time=change invented the concept of limit and the continuum to form a pedantic scaffolding of axiomatic truths which had nothing to do with the reality of calculus. So we have to clarify that concept. As all is perception a continuum is merely a sum of discontinuous stœps in which part of the whole process is hidden by the selection of information by a mental space. So the fantasy of the ‘continuum’ can be reached when the detail of each stœp is ignored, and we obtain only the measure of a relative infinity, ∝, of such steps summoned up, ∫, to calculate the whole change in the long time period, T, or total domain studied. And that is fine, as long as we understand that in detail the continuum is made of a sum of discontinuous stœps of change, we call finitesimals.
In a process of calculus thus the sum is ‘smoothed’, eliminating the ‘stop’ states, by reducing to the minimal 0’ the size or rate of change, so they cannot be seen in detail but can be ‘integrated’ in a long period of time, obtaining a meaningful result for the ∆0 scale of the experimenter, which only went down to ∆1 to be able to calculate with accuracy ∑∆1=∆º – the emergence of change in the size scale of the experimenter.
So in the same way we care little for the ∆1<∆ø stage of the seminal reproductive fetus – to the point humans have the right to ‘murder’ it, as long as it is not an emergent child, the observer cares little for the finitesimal, ¡ndifferent to him. And yet imprescindible for the whole to be=come.
This ‘goal oriented’ view of Nature is common to most beings; hence calculus became a much better method that discrete probabilities of change. Since the cumbersome reality of each stop and step of a dimotion that had to be analyzed in each individual stœp as it happens in the ∆1 reality could be simplified to facilitate the study of changes through a longer period of time.
So calculus was enormously advantageous to calculate long stretches of time periods and space volumes at the human scale with minimal loss of detail, providing that the philosophy of reality made of stops and steps was not forgotten. Or else we would enter into an idealization of reality confusing the mathematical mirror with the Universe as it is. But that was precisely what it happened. Since the essential error of all minds is to confuse the Universe they don’t fully perceive with the mind that reduces it.
So the wrong creationist solution was taken as truth, obscuring our understanding of the principles of reality. So pundits of calculus ignored the nature of finitesimals of change, and its small stop and motion, S<T>S beats.
So when you study calculus they apologize for reality as it is – a series of thin steps and stops, of small finitesimals to add. And consider discontinuity – the real broken spacetime an approximation to idealist simplification of a continuous graph that looses information about each finitesimal stœp. Instead the artifact, the continuous function is considered the truth and reality a method to reach the ideal truth.
As the graph of Descartes, the artifact, became the nature of space and time, the substances of which we are all made. And suddenly humans were not longer made of cyclical time and fractal space, broken by the limit of a membrain, but God had drawn a Cartesian graph below as background absolute Newtonian lineal spacetime – something most scientists still believe due to creationist mathematics, and the egocy paradox that confuses the mind language with reality itself; the simplification of calculus that smoothed the finitesimals became reality.
Fact is h finitesimals are real, and never reach a 0 that doesn’t exist as horror vacuum works, and so mathematicians should teach reality and then acknowledge that the experimental language of mathematics simplifies that reality; and that is Ok for practical reasons. Not the other way around.
When I was a wonderkid before high school I had a math professor who came one day worried about the fact he had to explain us the ‘limit’ of h>0 and thought his students wouldn’t understand. Of course students never understand a dot of it, just memorize and say they understand. But I understood it was all wrong. Thus he ended up throwing me out of the class (: when I told him, the limit can never reach 0. Since If nothing exists, there is no way to define it, hence calculate it. Nothing might be anything. Only if something is no longer zero but leaves a trace we can define it. So I told him undefined things should not be the realm of an ‘exact’ science. That blew him away and having no answer he resorted to authority (: So I was kicked out as I wouldn’t ever yield to authority but reason. It was my first realization that when humans don’t understand something or something is wrong but they want it anyway they put up a dogma, a postulate, a pedantic definition or ‘selfevident axiom’ (:
Fact is if h>0 makes an equation undefined, h must stop before it looses the quality of the whole – hence h is the last atom, last cell, last frequency, last temperature vibration.
As he just ignored me with the pretentious authoritas of an old man who doesn’t see an elephant in the drawing of a hat, le petit prince has to search for himself. Which is what I have always done. So the question is where to stop a finitesimal portion. And the answer is as we said, before it becomes random, ¡ndifferent.
For example e is defined as (1+1/n)^{n }. Yet if n is ∞, the parenthesis is 1+ 0 .0000000… up to infinity. And those are undefined limits of reality.
Physicists at least acknowledge physical laws are idealizations of reality and that is Ok (even if they deny nonmathematical properties that ‘cannot be measured as tge organic and sentient properties of particles, the units of life). So their egocy paradox is a bit different from that of mathematicians with its axiomatic truths.
Still more profound mathematicians do understand a continuous function as one in which S=T happens, so the X and Y coordinates do not make very different changes, and stœps of ‘present’ can be put one after another. Or as Leonardo said: “The instant does not have time; time is made from the movement of the instant. In rivers, the water that you touch is the last of what has passed, and the first of that which comes. So with time present. Observe the light. Blink your eye and look at it again. That which you see was not there at first, and that which was there is no more.”
As the second, the glimpse of an eye is the quanta of human present, there is always for all timespaces a quanta we shall call an instant of timepresent or a finitesimal of populations.
And when the ratio of change in time, its quanta, is balanced to the minimal form of space it change, so in an S=T graph each pointpositionstop of the movingstep function is close, with a ‘tangent angle, s/t’ that can be smoothed in a series… the sum of discontinuous stœps of a dimotion, can be measured as a continuous, larger stœp of spacetime of a larger ∆+1 scale.
Thus continuity is not the limit in which h>0, but the limit in which S≈T…. and hence an instant of present change that is harmonious between the S and T components of the being that doesn’t change internally but only externally (as S=T remains unchanged) takes place.
This has deep implications, as it implies that the system is continuous because it lasts, and it lasts because its ‘actions’ tend to zero internal change, becoming conserved cycles of energy for the inner structure of the being. Changes thus are always ‘returning to balance’. And when change is extreme, as in an internal x external TTentropic change, the system collapses, the ‘tangent’ tends to zero or infinite in the Y(s) or X(t) axis and the ‘function ends’. Which implies that calculus works on the St, ST and Ts dimotions of locomotion, information and energy=reproduction, NOT on SSform with no change or TTabsolute change with no form.
For example, the first dimotion to calculate as change was the space traversed by a system which in detail is just a series of steps, which add to the space traversed, l(s) x ƒ (t)=S. This give us a lot of detail in numerical approximations if we further break it to a sum of steps, ∑ l(s¡) x ∑ƒ (te)=S.
So the more information we want, the more detailed the discontinuity becomes, till we can indeed add each ‘fractal step’ with all the information, length and time duration of each step. But we don’t want that much information, so we can simplify with ‘statistical means’ since as we have seen in statistics, the law of great numbers bring a mean for each stœp and that is the justification of simplifying into continuity – NOT the non experimental idealist, mindgenerated, creationist hypothesis of the continuum.
So it is fine to be humble and marvel at the fact we can ‘transcend’ the ∆1 finitesimal scale, into parameters of the whole by making ¡ndifferent the information at ∆1 through statistical methods that average each step, erasing the uneeded information on each stop (when the motion touches the floor, or the mover looks and gathers information). But when trying to understanding paradoxes such as the speed of light constancy, as each electron emits light in a relative entangled stop position to the perceiver that measures its speed, hence in a stop distance, (Lorentz transformations) it is good to know that there is no magic on it. And the ‘idealized’ form is the mathematical transformation that eliminates the stop state of the electron for a continuous motion we do NOT observe in Nature (the electron is always observed as a stop particle when emitting light, and moves in zigzag as if it were all the time calculating its trajectory in a stop position).
The methods of calculus are awesome and once we realize the cruelty of the Universe that cares nothing for the differences between the ¡ndifferent finitesimals of a massgroup herded and ruled by the ∆0 larger scale by massmethods, including humans as we show in the models of history and economics herded today by financiers with credit ratios established by anonymous big data computers, herded by politicians with equalizing laws for each ‘social class’, herded by military as soldiers or numbers of a concentration camp… and past over the thought that each atom might be a galaxy of its own, and dare to explode it in an accelerator with other atom, which might on the microscopic scale provoke the biggest genocide of infinite relative planets, and connect with the Tao and feel just to be, ∆@st, which matters nothing and worship the ¡ogic of GoÐoG the inverse Dimotions of existence, we can rest in peace, R.I.P. as dust of space time that dust shall become. And so calculus…
The methods then are well known and we cannot but make a few comments beyond the philosophy of its science, which is the main purpose of 5D mathematics in this simplified texts. We just need to calculate a finitesimal, which was the first thing discovered in the ‘1^{st} age of calculus’ and summon them up to get the whole, which was first done as in reality, with the exhaustion method by the Greeks.
Yet the finitesimal 0’ in itself is important as it give us information about the rate of change with a single number, the tangent to the curve. Since we can apply to curvature in space its synonymous in time – speed – when we realize that s=T means in terms of curves the curve’s tangent representing the ‘speed of change’ S/T of the function. Further on as SxT=C, means that the smaller space is faster in time cycles, the more curved a cycle is the smaller it becomes and the faster it moves (vortex equation, acceleration principle in Relativity over a curved space). So we also realize, the second tangent of the curve is the acceleration of change, y”, increased as the system curves further.
Many more wonders then kept appearing in calculus as we play with S=T, SxT=C 5 Metric laws, and see them through calculus; specially considering that most curves are the functions of existence of an event between its 0’ points of initial and final conditions. And so what we shall do in this brief introduction to 5D calculus is to highlight for the seemingly most simple equations of calculus the underlying insights they provide on the processes of time change of the 5 DImotions of the Universe.
We shall do it in a historic, easier to understand narration, as indeed, the first thing a language sees is ‘space’, and ‘reality as It is’ in its simpler terms. So the first age of calculus was that of the search for finitesimals both in praxis and meaning, wrongly resolved in favor of the concept of absolute zero and limit, then mathematicians erased uneeded information on those finitesimal steps and stops establishing the method of tangents and continous sums, ∫, and then applied those functions of breaking a whole into parts to rebuild it to specification to the different dimotional operands from the sin to the exponential of maximal change, marveling that the limit of change per unit of frequency time was the change of the whole, that is the equation of death in which the whole dissolves into its parts in a single quanta of time, the negative exponential so ever pervading in studies of entropic death. Of course they understood none of it – they still don’t, but the method worked to mirror the dimotions of reality, which they neither understood in an orderly manner. So calculus became magic. And as it got more complex, as the Universe does by repetitions, transformations, scaling into more complex ‘packages’ of parts, to the point that I could say, as Einstein put it – I don’t understand relativity since mathematicians got into it (: I don’t understand calculus since mathematicians got to it 🙂 . That is calculus today is so complicated in its more powerful and detailed analysis that only a computer can calculate its results. Which ultimately try to anticipate the future of change in a synoptic manner by transforming sequential patterns of change into parallel simultaneous spatial components (multiple variables in PDEs), happening in multiple points of view at the same time, to gather into a whole result, which is still impossible for the most intelligent, liquid states of multiple changes (Navier stokes equations) unless you trick it with a faster digital mind, which will do those calculus for the slower humind, jumping as we always do past the intermediate sequential steps from beginning to end.
The proper concept for finitesimals. Reproductive unit of change.
The great advance of calculus in the understanding of ∆ST changes is the concept of a finitesimal of change, h, which in the symmetry between scales, populations in space and time frequencies has the same role: to increase a ‘seminal’ unit, the system (or decreate in inverse fashion), becoming the Unit of reproduction of an ∆st system, at a point in which s=T.
In the Universe the fundamental form of change happens when S=T, the function of present time finds a balance and symmetry in scal,e form and motion that triggers the reproduction of the sytem and its 3 parameters.
Latter we will study, the simplest case of polynomial reproduction, whereas an X^{2} has as unit of change, 2x; which means the square grows through both sides reproducing its form, in ‘s=t’ balance, in a manner that the square preserves its form.
The method of calculus.
How differential equations show us the different actions of the Universe?
The correspondence to establish is between the final result, the åction, and the finitesimal quantas, the system has absorbed to perform the action, ∫∂x, such as: å= ∫ ∂x, whereas x is a quanta of time or space used by ∆ø, through the action, å to perform an event of acceleration, energy feeding, information, offspring reproduction or universal social evolution.
It is then when we can establish how calculus operations are performed to achieve each type of actions.
First we notice that the space between the actor and the observable quanta is relative, so even if there are multiple ∆planes between them the actor will treat the quanta as a direct finitesimal, pixel, bit, or bite which it then will integrate with a polynomial derivative or sinusoidal function that reflects the changes produced.
We will consider in this introductory course only a few of the finitesimal ∫∂ actions where the space state is provided by the integral and the ∂ finitesimal action by the derivative.
Derivatives point out to the main consequence of the sum of those actions in any being in existence, namely the fact that its sums tend to favor growth of information on the being and then signal the 3 stages and/or states of the being through its world cycle of existence, which in its simplest physical equations is the origin of… the maximal and minimal points of a wellbehaved function.
So to establish the action – the final result – we have to isolate the finitesimal quanta/moment of spacetime the system has absorbed to perform the action, ∂x, and integrate them over a surface of space or a length of time, such as: å= ∫ ∂x, whereas x is a moment/quanta of time or space used in repeated frequencies or quantities, ∫∂x, by ∆ø, through the action, å to perform an event of acceleration, energy feeding, information, offspring reproduction or universal social evolution.
We can then establish which operand is best suit to perform each type of actions. I.e. the action of reproduction, most often is expressed for quantitative simple physical systems through the operation of re=production.
We ascribe each operand to a single dimotion, but they are ‘once more’ entangled operations, which besides its preferential Dimotion, do participate of all the others – remember languages as mirrors of reality have also the same entangled properties of the pentalogic, ¬∆@ST universe, looking at all its elements. So we shall now analyze them in more depth.
We establish direct relationships of operands and actions taking into account that for each operand we must also distinguish the dualities of ‘spacelike integral of volumes and its derivative quanta’ and ‘timelike moments of motions and its frequency sum to complete a osum worldcycle’. And to achieve those balanced 0’ sums finally we need to define inverse operations for all actions. So we depart from a ceteris paribus analysis and search for a finitesimal, and then we must study how they merge and entangle in space and time. This is done generally speaking, with a first partial derivative in space or time (PDE) defines those dimotions only as S or T, while the integral of double derivatives put both processes together, to find the whole action: å(st) = ∫∫dsdt
However as all planes of existence have discontinuities beyond its minimal quanta and larger whole, analysis through multiple Planes beyond those of ∆ø<<∑∆1 entropic death, tend to be distorted.
Still they can be studied with power polynomials and further approached (Taylor series) with ∫∂ operators that cross planes of existence for certain highly symmetric actions across.
But again it is best to use existential algebra, as the fundamental limit of the mathematical language is one of synthetic understanding of the organic vital laws of the Universe, reason why theories that are only mathematical in the largest scales and do NOT understand that there are not equations that go to infinity, as all have a limit that brings a change of state, such as the bigbang theory of the universe, are false.
Reason why systems do have besides spatial mental spaces of ‘calculus’, a longtime range language of logic nature to express the vital games of worldcycles, and this is the function of existential algebra, we study first to then consider the basics of Calculus.
Finally in this brief introduction to notice a ‘revealing’ fact of the inversion between finitesimals and integrals. As the absolute arrow of timefuture is social evolution of parts into wholes, while a function has only a derivative, that is, all molecules can be reduced to a set of atoms; all living beings to the cell; the opposite is not truth: creation of complex futures is multiple. So an integral has a C variational constant and a differential equation multiple solutions: the future is open, the past is only one.
RECAP. Analysis studies the finitesimal quanta of time, space and scale; NO ∞ in its smallness – an error of the mind Px. searching for continuity. A true philosophy of calculus thus deals with the meaning of ‘finitesimals’ in space, time and scale, as a first ‘seed’ of a ‘clone species’ multiplies, creating the regularities of ‘social numbers’ that make ‘analysis’ to work its ‘magic. Integrals in inverse fashion act after ‘calculating’ this minimal point, often as a ‘lineal shortest step’ (differentials), to reach the final ‘whole value’ of the system. The beauty of the field revealing the nature of dimotions and its wide applications, thus will require an entire II book on 5D mathematics, which should r=evolve the discipline.
Thus only the integral and derivative can study all those dimotions of spacetime, hence they are the king and queen of the operators of ¬Algebra, reason why analysis is so extended.
Below, Analysis’ multiple perspectives on the 5 Dimotions= functions of existence & 5 simultaneous structural elements, ¬∆@S≈T, that conform all systems in time and space. So 3±D¡ points of view (trinity or pentalogic) finds a higher truth & applies to all languages mirroring reality as analysis does:
Stopology: Analysis is used to study (left) structurally the role of the 3 elements of a topologic spatial superorganism: Its membrane’s curvature and tangential value (line integrals), its vital space (surface integrals) and singularity (derivatives).
∆Planes: Its inverse operands study 5D Planes: derivative measure the value of 1 of its infinitesimal ‘cells’. Integrals give us its internal volume of spatial energy. While double derivatives peers down 2 ∆±1 planes
@: We extract information on its central @singularity, which commands the lineal motion of the whole system.
Timecycles: it can model the standing points, maximal and minimal, which signal the changes between ages, where the derivatives, become null, as the ‘world cycle of existence’ changes its ‘phase’.
DIFFERENT GEOMETRY AND METHODS DEPENDING ON DIMOTIONS.
Because calculus is about Dimotions, it studies mostly the 2 dimotions with a larger content of T, TTentropy and Tslocomotion. And it is also suited to study S=T, iterative change, in a presentreproductive wave of growth.
It is of lesser use to study SS and St, changes in spatial information, though sinusoidal operands give us insights on it. It is then obvious that for a future 5D researcher, if there is ever anyone besides this writer, the observance of the structure of an equation of calculus and its geometry will give insights about the type of dimotion it studies; since there are some basic differences between them:
Tslocomotion: A clear difference happens at first sight between a simpler analysis of locomotion, which concerns a single pointlike form through space in sequential, lineal time, as there is no internal dimotion, and the T.œ can be treated from the point of view of its mindwhole singularity (so for example a moving rock, regardless of rotations in its lineal motion can be treated as the motion of its gravitational center)
– Entropic motion on the other hand is a dual motion, internal and external to the being. Entropic motions then are easier treated if we consider the point of explosion or death as a fixed point (which is often the case as entropy happens after death, which leaves the whole system unchanged in motion). Then we shall observe that the integral of all the motions of the entropic system remains zero, because the negative sides of the frame of reference cancel those dimotions in the positive side, which is essentially the meaning of death, and so the fundamental change of an entropic motion happens in the volume of space of the system which is where the internal dimotion of the being ends up, transformed into an external dimotion.
– Reproductive growth coincides with entropic motion in the factor of expansion in space. However reproductive growth is a real growth that fills space, NOT merely expands the distribution of its ∆1 elementary parts on the background ∆¡ space. So the differences with entropic motion are easy to spot: Reproductive growth does NOT change the density of form in the vital space it fills. Entropic motion becomes rarefied in its dwindling density, a bubble that expands and then dissolves. Reproductive growth is far slower, unless it happens in a truly friendly dense in energy placental world where it happens in a geometric 2^{x} factor of maximal growth; but even then it will seem slower than a bigbang if the speed of death is fast. And as death is a collapse in a single quanta of time, two scales down, ∆º«∆2, almost all process of death and decay expand faster in space.
What about systems of multiple timechanges, ‘PDEs’ so to speak not in its how but why existential processes?
Combined reproductive and entropic motion. There is the most important case when we observe a dual sequential process, in which first the death of the system does not seem to change in space, as growth is internal through the radiation of the ‘predator’ species.
This happens in cosmological bigbangs (beta decays, quasar bigbangs, novas and the hypothetical false cosmic bigbang, studied on physical papers), when the death of the system is due to the birth of a denser form of matter (strangelets in sillynilly planets like Earth that do accelerator experiments or star novas, top quark quasars in BCB stars=black holes. A similar processes, whereas death is parallel to the growth of the predator species, inside out (organic death). In all those cases; in its first time sequence, the system becomes less motile and often shrinks in size, as it is being carved inside out, and then in the second phase it explodes in a single quanta of time, as the faster, smaller form or herd of forms spreads on a larger space.
In praxis then you can act as partial differential equations do, just performing two sequential calculus because and that is the beauty of the Universe that facilitates its comprehension as we have repeatedly stated in all our paragraphs on existential time, at the level of actions, we follow a series of finitesimal steps which seem to be continuous (concepts clarified in the next paragraphs). So the dua dimotion of a new form feeding in a T.œs body, to then explode it an expand, ST_{x}»SS_{y}«TT_{x}+Ts_{y }written as a sequence of existential algebra (a body ST_{x }feeds a new species in its seed form, SSy… that will walk away Tsy as the form collapses in entropy TTx); becomes a series of partial derivatives.
In the deepest sense this is the existential why of the methods of calculus of partial derivatives and the key difference in the concepts of continuity in classic calculus and ‘stœps’ of discontinuity, in 5D calculus, bridged with the common concept of a smooth transition through finitesimal changes, which in reality are discrete (the body corrupts in discrete steps even if as the bacteria grow expenentially each step is larger), but from a higher point of view ¡ndifferent to the detail can be calculated as long as it is smooth.
It is also the reason why both, reproductive and entropic motions can be described by e^{±x} functions, which are the maximal ‘rate of change’ (as the derivative is equal to the function, and since ∂x≤x, is maximal).
The e function has so multiple meanings precisely because of the ‘horror vacuum’ and thirst for existence of its spatial fractal points. So it can also be used in its imaginary form, which as the name indicates is related to the creation of ‘mental SS=§paces’, in its e^{xi }form, connected to the sinusoidal functions, in which we can observe, as in AC currents, a back and forth motion=translation in space, coupled with a rotational perceptive motion. In those rotational motions, the complex plane and exponential function is so useful because what perception IS really doing is 1) collapsing at the fastest possible rate the Universe into the finitesimal mindmapping of the point (hence the e^{x} function involved); it does so with a clear bias in favor of the length dimension of the focused perception, while the idimension of height is greatly compressed to fit the system (hence the usefulness of a frame of reference where Y=√x; and finally it does so in a periodic pattern, scanning back and forth the same worldcycles to convert them into mental space; hence the recursive use of ±sin, cosine functions involved.
Let us consider this essential equivalence of mathematics in more detail.
Connection between exponentials and sinusoidal functions: derivatives as angles of perception.
One very realized role of a derivative as a tangential division of the height in the dimension of information and distancelineal motion to the observer is a measure of the angle of the being, which recedes in spacetime till reaching the nonperception as a relative finitesimal out of the territorial mind world of the observer, which connects directly derivatives with the 1D first dimotion of perceptive existence. The being might still be of certain size but as a fractal point he has receded in the mentalspace of the world of the perceiver.
The first ‘timespace’ numbers: Polygons as root of unity
By their very nature, as numbers that probe planes of the fifth dimension, exponentials are closely related to the complex plane. Let us consider only one case, de Moivre numbers, which are any complex number that gives 1 when raised to some positive integer power n:
An nth root of unity, where n is a positive integer (i.e. n = 1, 2, 3, …), is a number z satisfying the equation:
They are complex numbers (including the number 1, and the number –1 if n is even, which are complex with a 0’ imaginary part), and in this case, the nth roots of unity are:
This formula shows that on the complex plane the nth roots of unity are at the vertices of a regular nsided polygon inscribed in the unit circle, with one vertex at 1. This geometric fact accounts for the term “cyclotomic” in cyclotomic polynomial; it is from the Greek roots “cyclo” (circle) plus “tomos” (cut, divide).
Euler’s formula, e^{ix}= cos x + sin ix which is valid for all real x, can be used to put the formula for the nth roots of unity into the form: e^{2πi k/n }0≤k<n. Which is a primitive nthroot if and only if the fraction k/n is in lowest terms, i.e. that k and n are coprime.
We find therefore the first timespace numbers, in the roots of unity. And as such they will become ‘the creative process’ of dividing the ‘whole’, 1, into cyclical ‘tics of time’ of increasingly faster frequency, in a progression, for k = 1, 2, …, n − 1, which will generate the frequencies of all clocks of time, till reaching the circle, which can then be considered in bidimensional spacetime, the ‘Infinite clock, of infinitesimal time tics’. Those infinitesimal ticks have a ‘limit’ as all relative infinites do In physics it is believed the minimal tick will be 10^{43} or Planck’s time, which therefore would become the limit of ‘points’ that form a time clock.
Another fundamental theme being the reasons why the ‘clock’ is counterclockwise in its direction, as it will also be its complex representation in 4D relativity theory. The reason being that in Planes of the fifth dimension, as we create new dimensions from the lower planes with more entropy, the emergent dimension ‘sucks’ part of the entropy of lineal space of its lower dimensions, ‘contracting’ it as it rises on height. I.e. a pi circle is made of 3 ‘curved’ diameters (with open holes between them), but it does not measure 3 but 1 in the length dimension.
It also means we are adding a new time dimension, with a negative entropic property for the ‘dimension of real space’, which therefore can be written also with the number of entropy, e,
Dimotion of Entropy
The relationship between the sine and cosine functions allows an angular perception of the whole and the exponential function that reduces the whole to its decaying elements (Euler’s formula). We could say then that the whole is ‘split’ between the entropic negative exponential part that is discharged, and the sinusoidal, informative elements that are absorbed by the mathematical mirror mapping.
It is interesting to note the connection, which occurs between the exponential and trigonometric functions when we turn to the complex domain, through series, since both functions can be approached by exponential series. If we replace z by iz, we get:Grouping everywhere the terms without the multiplier i and the terms with multiplier i, we have:
Euler’s formulas solved for cos z and sin z, get:
As we said the key insight of 5D in power series is the understanding of them as a series of ‘sequential steps’ in the Ñ∆ dual scalar growth and diminution scales of the fifth dimension, whereas the ± summand element represents a step in timechange, a ‘period’ of a frequency of growth and diminution, but in the case of the use of an I factor it creates a sinusoidal process of a repetitive worldcycle of perception, short of an opening and closening glimpse on reality, a back and forth motion in an AC current, a life and death cycle in a fast time quantum particle. It is an essential insight to resolve the whys of all those hows of mathematical physics.
In the complex plane, 1D (sin/cos) combine to represent a full worldcycles, interesting enough through the e^{ix}, 4D exponential decay function. This is possible because he exponential function switches between growth and negative decrease, as the sine and cosine switch between informative and energetic perception; but the sine function, the informative Dimotion grows less, as it happens in nature, where height and information has less energy, and so in parameters of size and volume matters less.
2 new qualities make interesting to cast trigonometric functions in terms of the function of entropy: we are adding both cosine and sine ‘on and off’ SMH for a value of 1, the total value of a world cycle, so we can use frequency equations (as in electromagnetism) to represent this exponential world cycle. And we superpose both, the function of ‘spaceform’ the sine and timemotionlineal distance, the cosine, to observe a harmonic balance as the function goes up and down but never passes beyond the value of the whole.
The complex plane is real because the cos is related to lineal motion and the sin to perceptive height, whose action in stop mode can be seen as a negative slow down of motion for a continuous view (S=T); as in relativity (cT). But the deepest level of understanding of those functions and equivalences happens when we carry the worldcycle of exist¡ence to the complex plane.
Duality on calculus: ∆Newton v. Só T Leibniz
Finally to notice the extraordinary fact that the STcartesian graph and the complex plane coincide in the root of unity, which essentially divides the being into its internal and outer parts. Only then the membrain can assess with accuracy both realities in objective terms. But it will perceive them with different ‘volume’ of information.
For the membrain the external world will be measured in the complex plane, with a lesser dimension of height that will make the world ‘flat’ in its perceived geometry, as the Y(i) plane will be the √ root of the Xplane.
Internally though for the osingularity which ONLY observes the root of unity circle with an equivalent height and width, this ‘verbal, temporal mind’ observing the ‘spatial biased membrain’ NOT the universe, that has already made a selection of information; in the same manner your internal verbal temporal thought on the 01 unit temporal sphere or the quantum 01 particle on the biased information provided by the harmonic spherics of his electronic eye (remember all is the same, all is homology in function even if it changes in form); will think the Universe is ‘perfectly regular’ and favoring its biased dimotion of length, NOT realizing of the equal important of the flattened dimotion of heightinformation.
Essentially all systems have 2 brains, the spatial membrain and the temporal singularity at the center of the 01 temporal sphere. The spatial membrain already bias reality and as the singularity of time only sees the spatial membrain it will act upon it, as its ‘territory’ in which to enact its 5 dimotions of existence, qualifying reality as the membrain has already done. And that is fine because singularities are selfish selfcentered knots or else they will be preys of other selfcentered knots. But that makes so difficult objective knowledge as we shall see in the bias of huminds. What does then the humind brain observe? The distortion we know exists considering the membrain homunculus for which hand sensations (enzyman’s actions) and mouth (entropic feeding and social communication) occupy most of the space while legslocomotion regardless of physicists ego matter nothing.
Locomotion as reproduction and death. This usefulness of the derivative of maximal rate has a deep philosophical consequence of the many insights a proper understanding of the symmetries between existential algebra and calculus methods provides to the 5D researcher. Consider the graph, which is in fact a trinity sequence of events, as the photon particle (if the wave were to represent a light ray), dies every complete wave and the wave represents its entire worldcycle of existence. But in the process it also translates in space, and reproduces in the point of maximal existential momentum (Max. ST), which if the graph were one of ST not its derivative of change will be at the peak, when the photon in fact is as the ‘head’particle state in the top of the dimension of heightinformation (in static space); but if the wave represents its derivative of change, it will happen in the point in which it touches its axis; where further on the wave is ‘feeding’ on the ‘string of tachyon neutrino’ or quantum potential that guides the wave… So in that brief period between birth and death, what it amounts finally to a 0 ST change in the existential momentum of the wave, there are the points of SSbirth, TTdeath, TTentropic feeding on a lower plane, STreproduction, sTlocomotion and finally the cyclical perception that will happen at the maximal height in the photon state.
Those are the whys of existential algebra for the 5 Dimotions of existence of a wave of light; as even the smallest form of our light spacetime Universe has all the properties of 5 Dimotional life encoded in its mathematical equations; the bridge between those equations and existential algebra being the understanding of mathematics as an experimental science of vital spacetime.
TRILOGIC ON DERIVATIVES AND INTEGRALS:∆ST: THE 3 GREAT FIELDS OF CALCULUS.
Trilogic on calculus. 3 ±¡ ages, scales, and Dimotions mirrored by calculus operands.
As we are made of ∆ST elements, limited by the reach of a supœrganism, selfcentered and expressing the program of the 5 Dimotions of existence in @ mind; any systematic analysis of an organism or language that mirrors it departs from the one – the whole, then explains its Space and time states, its evident duality in a single plane, SóT, then its trinity, as ∆ST, and finally its pentalogic ensemble of ¬∆@st. Variations on those themes might deliver an ¡mmense number of explanations of a subject or species. In the case of calculus though as in most developments of a subject the best consideration is a ternary analysis of its ∆ST elements in the historic growing complexity natural to the evolution of a being through 3 ages.
On the other hand as reality is entangled in ∆±1 scales, Spopulations and Time Dimotions and ages, Calculus has in a synchronous analysis 3 great fields: the study on how systems changes in size and scale through the growth or diminution of its ‘finitesimals’, the study of growth of populations in space, and the study of Time dimotions, which are often based, and this is the miracle of Nature, in the same concept of a fintiesimals.
Let us consider a trilogic example on how analysis’ operands represent those 5 Dimotions.
Spatial view: Analysis as a tool to extract quanta of whole social populations.
The fundamental particle of the Universe is a T.œ. a fractal point or scalar timespace superorganism, which in its simplest, commonest form has the shape of a circle with 3 canonical regions:
@MindCenter, measured by its radius, its axial length=motion around the ‘Territory’ of the organic system.
A membrain of angular momentum, or external clock that we measure as its circumference.
And an area of vital energy, which can be measured by the area.
So we get the value of the 3 elements of a disk, and expanding it to 3d spheres (graph) we get a volume, we find a ‘volume’ for the vital energy, a surface of an sphere for the membrane and a perimeter for the wanderings of the singularity.
As it turns out, the circle’s area is π R^{2}, and the circumference is 2πR, which is the derivative.
The volume of a sphere is V=/3πR^{3}, and the surface area is S=πR^{2}, which is again the derivative.
And inversely, the integral of the circumference is a surface and the surface integral is the volume.
The example shows the main use of analysis in static space: to describe through 3 ‘∆±1 Planes’ the 3 parts of the being, which is the ultimate reason why only 2 derivatives are of practical use to ‘descend’ from the whole down two Planes, to the finitesimal quanta, beyond which an entire new ‘world’ within the quanta appears, with a different content, not suitable to be calculated within the same plane.
So analysis become the essential tool to understand the social dimotions of parts and its growing Planes into wholes of a higher ∆+1 scalar plane of the fifth dimension, in a correspondence between analysis and 5D Planes, motions and populations of space.
Temporal view: Analysis as a measure of a temporal motion.
Yet analysis is most often used in temporal terms. This was though likely its first use (to calculate volumes from areas). It is in fact used to study motion, change in time, and we shall argue also Planes; and in that sense, as we shall repeat ad nauseam, the entangled Universe which shows a clear correspondence between the mirror elements of 3 motions in time, 3 topologies of space and 3 Planes of size, wholes and parts that bring together the 3 x3 (+2 mental) = 11 Dimensions of reality is fully realized in the fact that analysis works to explain the 3 ‘ternary symmetries’.
In the example, we can consider the sphere to be the whole sum of parts, where each part is a circumference. So our planet is the sum of all its ‘parallels’ with center in the poles. And then the volume as each internal sphere can be in terms of 5D metric, $ X ð = K have the same coinvariant value, can be considered the sum of all those equal 5D valued spheres, so again we can talk of ∫∫ ¡1=circumference> ∫ ¡0=sphere>¡+1 = volume.
What about the third ‘ternary symmetry’, that of timechange? This again is the fundamental use today analysis has, to study the rate of changes of a system, and it can be seen easily that the 3 elements of the ‘t.œ’ ARE measures of timechange when we study not a mere locomotion, but the ‘changerate’ of ‘growth’ more proper of the worldcycle of existence from ‘seed’ (the internal minimal sphere’) to emergent system:
If you describe volume, V, in terms of the radius, R, then increasing R will result in an increase in V that’s proportional to the surface area. If the surface area is given by S(R), then you’ll find that for a tiny change in the radius, dR, dV=S(R)dR or dV/dR=S(R),
Increase in volume, dV, is the amount of new ‘cellular layers’ their system grows, and the amount of cells form the membrane, which is the surface area, S(R), times the thickness of the growth, where each unit is a layer, dR.
This same argument can be used to show that the volume is the integral of the surface area (just keep adding layer after layer of atoms or cells).
Finitesimals in Time vs. space
Space is symmetric; in its directions and they coexist together. Time is not symmetric and it is experienced as a sequential pattern of single Time cycles. So Time parameters are shorter in form, space is a more extended system. Of time we see only an instant, of space we integrate instants/cycles of time and sum them as frequencies which all play the same world cycle.
Time though often is just the reproduction of a new unit of space. Thus, time cycles become populations of a spatial herd due to its reproduction of a ‘seed’ form.
Space thus is the ‘mirror reproductive symmetry’ of ‘frequencies in time’, its tail of memories, by reproduction, expansion, and radiation along the path of the singular timeline of the wave.
So in broad strokes derivative and integrals cover a wide range of 5D themes: the infinitesimal units of time frequencies and complex herds of space populations.
Whereas given the simultaneity properties of space, integrals tend to be used to calculate space populations, and given the individual sequential structure of time frequencies, derivatives are best to calculate time motions.
Thus the key concept of 5D mathematical analysis is the finitesimal, which was rightly defined by Leibniz as:
∆: 1/n; the minimal part of a whole.
S: While in space is an individual unit of a social population.
T: While in lineal time duration is the minimal bit of a frequency ƒ=1/t, or quanta of time.
Thus a finitesimal is a discrete minimal unit in any scale of the fifth dimension – hplanckton, cellular units, atomic units.
And by the equivalence between spaceform and timemotion, S=T, as most time actions require a fractal reproduction of form, for each quanta of time, we shall se the existence of a reproduction of a quanta of space…
On the other hand its inverse Integral ‘integrate’ an amount of such units of time, space or scale to obtain a simultaneous whole, a supœrganism, a ‘T.œ’, ∫ds, ∫dt, ∫∆1.
Of those 3 types of derivatives and integrals, as frequency and time duration are inverse parameters currently used in all sciences, the less understood is ∫∆1, whereas ∆1 is taken to be the infinitesimal or minimal quanta of a whole, ∆º, (cell, atom, individual in a society), and its integral, a Social 4Ðimotion that mimics the creation of wholes.
A dual derivative, TT, ort SS, will then extract either an entropic unit 2 scales below the form or as we found in the analysis of the sphere The Point, NOT ANY point but the Center of mass or charge in a physical system, its mind singularity. Because derivatives ‘extract’ the first finitesimal quanta, or fractal point from a function of exist¡ence (T.œ), often directly as in log x’:1/x, it can lead directly to the value of the mind, or ‘center point’ of the system – the ‘finitesimal whole’; and its inverse, an integral, which ads finitesimals till reaching the whole, as in the case of a volume of populations, but also illuminates the dissolution of a whole into its integrating parts.
What kind of point a derivative gives us, depends on the configuration of the whole we analyze. I.e. In a heat equation the whole lacks a center, as it is a flux of kinetic energy, so derivatives will extract any unit…
In 5D analysis depending on what we study ‘motion’, or ‘space’ or ‘scale’ up or down the planes of the 5^{th} dimension we shall apply either an integral commonest for spatial sums of populations or a derivative, most often for instants of time, and double derivatives for reproductive functions. Since space and time are inverse, perpendicular functions, in its min. S x Max. T, and Max. S x Min. T states, but symmetric in S=T. So goes for the 2 different arrows of entropy, a dissolution downwards and social evolution upwards.
So the ∆ST trinity of integrals and derivatives gives a huge range of possible interpretations for the equations of mathematical physics. Infinities though don’t exist, as all has a finite membrane and a finite duration in time. Beyond the third derivative, as the scalar Universe is a ‘ternary game’, there is no significance to the mathematical operations of derivatives and integrals – a strong proof that 5D is truth as it limits reality to ternary Planes, topologies and time ages.
So a qualitative analysis is required to specify what dimotion we are ‘calculating’, with derivatives and integrals: time motions, space populations or reproductive motions.
5d ∫∆1 pentalogic on integrals
∑s1=S0: Integrals, on the other hand represent the growth of a space population, till it reaches a wholeness in a closed domain. So we can do ‘line integrals’, ‘surface integrals’, ‘volume integrals’, in simultaneous space.
Such integrals must be positive in its results, because we are as in the case of + v. – numbers calculating a ‘statistical population in space’.
(to>t) ∑a = 0: Integrals though are also related to a world cycle, as the continuous sum of steps in a sequential duration of time that must therefore have a 0’ final result as all worldcycles when chosen in the apppropiate parameters of ‘energy and information’ end up returning to its origin. Such integrals when properly written must therefore give us a 0 value. The classic case being a sinusoidal function of a wave with positive and negative sides for the worldcycle that ends in a 0 value, when we add the surfaces below and above the curve.
T=S: However when we express those ‘actions=dimotions=stœps’ of the worldcycle with the ‘simpler, first age’ formalism of probability; wheras an individual event is a ‘finiteismal’ of time, and the sum of all events a ‘1 value’ distribution, if we integrate the probability to get the sum of all events, whole entity as an event, which is by convention valued as ‘1’; the result of such integral must be ‘renormalized’ to 1.
This is a complicated way to calculate a 0’worldcycle but as it has become the formalism chosen in quantum physics, it is constantly carried out to calculate the sum of events of an electron that give birth in space to an statistical population of all the potential positions of the electron in space (themselves taken in ∆1 as dense photon points). As the electron in trilogic can be seen as a cloud of ∆1 dense photons, as an ∆o whole in space, or as the sum of the sequential points it occupies in time, but humans are monologic, a lot of confusion is natural to quantum physics, the more so with the addition of further complexity with renormalization methods and probabilistic interpretations.
T=S: Integrals are also necessary to add a locomotion of time, closer to the action of reproduction in space, as nature is ‘constantly building integrated wholes by the accumulation of single time actions of reproduction that become ‘clone’ cellsatomscitizens of an integrated supœrganism.
¬: Integration of any of those actions however needs to be ‘defined’ due to the uncertainty of infinities, by constrains (initial time and final time, or ab interval of domain in space), which act as the integral line membrain, becoming the Riemann integral or ‘Cauchy’ condition for it to have a solution. .
As a function of entropy integrals can also portray the growth or diminution of populations in space, with most of those growth/decay inverse functions, represented by e^{±x }or 10^{±x} which are the standard constants of growth.
They are maximal when a system decreases and the space is dying with no constrain at maximal speed in a quanta of time – hence using the maximal growth of efunction. However when it grows socially it does so slower, most often in decametric scales; so we find also different speeds on the two time dimotion of the 5th dimension.
Recap. Integrals are overwhelmingly the measure of change in a fictious mental space constructed.
1^{st} AGE SCALAR VIEW: FINITESIMALS . UNIVERSALS≈WHOLES
Universals
Perhaps the clearest historic proof of the nature of finitesimals as the parts of wholes is the fact that he beginning of calculus was not related to the study of rates of change in continuous motion but precisely to the relationship between parts into wholes.
So Greeks studied in philosophical terms the integration=growth of a social system from micro to macrocosms, from individuals into Universals, and mathematically through ‘finitesimal’ minimal quanta or parts of the whole, through ‘series’ and exhaustion methods.
This age extended from the Greeks to Newton, which was the last of the ancients, changing the use of those exhaustion methods from spatial series of growth to temporal series of change, but he failed to represent them properly through the space=time symmetry of Y(s)=X(t), in a Cartesian frame as Leibniz did, adding the property of ‘continuity’ as explained before, not the limit in which h>0, but the limit in which S≈T….
Plato maintained that exemplifying a property is a matter of imperfectly copying an entity he called a form, which itself is a perfect or pure instance of the property in question. Several things are red or beautiful, for example, in virtue of their resembling the ideal form of the Red or the Beautiful. Plato’s forms are abstract or transcendent, occupying a realm completely outside space and time. They cannot affect or be affected by any object or event in the physical universe. This is correct, though the error lies in positioning universals outside space and time. They are in fact the ultimate properties of SEspatial ‘kinetic energy+entropy’ and TO Temporal information, which ‘emerge’ in each new scale.
Few philosophers now believe in such a “Platonic heaven,” at least as Plato originally conceived it; the “copying” theory of exemplification is generally rejected. Nevertheless, many modern and contemporary philosophers, including Gottlob Frege, the early Bertrand Russell, Alonzo Church, and George Bealer are properly called “Platonic” realists because they believed in universals that are abstract or transcendent and that do not depend upon the existence of their instances.
They are closer to the truth, but they should substitute the word ‘transcendent’ for ‘emergent’ in the parlance of general systems.
For that matter General Systems (5D ST) reduces the meaning of ‘transcendence’ to its first semantic meaning:
Vb: L transcendere to climb across, transcend, fr. trans + scandere to climb.
vt : to rise above or go beyond the limits.
Indeed, Universals are found beyond the limits of its finitesimals, in the next n+1 scale.
Dimensional growth area finitesimals as: reproduction of spatial form
Finitesimals were first found in space, as the means to quantify a simultaneous areas as the sum of ∆1 discontinuous, fractal parts. Let us remember this concept, key philosophical discussion even with the greeks – it is the Universe continuous or discontinuous, made of Universal wholes or individual parts?
This concept was the earlier idea of Leucipus and Democritus regarding the composition of physical systems; and Anaximander, regarding the composition of life systems, with its ‘homunculus’ concept (we were made of smaller beings)
Anaximenes’ assumption that aer is everlastingly in motion and his analogy between the divine air that sustains the universe and the human “air,” or soul, that animates people is a clear comparison between a macrocosm and a microcosm.
It also permit him to maintain a unity behind diversity as well as to reinforce the view of his contemporaries that there is an overarching principle regulating all life and behavior. So here there is a first bridge that merges universals and finitesimals.
And of earlier mystiques, regarding the composition of a superior God, as the subconscious collective of all its believers’ minds, fusion in a ‘bosonic’ way into the soul of the whole.
The 3 were right as finitesimals are clone beings with properties that transcend into the Universal, being the homunculus the ‘future cell’.
Universal wholes and individual finitesimals.
Because the praxis of continuity was not yet ‘erased by idealism reality’ the Greeks accepted as real their exhaustion methods, but Pythagorism opened the road to idealism. So the first age of analysis had a great deal of philosophical disquisitions on the nature of wholes and parts, connecting directly with the greek logic arguments on the nature of individuals and universals.
The historical origins of analysis can be found in attempts to calculate spatial quantities such as the length of a curved line or the area enclosed by a curve.
As we know, a curve, is always part of a worldcycle, with a finite number of steps, and so the conclusions of those earlier studies can be extended to understand better the spacetime worldcycle in a general way: a circle can be calculated as a polynomial number, which becomes nearly undistinguishable, past the 1020100th ‘fractal points’ stœps of social scales of number all pervading in Nature.
This lead to the exhaustion method of calculating irrational numbers, from parts into wholes.
o1: ∆1: 1/n finitesimal scale vs. 1∞: ∆+1: whole scale.
So only a question of that section is worth to mention here, on how to ‘consider Planes’, which tend to be decametric, good! One of the few things that work right on the human mind and do no have to be adapted to the Universal mind, from d•st to ∆ûst.
Shall we study them downwards, through ‘finitesimal decimal Planes’ or upwards, through decametric, growing ones? The answer is an essential law of Absolute relativity that goes as follows:
‘The study of decametric, §+ Planes (10§≈10•^{10} ∆ ≈ ∆+1) is symmetric to the study of the inverse, decimal ∆>∆1 scale’.
Or in its most reduced ‘formula’: ( ∞ = (1) = 0): (∞1) ≈ (10)
Whereas ∞ is the perception of the whole ‘upwards’ in the domain of 1, the minimal quanta to the relative ∞ of the ∆+1 scale. While 1 is the relative infinite of a system observed downwards, such as ∆+1 (1) is composed of a number of ‘finitesimal parts’ whose minimal quanta is 0.
It is from that concept from where we accept as the best definition of an infinitesimal that of Leibniz: N (whole) = 1/N (Finitesimal).
So in absolute relativity the ∆1 world goes from 1 to 0, and the ∆+1 equivalent concept goes from 1 to ∞. And so now we can also extract of the ‘infinitorum thought receptacle’J a key difference between both mathematical techniques:
A conceptual analysis upwards has a defined lower pointquanta, 1 and an undefined upper ∞ limit. While a downwards analysis has an upper defined whole limit, 1 and an undefined ‘finitesimal minimum, +0).
Finally to notice that as all ∆Planes have relative finitesimal +0 and relative infinities (see ∞º to understand the limits and meaning of numbers and its Planes), essential to all theory of calculus is the study of the domain in which the system works, and the ‘holes’ or singularities and membranes which are not part of the open ballsystem. So functions can be defined with certain singularity points and borders; hence functions need not be defined by single formulas. This would be understood by Leibniz – who else 🙂
Unlike Newton, who made little effort to explain and justify fluxions, Leibniz, as an eminent and highly regarded philosopher, was influential in propagating the idea of finitesimals, which he described as actual numbers—that is, less than 1/n in absolute value for each positive integer n and yet not equal to 0’.
For those who insisted in infinities, Berkeley would reveal those contradictions in the book ‘The Analyst’. There he wrote about fluxions: “They are neither finite quantities, nor quantities infinitely small, nor yet nothing. May we not call them the ghosts of departed quantities?”
Definition of ∆t, ∆s, finitesimals: A quantum of time and space.
Berkeley’s criticism was not fully met until the 19th century, when it was realized that, in the expression dy/dx, dx and dy need not lead an independent existence. Rather, this expression could be defined as the limit of ordinary ratios Δy/Δx.
And here is where we retake it; before the formal age of mathematics, made a ‘pretentiously rigorous definition of infinitesimal limits and the the logician A. Robinson showed the notion of infinitesimal to be logically consistent, but NOT real.
As we believe mathematics must be real to be ‘consistent’ (Gödel’s theorem), we return to the finitesimal concept, ±∆y, either as a ‘real’ increase/decrease of a quantity, with a variation ±∆x of either the surface of space or the duration in time of the being.
Thus finitesimals depend for each species of the ‘quanta’ of space or ‘minimal cell’ and quanta of time or minimal moment, which the system can measure.
For man, for example time actions are measured with its minimal time quanta of a second, below which it is difficult to perceive anything; a nanosecond in that regard in the human plane of existence is NOT worth to measure, as nothing happening in a nanosecond will be perceived as motion or change. For an atom however a nanosecond is a proper finitesimal to measure changes.
In space, man does not perceive sensations below certain limits, which vary for each sense, a millimeter, 100 hertzs of sound, the frequency of infrared waves; and so on.
There was only at this stage a mathematical approach to the concept by Archimedes – the methods of exhaustion to calculate areas and ratios, notably the pi ratio.
The method of exhaustion… was first used by Eudoxus, as a generalization of the theory of proportions.
Eudoxus’ idea was to measure arbitrary objects by defining them as combinations of multiple polygons or polyhedral. In this way, he could compute volumes and areas of many objects with the help of a few shapes, such as triangles and triangular prisms, of known dimensions. For example, by using stacks of prisms (see figure), Eudoxus was able to prove that the volume of a pyramid is onethird of the area of its base B multiplied by its height h, or in modern notation Bh/3.
Loosely speaking, the volume of the pyramid is “exhausted” by stacks of prisms as the thickness of the prisms becomes progressively smaller. More precisely, what Eudoxus proved is that any volume less than Bh/3 may be exceeded by a stack of prisms inside the pyramid, and any volume greater than Bh/3 may be undercut by a stack of prisms containing the pyramid.
The greatest exponent of the method of exhaustion was Archimedes (c. 285–212/211 BC). Among his discoveries using exhaustion were the area of a parabolic segment, the volume of a paraboloid, the tangent to a spiral, and a proof that the volume of a sphere is twothirds the volume of the circumscribing cylinder. His calculation of the area of the parabolic segment (see figure) involved the application of infinite series to geometry. In this case, the infinite geometric series:
1 + 1/4 + 1/16 +1/64 +… = 4/3
is obtained by successively adding a triangle with unit area, then triangles that total 1/4 unit area, then triangles of 1/16, and so forth, until the area is exhausted. Archimedes avoided actual contact with infinity, however, by showing that the series obtained by stopping after a finite number of terms could be made to exceed any number less than 4/3. In modern terms, 4/3 is the limit of the partial sums.
His paper, ‘Measurement of the Circle’ is a fragment of a longer work in which π (pi), the ratio of the circumference to the diameter of a circle, is shown to lie between the limits of 3 10/71 and 3 1/7.
Archimedes’ approach to determining π consists of inscribing and circumscribing regular polygons with a large number of sides. It was followed by everyone until the development of infinite series expansions in India during the 15th century and in Europe during the 17th century. This work also contains accurate approximations (expressed as ratios of integers) to the square roots of 3 and several large numbers.
It is then interesting to consider Archimedes’ main role on the perception of problems today forgotten after the absurd dogmatic germanic ‘foundations under the axiomatic method’ of analysis.
2 problems troubled him and indeed they were very important problems: the comparisons of different pis, (it is the pi square with 2 dimensions the same than the pi of the perimeter) and its proper calculus by approximation.
Approximations in geometry.
The unit of space is the area and the unit of time the cycle, and so both are bidimensional, and hence the transformation of one into another is not always perfect, as there is not a perfect ‘quadrature’. But as this happens constantly a part is lost as ‘entropy’ in all timespace transformations, or as ‘a bit of a circle’, that is a motion or particle, as when in particle reactions there are always ‘forces’ escaping (neutrinos, gammar rays). So this means that pi is not exact, neither √2, the two key constants for the squaring… Yet that doesn’t mean the transformation happens all the time, and it was the way in which the game of analysis started with Archimedes:
The transformation of a circular region into an approximately rectangular region. In graph ∆ST theory eliminates all infinitesimals problems as infinities are limited, so are the 0s, which must be regarded as the +0 minimal quanta of the domain – the need for further infinities is an error of the mind, the dogmatic truth and the single spacetime ‘continuum). In that regard pi is not ∞, but its calculus becomes ‘chaotic’ beyond a limit of ±40 decimals, which is really all what the human mind can conceive n its largest finitesimal analysis.
It is then when the ‘Greek Age’ becomes just as in the Archimedean calculus of pi by exhaustion the same concept, just with less detail.
A simple geometric argument shows that both processes are similar with different degrees of approximation:
In the graph we see how ∆ST theory immediately eliminates all those problems of infinitesimals as all infinities are limited, so are the 0s, which must be regarded as the +0 minimal quanta of the domain – the need for further infinities is an error of the mind, the dogmatic truth and the single spacetime ‘continuum). In that regard pi is not INFINITE, but its calculus becomes ‘chaotic’ beyond a limit of ±40 decimals, which is really all
what the human mind can conceive n its largest finitesimal analysis.
It is then when the ‘Greek Age’ becomes just as in the Archimedean calculus of pi by exhaustion the same concept, just with less detail.
Indeed, a simple geometric argument shows that both processes are similar with different degrees of approximation:
The idea is to slice the circle like a pie, into a large number of equal pieces, and to reassemble the pieces to form an approximate rectangle (see figure). Then the area of the “rectangle” is closely approximated by its height, which equals the circle’s radius, multiplied by the length of one set of curved sides—which together form onehalf of the circle’s circumference. As the slices get very thin, the error in the approximation becomes very small.
The duality of free lines/planes v. closed order.
It is interesting to notice that in general when we grow in scale, we change from freedom to order or vice versa – that is the fundamental  v. O, past vs. future, part vs. whole, form vs. motion, dualities of ∆@st changes. So when we integrate open lineal triangles, with its vortex as the @foreward mindfuture path, becomes an internal locked social, circular closed point of the singularity of a cyclical form.
The approximation of square space to cyclical points. Ratios and ir(ratio)nal numbers, its finitesimal limits.
A theme that will be soon casted on terms of number theory was also studied by Archimedes by exhaustion methods.
Before the invention of the new methods of calculation, it had been possible to find the area only of polygons, of the circle, of a sector or a segment of the circle, and of two or three other figures. In addition, Archimedes had already invented a way to calculate the area of curves by exhaustion, leaving a sound error according to the minimal step he took, which raises the question, does have a circle a finitesimal minimum step? It is then pi and all other S>t constant transformations and ‘ir(ratio)nal numbers/ratios, limited by a finitesimal error?
THE ANSWER IS YES!, Normally a decametric limit define the ‘valid value of an ir(ratio)nal numbers, which is not a number in strict sense (a social number) but a ratio of an S/T action/function. The examples of the two fundamental ir(ratio)nals will suffice:
– pi is really the ratio of 3 diameters that form a closed curve, whose value depends on the lineal ‘step sizes’.
So pi has a minimal value of 3, which is the hexagon with its 6 steps of 1/2 value (triangulation in 6 immediately gives the result, as the triangle is the radius, so are the 6 triangular sides: 1/2 x 6 =3); which happens to be the value of pi in extreme gravitational fields on relativity, which brings another insight: black holes decompose the circle into ultimate lineal flows of pure ‘dark energy’ shot through the axis, by converting the curvature of a light circle on the event horizon in a 6pi hexagon. But this is well beyond the scope of this intro.
So what is the ‘decimal limit’ of pi, before it breaks into meaningless (noneffective) decimal scales, with little influence on the whole?
While this is hypothetical I would say for different reasons explained in the article on number theory, as it is quite often the case it responds to the general ∆ ≈ S ≈ T ternary symmetries, so common in the perfect Universe.
So pi responds to the symmetry between its spatial minimal, 6 x 1/2=3 hexagonal steps, which means it breaks in the 6th ∆scaling decimals, 3,1415…9. So, 3,1416, which incidentally is basically what everybody uses is the ‘real value’ of pi, and why it is that value is studied elsewhere (deducing from it one of the most beautiful simple results of getmathematics, the value of dark energy in any system, of the Universe, as the part not perceived through the apertures of a pi cycle: π/π3 = 96% of ‘darkness’ which the singularity of a pi system cannot see as its apertures are only π3= 0.14
Now, the other constant e, which is the ratio of decay ACTIONS, or death processes (ST<<S), is a longer two ‘scales’ down process, of selfdestruction of a system, unlike the pi, single scaling process, S>T. So it breaks at 10 decimals:
2.718281828…459045
Indeed. Now, why 5 and not ten if the scales are 10¹º? Because 10 scales are in terms of spacetime actions, the ‘whole’ dual game of two directions of time up and down, which happens only in reproductive actions. And this connects with the S>T<S Rhythms of motion go/stop/go back and forth between two arrows which happens both in stsingle planes and ∆±motions.
There is then a limit for existential planes? The ‘meaningless’ breaking down of e, the ‘number of entropic functions’ seems to signal this. But IT WOULD BE AN ERROR TO CONSIDER THE LIMITS of eregularity as it only indicates the LIMIT of entropic death. Death happens and when a system breaks down its natural 10ˆ±10 scales to its finitesimal 1/n parts it stops as the system is dead.
The limit that matters is the limit of the picircle as an Archimedean spiral that lets information enter through its ±never closing spiral to perceive or feed in the external microbits and bites of the Universe. And as we cannot find neither a limit nor a regularity, we could conclude that the most important dimotions of angular perception, and creation of inner mirrors of the outer world by a pispiral have no limit.
What about locomotion? Can we exhaust the limit of a series of steps? Again, this is more evidently no, even though the Greeks thought so, in the so called…
Aquiles Paradox. Birth of the concept of series and limits.
In mathematics, a series is, roughly speaking, a description of the operation of adding infinitely many quantities, one after the other, to a given starting quantity.
^{}The study of series is a major part of calculus and its generalization, mathematical analysis.
For a long time, the idea that such a potentially infinite summation could produce a finite result was considered paradoxical by mathematicians and philosophers.
This paradox was resolved using the concept of a limit during the 19th century.
Zeno’s paradox of Achilles and the tortoise illustrates this counterintuitive property of infinite sums:
Achilles runs after a tortoise, but when he reaches the position of the tortoise at the beginning of the race, the tortoise has reached a second position; when he reaches this second position, the tortoise is at a third position, and so on.
Zeno concluded that Achilles could never reach the tortoise, and thus that movement does not exist. Zeno divided the race into infinitely many subraces, each requiring a finite amount of time, so that the total time for Achilles to catch the tortoise is given by a series.
The resolution of the paradox is that, although the series has an infinite number of terms, it has a finite sum, which gives the time necessary for Achilles to catch the tortoise.
The physical explanation of locomotion though defines it as a reproduction of for of the lower scale, so it establishes a finitesimal stœp, equivalent to the minimal ∆¡ quanta of the waveparticle dual motion states:
In modern terminology, any (ordered) infinite sequence (a1,a2,a3,…) of terms (that is numbers, functions, or anything that can be added) defines a series, which is the operation of adding the ai one after the other.
To emphasize that there are an infinite number of terms, a series may be called an infinite series. Such a series is represented (or denoted) by an expression like: a1+a2+a3+⋯ or, using the summation sign:
The infinite sequence of additions implied by a series cannot be effectively carried on (at least in a finite amount of time).
However, if the set to which the terms and their finite sums belong has a notion of limit, it is sometimes possible to assign a value to a series, called the sum of the series. This value is the limit as n tends to infinity (if the limit exists) of the finite sums of the n first terms of the series, which are called the nth partial sums of the series. That is:
What this means in 5D though is slightly different: because the infinite number of timesteps will make impossible to do any calculus, all limits must have in ‘reality’ beyond the idealized mirror of mathematics, a limit of steps and a limit of size of those steps. Which is indeed what happens in reality.
The turtle has a timecycle and a size of steps, measurable. And when explaining the reproduction of motion, we shall see that limit is the reproduction on the lowest plane of light and particle forces of the entire form of the being in discontinuous adjacent spaces.
In other worlds, the word ‘limit’ in the formulae should not be infinite. But a ‘finite infinite’, for which we shall use a different symbol:
∞:
Relative infinities and finitesimals
The new symbol for a ‘relative infinity’ and its inverse 1/∞, ‘finitesimals’, become then essential to 5D Analysis and it gets rid of all infinite paradoxes from Zeno’s to Cantor, further showing the idealized mirrorimage nature of mathematics; as a mirror recedes apparently into infinity but at a certain point it ceases to be observable and hence it does NOT exist anymore.
The meaning of series then in real existences becomes clear as it is ANOTHER WAY TO DESCRIBE IN DISCONTINUOUS MANNER, WHAT DERIVATIVES ON THE CONTINUOUS PLANE (REMEMBER THE DUALITY OF DISCRETE NUMBER VIEW VS. CONTINUOUS GEOMETRIC VIEW), SHOWS:
A TRAVEL UP AND DOWN THE SCALES OF THE FIFTH DIMENSION.
The problem of equivalences confused as identities between lines and areas.
It is absurd to talk about continuity of a real number, pi, e, and √2, beyond the 10 decimal. This is easily proved because those ratios are normally obtained by limits in which certain terms of the infinitesimal are despised, by postulating the falsity that there are infinite smallish parts, and so x/∆ can be throw out when ∆>∞. But since x/∆, the finitesimal has a limit, the pretentious exactitude does not happen.
This in turn leads to questions about the meaning of quantities that become infinitely large or infinitely small—concepts riddled with logical pitfalls in a simplified world of a single spacetime continuum, where on top humans LOVE to consider ‘identities’ of the mind absolute identities in the larger information of the detailed Universe, which are never so, as d@st ≈ ∆ûst (the mind, world view is merely similar to the Universal view) .
In our example example, a circle of radius r has circumference 2πr and area πr^{2}, where π is the famous constant 3.14159…. Establishing these two properties is not entirely straightforward, although an adequate approach was developed by the geometers of ancient Greece, especially Eudoxus and Archimedes. It is harder than one might expect to show that the circumference of a circle is proportional to its radius and that its area is proportional to the square of its radius. The really difficult problem, though, is to show that the constant of proportionality for the circumference is precisely twice the constant of proportionality for the area — that is, to show that the constant now called π really is the same in both formulas.
This boils down to proving a theorem (first proved by Archimedes) that does not mention π explicitly at all: the area of a circle is the same as that of a rectangle, one of whose sides is equal to the circle’s radius and the other to half the circle’s circumference.
However in GST theory, those 2 pis are not the same, because they belong to two discontinuous, ‘different species’ of topology, the St area, and the SMEMBRANE.
An easy, immediate proof. If we make them identical, then we can find a circle, where: 2πr ≈ πr^{2}. So 2r=r^{2 }. Hence 2=r and we get to the conclusion that the thin membrane of an open ball is identical in area to the internal ST volume of the being, which is ‘conceptually absurd’ (the area intuitively has more surface, as it is bidimensional, the line, infinitely thin).
What’s the problem here? We cannot in true form, unless we deal always with less dogmatic concepts of relative similarities with ‘lines as if they were squares’. They are different realities. In the first equivalence, we compare a line radius with a circle perimeter, in an S>t structure.
In the second as we compare π², a cyclical area with the square of the radius we are also in good footing. But when we do the S>ST comparison, we are in a Dynamic transformation of ∆scales, from ∆, the world of lines, to ∆+1 the world of squares (as a polynomial square is obviously a growth from a complete ∆entity the line, into an ∆2=∆+1 one, the area). It is then when we can do some ‘dynamic equivalence’ analysis, and the equivalence has meaning, stating that for a ‘perfect cycle’ of relative radius 2, the membrane absorption of bits an bites of energy and information, can fully, fill, the internal area, making equivalent, a ‘line and a surface’ integral. And finally state that all ‘dynamic vortices of force’ ruled by Newtonian/Coulombian equations on the ∆1 and ∆+1 scales, are relative perfect systems of radius 2.
And here we find the ‘whys’ of the dualities of Maxwell’s laws which can be written both ways:
Or in simpler terms, we are talking when doing those equalities of properties that become dynamic and transcend the static mind of mathematics into the reality of physical systems.
Finally as we defined real numbers as non existent (see ∞ posts), but approximations to a ±0 infinitesimal, in the measure of a square, uncertainty grows further, π^{2}, thus have the square ‘error’ of pi.
All this of course is important to conceptualize reality, in praxis as we know we always work in an uncertain game with errors and deaths. So analysis does work, and all this ‘search for dogmatic proofs’ is just ‘absolute bull$hit’ for absolute egocentered scholar huminds.
But on the other hand the graph also shows that both pis, the one of the ‘surface’ and the one of the ‘perimeter’ ARE NOT equal, as there will be a limit on the number of ‘bidimensional triangles’ we can cut.
As a triangle is indeed the bidimensional line, that is: $t (onedimension); ∆$t (2 dimension).
So it is not the line.
And so as the approximation will find a finitesimal quanta or limit of detail, prove the theorem, this error, however tiny, remains an error. THIS MINIMAL QUANTA THUS EXIST IN ALL RELATIVE ∆>∆+1 measures of scales as the minimal uncertainty of all mathematical calculus, and justifies in physics (∆1 quantum theory) that thee is always an uncertainty of a minimal quanta, which is precisely h/2; that is h/2π; the minimal quanta of our light spacetime.
Only in the absolutist imagination of dogmatic axiomatic mathematicians it made sense to talk of the slices being infinitesimally thin, so the error would disappear altogether, or at least it would become infinitesimal.
As it happens quantum theory proved experimentally the case to be wrong. And as we stress (Lobachevski, Godel, Einstein) mathematics must be confronted with reality to realise what is ‘real’ in maths.
QUEST FOR FINITESIMALS. FROM ARCHIMEDES TO NEWTON.
We change following the transformation of sciences into slightly different stiences, the concept of infinite from a relative infinite, ∞, and an infinitesimal for a finitesimal. The first being the whole of an ¡plane of reality, the second its minimal part.
It is then obvious that the discrete, geometric, spatial, static numerical analysis of calculus is the of power series, which can be taken as discrete stœps (stops + steps) in a motion down the fifth dimension from the whole to the 1/n part, whereas we count also the static form (as we see only in a movie the static frame) NOT the step of motion.
This was then the work from Archimedes and earlier Greeks to Newton, which can in that sense be considered the last of the ancients.
While as all S=T, that is there is always a symmetry between discrete numbers and continuous motions, Leibniz with its geometric interpretation and far more profound understanding of finitesimals, which he rightly defined as 1/n, represents the first step in the future of the discipline, the renovator and deep understanding of it – which Newton, which can be considered merely an automaton mathematician, specialized brain, as most modern scientists is – he is indeed the father of the wrong view of science – understood NOTHING OF IT.
Indeed, Leibniz, the closest predecessor of this blog IS the genius, Newton the talent.
Rates of change. The stop and go motion: stœps.
Finitesimal changes are related to the fundamental beat of the Universe, the stopformspaceperception, gomotiontime, beat of the Universe, which we shall call a stœp, the discrete way of motion of tœs through SPace, which often as in movies we perceive in continuous mode eliminating the stop element:
∆S(top)>∆t>∆S≥∆(S)t(ep).
Moreover most of those Stœps will have either in a travel through 5D, or through a single ST, a unit of ‘expenditure of vital energy’, transformed in the lengthmotion of the lower scale in which the imprinting of motion as reproduction of form, happens (studied in 2D locomotion). So each stœp becomes an ∆4 unit of locomotion.
Thus if we consider a relative constant or function of the existence, ∆1:œ, as a finitesimal of its larger whole, ∆Œ, we obtain 2 simple functions:
œ=∆s/∆t and œ=∆t/∆s as the mathematical measure of a ‘time stœp’ or locomotion and ‘volumedensity stœp’ or finitesimal quanta.
We shall call the first form a spatial finitesimal or step in space, a quanta of constant speed that moves, reproduces the being in space.
And if we again change this quanta, a quanta of constant acceleration
And we shall call the second function, a time finitesimal, a change in the density of information or cyclical speed of the being and a second change in relation sop
Infinite series
Graphical illustration of an infinite geometric series. Before understanding calculus mathematicians were concerned with relative infinitesimal series.
Since similar paradoxes occur in the manipulation of infinite series, such as: ^{1}/_{2} + ^{1}/_{4} + ^{1}/_{8} +⋯
This particular series is relatively harmless, and its value is precisely 1, the whole, which is the conceptual meaning of infinity.
To see why this should be so, consider the partial sums formed by stopping after a finite number of terms. The more terms, the closer the partial sum is to 1. It can be made as close to 1 as desired by including enough terms. Yet once we arrive to the Minimal quanta of the physical reality we describe (cell, atom, individual, etc.) there is NO need to go beyond except in errors of the mind.
In the graph, 1/±10² is the limit considered the finitesimal of this particular ‘graph perception’. And also the error of our measure, as if we add another 1/±10², the series becomes a whole.
Thus most paradoxes of mathematics arise from not understanding those simple concepts, as well as the meaning of ‘inverse negative numbers’ .
For example an infinite series which are less wellbehaved are the series: 1 − 1 + 1 − 1 + 1 − 1 + ⋯
If the terms are grouped one way: (1 − 1) + (1 − 1) + (1 − 1) +⋯, then the sum appears to be: 0 + 0 + 0 +⋯ = 0.
But if the terms are grouped differently, 1 + (−1 + 1) + (−1 + 1) + (−1 + 1) +⋯, then the sum appears to be 1 + 0 + 0 + 0 +⋯ = 1.
It would be foolish to conclude that 0 = 1. Instead, the conclusion is that the series has a due value, and so it is creative oscillatory series with a time dynamic that cannot be merely said, not to have a solution, but has 2.
It has therefore an internal dual structure, which in modern algebra is the group:
‘a’: 11=0. And so if we accept that internal ∆1 unit for the series grouping and its ‘real value is:
a+a+…. = 0+0+0…=0.
So we can write it in terms of the generator as:
∑ $t (+1) <≈> ∑ðƒ (1), which defines generically a feedback ‘world cycle’ whose sum is zero.
In classic maths of a single spacetime continuum, the difference between both series is clear from their partial sums. The partial sums of 1/2+1/4… get closer and closer to a single fixed value—namely, 1. The partial sums of a+, without its internal ∆1 (a) structure, alternate between 0 and 1, so the series never settles down.
A series that does settle down to some definite value, as more and more terms are added, is said to converge, and the value to which it converges is known as the limit of the partial sums; all other series are said to diverge. But in GST many diverging series become when considered also its internal structure, convergent and wellbehaved.
Actually, without even experimental evidence, there exist subtle problems with such ‘infinite’ construction. It might justifiably be argued that if the slices are infinitesimally thin, then each has zero area; hence, joining them together produces a rectangle with zero total area since 0 + 0 + 0 +⋯ = 0. Indeed, the very idea of an infinitesimal quantity is paradoxical because the only number that is smaller than every positive number is 0 itself.
The same problem shows up in many different guises. When calculating the length of the circumference of a circle, it is attractive to think of the circle as a regular polygon with infinitely many straight sides, each infinitesimally long. (Indeed, a circle is the limiting case for a regular polygon as the number of its sides increases.) But while this picture makes sense for some purposes—illustrating that the circumference is proportional to the radius—for others it makes no sense at all. For example, the “sides” of the infinitely manysided polygon must have length 0, which implies that the circumference is 0 + 0 + 0 + ⋯ = 0, clearly nonsense.
SO BY REDUCTIO AD ABSURDUM, the limits of infinitesimals are shown to be always an ∆1 quanta. THIS of course also resolves all the Cantor’s nonsense of different infinities and its paradoxes. It is just ‘mathfiction’ and worthless to study.
The interest of those works for 5D maths, lies on the fact that THE EXHAUSTION METHOD DOES LIMIT the parts to finitesimals, as a realist method, which implies nature also limits its divisions. This concept would be lost in the 3rd formal age, also with the ‘lineal bias’ introduced on Dedekind’s concept of a real number NOT as a proportion/ratio, between quantitative parameters of the ‘parts’ of a whole, or the ‘actions’ of a system and its SE<STI>TO parameters, which is what IT IS, but as an ‘abstract cut’ in a lineal sequential order of ‘abstract numbers’.
To notice that in the classic STi balanced age, both the limits method and finitesimal method of Leibniz considered infinitesimals, finitesimals, that is with a ‘cutoff limit’ and real nature.
Those limits are minimal ‘steps’ of any scale (in timemotion), or minimal parts (in spaceforms).
Let us now deal with all this in a cleaner way in terms of polynomials, as SERIES are indeed the justification for POLYNOMIALS beyond the simplest spatial view of them in 3 steps of dimensions of space (point, line, volume) or motions of time (distance, motion, acceleration):
Polynomials as divergent or convergent scalar series.
In mathematics, a power series (in one variable) is an infinite series of the form
where a_{n} represents the coefficient of the nth term and c is a constant. a_{n} is independent of x and may be expressed as a function of n (e.g., an=1/n!). Power series are useful in analysis since they arise as Taylor series of infinitely differentiable functions.
In many situations c (the center of the series) is equal to zero, for instance when considering a Maclaurin series. In such cases, the power series takes the simpler form
Any polynomial can be easily expressed as a power series around any center c, although most of the coefficients will be zero since a power series has infinitely many terms by definition. For instance, the polynomial f(x)=x²+2x+3 can be written as a power series around the center c=0 as
or around the center c=1 as
or indeed around any other center c One can view power series as being like “polynomials of infinite degree,” although power series are not polynomials.
The geometric series formula
which is valid for x<1 is one of the most important examples of a power series, as are the exponential function formula
 =
and the sine formula
valid for all real x.
These power series are also examples of Taylor series.
We shall then in other posts consider their relationship with those functions, which are the key DIMOTIONS of scalar motion (1/1x), entropy (exponential) and 1Dimotion (Sin).
Geometric series
A geometric series is a series with a constant ratio between successive terms. For example, the series
is geometric, because each successive term can be obtained by multiplying the previous term by 1/2.
Each of the purple squares has 1/4 of the area of the next larger square (1/2×1/2 = 1/4, 1/4×1/4 = 1/16, etc.). The sum of the areas of the purple squares is one third of the area of the large square.
We can then consider to be a series that diminishes till it reaches the ‘finitesimal’ 1/n part of the whole. And it can easily be casted as a polynomial; since the terms of a geometric series form a geometric progression, meaning that the ratio of successive terms in the series is constant. This relationship allows for the representation of a geometric series using only two terms, r and a. The term r is the common ratio, and a is the first term of the series.
In the example we may simply write:
 , and .
The behavior of the terms depends on the common ratio r:
 If r is between −1 and +1, the terms of the series become smaller and smaller, approaching zero in the limit and the series converges to a sum. In the case above, where r is one half, the series has the sum one.
 If r is greater than one or less than minus one the terms of the series become larger and larger in magnitude. The sum of the terms also gets larger and larger, and the series has no sum. (The series diverges.)
 If r is equal to one, all of the terms of the series are the same. The series diverges.
 If r is minus one the terms take two values alternately (e.g. 2, −2, 2, −2, 2,… ). The sum of the terms oscillates between two values (e.g. 2, 0, 2, 0, 2,… ). This is a different type of divergence and again the series has no sum; for example in Grandi’s series: 1 − 1 + 1 − 1 + ···.
Geometric series are among the simplest examples of infinite series with finite sums, although not all of them have this property.
Historically, geometric series played an important role in the early development of calculus, and they continue to be central in the study of convergence of series.
Geometric series are used throughout mathematics, and they have important applications in all sciences, as all of them are obviously scalar in its form, and respond to any of the 3 possible behaviors of systems, ‘convergent information’, divergent entropy and repetitive=reproductive oscillation.
Of the many mirror correspondences between series and 5D we want now to stress the relationship between the part and the whole, as elements the ternary structure of any T.œ with its singularity, that can be considered the a, initial term, the FINITESIMAL above all other finitesimals, the king of the hill so to speak, its membrane and the space between them.
This relationship is truly enlightening of the symmetry between the 3 regions in space of a being, and its 3 regions in scale. Whereas the central finitesimal @mind is the finitesimal of the lower plane, the external membrane the ‘larger term’ arˆn of the series, and vital energy within them, the intermediate terms of the series which are irrelevant. So as the singularity @=a, of the series expands through the vital energy elements in growing ‘circles’ to reach the final ‘membrane’ arˆn, magically those irrelevant vital space cells will disappear in the final calculus of the value of the series.
Further on, those sums will be limited by n, which IS THE value of the NUMBER OF ‘scales’ within the vital energy (concentric circles) required to arrive to the surface of it.
So a can also be viewed as the relative ‘radius’ of the singularity mind, which gives conceptual birth to the formula of the angular momentum of the series, where rmv, signifies r=sum of singularity radius (imagine the inner region of the system as an Archimedean spiral) m the vital energy mass, and v the membrane.
All this is expressed in terms of discrete numbers – not geometric continuous motion – by the classic formula:
For r≠1, the sum of the first n terms of a geometric series is
 ,
where a is the first term of the series, and r is the common ratio. We can derive this formula as follows:
As we see, the @ singularity value and its final term, arˆn are the ONLY values that matter, with all the intermediate terms ‘absorbed’ in the dynamic relationship between membrane and singularity by them. If s, is the value of the series for the singularity, without the membrane, rs is the value of the system for the membrane, without the singularity. As the vital energy within has both the singularity and the membrane as its ‘Klein’ limits of a noneuclidean sphere, which they never reach. And so we rest from the ‘Singularity’, S plus ITS perception of the vital energy, the membrane, ‘rs’, and its feeding (negative value) of the vital energy, SrS, to search for the Solution of the power series which IS NOT THE MEMBRANE view but the singularity view, s:
And so the solution as always is that of the mind view (in any discrete, numerical selfcentered analysis) s=a (value of the singularity) multiplied by the parenthesis.
Then we can easily see the symmetry of that topological explanation of the series, with its scalar translation as a travel down a scale from the whole to the finitesimals. Since as we differentiate those series to converge and make sense, because we are traveling down the scale to the finitesimals, r, as n goes to infinity, must be less than one for the series to converge. The sum then becomes:
When a = 1, that is the singularityminds as the view=value of the whole this can be simplified to:
the lefthand side being a geometric series with common ratio r.
The beauty and simplicity of the formula shows by occam’s razor principle indeed its ‘essential nature’ in terms of timespace laws.
Now the type of series define the different
The behavior of the terms depends on the common ratio r:
 If r is between −1 and +1, the terms of the series become smaller and smaller, approaching zero in the limit and the series converges to a sum. In the case above, where r is one half, the series has the sum one.
 If r is greater than one or less than minus one the terms of the series become larger and larger in magnitude. The sum of the terms also gets larger and larger, and the series has no sum. (The series diverges.)
 If r is equal to one, all of the terms of the series are the same. The series diverges.
 If r is minus one the terms take two values alternately (e.g. 2, −2, 2, −2, 2,… ). The sum of the terms oscillates between two values (e.g. 2, 0, 2, 0, 2,… ). This is a different type of divergence and again the series has no sum; for example in Grandi’s series: 1 − 1 + 1 − 1 + ···.
It is quite interesting then to understand in terms of the 5 Dimotions and o1=1∞, TIMESPACE dual sphere (essential for quantum physics) the variations of the power series. As they work for the o1 sphere, in which the series travels a SCALE OF THE FIFTH DIMENSION: FROM 1∆ down to ∆1 vs. its entropic divergent expansion when r is larger than ±1, as it travels in the 1∞ sphere, which should HAVE A SOLUTION, WHEN we define ∞, a relative infinite as the value of the whole perceived from the finitesimal point of view, which means a relative infinite. Then we make a travel upwards from the ∆1 finitesimal or ∆being to the ∆+1 world.
So those series represent the 1D and 45Dimotions, while the 3rd reproductive dimotion happens when r=1, as the reproductive sum that creates terms of a reproductive wave, which in a lineal sum of steps will represent the 2D locomotion of the being. Finally if r is 1 the series forms a ‘steady state’ zero sum world cycle, an oscillation of two values.
So the key concept of a proper 5D scalar interpretation of series (this analysis on the simplest of all series for 5D advanced theory would obviously expand to power series Taylor series etc, but we leave this work for the future pouring of my notebooks or in case I likely die earlier, for future researchers) is the concept of finitesimals and relative infinites, ∞.
The limit of a sequence
In that regard we amend the work of the German mathematician Karl Weierstrass and its formal definition of the limit of a sequence as follows:
Consider a sequence (a_{n}) of real numbers, by which is meant an infinite list: a_{0}, a_{1}, a_{2}, ….
It is said that a_{n} converges to (or approaches) the limit a as n tends to infinity, if the following mathematical statement holds true: For every ε > 0, there exists a whole number N such that a_{n} − a < ε for all n > N. Intuitively, this statement says that, for any chosen degree of approximation (ε), there is some point in the sequence (N) such that, from that point onward (n > N), every number in the sequence (a_{n}) approximates a within an error less than the chosen amount (a_{n} − a < ε). Stated less formally, when n becomes large enough, a_{n} can be made as close to a as desired.
For example, consider the sequence in which a_{n} = 1/(n + 1), that is, the sequence: 1, ^{1}/_{2}, ^{1}/_{3}, ^{1}/_{4}, ^{1}/_{5}, …, going on forever.
Every number in the sequence is greater than zero, but, the farther along the sequence goes, the closer the numbers get to zero. For example, all terms from the 10th onward are less than or equal to 0.1, all terms from the 100th onward are less than or equal to 0.01, and so on. Terms smaller than 0.000000001, for instance, are found from the 1,000,000,000th term onward. In Weierstrass’s terminology, this sequence converges to its limit 0 as n tends to infinity. The difference a_{n} − 0 can be made smaller than any ε by choosing n sufficiently large. In fact, n > ^{1}/_{ε} suffices. So, in Weierstrass’s formal definition, N is taken to be the smallest integer > ^{1}/_{ε}
This example brings out several key features of Weierstrass’s idea. First, it does not involve any mystical notion of infinitesimals; all quantities involved are ordinary real numbers. Second, it is precise; if a sequence possesses a limit, then there is exactly one real number that satisfies the Weierstrass definition. Finally, although the numbers in the sequence tend to the limit 0, they need not actually reach that value.
Now this n > 1/ε is exactly what Leibniz without so much pedantic formalism considered the finitesimal, what we call the quanta of an ∆1 scale and what physicists call in its study of different scales, the minimal ‘errorquanta’ h/2π, kentropy, or ‘Planck mass’ (Black hole of a compton wavelength volume, or minimal quanta of gravitational ∆+1 scales).
Continuity of functions
All this understood we can then return to the inflationary nature of languages, which in the case of mathematics means that without a mirror reflection in reality, it tries to introduce false concepts of infinity and continuity with pedantic axiomatic methods, origin of the concept of absolute continuity of a function; when the true concept is the ‘stop and step’ nature of motions, and dark, non perceived regions between continuous points, or finitesimals. So it is irrelevant if the finitesimal is a natural number to talk of discontinuity, as the system will have contiguous finitesimals of 1 number size. We then talk of measure more than continuity and errors of measure, from an upper ∆º mind”
Intuitively, a function f(t) approaches a limit L as t approaches a value p if, whatever size error can be tolerated, f(t) differs from L by less than the tolerable error for all t sufficiently close to p.
Just as for limits of sequences, the formalization of these ideas is achieved by assigning symbols to “tolerable error” (ε) and to “sufficiently close” (δ). Then the definition becomes: A function f(t) approaches a limit L as t approaches a value p if for all ε > 0 there exists δ > 0 such that f(t) − L < ε whenever t − p < δ. (Note carefully that first the size of the tolerable error must be decided upon; only then can it be determined what it means to be “sufficiently close.”)
But what exactly is meant by phrases such as “error,” “prepared to tolerate,” and “sufficiently close”?
Again it is the relative ¡1 quanta of the system studied. The ‘error’ of measure will then become ESSENTIAL to the explanation of the Uncertainty principle of Heisenberg, which indeed can be obtained from theory of measure and error, by pure mathematical methods.
So in ideal mathematics, having defined the notion of limit in this context, with no limit to the infinitesimal size of the error, it is straightforward to define continuity of a function. Continuous functions preserve limits; that is, a function f is continuous at a point p if the limit of f(t) as t approaches p is equal to f(p). And f is continuous if it is continuous at every p for which f(p) is defined. Intuitively, continuity means that small changes in t produce small changes in f(t)—there are no sudden jumps.
But as that small change will always be in detail an εquanta, in great detail THERE ARE QUANTUM JUMPS. In fact, as there is always an εquanta, in any process in space or time, in form and motion (as we have shown when considering the nature of motion as reproduction of form in adjacent spaces) there will always be a quantum jump for all motions. And motion will be the reproduction of form in quantum jumps of ε, nature.
∆1: Lebiniz’s definition of Finitesimals: 1/n
So we accept Leibniz’s concept of a finitesimal, as ALL organic systems have a minimal cellular quanta and a maximal enclosure, which in mathematics can be represented in the o1 finitesimal circle, closed above, as it becomes the 1 element in ∆1 of the ∆º whole, which is represented by the 1∞ equivalent graph, WHICH is opened above into the wholeness of a larger Universe (but will have also a limit normally in the decametric logarithmic scale of the ∆º whole world embedded in the ∆+1 truly infinite Universe).
Perhaps the most fascinating part of number theory is the finitesimal, as infinitesimal do not exist – being space quantic, there will be always a limit, a microcycle of time or quanta of population in space, to signify the finitesimal point, as Leibniz rightly understood and defined it with a simple powerful form: 1/n.
And indeed in the Universe finitesimals tend to be structured as in a russian doll, such as the biggest wholes, n>∞ have the smallest finitesimals, 1/n>∞>0.
The absolute zero size is thus the finitesimal of the largest possible Universe. In praxis, we humans only observe a finitesimal from our mindo perspective, of the planck scale, and accordingly we see a Universe of inverse relative size, being humans in the ∆º middle view (at cellular level) as physicists wonder without realizing this is NOT a coincidence, but a natural law of the scalar, fractal organic structure of the Universe:
1/n: the (in)finitesimal (in)finite
With the convention that ƒ (x) is normally a function of time frequencies, ƒ (t), of motions of time, whose synthonies of synchronicity in space are expressed by an algebraic equation, we bring the following understanding:
Infinitesimal quanta in any scale is the departure point to build any function, as such it must have a minimal size, and ƒ'(t) is normally a good measure.
The infinitesimal study as perceived from the finite point of view is the view of fractals, when in detail and observing the closed worldcycles that separate and make each infinitesimal a whole.
A derivative is the finitesimal of the function observed, and so when we go even further and study as enlarged into our scalar view tin maximal information we are in the fractal view of reality.
So as we expand our view the fractal view becomes more real, till finally the enclosures observed ∆1 become fractal and we recognise its selfsimilarities: ∆1 ≤ ∆º.
For each derivative thus a function shows its 1/n infinitesimal (not necessarily this function, which is the derivative of the logarithm).
It follows that functions which grow ginormously have a ‘quanta of time’ reproduced and so its minimal derivative finitesimal is the function itself, eª.
In the next graph we see inverse equations of exponentials and logarithms.
Exponentials express better decay than exponential growth, with the exponent “negative”.
Mathematics is a reflection of nature. A small mirror of its ∆º±i Structure and so we need for exponential growth that Nature provides unlimited energy for growth, which happens only in the 01 generational dimension of the being, or in its inverse decay/ in its 4D entropy age. of death.
On the other hand the limit of logarithmic growth maps out better in logistic curves real growth being a good function to express ∆§cales.
So numbers reflect those processes in their inverse exponential/logarithm mathematical graphs and numerical series.
ST: As the three coordinate systems, selfcentred into an ∆º pov, which reflects each of the three ‘topologies of spacetime’ (Cylindrical: lineal, polar: cyclical and cartesian: Hyperbolic); while the infinitesimal o1 scale, and the infinite 1∞ scale divided by the ‘1’ ∆º relative element, represent perfectly the ∆scalar nature of super organisms.
∆º±1: Further on, we can ‘reduce’ each relative infinity to those 3 scales, and represent all timespace phenomena with the different families of numbers that close algebra (entropic, positive numbers, informative, negative numbers, present spacetime, complex bidimensional numbers, s/t irrational numbers, etc.), mathematics becomes essentially the more realist language to represent the scalar, organic, ternary Universe.
The 01 scale is equivalent to the 1∞ scale for the lower ∆1 Universe, where 1=∆º, the whole and 1∞ is the ∆+1 eternal world.
And this is the symmetry to grasp the consequences of the o1∞ fundamental graph of the fifth dimension. Let us see how with a simple example:
Now the mirror symmetries between the 01 universe and the 1∞ are interesting as they set two different ‘limits’, an upper uncertain bound for the 1∞ universe, in which the 1world, ∆º exists, and a lower uncertain bound for the 01 Universe, where the 1 does not see the limit of its lower bound. Are those unbounded limits truly infinite?
This is the next question where we can consider the homology of both the microscopic and macroscopic worlds.
Of course the axiomatic method ‘believes’ in infinity – we deal with the absurdities of Cantorian transinfinities in articles on numbers. But as we consider maths, after lobachevski, Godel and Einstein, an experimental science; we are more interested in the homologies of ∆±1. For one thing. While 0 can be approached by infinite infinitesimal ‘decimals’, so it seems it can never be reached, we know since the ‘violet catastrophe’ that the infinitesimal is a ‘quanta’, a ‘minimum’, a ‘limit’. And so we return to Leibniz’s rightful concept of an 1/n minimal part of ‘n’, the whole ‘1’.
This implies by symmetry that on the upper bound, the worlduniverse in which the 1 is inscribed will have also a limit, a discontinuity with ∆+2, which sets up all infinities in the upper bound also as finite quanta, ‘wholes of wholes’.
So the ‘rest’ of infinities, must be regarded within the rest of ‘theory of information languages’ and its inflationary nature, inflationary information. What is then the ‘practical limit’ for most infinities and infinitesimals? In GST, the standard limit is the perfect game of 3 x 3 + 0(±1) elements, where the omind triples as it is an ∆1 ‘god of its infinitesimals it rules subconsciously, as you brain rules you cells’, ∆º, consciousness of the whole and ∆+1 infinitesimal of the larger world.
An o1 time mirrored quantum world of probabilities of existence, as indistinguishable infinitesimals through the surface limit of its statistical description in the thermodynamic scale of atomic beings end in the 1 unit of our human cellular space, where thermodynamic considerations are reduced to temperature gradient towards the homeostatic mass based forces of our human level of existence, ∆º.
So we consider as usual the Kaleiodoscopic, multiple function of analysis, and the multiple meanings of its inverse, ∆±1 operations, derivatives and integrals; since as usual the potency of ∆st is on the search of whys, not on the discovery of new equations, which humans always exhaust by the Monkey Method of trials and errors, sweat and transpiration more then the inspiration of pure logic thought…
Conclusion.
The Universe is discontinuous. To differentiate a function we do NOT need absolute continuity but the existence of an infinitesimal 1/n, and no jump between ‘neighbourhoods’, which should be no further than 1/n distance either in the X or Y coordinates. ‘Adjacency’ of the function then is defined by discrete 1/n intervals, which suffice in Nature=reality, REGARDLESS of mathematical methods to define them.
2nd AGE: CALCULUS
OPERATIONS: ∫∂
Its inverse symmetries on the cartesian plane: merging all the elements of ∆@s=t maths.
Descartes idea of representing solutions to equations with a larger dimension – the variable letter that represented all the ‘§ets’ of dual X, Y possible solutions; and to ‘imagine’ them in a graph to plot them, forming a visual ‘informative’ geometric figure, the new ‘scalar dimension‘ that gathered all the X(S)<≈>Y (t) pairs of possible ‘variations’ on the spacetime construct.
Up to the time of Descartes, where an algebraic equation in two unknowns F(x, y) = 0 was given, it was said that the problem was indeterminate, since from the equation it was impossible to determine these unknowns; any value could be assigned to one of them, for example to x, and substituted in the equation; the result was an equation with only one unknown y, for which, in general, the equation could be solved.
Then this arbitrarily chosen x together with the soobtained y would satisfy the given equation. Consequently, such an “indeterminate” equation was not considered interesting.
Descartes looked at the matter differently. He proposed that in an equation with two unknowns x be regarded as the abscissa of a point and the corresponding y as its ordinate. Then if we vary the unknown x, to every value of x the corresponding y is computed from the equation, so that we obtain, in general, a set of points which form a curve.
The deepest insight on what Descartes did is then evident:
HE GAVE MOTION=CHANGE TO GEOMETRY, ADDING ITS TIMEDIMENSION; AND SO its method could be used to study the actions/motions of a ‘fractal point’ whose inner geometry of social numbers was NOW ignored, in the ∆+1 scale of its world.. And so the graph would be a perfect graph to study all the ACTIONS=MOTIONS external to a given being, becoming for that reason the foundational structure of mathematical physics.
Thus analysis we will find that the curves DO represent key features of the ‘arrows of change’ of the Universe, specially the ‘standing points’ of change of parameters of Space=Information, ST=energy and Time=entropy (or any other kaleidoscopic combination of ST), in essence they represent the world cycle of the action or motion we study, with its 3 phases of starting motion, steady state, and 3rd informative age coming to a halt.
Historic view.
Particularly important here is the theorem of Newton and Leibnitz to the effect that the problem of quadratures is the inverse, in a wellknown sense, of the problem of tangents.
For solving the problem of tangents, and problems that can be reduced to it, there was worked out a suitable algorithm, a completely general method leading directly to the solution, namely the method of derivatives or of differentiation.
It turned out that if the law for the formation of a given curve is not too complicated, then it is always possible to construct a tangent to it at an arbitrary point; it is only necessary to calculate, with the help of the rules of differential calculus, the socalled derivative, which in most cases requires a very short time. Up till then it had been possible to draw tangents only to the circle and to one or two other curves, and no one had suspected the existence of a general solution of the problem.
If we know the distance traversed by a moving point up to any desired instant of time, then by the same method we can at once find the velocity of the point at a given moment, and also its acceleration. Conversely, from the acceleration it is possible to find the velocity and the distance, by making use of the inverse of differentiation, namely integration. As a result, it was not very difficult, for example, to prove from the Newtonian laws of motion and the law of universal gravitation that the planets must move around the sun in ellipses according to the laws of Kepler.
Of the greatest importance in practical life is the problem of the greatest and least values of a magnitude, the socalled problem of maxima and minima.
A note of importance, specially in such calculus of variations, will then be the nature of that minimal fractal step, which is the point of a tangent, as a point has always parts (it is a fractal point) the finitesimal is the fractal point and it is not a single point but the point and a very ‘small’ surrounding (the previous or next points). So a maximum and minimum will be a dual point, so to speak, with a zero tangent (flat line), or in terms of motion a still moment in the summit of the function, which justifies such 0 value (or else if it was a single point and before it it was upwards and then downwards, or viceversa in a minimum, the value of the derivative in that point will be undetermined).
At various points of a curved line, if it is not a straight line or a circle, the curvature is in general different. How can we calculate the radius of a circle with the same curvature as the given line at the given point, the socalled radius of curvature of the curve at the point? It turns out that this is equally simple; it is only necessary to apply the operation of differentiation twice. The radius of curvature plays a great role in many questions of mechanics.
Now, we observe a curious duality between mathematical mind solutions vs. reality check: while classic science differentiates ‘twice’ to know if the point will ‘fall’ or ‘rise’ after the standing point, (∆nalytical solution) we obtain the same knowledge by ‘seeing’ in ‘reality’ how the ‘2 sequential points’ that surround the flat step behave in spacetimes. It is this ‘time interval’ of 4 sequential ‘steps’, what the derivative method, which can be considered a reduction of the curve to the essence of its time sequence solves. And we shall see often this inverse GST reduction, from reality and its complex actions to its sequential origin.
Indeed, in our study of sequential actions of world cycles we noticed that the steps of actions are always the same:
1D: ïperception > 2D: A: motion towards energy > E: feeding > 3D:wide storage of food, or 3Dx5D: O: reproduction and U: social evolution.
1D: As we move from the first ‘action’: to open you ‘eyes’ perceive and be perceived as a function in existence (with a quantitative parameter, which is a scalar ‘point’)…
2D: We then move into a motion, (with a more complex quantitative parameter, a bivector o 3vector parameter).
Locomotion physics is thus a 2 Steps, ST action which we can measure as momentum with 2 parameters; the ‘tiƒ parameter, frequency, mass, temperature’ for the 1D point and the spatial location, which will break into 3 parameters for a vector x,y,z parameter (x,y,z + t).
It is though still a simple S=T, though the time parameters are reduce to an external measure of spaceotion.
We though will depart from this simplest I>Analysis of locomotion to include not necessarily in quantitative terms the description of the other ‘motions/actions’ of reality, Energy feeding, O & U.
Mathematicians were greatly pleased when it turned out that the theorem of Newton and Leibnitz, to the effect that the inversion of the problem of tangents would solve the problem of quadrature, at once provided a method of calculating the areas bounded by curves of widely different kinds. It became clear that a general method exists, which is suitable for an infinite number of the most different figures. The same remark is true for the calculation of volumes, surfaces, the lengths of curves, the mass of inhomogeneous bodies, and so forth.
And this as most pure spatial questions is straight forward: you put finitesimal linesteps or square areas (which would also have the absolute limit of triangular Planck’s areas, which according to the bidimensional holographic principle, are the minimal area of information of a black hole; as indeed the black hole converts the spherical event horizon into ‘static hexagonal π=3 shrunk curvatures, incidentally the strongest most stable ‘buckminster domes’ and graphenes).
The new method accomplished even more in ‘time’ mechanics, because unlike easytofigure out approximations of areas, the staple food that started up mathematics in agricultural measure, time was NOT, it is still NOT understood – so alas, the ‘magic’ method of LN (ab. for Leibnewton, Leibniz first:), got solved questions of timechange without knowing much about time.
As it seemed that there was no problem of locomotion and ratios of change the new calculations would not clarify and solve.
Not long before, Pascal had explained the increase in the size of the Torricelli vacuum with increasing altitude as a consequence of the decrease in atmospheric pressure. But exactly what is the law governing this decrease? The question is answered immediately by the investigation of a simple differential equation (the deep philosophical insights on S≈T transformations ignoramus – who cares would say Feynman ):
It is well known to sailors that they should take two or three turns of the mooring cable around the capstan if one man is to be able to keep a large vessel at its mooring. Why is this? Of course, you need two and better 3 elements for a ‘system’ to become a stable whole, so there are always a sailors said ‘3 saint marys… 3 huge waves, and 3knots are best’… but alas, a similar differential equation to that of Torricelli solves it magically.
Thus, after the creation of analysis, there followed a period of tempestuous development of its applications to the most varied branches of technology and natural science. Since it is founded on abstraction from the special features of particular problems, mathematical analysis reflects the actual deeplying properties of the material world; and this is the reason why it provides the means for investigation of such a wide range of practical questions. The mechanical motion of solid bodies, the motion of liquids and gases of their particular particles, their laws of flow in the mass, the conduction of heat and electricity, the course of chemical reactions, all these phenomena are studied in the corresponding sciences by means of mathematical analysis.
At the same time as its applications were being extended, the subject of analysis itself was being immeasurably enriched by the creation and development of various new branches, such as the theory of series, applications of geometry to analysis, and the theory of differential equations.
So among mathematicians of the 18th century, there was a widespread opinion that any problem of the natural sciences, provided only that one could find a correct mathematical description of it, could be solved by means of analytic geometry and the differential and integral calculus. And so the flurry of activity in the next centuries would be to extend its practical uses
Discovery of the calculus and errors in dogmatic foundations
Two ‘S≈t’ and ∆±1 major steps led to the creation of analysis:
S≈t: The first was the discovery of the surprising relationship, known as the fundamental theorem of calculus, between spatial problems involving the calculation of some total size or value, such as length, area, or volume (integration), and problems involving rates of change in time, such as slopes of tangents and velocities (differentiation). ( Gottfried Wilhelm Leibniz and Isaac Newton.)
 While the utility of calculus in explaining physical phenomena was immediately apparent, its use of infinity in calculations (through the decomposition of curves, geometric bodies, and physical motions into infinitely many small parts) generated widespread unease… as only Leibniz got the understanding of a ‘fractal point which is a world in itself’ and the finitesimal nature of derivatives (1/n).
So the dogmatic zealot, an Anglican bishop George Berkeley published a famous pamphlet, ‘The Analyst; or A Discourse Addressed to an Infidel Mathematician’ (1734), pointing out that calculus—at least, as presented by Newton and Leibniz—possessed serious logical flaws on the arrogant pov of the human mind son of god, who must access absolute truths. LOL.
Analysis the grew out of the resulting painstakingly experimental close examination of concepts such as function and limit, which are still improperly defined with axiomatic zealots of the humind (ab. Human mind) rights to more than humid truths: ‘man is a mush over a lost rock of the Universe, departing from this (relative) principle, we can talk about him’ Schopenhauer.
As all entities have a causal development from a spatial, first entropic age into complex time analysis to end in the ‘awareness of an ∆±1 dimension to it’, the pioneers, Newton’s and Leibniz’s approach to calculus had been primarily geometric, involving ratios with “almost zero, +0” divisors—Newton’s “fluxions” and Leibniz’s “infinitesimals.”
During the 18th century calculus became increasingly temporal, algebraic, as mathematicians—most notably the Swiss Leonhard Euler and the Italian French JosephLouis Lagrange—began to generalize the concepts of continuity and limits from geometric curves and bodies to more abstract algebraic functions and began to extend these ideas to complex numbers, which are the ideal elements to study ∆spacetimeprocesses in its more complex interrelationships oftenreduced to the 01 ‘infinite/simal domain’.
Then in a useless attempt to show the humind absolute in its truth, as these developments were not entirely satisfactory from such deluded foundational standpoint, the so called ‘rigorous’ (: basis for calculus was ‘invented’ by the AugustinLouis Cauchy, Bernhard Bolzano, and above all the idealist squared, usual suspect of total false truths – a cultural, simpleton German Karl Weierstrass in the 19th century.
In the regard, (see ∞^{0} post on the meaning of numbers, infinites and infinitesimals), the logical difficulties involved in setting up calculus on a sound basis are all related to one central problem, the notion of continuity.
Newton and Leibniz.
NOW THERE has been much irrelevant argument about who was first Newton or Leibniz on the discovery of calculus. To me it has always been obvious at all levels the enormous superiority of Leibniz over Newton, ethically and intellectually.
And it can be resumed in this: Newton is NOT really a modern 2nd age researcher of calculus but rather the culmination of the Arquimedes’ method of exhaustion of limits – he didn’t understand anything about the true meaning of calculus. And for that reason his notation is so convoluted (plus his nauseating treatment of Leibniz makes him a complete a$$hole). Leibniz on the other hand understood more than all what would come after him when he said ‘a point is a world in itself, defined the ‘finitesimal’ as 1/n – which latter abstract mathematicians forgot and build an entire philosophy of the Universe (monads), right in the spot, clear predecessor of all our work.
And as usual the a$$hole, making military instruments for the Navy, bullying and calumniating Leibniz carried the day. But we use Leibniz’s notation, the modern view…
So let us first close the ‘Greek era’ with the last of the Greek Alcibiades’, Mr. Newton.
In the second age of mathematics, the question of infinitesimals was resolved but not accepted by Leibniz, which used geometrical concepts on the Cartesian plane to understand them, as opposed to Newton which used algebraic concepts in his study.
INVERSION OF SYMMETRIES.
A key concept as it belongs to the fundamental structure of the Universe is that of inverted entropic numbers.
Entropy, ¬, in mathematical systems are the inverse operations that eliminate the information of a system. As it happens entropy can then take the general format of the negative operand of the systems. So for each positive system there is a negative one. And among all the operand, there is one which is the most entropic of them all, the exponential, notably eˆx, whose massive negative growth signifies the growing dissolution of a form into its finitesimal parts. As systems are in general decametric, such exponential entropy also affect the very same number, which looses its ‘meaningful series form’ after 10 decimals.
Entropy then in calculus is represented by the inverse function of a positive social growth, normally a derivative that extracts the infinitesimal, and when we work on entangled series of Dimotions, a differential equation, perpendicular or inverse to other equation.
Geometric Interpretation of the Problem of Integrating Differential Equations; Generalization of the Problem
For simplicity we will consider initially only one differential equation of the first order with one unknown function dy/dx = ƒ (x,y) where the function f(x, y) is defined on some domain G in the (x, y) plane. This equation determines at each point of the domain the slope of the tangent to the graph of a solution of equation (29) at that point. If at each point (x, y) of the domain G we indicate by means of a line segment the the direction of the tangent (either of the two directions may be used) as determined by the value of f(x, y) at this point, we obtain a field of directions. Then the problem of finding a solution of the differential equation (29) for the initial conditon y(x0) = y0 may be formulated thus: In the domain G we have to find a curve y = ϕ(x), passing through the point M0(x0, y0), which at each of its points has a tangent whose slope is given by equation (29), or briefly, which has at each of its points a preassigned direction.
From the geometric point of view this statement of the problem has two unnatural features:
1. By requiring that the slope of the tangent at any given point (x, y) of the domain G be equal to f(x, y), we automatically exclude tangents parallel to Oy, since we generally consider only finite magnitudes; in particular, it is assumed that the function f(x, y) on the right side of equation (29) assumes only finite values.
2. By considering only curves which are graphs of functions of x, we also exclude those curves which are intersected more than once by a line perpendicular to the axis Ox, since we consider only singlevalued functions; in particular, every solution of a differential equation is assumed to be a singlevalued function of x.
So let us generalize to some extent the preceding statement of the problem of finding a solution to the differential equation (29). Namely, we will now allow the tangent at some points to be parallel to the axis Oy. At these points, where the slope of the tangent with respect to the axis Ox has no meaning, we will take the slope with respect to the axis Oy. In other words, we consider, together with the differential equation (29), the equation: dy/dx = ƒ (x,y)
where f1(x, y) = 1/f(x, y), if f(x, y) ≠ 0, using the second equation when the first is meaningless. The problem of integrating the differential equations (29) and (29′) then becomes: In the domain G to find all curves having at each point the tangent defined by these equations.
These curves will be called integral curves (integral lines) of the equations (29) and (29′) or of the tangent field given by these equations. In place of the plural “equations (29), (29′)”, we will often use the singular “equation (29), (29′)”. It is clear that the graph of any solution of equation (29) will also be an integral curve of equation (29), (29′). But not every integral curve of equation (29), (29′) will be the graph of a solution of equation (29). This case will occur, for example, if some perpendicular to the axis Ox intersects this curve at more than one point.
In what follows, if it can be clearly shown that: ƒ(x,y)= M (x,y)/N (x.y) then we will write only the equation: dy/dx = M (x,y)/N (x.y) and omit dx/dy = N (x,y)/M (x.y). Sometimes in place of these equations we introduce a parameter t, and write the system of equations: dy/dt = M (x,y), dx/dt = N (x,y) where x and y are considered as functions of t.
Example 1. The equation:dy/dx = y /x defines a tangent field everywhere except at the origin. This tangent field is sketched in figure 7. All the tangents given by equation (30) pass through the origin.
It is clear that for every k the function: y=kx is a solution of equation (30). The collection of all integral curves of this equation is then defined by the relation ax=by=0 where a and b are arbitrary constants, not both zero. The axis Oy is an integral curve of equation (30), but it is not the graph of a solution of it.
Since equation (30) does not define a tangent field at the origin, the curves (31) and (32) are, strictly speaking, integral curves everywhere except at the origin. Thus it is more correct to say that the integral curves of equation (30) are not straight lines passing through the origin but half lines issuing from it.
Example 2. The equation: dx/dy= – x/y defines a field of tangents everywhere except at the origin, as sketched in figure. The tangents defined at a given point (x, y) are perpendicular to each other. It is clear that all circles centered at the origin will be integral curves of equation (33). However the solutions of this equation will be the functions: Now this duality is ESSENTIAL as any undergraduate student knows to the duality of potential fields vs. charge singularity forces; and if he has understood anything it will see they are the 2 views of a T.œ control of its vital energy, from the perpendicular=predatory view of the singularity (4th nonE postulate) vs. the parallelism of the membrane that encircles it.
So it also ultimately reflects the DEEPEST meaning of the potential, parallel, stable vs. kinetic, perpendicular, unstable duality of the 2 existential states of a vital energy, which naturally will tend to a potential state of minimal kinetic disturbance, to the eternally wished for state of informative 1D curved eternal existence over the lineal, destructive entropic motion.
So deep it is the duality of y=kx vs. dx/dy=x/y (:
And so we shall bring it back in many posts.
LEIBNIZ: GEOMETRIC FINITESIMALS
Leibniz along Aristotle and Leonardo are the triad of great Genius of the Western, visualdominant civilization, whose ternary structure in ages an regions we treat on the posts on the European civilization… As this writer, though, Leibniz and his equals had the problem of wanting to understand it all and that gets more complex than a simple lineal sword, as the one of Newton, who had a very specialized mathematical mind, without asking further whys filled with the myths of the Bible. By all means then Leibniz along Descartes and its parallel, artist fellow Basque countryman Fermat can be considered for its insights on the whole mirror of mathematics the fathers of the second age of mathematics – Newton being greatly overrated.
Its insights were on the S=T, point=number symmetries of the mathematical mirror; on the fractal point with breath – a world in itself, on the finitesimal with a limit 1/n; and its ‘curvature’ respect to the lineal radius, and hence on the duality between lineal freedom and cyclical order.
Enter Leibniz. Derivatives as spatial tangents: line vs. curve. Its pentalogic.
All this said the true innovation on calculus was to understand that an infinitesimal ‘h’, can be represented by the lineal tangent to a curve.
This point is NOT infinitesimal but a finitesimal, since in I nonEuclidean maths, points have breath. So the finitesimal, which Leibniz defined also in a algebraic terms as 1/n does have breath. The pretentious search for ideal exactitude in the axiomatic method is what makes the lack of empirical knowledge reject him in modern mathematics.
S<=>T. Pentalogic then gives the derivative multiple functions in a single plane: as it can be used when applied to functions that define entropy (negative exponential), angles of perception (trigonometric functions) or simple locomotion (lineal and angular momentum, energy functions)
∆±¡: The derivative of a function of a single variable at a chosen input value, when it exists, is the slope of the tangent line to the graph of the function at that point. The tangent line is the best linear approximation of the function near that input value. So it is a finitesimal and it is lineal.
So in pentalogic we say that the ‘FINITESIMAL’ feel always free, unbounded, ‘lineal’, as your steps in the flat earth, but the whole is curved, ordered, closed. So we could state that a derivative by giving us the finitesimal part, misses the order of the whole and becomes a lineal transformation, and this holds also for multiple variable derivatives.
Let us then introduce derivatives in simple ‘wikiterms’ with 5D insights:
The derivative of a function of a real variable measures the sensitivity to change of the function value (output value) with respect to a change in its argument (input value). It is then all what type of change does take place? As normally a space function is one with form, within change, this should be the input argument; and the yvariable that of change. But when study locomotion we shall observe a clear variation in 5D. It is NOT longer space but time, the frequency of motion what changes, the speed of the wheel is not a change in ‘position’ – the external change – but a change in the internal frequency of turn of the car’s wheel. Time speed thus is what changes, not only in relativity but also in normal motion.
Can then assess change in space? Only externally in the position respect to a background larger ∆+1 world, but not in the internal vital space of the being, which remains stable, changing also the frequency of its stœps.
So while the derivative of the position of a moving object with respect to time is the object’s velocity: this measures how quickly the position of the object changes when time advances, in 5D is inverse – change in the frequency of motion.
@Mind. Derivatives may be generalized to functions of several real variables. And again in this generalization, the derivative is reinterpreted as a linear transformation whose graph is (after an appropriate translation) the best linear approximation to the graph of the original function.
The Jacobian matrix is the matrix that represents this linear transformation with respect to the basis given by the choice of independent and dependent variables.
It can be calculated in terms of the partial derivatives with respect to the independent variables.
For a realvalued function of several variables, the Jacobian matrix reduces to the gradient vector.
Why the Jacobian Matrix defines the change of mind perspective is clear, when we use it to transform planar coordinates (Cartesian) Into polar spherical ones:
The transformation from polar, particle/head/informative coordinates (r, φ) to Cartesian, body/wave hyperbolic coordinates (x, y), is given by the function F: ℝ^{+} × [0, 2π) → ℝ^{2} with components:

 x=r cosρ; y=r sin ρ.
The Jacobian determinant is equal to r. This can be used to transform integrals between the two coordinate systems:
Change of a derivative is a measure of a finitesimal change in any of its Dimotions, and a lineal free approximation to a larger order.
Differential of a function..
This deep thought fact – that small steps are ‘lineal’ and longer ones are curved and ultimately zero sum closed paths is the justification for the use of differentials instead of derivatives in most applications of calculus to reality.
Differentials in essence are ‘lineal’ rates of change in small ‘intervals’ of any function that is curved, and whose exact, ideal, nonlineal rate of change in a long stretch is difficult to calculate. And in reality is used everywhere instead of the ideal derivative. And the justification in 5D is the concept of a finitesimal minimal quanta, and the fractal nature of points and stœps, the minimal quanta of change. That is change IS NEVER infinitesimal, but a change implies a minimal ¡1 unit of the being, either its frequency step or reproductive cell, etc. So that ‘quanta’ of change, which is better measure by the ‘diameter’ or ‘height’ or length of the spherical or tall or flat form (cell, atom, individual) is a differential.
The maths of it, are well known to any student:
Let us then consider a function S = ƒ(t) that has a derivative. The increment of this function: ∆s = ƒ (t+∆t) – ƒ(t) corresponding to the increment Δt, has the property that the ratio Δs/Δt, as Δt → 0, approaches a finite limit, equal to the derivative:
∆s/∆t>ƒ'(t)
This fact may be written as an equality:
∆s/∆t>ƒ'(t) +a
where the value of a depends on Δt in such a way that as Δt → 0, a also approaches zero; since in ∆st the minimal step of any entity always has a lineal form.
Thus the increment of a function may be represented in the form:
∆s=ƒ'(t)∆t + a∆t
where a → 0, if Δt → 0.
The first summand on the right side of this equality depends on Δt in a very simple way, namely it is proportional to Δt. It is called the differential of the function, at the point tn corresponding to the given increment Δt, and is denoted by:
ds=ƒ'(t)∆t
The second summand has the characteristic property that, as Δt → 0, it approaches zero more rapidly than Δt, as a result of the presence of the factor a.
It is therefore said to be a finitesimal of higher order than Δt and, in case f′(t) ≠ 0, it is also of higher order than the first summand.
By this we mean that for sufficiently small Δt the second summand is small in itself and its ratio to Δt is also arbitrarily small.
In the graph, practical stience only needs to measure a differential either in space dy=BD+BC or in time, as a fraction of the unit world cycle, ƒ(x)=cos²x+sin²x=1 which becomes a minimal lineal step or action, ƒ(t)=S step.
In graph, decomposition of ΔS into two summands: the first (the principal part) depends linearly on ΔT and the second is negligible for small ΔS. The segment BC = ΔS, where BC = BD + DC, BD = tan β · ΔT = f′(t) Δt = dS, and DC is an infinitesimal of higher order than Δt.
For symmetry in the notation it is customary to denote the increment of the independent variable by dx, in our case dt, and to call it also a differential. With this notation the differential of the function is:
ds= ƒ'(t) dt
Then the derivative is the ratio, f′(t) = ds/dt of the differential of the function, normally a ‘whole spatial view’ to the differential of the independent variable, normally a temporal step or minimal changemotion in time.
The differential of a function originated historically in the concept of an “indivisible”, similar to our concept of a finitesimal and so much more appropriate for ∆st than the abstraction of an infinitesimal with ∆t>0, since time is discrete and there is always a minimal step of change, or reproductive step in a motion of reproduction of information.
Differentials of calculus are practical infinitesimals and its knowledge for any function acts as an ∂st limit.
On the other hand, there is for any group that we can take as vital spacetime, finds us a middle point.
Rightly then the indivisible, and later the differential of a function, were represented as actual infinitesimals, as something in the nature of an extremely small constant magnitude, which however was not zero.
According to this definition the differential is a finite magnitude, measurable in space, for each increment Δt and is proportional to Δt. The other fundamental property of the differential is that it can ONLY be recognized in motion, so to speak: if we consider an increment Δt which is approaching its finitesimal limit then the difference between ds and Δs will be arbitrarily small even in comparison with Δt – till it becomes zero. The error of interpretation in classic calculus being that it is THE DIFFERENCE what approaches 0 as finally the function will be also lineal, not ∆t, which will become a ‘quanta’ – as quantum physicists would latter discover.
As this is the ‘real’ model, the substitution of the differential in place of small increments of the function forms the basis of most of the REAL applications of the nowcalled ‘finitesimal analysis’ to the study of nature.
Finitesimals of minimal Dimotions=actions
For starters the word to use is ‘finitesimals’, not infinitesimals. Infinite does not exist in a single continuum, but through multiple discontinuities as all systems in time and space are limited in space and time, both in a single membrane, and in within the scales of the 5th dimension (as information and energy doesn’t flux between those scales without loss of entropy).
It would be in that sense important to understand the need for a finite limit, solving the paradox of Zeno, with the concept of a quanta or limit of a finitesimal.
Now the HUGE QUESTION TO CONNECT DERIVATIVES AND DIFFERENTIAL CALCULUS WITH REALITY IS THIS: WHAT KIND OF FINITESIMALS RATE OF CHANGE ARE WE CALCULATING?
The answer is deeper in 5D, as a finitesimal Dimotion of a being is called an action. A finitesimal is an action of spacetime of any of the 5 Dimensions, an action being a spacetime cycle hence a bidimensional holographic quanta, which can be expressed in any of the graphs, lineal, cyclical, cylindrical or polynomial (complex).
What we will measure then is an action of spacetime, which classifies time cycles in 5 subspecies by its complexity:
Acelerations, lineal motions, entropic motions, energy flows, informative vortices and Social evolutions (a,e,i,o,u).
And the sum of actions is what creates a sequential world cycle. So we must conclude that in ‘time’, a finitesimal is a quanta of a Dimotion; and in space it is a quanta of a volume. But in general terms, derivatives and differentials will be better suit to calculate temporal quanta, and THE TYPE OF DIMOTION STUDIED WILL DEPEND ON THE ‘FUNCTION’ to which the differential is applied:
Why finitesimals and derivatives are so important becomes then evident as they MEASURE IN THE REAL UNIVERSE, THE ACTIONS, which are ∆1 minimal Dimotion units of any system of nature.
So the full theory of CALCULUS becomes related to the actions of exist¡ence of a supœrganism, and as such depending on the functions we use (related themselves to entropic, reproductive, locomotive etc. actions), the derivative will calculate the minimal actions of the being.
Of course this is of not use for complex systems of which we have languages less synoptic than maths to evaluate actions – we do not for example measure a bite of food or an amino acid with a ‘derivative’ but just explain it with full morphological detail of the being involved.
But in physical systems, where the finitesimal values are quite homogenous, actions of spacetime can be measured with derivatives. And theses the main application.
Then we can evaluate sums of action with integrals and in this manner, mathematical physics and 5D connect to each other, through the use of calculus, which is widespread in all the analysis of atomic systems, where the regularity and determinism of its Dimotions rests assured.
So what 5D will introduce in mathematical physics is the connection of equations of physics with the vital actions of its physical systems.
And so in this constant merge of ‘vital noneuclidean geometry’, actions of organic spacetime beings, we can give also vital meaning to the different use of the operand of calculus in FUNCTIONS THAT REPRESENT the 5 Dimotions of exist¡ence.
In that sense the MOST important ad on that ∆st will bring to the use of differentials is its temporal use as the ‘minimal action in time’, of a being, a far more expanded notion that the action of physics (which however will be related to the lineal actions of motion on 1D).
IN THE GRAPH, the general 5 actionsdimensions of existence of different ∆±i species, from above down – a view of them all, one of the physical simplest light and electronic i<eye minds, and below the human being.
Mathematically it is quite irrelevant to make derivatives in time of human actions beyond some quantitative results – so the quanta of minimal human time, the second is the minimal informative action for the 3 synchronous tsts parts, an eye glimpse of mindperception, a limbate of motion and a heartbeat of the body.
Nothing to blame on Nature, rather as Lindau put it, a feature of the human ego: ‘what time uncertainty, I don’t see any time uncertainty in quantum physics, I look at my clock and I know what time is’ (:
Yet mathematical operations are better for simpler social numbers forming herds, moving=reproducing simple information in a few scales of existence, with topological evolution at the height on its capacity to describe complex simultaneous super organisms in more detail. This means the actions best explained by operators are those simpler, and quantitative operations will give then the simplest of the interactions between the elements of those smaller, ‘larger’ ensembles of beings.
So to escape the limits of huminds and mathematical reductionism, it will be important even for physical systems today only described quantitatively in abstract mathematical terms, to vitalise and explain the organic whys of its spacetime events by adding its existential actions to those ‘analysis’:
It will be important for physical systems today only described quantitatively in abstract terms, to vitalise and explain the organic whys of its spacetime events.
The connection on qualitative terms though is selfevident, for all scales, as most actions of any being are extractions of motion, energy and form from lower ∆i scales.
So we and all other beings perceive from ∆3 quanta (light in our case), feed on amino acids, (∆2 quanta for any ∆º system), seed with seminal ∆1 cellular quanta (electrons also, with ∆1 photon quanta).
So derivatives are the essential quantitative action for the workings of any Tœ, spacetime organism.
And so we study in depth the connection of the a,e,i,o,u actions between Planes (qualitative understanding) and its mathematical, analytic development (quantitative understanding of 1st second and 3rd derivatives – the late extracting ‘1D motion’ from the final invisible gravitational and light spacetime scales).
SO THE FUNDAMENTAL LAW OF OPERANDS TO VITALIZE THEM IS THIS:
In Pentalogic ALL differential OPERANDS CAN BECOME AN ACTION IN ONE OF THE 5D DIMENSIONAL VOWELS (A,E,I,O,U) THAT DEFINE THE FIVE dimensions OF EXISTENCE, AS VITAL QUANTAACTIONS OF THE BEING.
THIS IS THE LOGIC CONCEPT THAT TRULY VITALIZES THE OPERANDS OF calculus.
So analysis allow us to extract actions from wholes, among many other uses, reason why THERE ARE NOT REALLY use of the third derivative of a being, as super organisms coexist in 3 only scales. So to speak, if you derivate a world, you get its organism, and if you derivate it again you get its cell and then its molecular parts. And then if you do that in time, you get its speed and then its acceleration and then its jerk.
Of course, this is NOT how simplifying maths work – but it works in terms of a volume, then you get its plane, then its unitcell and its point…
The connection on qualitative terms is selfevident, for all scales, as most actions of any being are extractions of motion, energy and form from lower ∆i scales.
So we and all other beings perceive from ∆3 quanta (light in our case), feed on amino acids, (∆2 quanta for any ∆º system), seed with seminal ∆1 cellular quanta (electrons also, with ∆1 photon quanta).
So derivatives are the essential quantitative action for the workings of any Tœ, spacetime organism.
And so we study in depth the connection of the a,e,i,o,u actions between Planes (qualitative understanding) and its mathematical, analytic development (quantitative understanding of 1st second and 3rd derivatives – the late extracting ‘1D motion’ from the final invisible gravitational and light spacetime scales).
SO THE FUNDAMENTAL LAW OF OPERATIONS TO VITALIZE THEM IS THIS:
‘BY THE RASHOMON EFFECT ALL differential OPERATIONS CAN BECOME AN ACTION IN ONE OF THE 5D DIMENSIONAL VOWELS (A,E,I,O,U) THAT DEFINE THE FIVE dimensions OF EXISTENCE, AS VITAL QUANTAACTIONS OF THE BEING.
THIS IS THE LOGIC CONCEPT THAT TRULY VITALIZES THE OPERANDI OF ALGEBRA.
Derivatives allow us to integrate, a sum of the minimal quanta in space or actions in time of any being in existence, namely the fact that its sums tend to favor growth of information on the being and then signal the 3 stages and/or states of the being through its world cycle of existence, which in its simplest physical equations is the origin of… ITS spacetime beats.
Actions in timespace are the main finitesimal part of reality, its quantity of time or space if we consider tridimensional actions as combinations of S and T states, stt, tst, tss, sss and so on…
So how differential equations show us the different actions of the Universe?
To fully grasp that essential connection between ∆st and mathematical mirrors, we must first understand how species on one hand, and equations on the other, probe in the scales of reality to obtain its quanta of spacetime converted either in motion steps or information pixels, to build up reality.
So for each action of spacetime we shall find a whole, ∆ø, which will enter in contact with another world, ∆±i, from where it will extract finitesimals of space or time, energy or information, entropy or motion, and this will be the finitesimal ∂ ƒ(x), which will be absorbed and used by the species to obtain a certain action, å.
So the correspondence to establish is between the final result, the åction, and the finitesimal quantas, the system has absorbed to perform the action, ∫∂x, such as: å= ∫ ∂x, whereas x is a quanta of time or space used by ∆ø, through the action, å to perform an event of acceleration, energy feeding, information, offspring reproduction or universal social evolution.
It is then when we can establish how operations are performed to achieve each type of actions.
The first element to notice is the fact that the space between the actor and the observable quanta is relative, so even if there are multiple ∆planes between them the actor will treat the quanta as a direct finitesimal, pixel, bit, or bite which it then will integrate with a polynomial derivative or sinusoidal function that reflects the changes produced.
We will consider in this introductory course only a few of the finitesimal ∫∂ actions where the space state is provided by the integral and the ∂ finitesimal action by the derivative.
Further on derivatives will allow us to point out the main consequence of the sum of those actions in any being in existence, namely the fact that its sums tend to favor growth of information on the being and then signal the 3 stages and/or states of the being through its world cycle of existence, which in its simplest physical equations is the origin of… the maximal and minimal points of a wellbehaved function.
∫∂ Operands reflect 5 D¡
The purpose of operands is obvious: to explain the most general laws of Ðimotions of spacetime beings. We put them in correspondence with those Dimotions in the posts of Algebra, and the introduction to Maths. So here we just plunge on the use of analysis makes of this operators when it derivates them to obtain a quanta of a Ðimotion; since depending on which operator derivates it will be a quanta of a different ‘function=ƒn(action) of exist¡ence’.
The all pervading use of ∫∂ is then clearly because it reflects ALL forms of change. And so analysis is the most extended subset of Algebra.
Why there is a negative inverse operand to each of them, is the first question to be considered: simply stated, the 5 Dimotions of reality have its destructive entropic 5th dimotion for all of them. So negative/inverse operand balance positive ones.
∫∂: Finally it came calculus, with its inverse operand, which represent the scalar next social gathering of elements of algebra, as it IS APPLIED TO THE PREVIOUS OPERANDS, AS WHOLES, except for the trivial xª – the previous more complex polynomial, and so we must regard calculus not only as the operand of all dimotions of change, but also as the operand between planes of existence, since the logarithimic/power previous expression, just reaches between the two limits of two planes of the fifth dimension, but calculus allow us to ’emerge’ and transcend between planes.
Indeed, consider the ‘derivative’ of the exponential, which is also the exponential, and that of the logarithm, which is 1/x, where 1 is the whole and x its parts. It gives us the ratio of change of an individual that form a whole or a radiation of a species, with two clear phases; the first of maximal possible growth as the E number is the base with maximal exponential growth of all numbers, but then that growth reduces to a ‘cell’ after a ‘cell’ of the whole, which means it merely maintains the being as the system will also keep loosing its 1/ units slowly.
Thus while all operand are in correspondence mainly to a Dimotion=action of a being, Analysis IS the operand ∫∂ which applies ‘in a secondary level’ to the previous simplex study of Ðimotions in a single plane. So Analysis study all Ðimotions of spacetime, all dimensional motions, all forms of change.
Let us THEN SEE THe analysis of the main functions THAT EXPRESS THE 5 DIMOTIONS of existence and its derivatives:
In the graph the main functions, its derivatives and integrals. Each of them would define under the pentalogic uses of calculus, multiple true insights in the structure of reality, the finitesimals of wholes and its actions. We shall just consider some elements of them.
The simplest ‘mathematical form’ is a ‘social number’, a constant group. But a constant doesn’t change, which also means it has no ‘hair’ that is no scales of 5D depth. It never emerges or degrades. It is a static form, eternal unchanged, and its derivative is zero.
0 then must be interpreted as an infinitesimal for the whole, x/∞, which is the ultimate meaning of a derivative.
But a zero can have WHEN OBSERVED AS A FRACTAL POINT ANY NUMBER OF ‘INTEGRAL’ PARTS. So its integral which give us the parts of the system, while the derivative gives us the value of the whole, is any number, c.
Next the whole x, indeed shows its derivative to be 1, which shows really the meaning of a derivative in terms of scales – the whole, 1. So if 0 is the infinitesimal, which is never truly zero but the minimal element that can be perceived in a plane of existence, 1 is a relative infinite, the whole of the 01=1∞ dual symmetric scales represented in mathematics.
Yet a whole, again, can give us, besides its number of parts (integral), any relative ad on, C.
To notice at this stage that as in so many cases the words are confusing. Derivatives should be call integrals as they give us the number of parts the whole has, which are always less in quantitative terms (the whole is 1, the parts are x, each part is 1/x in terms of the whole total value).
POLYNOMIAL ACTIONS
This understood then a whole expressed as a power of parts, xª, will be reduced if we want to ‘perceive’ its parts as a whole into aX ª¯¹, which simply means there will be a number of parts, converted into X ª¯¹ wholes and vice versa.
Polynomial Algebra as an approximation to Analysis.
The concept of differential and finitesimal makes us possible to consider a different ‘finite view’ of continuity, as the region in which the function DO actually has a meaningful differential, meaning, the region where ∆t truly comes to zero, instead of provoking a huge gap, making DC very large, towards a hyperbolic form:
In those ‘verges’ of the Plane or the T.œ, (singularity center opening to 5D, still mind and membrane, opening to 4D entropic world) continuity breaks because the change in ∆S is huge for small increments of ∆t (time age discontinuity) in the simplest obvious case of 1D analysis or if we are measuring a different type of dimensional change, for example, that of topological form, we find a ‘change’ of state, or form, or region of the being – a topological tearing and transformation.
What happens on those limits of the ‘entity’, its membrane and singularity? Simply we change of state, of topology, of region of the being, and as the limits where the function looses its meaning they are no longer of use.
And indeed, we shall see that in the real use of calculus to resolve problems of mathematical functions, SOLUTIONS TO DIFFERENTIAL EQUATIONS are limited to the regions between the Limits of the functions in spaceform, and in time form most SOLUTIONS required to KNOW THE initial and final ‘conditions’, that is the value of the function in its t=0 and t=t, begining and end of its worldcycle.
Continuity therefore is not always quantitative, but also topological, qualitative:
Powerroot systems and integralderivatives operate fully on the ∆§cales and planes of the system, which require two slightly different operandi. As §¹º ‘social decametric scales’ are lineal, regular, so we can operate them with potencies, roots and logarithms.
∂∫ But when we change between scales into new wholes and new planes of existence we are into ‘a different species’ and so we need to operate with the magic of finitesimal derivatives and analytical integrals, which keep a better track of the infinitesimal ‘curved’ exponential changes that happen between two planes, where linearity is lost.
In that regard the main difference between polynomials/logarithms vs derivatives and integrals is dual:
Algebra is the analysis of systems which focus in numerical social quantities and symmetries between dimensions, than the organic ‘fluid’ properties of systems described through analysis.Now departing from the general rule that f(x) is a function of ‘time motions’ as all variables are by definition time motions, and the Y function its spatial views as ‘a whole’, we can take this as the rule of interpretation ƒ(x)=t=S=Y as a general rule, and as so often we have a function of the type ƒ (x)=ƒ(t)=0, we consider the polynomial a representation of a world cycle. And from that we can differentiate factors through ∆ scales such as… ∆±1= 01 probability sphere (∆<0) and Polynomial (Xª=∆ª).
It is then obvious that one of the key equations of the Universe, the equation that relates polynomials and derivatives, space and time views of complex symmetric bundles must be reinterpreted on the light of those disomorphisms between the mathematical mirror, and the 5D³ Universe.
So polynomials are the rough approximation to the more subtle methods of finding dimensional change proper of analysis – even if huminds found first the unfocused polynomials and so we call today McLaurin & Taylor’s formulae of multiple derivatives, approximations to Polynomials.
Taylor’s formula resumes the main spacetime symmetries and its development, left polynomials, right derivatives fills in the content of algebra in the measure of spacetime systems.
In a given point it can then be understood as a differential value and then consider Dc the polynomial vs. ∫∂ep differential. Lineal functions in short distances’ view that become curved in larger more accurate spatial views, make us think that the ƒ(t) time function is step by step building the ƒ(y) spatial worldycycle, which dId all those step curvatures.
The actions it describes.
The minimal unit for any T.Œ are its a,e,i,o,u actions of existence, its accelerations energy feedings, information processing, offspring reproduction and universal evolution. So the immediate question about mathematical mirrors and its operations is what actions reflect. We have treated the theme extensively in the algebraic post, concluding that being mathematics a mostly spatial, social more than organic language, its operations are perfect to mirror simple systems of huge social numbers=herds; and as such to describe the simpler accelerations=motions, which are reproductions between two continuous scales of the fifth dimension; informative processes, where the quanta perceive are truly finitesimal ∆i elements pegged together into the mirror images of the singularity and so we talk of motions, simple reproductions and vortices of information, and time>space processes of deceleration of motion into form, as the key actions reflected by mathematical operations.
It also follows that when we study the more complex systems and actions of reality, reproduction and social evolution of networks into organisms, mathematics will provide limited information, and miss properties for which illogical biological and verbal languages are better.
And it follows that physical and chemical systems are the best to be described with mathematical equations, either in algebraic terms or analytic terms, which fusion together when we try to describe the most numerous, simpler systems of particles and atoms (simpler because by casting upon them only mathematical mirrors we are limited to obtain mathematical properties).
1D: PERCEPTIVE ACTIONS
Next comes from the bottom of that list, the functions of perception, sin and cos angles; and the result have some ‘metaphysical’ meanings. Indeed, the rate of change of our informative angle measure (the sine), becomes the cos, the rate of change of our motion, or in other words, we SWITCH from sinstop states to cosmoving states, in stœps. We go from stopsin to stepcos; but the inverse doesn’t hold. That is if we go from motion cos to stop sin, this will be perceived from the perspective of cosmotion as a ‘negative’ reduction motion – sine.
Finally the logarithmic function and exponential function the ratio of change (derivative) diminishes from the absolute maximal, eª, which is its own derivative, to the absolute minimal 1/a the log derivative which is the definition of an infinitesimal part (Leibniz(, till it peaks, converting an ∆1 first unit into an ∆º whole in the peak of an existential world cycle that then will start an inverse function of decay with 1/x diminution and a final fast collapse in the 3rd age<<death moment at eˆa speed.
So the combination of ± exponentials and logarithm curves are also the best way to graph as a bell curve the worldcycle of existence in lineal terms.
PRODUCT AND INVERSE DIVISION INTEGRALS
We defined in Algebra the product as the king of all operations, since it ‘merges’ into ‘a new entity of spacetime’, two parameters disjoined previous to the product, proving the very same existence of a holographic Universe, often through a merge of the ‘¡1’ elements or ‘cellular parts’ of the being, which create ‘axons’ of communication with all the other parts of the being, such as:
X(5¡1) x Y (4¡1) = {X(5¡1)Y(4¡1)}20¡2
That is, a multiplication that merges two elements gives us the number of i2 axons of communication connecting at a deeper level the two parts.
In a momentum the mass stop state and the wave step state merge in the potential I2 level that holds them together.
The most abundant of all operand, the merging product requires therefore a more complex rule than a direct sum, which acts by ‘superposition’ of EQUAL BEINGS.
IT IS ALSO susceptible to be operated by calculus and ‘derivatives’ as now we INVOLVE FOR THE FIRST TIME, BOTH, a SCALAR LEVEL, since multiplication tends to happen in the lower scale of the being and different states of time and space. So we no longer operate as in additions, with the same type of T.œs in the same plane.
The most abundant of all operand, the merging product requires its own rule which interesting enough shows how indeed product is a merging operation, as the derivative of a product of functions merges first each function with the change rate of the other, and then once both are merged, superposes them by addition:
The Product Rule used to find the derivative of a product of two functions, is thus more complex than the sum even though it also keeps as in polynomials the distributive property – which shows once again that the product is a ‘democratic merging’ that can go both ways.
So h'(x) = [ƒ(x) x g(x)]’ = ƒ(x) • g'(x) + ƒ'(x) • g(x).
The rule, interesting enough shows how indeed product is a merging operation, as the derivative of a product of functions merges first each function with the change rate of the other, and then once both are merged, superposes them by addition.
In that sense it keeps with the ‘rule’ of merging at the lower ‘plane level’ of its infinitesimal parts, in this case, taking instead of the spatial elements of X and Y, its ‘temporal’ quanta of change, f(x)’ and g(x)’, MERGING them with the other wholes, before a ‘superposition’=addition can be effected.
In the product rule thus Derivatives act in inverse fashion to power laws, searching for the infinitesimal.
While power wholes (integrals) search the wholeness, and as we know the two directions of spacetime are different in curvature, quantity of information and entropic motions.
Here we shall bring a little explained fact – derivatives act in the inverse fashion to power laws, searching the infinitesimal, while power wholes (integrals) search the wholeness, and as we know the two directions of spacetime are different in curvature, quantity of information and entropic motions.
So an external operation that reduces a whole which is NOT integrated as such but a lineal product of two wholes, ƒ(x) and g(x), a COUPLE, is mixing the infinitesimals of one, with the other whole before herding them; in a process of ‘genetic mixing’ of the parts of the first shared with the second whole and the parts of the second shared with the first whole.
This law of existential algebra simplified ad maximal as usual in mathematical mirrors surprisingly enough is the origin also of genetic ‘reproduction’, which occurs at two levels, mixing the ‘parts’ – the genes of the whole – in both directions to rise then the mixing to the ∆º level of the G and F gender couple.
Then WHAT WILL COME out of that genetic multiplication is its division into two equal parts, showing how the INTERACTION OF INVERSE operandi DOES NOT CANCEL REALITY but MERELY COMPLETES A DIMOTION moving ahead the eternal time space Universe.
So if a power followed by a logarithm brings the infinitesimal seed into a whole herd, the multiplication followed by a division of the reproduced new layer of mixed ‘axons, genes’ or parts, brings the replication of identical forms.
While the simplest definition of a division is as usual in huminds an entropic destructive feeding action, the complex view from the perspective of information is a genetic mitosis. And both are reflected in the derivative of a division, which is impossible for two equal functions (resulting in 0 constant) and viceversa can give us any constant value in its integral – so it does not give us any information.
While in most cases is NOT a positive communicative act but a perpendicular negative reducing game, where the DOMINANT element is the ‘predator’ larger denominator that cuts the function, multiplying its infinitesimal f'(x) parts, to which it will deduce the lesser parts absorbed by the f(x) function from it, and then cut it at the ‘lower’ level of its potential elements (G(x)²) :
So the numerator, the victim, shared by the denominator the predator so to speak is first absorbed in its ƒ'(x) parts, g(x) ƒ'(x), subtracting the g'(x) parts that the prey has absorbed in the ‘fight’, ƒ(x) g'(x), and then shared by the parts, g(x)² of the whole as entropic feeding.
So we can consider the derivative of a divisive function as an ‘idealized’ expression of the process of killing and feeding of a system, whereas the predator absorbs the infinitesimal parts of the other being, and feeds its cellular, i1 elements with it.
Which obviously is NOT a commutative process.
Of course, we love to bring vital interpretation to abstract math, but as we apply such rules to particular cases, the interpretations vary but in all cases will be able to be interpreted in terms of subequations of the fractal generator.
What might be notice in any case is that unlike in our rather ‘abstract’ dimensional explanation of the rules of power laws, here we are able to bring real vital analysis of those roles in terms even of biological processes, showing how much more sophisticated is the ∫∂ operandi, the king of the hill of mathematical mirrors on real step motions and actions, reason why its use is so wide spread.
SO THE FUNDAMENTAL LAW OF OPERANDS TO VITALIZE THEM IS THIS:
‘BY THE RASHOMON EFFECT ALL differential OPERANDS CAN BECOME AN ACTION IN ONE OF THE 5D DIMENSIONAL VOWELS (A,E,I,O,U) THAT DEFINE THE FIVE dimensions OF EXISTENCE, AS VITAL QUANTAACTIONS OF THE BEING.
THIS IS THE LOGIC CONCEPT THAT TRULY VITALIZES THE OPERANDI OF ALGEBRA.
So those properties tell us new things about the meaning of ∫∂.
Finally the chain rule WHICH IS TRULY the one that encloses all others is
used in the case of a function of a function, or composite function writes:
And this truly an organic rule, as we are not derivating on ‘parts’ loosely connected by ± and x÷ herds and lineal dimensional growth, but the ‘function’ is a function of a function – a functional, as all ∆+1 is made of ∆º which are also functions of xo fractal points.
So this is the most useful of all those rules to mirror better reality. And we see how the derivative, the change process deeps in at the two levels, at the ∆º=g(xo) level, which becomes g'(xo) and at the whole level, which becomes ƒ'[g(xo)], which tell us we can indeed go deeper with ∫∂ between organic scales, which is what we shall learn in more depth when consider partial derivatives and second derivatives and multiple integrals.
We are getting so to speak into the infinitesimal of the parts of a whole from its ∆+2 perspective, and thsi rule encloses all others, because it breaks into the multiplication of its parts – DWINDLING TRULY A SCALE DOWN, AND SEPARATING THE WHOLE AND THE PARTS DERIVATED INTO LOOSE PARTS AND FINITESIMALS NOW MULTIPLIED.
And what will the parts do when they see their previous finitesimals now camping by themselves but ‘at sight’ to get them to ‘produce’ an operative ‘action’ (a,e,i,o,u actions are ALL subject to the previous operandi), ON them.
AND WHAT WILL COME of that multiplication. Normally it will capture them all again and then normally will not re=produce on them (one of the operandi actions which are possible under the rashomon effect) but divide and feed on them the last operation to treat:
And its inverse, which is NOT a positive communicative act but often a perpendicular negative reducing game also consequently differs.
In that sense the MOST important ad on that ∆st will bring to the use of differentials in EXISTENTIAL ALGEBRA, is its temporal use as the ‘minimal action in time’, of a being, a far more expanded notion that the action of physics (which however will be related to the lineal actions of motion on 1D).
Time evolution equations.
Time evolution is the change of state≈age brought about by the passage of time, applicable to systems with internal state≈age distribution (also called ageful systems).
In this formulation, time is not required to be a continuous parameter, but may be discrete or even finite. And so we can use frequencies and densities, fluxes and all the elements required for a real description of the ∆ST universe.
In classical physics, time evolution of a collection of rigid bodies is governed by the principles of classical mechanics. In their most rudimentary form, these principles express the relationship between forces acting on the bodies and their acceleration given by Newton’s laws of motion. These principles can also be equivalently expressed more abstractly by Hamiltonian mechanics or Lagrangian mechanics; which themselves use the ∫∂ jargon.
The concept of time evolution may be applicable to other state systems as well. For instance, the operation of a Turing machine can be regarded as the time evolution of the machine’s control state together with the state of the tape (or possibly multiple tapes) including the position of the machine’s readwrite head (or heads). In this case, time is discrete.
State systems often have dual descriptions in terms of states or in terms of observable values. In such systems, time evolution can also refer to the change in observable values. This is particularly relevant in quantum mechanics where the Schrödinger picture and Heisenberg picture are (mostly) equivalent descriptions of time evolution.
Calculus of finitesimals, ∂, and its integrals, ∫ are a dual dimotion, back into ∆1 to extract a part, and forwards into its integral ∫ through an spatial or temporal finite domain. So the inverse operations of analysis have multiple functionality in terms of actions performed through them, because of its perfect mirroring of the action itself, which consists in using ∆i finitesimals to absorb energy, motion or information for the 3 simpler actions of motion, informative perception, and energy feeding for which paradoxically, the more complex ‘organic’ operations are the most useful.
This paradox though has a less ‘motivating’ cause for those involved with the mirror of mathematics – essentially that mathematics is NOT the best language to describe the complex actions and relationships that appear out of the reproductive biological and social, engaging processes of organisation, at least algebraic operations.
NEXT OPERAND IN COMPLEXITY IMPROVES UPON THE PREVIOUS ONE (POWERS)
It is a rule of the scalar Universe that all the action are chained such as to effect a more complex action we need the previous ones, as the basic chain of 1D perception>2D locomotion>4D entropic feeding>3D reproduction>5D social evolution into a larger whole shows.
This chain can then be expressed with the classic operands from angle to sum to division and product to power law. But then once the carry capacity of the system is completed at 5D which is a larger 1D we emerge into a larger world and here we get results only with the ∫∂ operands, which therefore must be related to a degree of accuracy in its simplest forms to the more complex forms of the polynomial, and this is the 5D why explanation of a well known rule to approximate functions by its higher derivatives.
Indeed, in Algebra, the third and higher derivatives are used to improve the accuracy of an approximation to the function:
f(xo+h)=f(xo)+f′(xo)h+f″(xo)h²/2!+f‴(ξ)h³/3!
Thus Taylor’s expansion of a function around a point involves higher order derivatives, and the more derivatives you consider, the higher the accuracy. This also translates to higher order finite difference methods when considering numerical approximations.
Now what this means is obvious: beyond the accuracy of the three derivatives canonical to an ∆º±1 supœrganism, information as it passes the potential barrier between scales of the 5th≈∆dimension, suffers a loss of precision so beyond the third derivative, we can only obtain approximations by using higher derivatives or in a likely less focused=exact procedure the equivalent polynomials, more clear expressions of ‘dimensional growth.
So their similitude first of all proves that both high derivatives and polynomials are representations of growth across planes and scales, albeit loosing accuracy.
Let us briefly then deal with the operands of the 5th dimension treated extensively in ∆nalysis:
However in the fifth dimensional correct perspective is more accurate the derivativeintegral game; as it ‘looks at the infinitesimal’ to integrate then the proper quanta. Let us briefly comment the operands treated with the 5th dimension of ∫∂ treated extensively in ∆nalysis:
As we did with the other operands, we need to consider the properties of calculus and its two operandi. This poises a problem, as there is not a ‘bottom operation, such as ±, x÷, directly related as with powers as the third dimension of calculus. But as calculus is a refined analysis of power laws, the direct connection is not exact.
Hence a certain discontinuity is established what implies that ∫∂ equations have been solved by the obvious method of applying the function h’ (x)= lim h>o h (x+h) – h (x)/h. We are not though to repeat here that procedure to get the results but merely analyse from ∆st perspective as we did with power laws, and x, the properties of derivatives, to see what they tell us in the higher T.œ language and then consider some specific functions and its integral and derivatives to learn more of it.
Those key properties are expressed in its rules of calculus, starting from the ‘derivative’ of a polynomial:
Xª= a Xª‾¹
So we are NOT fully lineally diminishing a polynomial dimension despite being derivatives a reduction of dimensionality – the search for the finitesimal 1/n quanta. Why? Obviously because in the rough view from a quanta, xª into its whole xª+¹, we grow lineally (polynomial), but as we repeat ad nauseam, the lineal steps curve into geodesic closed wholes, in the ∆+1 scale (NonE geometry), from the lineal spatial mind to the wholeness cycle of the closed being, and so as the ‘curve of a parabola’ diminishes the distance of a cannonball, growth is NEVER lineal but falls down as we approach the ‘(in)finite limit’.
IN THE GRAPH, the wholeness is curved upwards, the parts spread scattering entropically. The whole is a mind circle, @. So it curves/diminish the quantity of energy available, for the whole, as it really must be an addition of all the planes that share that vital energy to build ever slower, curved larger wholes.
Or in terms of the integral function:
And here we find the second surprise. There are ∞ integrals with the addition of a constant. As a constant by definition does not change.
Let us express this in terms of past (∆1: derivative )< Present (Function) > future (∆+1: Integral)
The past is fixed, the infinitesimal enclosed, only one type of species, ‘happening already’, as the parts must exist before the wholes to sustain them. But from the pov of the present function, the future integral into wholes is open, with ∞ variations on the same theme; unless we have already enclosed that whole, limiting its variations, which happens with the definite Integral.
So if the function f(x) is given on the interval [a, b] and, if F(x) is a primitive for f(x) and x is a point in the interval [a, b], then by the formula of Newton and Leibniz we may write:
Here the integral on the right side differs from the primitive F(x) only by the constant F(a). In such a case this integral, if we consider it as a function of its upper limit x (for variable x), is a completely determined primitive of f(x). That is the importance of the enclosure membrane to define a single organism, and establish its order, as opposed to the entropic, multiple open future of a nonenclosed vital function which will scatter away.
Consequently, an indefinite integral of f(x) may also be written as follows:
where C is an arbitrary constant, the enclosure will eliminate.
Linearity: Yet and this seems to contradict the previous finding, when we operate derivatives with the ‘basic dimension’ of social herding, ± operators, linearity comes back, and so the minimal Rashomon effect give us two explanations:
Γ(st): We are INDEED herding in the base dimension of a single plane, where each derivative will now be considered a fractal point of its own:
∆+1 perspective: Suppose f(x) and g(x) are differentiable functions and a and b are real numbers.
Then the function h (x) = aƒ(x) + bg (x) is differentiable and h’ (x) = a ƒ'(x) + b g'(x), which is really the distributive law already studied in algebra’s post for x and power law. So the interpretation of the sum rule from ∆+1 is one of ‘control’:
WHEN operating from a whole perspective, the whole breaks the ‘smaller’ parts and its simpler dimensional operandi, +, to treat each part with its ‘whole action’ (in this case ∂). In brief the whole totally control the parts.
5D FINITESIMALS IN SPACE, TIME AND SCALE
TIME FINITESIMALS: ACTIONS
Let us recall then how actions are added through frequency integrals of time instead of area integrals of space, through a mathematical method called…
2nd dimension line=sum of points:
Why both definitions work? In pure equations, T=1/ƒ=1/ð.
In depth, because both are isomorphic definitions, albeit in ‘different scale’:
The continuous definition focuses on the ∆1 ‘potential field of forces’ over which the system reproduces its wave of form. So the ‘frequency steps’ are substituted by the external ‘nanoscopic’ continuous (indistinguishable) gravitational and electromagnetic fields over which the ‘being’ slides unaware of the invisible=indistinguishable field over which glides. In the previous equation we adopt the ∆º point of view, internal to the being, where its quanta are much larger, and not subject to derivatives.
It is then important to notice that the need for a ‘function to be continuous’ implies to study STeps in which the actions happen in a lower ‘scale’ of being; hence we talk of the primary actions of motion (Max. S) and perception (Max. I), of minimal forms (∆Max. i) in relationship to the actor and/or death processes of entropy. We can hardly establish ∫∂ operations for the complex social actions of the 5th dimension, and many of the 3D reproductive actions of seminal ∆1 seeds, for where a qualitative analysis of evolutionary topology is more proper.
Reason why the operation of ∫∂ is more proper of the 1D, 2D and 4Dimensions of existence.
Speed is important, on the other hand, as the ratio (s/t) continuous speed or s x t (discontinuous: step x frequency) it is one of the 3 fundamental ratios, s/tspeed, t/sdensity and txsmomentum that define in its simplest form, the singularity vital energy and angular momentum of the 3 parts of an organism, which for the perfect being, s=t, are, 1, 1, and 1.
S=T: ANALYSIS ON SPACE
The second consideration on the rashomon effect should be on analysis of SPACE and transformations between space=form and motion=time states. SO FIRST we shall remember what space is made of – namely NonEuclidean points:
Dimensions and analysis are possible because points have volume.
The fundamental truth derived from this simple analysis of derivatives is profound. First it connects them immediately with the pure geometric nature of dimensions, which in nonEuclidean geometry (graph) are relevant in as much as they represent motions in time but also dimensions in space.
In that regard, it is important to understand that in the fractal Universe a dimension has ‘always inner breath’ as the points grow when we see them closer.. So it is very simple to consider a single dimensional being, simply as one, whose preferential Xdimension is much larger than the others, but still the other exist as the particlepoint in detail is big:
1D being: X>>Y ≈ Z, for example a string, a lineal momentum…
And then a two dimensional being one whose two D are larger than Z:
2D Being: X ≈ Y >> Z; for example a graphene sheet; a plane wave.
Whereas a 3Dimotion being has volume, motion on the 3, for example a spherical being, an entropic explosion.
A derivative then merely ‘annihilates’ one dimension or one motion in space or time – we have here to split dimotions, as humans do, even if it is not the proper unit of the Universe, which is always bidimensional. I.e. even in a motion there is a particle that moves, so you have a pointdimension for the particle and one for the motion in time…
So indeed analysis IS the main mathematical instrument to study the 5D Universe and its ternary mirror symmetries between scales, topologies and modes of timechange. And we can consider a general formulae for analysis, as a specific version of the fractal generator:
∂(Bodywave of vital energy) = Membrane; ∂Membrane = Singularity path and its inverse, better known as line integrals, surface and volume integrals.
Because analysis is mainly used in mathematical physics, in praxis, the previous relationship is connected to the 3 elements of a physical system:
Field (entropic, locomotion source) < wave (reproductive body) > Particle.
So we make double derivatives to obtain the field (Laplacians), and single derivatives to relate particles and waves – ‘onedimensional species.’ (Fourier series). And those are the all pervading analytical functions of the 3 parts of the being:
The result are spherical harmonics, a set of functions used to represent functions on bidimensional membranes – surface of the sphere – the higher dimensional homology of Fourier series – periodic, single variable functions on the circle.
Spherical harmonics are thus the eigenfunctions of the angular part of the Laplacian, representing solutions to partial differential equations in which the Laplacian appears. Since the Laplacian appears frequently in physical equations (e.g. the heat equation, Schrödinger equation, wave equation, Poisson equation, and Laplace equation) ubiquitous in gravity, electromagnetism/radiation, and quantum mechanics.
The orbitals of the hydrogen atom in quantum mechanics in fact are totally undistinguishable from spherical harmonics, showing indeed that we are all topologic beings, and mathematical functions for the simplest forms of spacetime as the electron is – a dense function of ‘light spacetime particlepoints.
The intimate connection between the 3 elements of the being thus is perfectly explained by the dual ∫∂ functions.
In that regard, variations over the same theme respond to the ternary structure of all T.œs:
In the graph, when deriving and integrating, most operations refer to a ‘limited’ system, in which first we extract the finitesimal partelement, and then we integrate it to obtain a whole; so most likely the system described with depart from a timechangingvariable quanta, and integrate it to obtain a ‘static wholespatial view’.
But variations on the same theme happen by the natural symmetry of space and time states.
So we can also start with a quanta of space integrated over time to get a spatial area or volume.
What we shall always need to find ‘single solutions’ is the parameters that describe in time or space the 3 elements of the T.œ: So we shall start with initial or final conditions (definite integrals), and define mostly in space as a whole, the enclosure or membrane thate limits the domain of the function (which might include as a different limit the singularity).
All in all the analytical approach will try to achieve a quantitative description of the unit/variable of ‘change’, the ‘finitesimal quanta of space – interval, area, volume’ or the ‘steps of time’ (frequency), and then integrate it over a super organism of space or an interval of time, we wish to study, often because it forms a whole or a zero sum world cycle.
Galilean Paradox. LINEAL vs. Cyclical view.
In that regard, the S=T symmetry will once more become essential to the technical apparatus of analysis as it has done in all other sub disciplines.
Of them the 3 key ‘dualities’ between lineal perception in short and cyclical perception in large, is the key to obtain solutions, as the mind of measure is lineal made of small steps that approximate larger cyclical wholes. It is in essence the method of differential equations, where the differential dy= ƒ'(x) ∆x + α∆x, approaches to a lineal derivative, ƒ'(x) ∆x in short increases, and so we can get away with the smaller element that curves in longer distances the solution.
Finally the third Galilean paradox between continuity and discontinuity is also at the heart of analysis (and most forms of dual knowledge). Analysis has accepted as a dogma the continuity of the real number and so it considers continuity a necessary condition for differentiability but we disagree in a discontinuous Universe, continuity has a loose definition (as neither the axiomatic method is the proof of mathematical statements but experience also matters). So continuity is defined by a simpler rule: that the term α∆x of the discontinuity between the lineal and cyclical view of an infinitesimal derivative does indeed diminish faster the closer we are to the point ‘a’ in which the differential equation is defined. In brief, continuity means no big jumps and big changes in the direction of a function and the T.œ it reflects.
∆±¡: SCALES, PLANES
Following the Rashomon effect we shall thus consider now some basic themes of analysis to calculate scales and planes.
Galilean px in analysis: finitesimal steps (derivatives) integrated to calculate a cyclic whole.
Further on analysis has over all other branches of mathematics a special quality to study ‘changes’ between planes of the fifth dimension, as multiple derivatives ‘jump’ (albeit with different degrees of ‘focus’) better than mostly ‘lineal polynomials’ between planes, and the ‘curvilineal, Lorentzian’ variations, slow downs and accelerations on the S x T= K parameters happen between scales:
The formal stience of the 1st and 5th Disomorphisms in the mathematical mirror is analysis, which deals directly with the relationships between ∆1 ‘finitesimal’ parts’ and (in)’finities’. Two new terms we still accompany, with the lost inflationary term ‘in’; since infinitesimals and infinities are a Kantian paralogism; as all planes have a limit in its quantic units, and all wholes a finite circle that encloses them into a relative 01 ‘circle unit’.
Besides the duality of the 01 probabilistic mind unit which reflects the external 1∞ universe, a second duality that weights heavily in analysis is that of perception of lineal vs. cyclical form: We are minds of space that measure time cycles: ∫@≈∆ð.
Hence the equation of mindmeasure defines the understanding of differential calculus: As always in praxis, the concept is based in the duality between huminds that measure with fixed rulers, lineal steps, over a cyclical, moving Universe. So Minds measure Aristotelian, short lines, in a long, curved Universe.
So the question comes to which minimalist lineal step of a mind is worthy to make accurate calculus of those long curved Universal paths.
The general rule to identify both polynomials and analysis, is this:
Y=S= ƒ (x=t)
The difference between lineal polynomials and nonlineal analysis
In the graph, we explain the difference between a polynomial ‘regular’ description of a system as it changes in the “Newtonian’, social scale in which changes are not of quality as much as of quantity and the analytical region in which there is a change of structure and hence of quality with irregularities better shown by analytic operations.
It follows that more important than ‘variables’ are to algebra ‘operands’, whose encoded meaning and ‘magic’ way of relating systems to get a ‘future or present’ outcome by merging them according to certain rules of ‘creative engagement’, truly gives the power to algebra to mirror the a(nti)symmetries of the Universe.
The key connector of T.Œ with classic science is the full understanding of the dual algebra operands, ±, x/, ∂∫, √xª as part of the ¡logic, pentagonal game of reality in all its mirror symmetries; that is, as dimotions≈actions and structural elements, whereas an immediate correspondence between those operands and the ternary elements ∆@st can be established as follows:
 The sumrest are the inverse arrows of the simplest superpositions of dimensions between species which are identical in motion and form.
 The product/division rises the complexity of operands a first layer, and serves the purpose, besides the obvious sum of sums, of calculating the margin of dimensions, as combinations which are not purely parallel between clone beings, most likely through the recombination of its ∆1 elements, as the product of 2 Sœts inner elements give us all possible combinations. Ie. 5 x 4 = 20 IS also the number of connections between all the 5 elements and 4 elements of both sets. So multiplication ads either a dimension of multiple sums in the same plane, or probes for the first time in an inner scalar dimension.
 Then we arrive finally to the powersroot systems and integralderivatives, which operate fully on the ∆§cales and planes of the system, which require two slightly different operands. As §¹º ‘social decametric scales’ are lineal, regular, so we can operate them with powers, roots and logarithms.
 ∂∫ But when we change between scales into new wholes and new planes of existence we are into ‘a different species’ and so we need to operate with the magic of finitesimal derivatives and analytical integrals, which keep a better track of the infinitesimal ‘curved’ exponential changes that happen between two planes, where linearity is lost. The integral/derivative thus will be related to the closely connected ‘mind integration’ of information.
Ultimately a derivative of a larger world, measure in a still time point of zero latitude) and the processes of integration of parts into wholes that always discharges part of the being, reason why a derivative is essentially smaller than the power operand, as those processes eliminate part of the whole. This being a key technical element of analysis (which often is approached by binomial series – McLaurin, Taylor, etc. – connecting both operands, but reducing the power series to that ‘a’ constant timespace point in with the mental or whole integration takes place.
SPACE FINITESIMALS VS. TIME FINITESIMALS
We must ‘differentiate’ when differentiating (:
Space finitesimals, which are the minimal quantity of a closed energy cycle or simultaneous form of space, easier to understand, as they are ‘quanta’ with an apparent ‘static form’, which can be ‘added’, if they are a lineal wave of motionreproduction, along the path; or can be integrated (added through different areas and volumes), to give us a 3D reality.
Time finitesimals, which are the minimal period for any action of the being and will trace a history of synchronicities as the actions find regular clocks, which interact between them to allow the being to perform ALL their 5D actions needed to survive. So we walk (A(a)), but then eat energy (Å(e)), and we do not do them often together. Actions have different periodicities, for EACH species that perform 5 actions. So to ‘calculate’ all those periodicities in a single allencompassing function we have to develop a 5D variable system of equations.
– Spacetime finitesimals. But more interesting is the fact that Nature works simultaneously integrating populations in space and synchronising their actions in time. So we observe also spacetime finitesimals where the synchronicity consists in summoning the actions of multiple quanta that perform in the same moment the same ‘Dmotion’, which is ‘reinforced’ becoming a resonant action.
And for the calculus of those spacetime finitesimals the best way to go around is by ‘gathering the sum of ∆1 quanta’ into a ‘larger ∆º quanta’ treated as a new ‘1’ adding up its force. EVEN IF most of them are just complex ensembles of the simplest actions of many cellular parts – steady state motions, reproduction of new dimensions and vortex of curvature and information absorption.
All functions of analysis thus can be considered operations on actions of spacetime.
Groups of Finitesimals and their synchronous actions thus meet at ∆º in the mirror of mathematical operations, through the localisation of a ‘theoretical’ tangent≈ infinitesimal of the nanoscale (∂s/∂t proper) or an ‘observable’ differential, a larger finitesimal, which is the real element, as any finitesimal is a fractal micro points that have a fractal volume, expressed in the differential.
Then we gather them, in time or space and study their ‘inverse’ action in space or time.
So the first distinction we must do is between finitesimals expressed as functions of time frequencies and finitesimals expressed as areas of space. And the actions described on them. In practice though most finitesimals are spatial parts whose frequency of action is described by the ƒ(x)=t function.
The 3 parts of T.œ.
If we call energy, e, then:
∑$p x ðƒ = ∆e becomes the integral of the inner spatial quanta of the open ball, surrounded by the membrane of temporal cycles, which conserves its Energy and by the sum of all T.œs that of the Universe. Its calculus, after finding a ‘continuous derivative’, surrounded by the membrane is then an integral: ∫ Sp x ðƒ = Ke.
And inversely. If we consider a single quanta of space or a single frequency of time, a moment of lineal or angular momentum, the result is a derivative.
So Analysis becomes the fundamental method to study travels upwards and downwards of the 5th dimension.
In general if we call a spatial quanta a unit of lineal momentum of each scale and a time cycle a unit of angular momentum, the metric merely means the principle of conservation of lineal and angular momentum.
Thus analysis studies the process which allows by multiplication of ‘social numbers’ , either populations in space or frequencies of time, a system to ‘grow in size’; which is the ultimate meaning of travelling through the 5th dimension. For example, when a wave increases its frequency, it increases the quantity of time cycles of the system. When a wheel speeds up its increases the speed of its clocks. And vice versa, when a system increases its quanta, growing in mass, or increasing its entropy (degrees of motion in molecular space), it also grows through the 5th dimension.
And the integration along space and time, of those growths, is what we call the total Energy and information of the system
It is what physicists call the integral of momentum or total ‘Energy and information of the system ‘
So we shall only bring about here some examples of analysis concerned with the definitions of the fundamental parameters of the fractal Universe, that is the conservation principles and balances of systems which can be resumed in 2 fundamental laws:
Points of constrain, balance and limits of integrals.
Any equation with a real, determined solution must be a complete T.œ. Hence it will have limits either in space (membrane and singularity of the open ball), or in time, initial and final conditions, bridged by an action in the ‘least time’ possible.
This is the key ∆st law that applies to the search for solutions in both ODE and PDEs.
Maximise its ðƒ/Sp, density of information/mass, its Sp/ðƒ density of energy and hence, reach a balance at ðƒ=Sp
This simple set of equations: max. ðƒ x Sp > Tƒ=Sp: max Tƒ/Sp and Max. Sp/Tƒ are therefore the fluctuation points of systems that constantly move between the two extremes of information and spatial states across a preferred point of balance Sp=Tƒ as this is the max. Sp x Tƒ place.
Thus integrals, Lagrangians and Hamiltonians are variations of those themes. The motion of springs; the law of least time etc. all are vibrations along a point of balance, Tƒ=Sp, and 2 maximal inverse limits.
Dimensional integration. Dimensions of form that become motions and vice versa.
Now the key to fully grasp the enormous variety of integral and derivative results obtained in all sciences, is to understand that all space forms can be treated as instants in time, or events of motion, and all motions in time can be seen as fixed present moments in space.
These series of combinations of time and space, S>T>S>T, which leaves a trace of steps and frequencies and its whole integration, which emerges as an ∆+1 new scale of reality is at the core of all fractal, reproductive processes of reality.
For example the sT duality is at the core of the Galilean paradox of relativity (se mueve y no se mueve), of Einstein’s relativity, of zeno’s paradox.
So we can consider motion in time as reproduction of form in adjacent topologies of discontinuous space.
We can consider the stop and go motions of films, picture by picture, integrating those ‘spatial pictures’ into a time ‘motion picture’.
We consider the waveparticle paradox, as waves move by reproduction of form and particle collapse by integration of that form in space into a timeparticle.
In those cases integration happens because a system that moves in time, reproduces in space. And vice versa, steps in space become a memory of time.
Now it is important also to study case by case and distinguish properly what are we truly seeing population in space or events in time, as we can and often it happens that humans confuse in quantum physics where motion is so fast that time cycles appear as forms of space. We shall then unveil many errors, where a particle in time is seeing as a force in space (confusion of electroweak, transformative force as a spatial force,and so on).
All systems can be integrated, as populations in space to create synchronous super organisms and as world cycles in time, creating existential cycles of life and death. The population integral will be however positive and the integral in time will be zero.
Since. systems of populations in space do have volume. Yet the whole motion in time, can be integrated as closed paths of time, or conservative motions that are zero sums, and this allows us to resolve what is time integration and space integration.
Consider to fully grasp this, the reproduction of a wave, which constantly reproduces its form as it advances in space, and cannot be localised (Heisenberg uncertainty) because it is a present wave of time, as light moves NOT in the least space but the least time. Now, consider a seminal wave – you, which reproduces in time, but becomes a herd of cells that integrated emerges into a larger scale. In both cases the final result is in space and so it is positive.
So as I said, for each case the process must be studied but the results will give us the conclusion that we are observing a time event or a spatial organism.
In that regard the most important and hence first view of the Rashomon Effect on ∫∂ is:
∫≈∂ ARE TIME=SPACE BEATS/STEPS IN ANY D²
We have further defined the Disomorphism between the 5 Dimensions of spacetime and the 5 actions=motion=operators of mathematical space.
An operation or actor is thus a Disomophism of a language or form, which enacts through the operator mirrors of the language or form, which in mathematics are the operandi, ±, x ÷, xª log, ∫∂… but in other species as all of them encode, the social evolution, darwinian fight, decay, growth, functions of a super organism might be coded as 5D functions by genes, or words, or any other syntactic form.
It is the most fruitful ∫∂ symmetry, soon used by Leibniz and Newton to develop laws of lineal time motion in space, IN WHICH the full realisation of all other views BECAME THE BIGGEST MIRACLE OF magic mathematics.
Alas, the entire planet was astonished, when in a not yet fully understood scalar duality, derivatives in time turns out to be inverse to volumes of an integral of space?! This was the biggest surprise of mathematics since the finding by Pythagoras of the irrational pi. Why God had made coincide two different operations till then seemingly not related to each other; derivatives of time motions and volumes of spatial form?
Answer because according to the Galileo’s paradox, the first insight that prompted me 30 years ago to discover 5D², time and space are indeed the two sides of a holographic 2manifold dimension.
So a motion in time decelerates by becoming a new dimension of space, and indeed a moving curve is equivalent to a surface, it generates when we measure as space. So lineal timemotions produces space surfaces:
In the graph, a dimension of space volume transforms into a dimension of angular time motion, and so we can apply a derivative and integral duality as there has been an S>T stepmotion of spacetime.
4D: S∂
ENTROPIC MOTIONS: DERIVATIVES
ðIME VIEW
NEXT, to this realisation we must wonder the question of causality, which is expressed in terms of independence.
It seems then that most spatial functions are dependent on time the independent factor: $=ƒ(t)
Yet as we recall that time motions ‘stop’ into space we can interpret this independence in terms of order: as functions are first motions in time that stop and become ‘forms’ of space, leaving a pastmemorial trace. which if NOT erased becomes a population of space, which moves again and then becomes a population and in this manner reproduction of dimensions takes place, building a being of growing ∆Dst.
How this is expressed in ∫∂ terms becomes then clear: since time is discrete, discontinuous, made of a toe, moving, stopping (often perceiving), moving and stopping, we must first ‘encounter’ the minimal step of the time motion, and this is what we shall call dt, and then move, stop and move stop a number of steps, which we integrate, building in this manner a new dimension of spacetime.
So the combination of ∫∂ is in fact a process of creation of an ST dual dimension of spacetime. And that is the ultimate meaning of it.
So when study those simplest equations of physics, we shall consider those in which we make a ‘ceteris paribus’ rhythm of considering it first from the point of view of ‘time’ steps and then from the point of view of ‘space’ integrating them as a simultaneous space, when we have ‘traced’ enough steps to make that simultaneity meaningful .
And this is the meaning of a definite integral.
It follows then that we can escape the memorial creation, step by step of the spatial form, as something which for us is no longer needed, when we are interested only in integrating the space, and for that reason the integral work merely as an integral of a volume, a surface – whose creation in time has already happened.
But we still have to find a quanta of that ‘creation’ now a mere ‘population in space’.
The different timespace beats.
This of course must be done because reality is bidimensional and a dimension of space goes accompanied by a dimension of time, generating as in the previous graphs, the motions=changes, S≈T≈S≈T that shape reality.
And it is the justification on why differential equations that make systems dependant of such pair of variables happen.
BUT then it follows we shall be able to APPLY THE RASHOMON EFFECT and find a use for the pair ∫∂ as expression of an inverse beating for each pair of dimensions of spacetime.
And decompose both spacetime forms and timespace events in S>T<S beats.
And in the process of doing so, learn further insights about the symmetries between space and time.
The algebraic/graphic duality.
On view of our deeper departure from the ultimate essence of Analysis, which is to study steps of spacetime. That is to put algebraic S=T symmetries in motion; the algebraic vs. graphic interpretations of calculus responds to yet another symmetry of spatial vs. temporal methods, considered on our posts of @nalytic geometry and Algebra.
It does show more clearly what we mean by those ‘steps’ as basically the ‘tangent’ of the curve is in most cases a spacetime step expressed by the general function: X(s) = ƒ(t)
Obviously as s and t are ill defined, it was only understood for lineal spacedistance and timemotion. And so the ‘geometrical’ abstract concept remains, void of all experimental meaning… as a… Tangent… it was…
SPATIAL:GEOMETRIC VIEW.
We are led to investigate a precisely analogous limit by another problem, this time a geometric one, namely the problem of drawing a tangent to an arbitrary plane curve.
Let the curve C be the graph of a function y = f(x), and let A be the point on the curve C with abscissa x0 (figure 10). Which straight line shall we call the tangent to C at the point A? In elementary geometry this question does not arise. The only curve studied there, namely the circumference of a circle, allows us to define the tangent as a straight line which has only one point in common with the curve.
To define the tangent, let us consider on the curve C (figure up) another point A′, distinct from A, with abscissa x0 + h. Let us draw the secant AA′ and denote the angle which it forms with the xaxis by α. We now allow the point A′ to approach A along the curve C. If the secant AA′ correspondingly approaches a limiting position, then the straight line T which has this limiting position is called the tangent at the point A. Evidently the angle α formed by the straight line T with the xaxis, must be equal to the limiting value of the variable angle β.
The value of tan β is easily determined from the triangle ABA′ (figure up):
It is then clear that h is the frequency quanta of time, or if we are inversely using the ∫∂ method to measure space populations, the minimal unit. And so the ultimate concept here is that h NEVER goes to 0. And the clear proof is that if it were arriving to zero, x/h=∞.
So infinitesimals do NOT exist, and it only bears witness of the intuitive intelligence of Leibniz that he so much insisted on a quantity for h=1/n… (and the lack of it of 7.5 billion infinitesimals of Humanity, our collective organism, which memorise this h>o that so much abstract pain gave me when a kid – one of those errors I annotated mentally with the absurd concept of a nonE point with no breath, or else how you fit many parallels, of the limit of cspeed, how Einstein proved that experimentally?, and other ‘errors’ that ∆st does solve in all sciences).
But for other curves such a definition will clearly not correspond to our intuitive picture of “tangency.”
Thus, of the two straight lines L and M in figure below, the first is obviously not tangent to the curve drawn there (a sinusoidal curve), although it has only one point in common with it; while the second straight line has many points in common with the curve, and yet it is tangent to the curve at each of these points.
And yet such a curve is ultimately the curve of a wave, and we know waves are differentiable. So the tangent IS NOT the ultimate meaning of the ∫∂ functions – time/space beats are. The question then is what kind of st beat shall we differentiate in such a transversal wave?
A DIFFERENT DIMENSION, NORMALLY as waves are the 2nd dimension of energy, as in the intensity of an electric flow… a mixture of a population and a motion; or ‘momentum’ (the derivative of energy)…
And so the next stage into the proper understanding of ∫∂ operations is what ‘kind of dimensional spacetime changesteps’ we are measuring.
∆ VIEW
The inversion of the finitesimal calculus of ∆1 is the integral calculus of 5D.
The transition to ∆nalysis: new operations
The mathematical method of limits was evolved as the result of the persistent labor of many generations on problems that could not be solved by the simple methods of arithmetic, algebra, and elementary geometry.
The inverse properties of ∫pace problems and ∂temporal problems
What were the problems whose solution led to the fundamental concepts of analysis, and what were the methods of solution that were set up for these problems ? Let us examine some of them.
The mathematicians of the 17th century gradually discovered that a large number of problems arising from various kinds of motion with consequent dependence of certain variables on others, and also from geometric problems which had not yielded to former methods, could be reduced to two ST types:
Temporal examples of problems of the first type are: find the velocity at any time of a given nonuniform motion (or more generally, find the rate of change of a given magnitude), and draw a tangent to a given curve. These problems (our first example is one of them) led to a branch of analysis that received the name “differential calculus.”
Spatial examples: The simplest examples of the second type of problem are: find the area of a curvilinear figure (the problem of quadrature), or the distance traversed in a nonuniform motion, or more generally the total effect of the action of a continuously changing magnitude (compare the second of our two examples). This group of problems led to another branch of analysis, the “integral calculus.”
Thus two fundamental problems were singled out: the temporal problem of tangents and the spatial problem of quadratures.
Now the reader would observe that unlike the age of Arithmetics and Algebra, which stays in the same ‘locus/form’; here we observe a key property of analysis: the transformation of a temporal cyclical question, into a lineal spatial solution.
I.e. the solution of acceleration/speed by a lineal tangent, through an approximation; and the calculus of a cyclical, spatial area by the addition of squares. And the deep philosophical truth behind it, which only Kepler seemed to have realized at the time:
‘All lines are approximations or parts of a larger worldcycle‘
And so we can consider in terms of modern fractal mathematics, that ‘the infinitesimal is the fractal unit, quanta or step’ of the larger world cycle, and as a general rule:
‘All physical processes are part of a conservative 0sum world cycle’.
Which explains ultimately the conservation of energy and motion, as motions become ultimately world cycles, either closed paths in a single plane, or world cycles balanced through ∆±1 planes.
Such is the simple dual GST justification of Analysis, as always based in ∆… finitesimals and St… the inverse properties of ∫∂.
THE MAIN FUNCTIONS OF NATURE UNDER THE ∫∂ OPERATIONS.
Functions.
In simple terms, a function f is a mathematical rule that assigns to a number x (in some number system and possibly with certain limitations on its value) another number f(x). For example, the function “square” assigns to each number x its square x^{2}.
The common functions are thus definable by formulas, which are related to the ∆s ≈ ∆T duality, such as:
∆§: Polynomials of the type, f(x) = x^{2}. The logarithmic function log (x); & the exponential function exp (x) or e^{x} (where e = 2.71828…; and the square root function √x.
∆T: Trigonometric functions, sin (x), cos (x), tan (x), and so on.
Then there is the question of transformations between space and time and 5D a(nti)symmetries, which is an essential part of classic algebra and we resume in those terms:
 Integral transforms make possible to convert a differential equation of 5D spacetime within certain boundary values (time membrane, which limits the equation as a ‘real system’, not an infinity, into terms of an algebraic equation that can be easily solved (a polynomial which is a result in a single spacetime plane). And this transformation obviously should be of two canonical forms. And as it happens there are 2 canonical transforms:
 – A spatial, lineal transformation, and this is the Laplace transform: f(p), defined by the integral:
F(p)=∫∞0 ept F(t)dt .
The linear Laplace operator L thus transforms each function F(t) of a certain set of functions into some function f(p) and it is used most frequently by electrical engineers in the solution of various electronic circuit problems.
 A temporal transformation and this is the Fourier analysis, which proved that a function y = f(x) could be expressed between the limits x = 0 and x = 2π by an infinite series of waves:
F(x)=1/2 α ∑a cos kx + b sin kx.
That is an equation could become a cyclical time dependent equation developed as a sum of harmonic waves.
And finally the inverse, the fact that a function could be converted into a 5D analytical equation between scales of the 5th dimension is proved by the third most used approximation of functions, the Taylor series, which expresses a function f—for which the derivatives of all orders exist—at a point a in the domain of f in the form of the power series:
∑∞∆=0 f (∆)(a)(za)∆/∆!
In which Σ denotes the addition of each element in the series as ∆ ranges from zero (0) to infinity (∞), f (∆) denotes the nth derivative of f, and ∆! is the standard factorial function.
So this 3 transformations a means – and its applications enlighten an infinite number of real equations that the different 5D scales of reality can transfer energy or information.
That is, a 5D flow of energy and information can travel into a single membrane with absolute accuracy (no loss of entropy, no need of transforms or groups to resolve them.
But there is a minimal loss of entropy when we transform between planes back and forth (of information or energy) as the transform is NOT absolutely exact – as for it to be exact the number of terms normally tend to infinity, which is not possible in the finite duration of any flow between ∆±1 scales of the 5th dimension.
Further on it is important to understand the meaning of the operandi and the laws of relative equality and dynamic transformations of ¬Æ where equality never fully exists, but we transform, F(t)<=> F(s) as in E<=>Mc², or we approximate values through an evident property ≈.
We are not here extensive but just showing some ∆st insights to those inversions.
LET US do some comments of the main functions with fundamental roles in ∆st and its derivatives BY DIVIDING THEM IN 3 GREAT ∆st ‘groups’:
@: ∫∂ of IDENTITY ELEMENTS – FORMS THAT DO NOT CHANGE
The interest of those results refer to the concept of an identity number, as 0 is the identity of sum and 1 of product. But they also have a clear meaning as the interval 01 of the generation ‘seed’ dimension from ∆1 to ∆º.
And indeed, the surprising result that ∫o dx = C does indeed suggest that the 0point is a fractal point that ‘has volume’, or else how integrating the nothingness of existence shall we get a ‘constant’ which is a social number? But if we do start from a o1 unit its ‘integral’ sum will give us a reproductive group, or ‘social number’.
And if we integrate the full ‘1 being’, we shall get a new dimension, the variable plus the constant, which suggest also a little understood process related to the operations of derivatives and integrals, the switching CAUSED by OPERATIONS on motions of sets (our definition of analysis), which CHANGE a spatial state into a time state and viceversa. So the spatial 1formwhole becomes a timevariable X, while the variable X becomes a spatial derivative constant.
Since constant number does NOT change. So a time variable gives us the spatial identity number.
Finally, the deepest thought on those seemingly well known operations regards the subtle difference between both operations: the derivative localises a single ‘finitesimal solution’, or minimal ∆1 past part of the system…
But the inverse, ‘integral’ or ‘future 5th Dimensional arrow’ of social wholes opens up the possibility of multiple constant solutions to add to the variables, as the future is open to subtle variations (∫) but the past is fixed by the infinitesimal identity number (∂).
Of course if we instead consider the integral not in time but as a fixed spatial path, this concept of future vanishes and we get a determined single solution to the integral where the constant is just the starting point.
Other way of seeing it though is to consider the identity element @, the constant mind that does NOT change.
∫∂ of POLYNOMIAL GROWTH:
The first result already considered are the polynomial ‘reduced’ dimension by means of searching its infinitesimal, which however is for simple polynomials quite larger, compared to a direct xˆn1 reduction.
Further on, the logarithm IS clearly the 5D social scaling operation and its derivative is indeed the absolute finitesimal, 1/n.
And inversely the maximal growth is its inverse, the absolute decay of e¯ª.
It is worth to talk of those 3 corelated results from the philosophical pov: the maximal expansion of an event is an absolute future to past, ∆+1 <<∆1 entropic death expressed by the exponential:
The minimal process of growth (Log) is an infinitesimal, the maximal process of decay (e¯ª) is equivalent to the whole, in a single quanta of time. We state in the general law that death happens in a single quanta of time, in which the entire network that pegged together the being, disappeared.
Γst functions.
The third type of functions are concerned not WITH ∆±1 past to future to past d=evolutions but with present sinusoidal wave repetitions of the same timecycle, hence change is cyclical repetitive, and so those functions are very useful for the 3rd reproductive dimotion in space, but also for a time dimotion or cycle:
Both functions thus are clearly inverse not only in Γst but also in the ∆±1 scales – being the negative symbol one of conventions regarding the chosen ± direction of the cyclical, sinusoidal motion.
Here though the interest resides in comparing both type of present vs. ∆ pastfuture functions: the present derivative is self repetitive, as we return to the sin after 4 quadrant derivatives; and ideed we return to the present considering also the generational cycle, after 4 ages of life. So we can model a sinusoidal function as a world cycle of existence in its 4 quadrants.
PDEs and ODEs
A differential equation is a mathematical equation that relates some function with its derivatives. In applications, the functions usually represent physical quantities, the derivatives represent their rates of change, and the equation defines a relationship between the two. Because such relations are extremely common, differential equations play a prominent role in many disciplines including 5D stience, concerned with the Dimotions of spacetime.
Differential equations can be divided into several types. Apart from describing the properties of the equation itself, these classes of differential equations can help inform the choice of approach to a solution.
Experimental justification vs. Axiomatic method.
The axiomatic method, which is valid as all mirror languages have a similar consistency to the reality it mirrors, justifies and classifies them with Group theory – an instantaneous picture of all its varieties put in relationship with somewhat confuse concepts of symmetry.
We prefer as said ad nauseam the experimental method to limit the inflationary mirror to what is useful as reflection of ‘real spacetime properties’.
So the commonly used distinctions of O/PDES include 3 DUALITIES which we put in correspondence with THE 3 elements, ∆ST according to pentalogic. So IF the equation studies:
T by its Number of Dimotions can be Ordinary (1 Dimotion) /Partial (multiple demotions): An ordinary differential equation (ODE) is an equation containing an unknown function of one real or complex variable x, its derivatives, and some given functions of x. The unknown function is generally represented by a variable (often denoted y), which, therefore, depends on x. Thus x is often called the independent variable of the equation. The term “ordinary” is used in contrast with the term partial differential equation, which may be with respect to more than one independent variable.
ODES therefore imply Partial differential equations (PDEs) are equations that involve rates of change with respect to continuous variables. The position of a rigid body is specified by six parameters, ^{}but the configuration of a fluid is given by the continuous distribution of several parameters, such as the temperature, pressure, and so forth. The dynamics for the rigid body take place in a finitedimensional configuration space; the dynamics for the fluid occur in an infinitedimensional configuration space. This distinction usually makes PDEs much harder to solve than ordinary differential equations (ODEs), but here again, there will be simple solutions for linear problems. Classic domains where PDEs are used include acoustics, fluid dynamics, electrodynamics, and heat transfer.
STopology, according to its form can be Linear/Nonlinear=cyclical (entangled by product).
It follows from what we have said of ¬Ælgebra that Odes and lineal PDEs are those in which the ratio of change of the being adds to the function but does NOT entangle through multiplication with them.
This is really what makes nonlineal PDEs so difficult to solve as the entanglement which will happen in other scales of reality will make it almost impossible to get all the information needed, and multiply its solutions, themes those of 5D analysis.
Only the simplest differential equations are solvable by explicit formulas; and most have multiple solutions, implying the future is pentalogic – it can go different ways. Which ones are solvable then helps to understand the philosophy of time:
T=S symmetries.
A Cauchy problem in mathematics asks for the solution of a partial differential equation that satisfies certain conditions that are given on a hypersurface in the domain. ^{}A Cauchy problem can be an initial value problem (Time symmetry) or a boundary value problem (spacesymmetry or Cauchy boundary condition) or it can be neither of them.
The Cauchy problem consists of finding the unknown functions and solutions only will exist if there is an initial FINITE TIME (singularity related, as the will of the system and its dimotions) OR FINITE SPACE (membrane related), hence a formed T.œ structure for the spacetime event/being studied.
∆±¡: Equation order Differential equations are described by their order, determined by the term with the highest derivatives. An equation containing only first derivatives is a firstorder differential equation, an equation containing the second derivative is a secondorder differential equation, and so on. Each order representing them a scale of reality. And since most systems just extend through 3±¡ planes Differential equations that describe natural phenomena almost always have only first and second order derivatives in them.
Also a scalar division is that between Inhomogeneous/Homogeneous, which studies those in which its scaling by multiplication is conserved.
Since a homogeneous function is one with multiplicative scaling behaviour: if all its arguments are multiplied by a factor, then its value is multiplied by some power of this factor: for some constant k and all real numbers α. The constant k is called the degree of homogeneity.
Lineal, affine functions of the type y = Ax + c are not HOMOGENEOUS, which again brings us the duality of ± in the same plane and x/ in different planes of existence.
We can then with those simple concepts understand intuitively many properties of physical equations and parameters by the type of ‘rates of change that take place’.
I.e. products are NOT reproductions but entanglements in a lower plane. So lineal equations will study NONentangled additions in a single plane, and follow the superposition principle. They are the only solvable, as we have all the parameters.
Most ODEs that are encountered in physics are linear, as they deal with the 2nd Dimotion, lineal locomotion, and, therefore, most special functions may be defined as solutions of linear differential equations.
Partial differential equations
is a differential equation that contains unknown multivariable functions and their partial derivatives, used to describe a wide variety of phenomena in nature such as sound, heat, electrostatics, electrodynamics, fluid flow or quantum mechanics. These seemingly distinct physical phenomena can be formalised similarly in terms of PDEs which in general will correspond TO SYSTEMS THAT ARE NOT particle/head controlled, and hence hierarchical with definitive ‘stillness’ in position and single @ristotelian logic. THIS BASICALLY leaves two type of PDEs, those related to entropic, memoriless states, which will tend to be ‘lineal’ as a superposition of nonentangled elements, and those related to complex fluids that interact among its particles and have a complex, variable internal structure which tend to be nonlineal and partial and hence irresolvable. For example:
Lineal PDE: The position of a rigid body (ð§) is specified by a few parameters and it is a lineal ODE.
but the configuration of a fluid is given by several parameters, such as the temperature, pressure, and so forth. Classic domains where such PDEs are used include acoustics, fluid dynamics, electrodynamics, and heat transfer: the heat equation, the wave equation, Laplace’s equation, Helmholtz equation, Klein–Gordon equation, and Poisson’s equation.
Nonlinear differential equations finally are formed by the products of the unknown function and its derivatives are allowed and its degree is > 1. TNonlinear differential equations can exhibit very complicated behavior over extended time intervals, characteristic of chaos, as they are BOTH coexisting in several scales and interacting in its parts on a single scale.
So even the fundamental questions of existence, uniqueness, and extendability of solutions for nonlinear differential equations, are hard problems as the. Navier–Stokes differential equation of fluids show. ^{}
Linear differential equations frequently appear as approximations to nonlinear equations. These approximations are only valid under restricted conditions. For example, the harmonic oscillator equation is an approximation to the nonlinear pendulum equation that is valid for small amplitude oscillations.
So the conclusion is obvious: Nature with its infinite monads and scales IS NOT ALWAYS reflected in a mathematical mirror, which cannot be the origin of Nature (false creationist theories).
Generally speaking the techniques of differentiation distinguish between ODE ordinary equations with a single ST variable, which probe in depth on either space or time consecutive derivatives, but have a limited use as reality only allows 3 multiple derivatives into the single time or space dimension (beyond 3 the results are not essentially not related to the direct experience of how spacetime systems evolve through scales). Multiple derivatives though are the tool to approximate two of the 3 great fields of observance of the scalar Universe, through mathematical mirrors, which we can write as generator equation:
∆i: Fractal Mathematics (discontinuous analysis of finitesimals) < Analysis – Integrals and differential equations (∆º±1: continuous=organic space) < ∆+i: Polynomials (diminishing information on wholes).
It is important in that sense to understand the difference focus of the 3 approaches of mathematical mirrors to observe reality. We shall study in the usual order in which they were born, first ODE then PDE and finally fractals.
Now the mathematical elements of analysis are all well known and standard. Leibniz started them with the symbol ∫ that means summations.
Ordinary differential equations
An ordinary differential equation or ODE is an equation containing a function of one independent variable and its derivatives. The term “ordinary” is used in contrast with the term partial differential equation which may be with respect to more than one independent variable.
They are basically analysis of single ‘steps’/symmetries of S≈T systems, but ODE can go ‘deeper’ into the spatial or temporal structure of the system by establishing multiple derivatives on the original parameter, thus they are perfect systems to ‘study’ the ternary dimensions of ‘integral’ s pace (1Distance, 2D area and 3D volume) and ‘derivative’ time (steady motion, acceleration and deceleration).
And as the symmetries between those 3D of space and time are not clearly understood, ∆st can bring some insights in its analysis.
To notice finally that the best use of mathematical equations and its operations are the simplest actions of motion as reproductions of information in its 3 states/varieties (potentials, waves and particles); but for complex social and reproductive processes very few internal characteristics can be extracted with mathematical tools.
And yet even in those simple cases, exact solutions are not always possible, regardless of the dogmatic myths of mathematical accuracy. This happens as usual because humans measure ‘lineal distances’ and reality is curved, so we approximate lineal quanta/finitesimal and then ad them to find the whole curved state, making use of one of the 3 ‘primary Galilean dualities’ between continuity and discontinuity, linearity and cyclicality, large and small.
So what are the key elements for finding ‘solutions’, that is descriptions of the full T.œ, its state and simpler actions of 1Dmotion/reproduction in space, and topological ≤≥ change from lineal to cyclical form? Basically to have enough data about the ‘boundary conditions’ of the vital energy open ball (that is a parameter for the singularity if it exists, and for the membrane that encloses the system). As both are 1D, 2D hence lineal forms of the type A+Bx, then it is possible to measure and find determined solutions.
Linear differential equations, which have solutions that can be added and multiplied by coefficients, are welldefined and understood, and exact closedform solutions are obtained. By contrast, ODEs that lack additive solutions are nonlinear, and solving them is far more intricate, as one can rarely represent them by elementary functions in closed form: Instead, exact and analytic solutions of ODEs are in series or integral form. Graphical and numerical methods, applied by hand or by computer, may approximate solutions of ODEs and perhaps yield useful information, often sufficing in the absence of exact, analytic solutions.
ODEs are thus symmetric to simple Spacetime steps, which correspond themselves to the simplex 3 actions of 1D, 2D and some possible 4D simple entropy deaths and some simple 3D reproductive steps (3D however when combining space and time parameters and most combined steps of several dimensions and 5D worlds will required PDEs. )
Let us illustrate by a simple example. Consider a material particle of mass m moving along an axis Ox, and let x denote its coordinate at the instant of time t. The coordinate x will vary with the time, and knowledge of the entire motion of the particle is equivalent to knowledge of the functional dependence of x on the time t. Let us assume that the motion is caused by some force F, the value of which depends on the position of the particle (as defined by the coordinate x), on the velocity of motion ν = dx/dt and on the time t, i.e., F = F(x, dx/dt, t).
According to the laws of mechanics, the action of the force F on the particle necessarily produces an acceleration ω = d²x/dt² such that the product of w and the mass m of the particle is equal to the force, and so at every instant of the motion we have the equation:
2. m d²x/dt² = F (x, dx/dt,t)
Where we find the first key ‘second derivative’ for the dimension of time acceleration, which requires a first insight on the Nature of physical systems and its dimensions in space vs. time.
In space, the dimensions seem to us in an easy hierarchical system of growth, 1D lines (2D if we consider them waves of NonE fractal points with a 01 unit circle dimension for each point), 2D areas and 3 D volumes, but in time the 3 arrows depart from 1D steady state motion and can be considered as opposite directions, when volumes of space grow, through the scattering arrow of entropy diminishing its speed, vs. the acceleration of speed that diminishes space as the system collapses into a singularity:
So the 3D of classic space, ‘volume’ actually belongs to the entropic arrow of decelerating time that creates spacevolume, vs. the opposite arrow of imploding time vortices that diminish space volume and increases speed, Vo x Ro = k.
So what seems in space a natural growth of volume in space, in time has a different order:
Entropic ≈decelerating volume < steady state ≈ distancemotion > Informative, cyclical area ≈ accelerated motion.
This different ‘order’ of dimensions when perceived in simultaneous space and cyclical time is the main dislocation in the way the mind perceives both (which is sorely painful when we consider the order of a world cycle, always starting in the ∆1 scale of maximal information to decline as it grows and reproduces into less perfect, more entropic volumes of iterative forms that finally decline and die in the arrow of entropy; which the mind that has a SPATIALVOLUME INCLINED nature of evergrowth, does not understand).
This is the differential equation that must be satisfied by the function x(t) describing the behavior of the moving particle. It is simply a representation of laws of mechanics. Its significance lies in the fact that it enables us to reduce the mechanical problem of determining the motion of a particle to the mathematical problem of the solution of a differential equation.
Later the reader will find other examples showing how the study of various physical processes can be reduced to the investigation of differential equations.
The theory of differential equations began to develop at the end of the 17th century, almost simultaneously with the appearance of the differential and integral calculus. At the present time, differential equations have become a powerful tool in the investigation of natural phenomena. In mechanics, astronomy, physics, and technology they have been the means of immense progress. From his study of the differential equations of the motion of heavenly bodies, Newton deduced the laws of planetary motion discovered empirically by Kepler. In 1846 Leverrier predicted the existence of the planet Neptune and determined its position in the sky on the basis of a numerical analysis of the same equations.
To describe in general terms the problems in the theory of differential equations, we first remark that every differential equation has in general not one but infinitely many solutions; that is, there existsan infinite set of functions that satisfy it. For example, the equation of motion for a particle must be satisfied by any motion induced by the given force F(x, dx/dt, t), independently of the starting point or the initial velocity. To each separate motion of the particle there will correspond a particular dependence of x on time t. Since under a given force F there may be infinitely many motions the differential equation (2) will have an infinite set of solutions.
Every differential equation defines, in general, a whole class of functions that satisfy it. The basic problem of the theory is to investigate the functions that satisfy the differential equation. The theory of these equations must enable us to form a sufficiently broad notion of the properties of all functions satisfying the equation, a requirement which is particularly important in applying these equations to the natural sciences. Moreover, our theory must guarantee the means of finding numerical values of the functions, if these are needed in the course of a computation. We will speak later about how these numerical values may be found.
If the unknown function depends on a single argument, the differential equation is called an ordinary differential equation. If the unknown function depends on several arguments and the equation contains derivatives with respect to some or all of these arguments, the differential equation is called a partial differential equation. The first three of the equations in (1) are ordinary and the last three are partial.
The theory of partial differential equations has many peculiar features which make them essentially different from ordinary differential equations.
Let us now consider the ∫∂ operations for the different dimensions of reality, starting in this case with the simplest cyclical clockmotions, which as they do NOT move in space, and repeat its form in time, are in fact not operated by ∫∂ measures of change:
1D: cyclical clocks, angular momentum
In the graph, in the simplest physical systems 1D is merely the angular momentum of its cyclical clocks of time, maximised in the membrane that encloses the system. Strictly speaking it does not change but becomes the ‘present function’ of a repetitive frequency clock without a derivative of change as the timespace steps seem not to vary. When we introduce a torque, change happens, called ‘acceleration’, the second dimension of time motion in physics, which we shall latter study when analysing in 5D with the Galilean Px. Newton’s laws. Here we just shall briefly explain why in lineal time, as humans only use t to measure change, the 1D is the invariant one and its derivative is zero.
What about ‘higher’ more complex, cyclical, and scalar Dimensions? The answer is that as we change the form of the dimensions, we have to change the operandi we use; and specifically when we study the Dimensions of change, which is the one differential/integral equations quantify, those equations MUST ADAPT not the other way around as mirrors of reality to the FORM of the dimensions of spacetime they describe.
So as 1D is A STEADY STATE ROTARY MOTION, strictly speaking it does NOT change in spacetime locomotion (which is what humans with its lineal single time express in derivatives). Hence basically the derivative of those angular momentums is zero. It is conserved.
Let us recall briefly those classic definitions and maths:
Angular momentum is a vector that represents the product of a body’s rotational inertia and rotational velocity about a particular axis. In the simple case of revolution of a particle in a circle about a center of rotation, the particle remaining always in the same plane and having always the same distance from the center, we discard the vector nature of angular momentum, and treat it as a scalar proportional to moment of inertia, I and angular speed, ω:
L= Iω: Angular momentum = moment of inertia × angular velocity, and its time derivative is
dL/dt =dI/dt ω +I dω/dt is zero, and dL/dt=0+I dω/dt, which reduces to dL/dt =Iα.
Therefore, angular momentum is constant, dL/dt=0 when no torque is applied. And this is the essence of its conservation law, a specific case of the conservation of the 5Dimensions of spacetime of the Universe:
‘In a closed system, no torque can be exerted on any matter without the exertion on some other matter of an equal and opposite torque. Hence, angular momentum can be exchanged between objects in a closed system, but total angular momentum before and after an exchange remains constant’.
But when a torque is applied in a single present plane, or much more relevant to our inquire: when a system is submitted to the organising or disorganising entropic force of a higher or lower plane of existence, and acceleration exists, a vortex of timespace happens and we enter into the social dimensions of evolution – the 5th Dimension of the mind.
Social Number = first dimension that defines regular ‘points’ which are undistinguishable, as societies in regular polygons, where prime polygons have the property of ‘increasing inwards’ its numbers through reproduction of vortexpoints (ngrams), as the graph shows, studied in Theory of Numbers. So a number in its geometric interpretation is a ‘cyclical point’ of regular ‘unitpoints’ of growing ‘inner dimensional density’ a point with a volume of vital energy and information, a fractal point.
The bottom line though of this brief analysis of a system with a single Dimension of timespace, either as a fractal point emerging through a parameter visible to the ∆+1 observer, notably angular momentum as in quantum physics (h), is that it IS A CONSTANT, not a differentiable parameter for which a first step in spacetime, SS ˆST ˆ TS ˆTT is needed. Then we are in a 2D system, normally a TS motion, or ‘speed’ in which the quanta of space, ‘moves≈reproduces’ in a time trajectory, which allows to measure the change on one parameter, normally the spatial location, and accordingly ‘derivate’ the ‘ratio’ or ‘inverse product’ between them.
2D: LINEAL SPACETIME
Let us consider one example of each dual dimension, $t, the two samples mentioned – speed and area, which were the first 2 themes solved historically, with the classic notation, to keep the historical approach, to see how really the methods can be used equally for quanta=frequency=steps of time, or quanta=populations=finitesmals of space:
Speed and acceleration: 2D TT
The next possible steps or motions in spacetime are given by a dual timetime motion, which is acceleration, or a similar dual motion in space which is volume. As such those 2D motions have diametral opposite consequences shrinking a system in time, towards a mind zero point (TT>5D) or expanding it in space towards a extension of free, entropic ∆1 elements (SS>4D).
But they can be used in combined forms to extract the same equations of speed, density and momentum. Let us put the TT example:
The velocity of a point for which the distance s is a given function of the time s = f(t) is equal to the derivative of this function: v = s’ = ƒ ‘ (t).
So as it was established experimentally by Galileo, the distance s covered in the time t by a body falling freely in a vacuum is expressed in terms of TTacceleration by the formula: s=gt²/2
Whereas g is a constant that measure the acceleration on Earth, equal to 9.81 m/sec².
What is the velocity of the falling body at each point in its path?
Here as we are in two TT variation already we must do exactly the inverse operation to that of searching for speed departing from space:
Let the body be passing through the point A at the time t and consider what happens in the short interval of time of length Δt; that is, in the time from t to t + Δt. The distance covered will be increased by a certain increment Δs. The original distance is s1 = gt²/2.
From the increased distance we find the increment:
This represents the distance covered in the time from t to t + Δt. To find the average velocity over the section of the path Δs, we divide Δs by Δt:
Letting Δt approach zero, we obtain an average velocity which approaches as close as we like to the true velocity at the point A. On the other hand, we see that the second summand on the righthand side of the equation becomes vanishingly small with decreasing Δt, so that the average υav approaches the value gt, a fact which it is convenient to write as follows:
Consequently, gt is the true velocity at the time t, and so we can consider gt as yet another expression of the Sð equation of speed inn cyclical time, where now t is a ‘step’ and g the measure of its ‘feeding’ on gravitational space.
Let us make the following remark. The velocity of a nonuniform motion at a given time is a purely physical concept, arising from practical experience. Mankind arrived at it as the result of numerous observations on different concrete motions.
The study of nonuniform motion of a body on different parts of its path, the comparison of different motions of this sort taking place simultaneously, and in particular the study of the phenomena of collisions of bodies, all represented an accumulation of practical experience that led to the setting up of the physical concept of the velocity of a nonuniform motion at a given time. But the exact definition of velocity necessarily depended upon the method of defining its numerical value, and to define this value was possible only with the concept of the derivative.
In mechanics the velocity of a body moving according to the rule s = f(t) at the time t is defined as the derivative of the function f(t) for this value of t.
But now, as a result of our analysis, we have reached an exact definition of the value of the velocity at a given moment, namely the finitesimal, minimal action of a given time motion. This result is extremely important from a practical point of view, since our empirical knowledge of the velocity has been greatly enriched by the fact that we can now make an exact definition for the 5 different motions of time, greatly expanding our understanding in terms of analysis of those motions and its variations.
And we have use the method departing from timequanta, frequency of steps, ‘speed motion’…
And that’s good enough, keeping ∆t>o without reaching 0, as we shall always find a limiting ‘time unit’…
In the extreme of those limits (cspeed) the limit will be found on the gravitational field from where light extracts motion.
As it happens that limit can be in ∆4, according to the decametric 10¹º¹¹ scales between ∆planes and the 5D metric which accelerates clocks in smaller quanta, 4 planes down x 10¹º¹¹ time units between scales… faster in frequency, hence around ±10ˆ44 faster clocks/smaller bits of time.
So the gravitational infinitesimal is truly ∆t>o and hence irrelevant. (Incidentally physics discovered this value for the minimal clock of Nature, without knowing its scalar planes, social quanta, and 5 metric, by sheer chance. It was Planck and he call it the time of ‘God’ (:
How he did it? with the Universal constants; a fascinating theme treated further down this texts when done. In any case he was closer to the truth, as I always considered those numbers a solid quantitative proof along many other elements of the theory of grand numbers of ∆ST theory:
The theoretical importance of Tp in our argument over which type of continuous, S/t or discontinuous, λ(s) x ƒ (t) speed treat becomes now clear. As the absolute finitesimal of all the planes and scales in which time happens among human observers (regardless of possible ∆±≥4 planes beyond human observance) in as much as those are the minimal scales in which gravitational fields might exercise its forces over our atomic substance, continuity happens and makes possible finitesimal/integral calculus because by all means this is undistinguishable from our pov/scale.
3D: POPULATIONS
Now when we get into 3D, which are combinations of 1D + 2D, the vibrations of different ST combinations multiply our possibilities.
If there is only a type of 1D the fractal point or the invisible distance with no form, 2D GIVE us 4 fundamental variations, SS tt and st, ts. Now with 3D we can combine 2 and 1 varieties to give us the following orations:
1D + 4 2D, 2D + 4 1D and if they are not commutative as it seems the case, the inverse case, for 20 combinations of 3D populations.
It is then a whole encyclopaedia what you need to explain all the practical cases in which variations of 3 D integral Derivatives can be used to explain different vibrations of Sts motions of all kinds.
ADDING A NEW DIMENSION OF ‘WIDTHENERGY’ INTENSITY
Once this concept is fully understood we then need to deal with ‘finitesimal quantities’, either in time or in space, as the previous argument on ‘changes of speeds and frequencies of time motion steps’, ∆s/∆t , can be reversed to study changes of volumes of space and populations of simultaneous spacebeings. And so we apply all those concepts to the analysis of 2D populations. Let us put an example and resolve it in terms of spacequanta (method of limits) and in terms of its change with differentials.
Quanta of space.
Now a spatial use of the limit concept to calculate not a time but a space volume, forebear of differential calculus:
Example 2. A reservoir with a square base of side a and vertical walls of height h is full to the top with water (figure 1). With what force is the water acting on one of the walls of the reservoir?
We divide the surface of the wall into n horizontal strips of height h/n. The pressure exerted at each point of the vessel is equal, by a wellknown law, to the weight of the column of water lying above it. So at the lower edge of each of the strips the pressure, expressed in suitable units, will be equal respectively to:
We obtain an approximate expression for the desired force P, if we assume that the pressure is constant over each strip. Thus the approximate value of P is equal to:
To find the true value of the force, we divide the side into narrower and narrower strips, increasing n without limit. With increasing n the magnitude 1/n in the above formula will become smaller and smaller and in the limit we obtain the exact formula:
P = ah²/2
In the example again the finitesimal limit is extremely small. How much? We should consider statistical mechanics to find it is the size of molecules of water, which form bidimensional layers of liquid to shape the 3D volume, and are indeed about 10¹º times smaller than the whole.
Now the error ε is so small as to be P=(ah²/2) • 1.00000000001 (1 +1/n)
And this is a general rule in most cases: the finitesimal error is as small as 1/n, where n is the quanta of the scale. So when we do ∆+1 calculations as in most cases it is irrelevant. But theoretically it is important and in fact it will give us a ‘realist’ concept for the uncertainty principle of Heisenberg.
Hence indeed unnoticeable, truly infinitesimal but absolutely infinitesimal and certainly never proved by the axiomatic method as maths must be experimentally proved to avoid inflationary errors, and certainly as always in Γst (I should write gst as ∆st or Γst, the proper acronym, but i am lazy with wordpress 🙂 we do not accept a mathematical result without experimental proof (for me the fundamental use of mathematical physics), following Lobachevski, Godel and Einstein’s dictums.
The idea of the method of limits is thus simple, accurate and amounts to the following. In order to determine the exact value of a certain magnitude, we first determine not the magnitude itself but some approximation to it. However, we make not one approximation but a whole series of them, each more accurate than the last. Then from examination of this chain of approximations, that is from examination of the process of approximation itself, we uniquely determine the exact value of the magnitude. by ignoring the finitesimal error.
The same practical problem can be resolved with the differential used as an approximate value for the increment in the function. For example, suppose we have the problem of determining the volume of the walls of a similar closed cubical box whose interior dimensions are 10 × 10 × 10 cm and the thickness of whose walls is 0.05 cm. If great accuracy is not required, we may argue as follows. The volume of all the walls of the box represents the increment Δy of the function y = x3 for x = 10 and Δx = 0.1. So we find approximately:
DIFFERENTIALS – ANY STDIMENSIONAL STEPS
The disquisition of which ‘minimalist finitesimal’ allow us to differentiate an S≈T algebraic symmetry, brings us to the ‘praxis’ of calculus techniques that overcome by ‘approximations’ the quest for the finitesimal quanta in space or time, susceptible of calculus manipulation, which gave birth to the praxis of finding differentials, which are the minimal F(Y) quanta to work with and obtain accurate results (hence normally an spatial finitesimal of change under a time dependent function). This was the origin of the calculus of differentials.
As always in praxis, the concept is based in the duality between huminds that measure with fixed rulers, lineal steps, over a cyclical, moving Universe. So Minds measure Aristotelian, short lines, in a long, curved Universe.
So the question comes to which minimalist lineal step of a mind is worthy to make accurate calculus of those long curved Universal paths.
It is then obvious that the derivative of a lineal motion has more subtle elements that its simplest algebraic form, the x ÷ lineal operation of ‘reproductive speed’ and so the concept of a differential to measure the difference between steady state lineal reproduction and the variations observed by a curve appeared as the strongest tool of approximation of both type of functions.
As we have considered that most differential equations will be of the form: F(s) ≈ g(t), where s and t are any of the 5 Dimensions of Space ($, S, §, ∫, •) or 5 Dimensions of time (t, T, ð, ∂, O), whose change respect to each other, we are bound to study… showing how a spatial whole is dependent on the change and form of a world cycle, we shall consider generally that y>s and x>t…
The result of this change will be a much more GENERIC CONCEPT OF SPEED OF CHANGE in any OF THE DIMENSIONS OF ENTROPY, MOTION, ITERATION, INFORMATION OR FORM that defines the Universe, letting us introduce its 3 fundamental parameters, S/t=speed, t/s=density and s x t = momentum/force/energy… in a natural way with its multiple different meanings, Ðisomorphic to each other – as we repeat the s and t of the general
TIME DIMENSIONS BECOME SPACE DIMENSIONS BECOME TIME…
Physical equations in differential form, a general overview of its main species. History
Differential equations first came into existence with the invention of calculus by Newton and Leibniz. Newton listed three kinds of differential equations: those involving two derivatives one of space and time (or fluxions) and only one undifferentiated quantity (space or time parameter); those involving 2 derivatives and two quantities of space and time; and those involving more than two derivatives.
Its analysis thus was right in the spot as he referred changes to change in space or time, thus ∫∂ with STeps – a fact latter forgotten and today thoroughly missed with the ‘view’ of time as a single dimension of space (1Dlineal motion confused with 4Dentropy in philosophy of science)
It is still a good classification of partial differential equations as ‘timelike’ (∂x, ∂²x, ∂³x), or space like (∂²y, ∂y, ∂³y) or spacetime like (∂x∂y, ∂y∂x) as the main variations that represent, T, TT, TTT; S, SS, SSS, ST, TS steps, which are the main 5D, 4D and 1,2,3D changes of the Universe.
And it speaks of the enormous range of real phenomena ∫∂ functions can describe as the essential operandi of mathematical physics and any ∆st phenomena.
What allow all those ∆st phenomena to enter the world of quantitative mathematics was the discovery of a pendulum clock to measure time in lineal fashion and a telescope to measure space. Both gave birth to the 2nd age of science, the mathematical/scientific method, added to the experimental Aristotelian method, which now the isomorphic GST age of stience completes.
In 1609 appeared the “New astronomy” of Kepler, containing his first and second laws for the motion of the planets around the sun.
In 1609 too Galileo directed his recently constructed telescope, though still small and imperfect, toward the night sky; the first glance in a telescope was enough to destroy the ideal celestial spheres of Aristotle and the dogma of the perfect form of celestial bodies. The surface of the moon was seen to be covered with mountains and pitted with craters. Venus displayed phases like the Moon, Jupiter was surrounded by four satellites and provided a miniature visual model of the solar system. The Milky Way fell apart into separate stars, and for the first time men felt the staggeringly immense distance of the stars. No other scientific discovery has ever made such an impression on the civilised world.
It also killed a method equally valid of thought represented by the Greeks and Leonardo: the idealised understanding of the canonical perfect GST game of existence, of which we were all impure platonic forms, bond to dissolve unlike the perfect game of the ∞ Universe, which is immortal.
Man never went back because alas! what really mattered was ballistics, mechanics and power. Idealism died away:
The further development of navigation, and consequently of astronomy, and also the new development of technology and mechanics necessitated the study of many new mathematical problems. The novelty of these problems consisted chiefly in the fact that they required mathematical study of the laws of motion in a broad sense of the word. And now we had machines to measure it better than the artistic SpeyeT=words of the human spacetime mind.
The mean value theorem.
The differential expresses the approximate value of the increment of the function in terms of the increment of the independent variable and of the derivative at the initial point. So for the increment from x = a to x = b, we have:
ƒ(b) – ƒ(a)≈ ƒ'(a) (ba).
It is possible to obtain an exact equation of this sort if we replace the derivative f′(a) at the initial point by the derivative at some intermediate point, suitably chosen in the interval (a, b). More precisely: If y = f(x) is a function which is differentiable on the interval , then there exists a point ξ, strictly within this interval, such that the following exact equality holds:
ƒ(b)ƒ(a)=ƒ'(ξ)(ba)
The geometric interpretation of this “meanvalue theorem” (also called Lagrange’s formula or the finitedifference formula) is extraordinarily simple. Let A, B be the points on the graph of the function f(x) which correspond to x = a and x = b, and let us join A and B by the chord AB.
Now let us move the straight line AB, keeping it constantly parallel to itself, up or down. At the moment when this straight line cuts the graph for the last time, it will be tangent to the graph at a certain point C. At this point (let the corresponding abscissa be x = ξ), the tangent line will form the same angle of inclination α as the chord AB. But for the chord we have:
tan α = ƒ(b) – ƒ (a) / ba. On the other hand at the point C: tan α = ƒ’ (ξ):
This equation is the meanvalue theorem, which has the peculiar feature that the point ξ appearing in it is unknown to us; we know only that it lies somewhere in the interval (a, b).
Its interpretation in ∆st is that ƒ'(ξ) corresponds to the value of a finitesimal lying between both.
FIRST, the fact that ‘membranes must determine the beginning and end point of any function for it to be meaningful and solvable. And indeed, only because we know when it starts and ends the domain, we are sure to find a mean point.
If we consider then a T.œ mean value theorem, where ƒ(b) > ƒ(a) if we are deriving in space, where f(b)=Max. S represents the parameter of the membrane, ƒ(a) will represent the singularity and so we shall find in between a finitesimal part of the vital energy of the T.œ with a mean value within that of Max. S (membrane) and Min. S (singularity). And viceversa, if we are deriving in search of the minimal quanta of time, ƒ(a) > ƒ (b), where ƒ(a) represents the time speed of the singularity and ƒ(b) the time speed of the membrane. And the mean value will be that of the infinitesimal.
But in spite of this indeterminacy, the formula has great theoretical significance and is part of the proof of many theorems in analysis.
The immediate practical importance of this formula is also very great, since it enables us to estimate the increase in a function when we know the limits between which its derivative can vary. For example:
sin b – sin a = cos ξ (ba) ≤ ba.
Here a, b and ξ are angles, expressed in radian measure; ξ is some value between a and b; ξ itself is unknown, but we know that cos ξ ≤1
Another immediate expression of the theorem which allow to derive a general method for calculating the limits and approximations of polynomials with derivatives is:
For arbitrary functions ϕ(x) and ψ(x) differentiable in the interval [a, b], provided only that ψ′(x) ≠ 0 in (a, b), the equation, holds where ξ is some point in the interval (a, b).
From the mean value theorem it is also clear then that a function whose derivative is everywhere equal to zero must be a constant; at no part of the interval can it receive an increment different from zero. Analogously, it is easy to prove that a function whose derivative is everywhere positive must everywhere increase, and if its derivative is negative, the function must decrease.
And so the ‘classic function of meanvalue theorem’ allow us to introduce an essential element of ∫∂ which will open up the ∆st calculus of worldcycles of existence, the standing points of a function.
The mean value sets for the region between the limiting points of the curve – which must be taken in higher steptimespace as two sections of a bipodal spherical line, part of the membrane of a 3D form, gives us then a value for the vital energy to be expressed with a scalar. And the initial and final point of the segment become the maximal and minimal of the function in F(f)=x values.
It is then between those two limits a question of find points of the vital energy among them the singularity Max. S x t, to have a welldefined TOE in its membrane (maximal minimal values) volume of energy, mean value and Maximal point of the Singularity.
Maxima and minimum. The 3 standing points of a world cycle.
The minimal reality is a 3D² form seen in a single plane, with a singularity @mind a membrane and a vital energy within. When we make a holographic broken image of this reality the simplest way to do it is in four cartesian regions, TT, ST, ts, and ss, which correspond to the +1 +1, +1 1, 1 +1 and 1 1 quadrants of the plane.
We can then dissect the sphere in antipodal points related to the identity neutral number o1 the sphere of time probabilities that the largest whole maximises in its antipodal points. If we consider the antipodal points the emergent and final death point, which imperfect motions still close to zerosums, the maximal middle point will be the singularity, Max. S x Max t.
One of the simplest and most important applications of the derivative in that sense is in the theory of maxima and minima.
Let us suppose that on a certain interval a≤t≤b we are given a function S = f(t) which is not only continuous but also has a derivative at every point. Our ability to calculate the derivative enables us to form a clear picture of the graph of the function. On an interval on which the derivative is always positive the tangent to the graph will be directed upward. On such an interval the function will increase; that is, to a greater value of t will correspond a greater value of f(t). On the other hand, on an interval where the derivative is always negative, the function will decrease; the graph will run downward.
We have drawn the graph of an ∆st function of the general form, S (any dimension of a whole world cycle or T.Œ) = f(T) – Any time motion or action.
It is defined on the interval between a minimal quanta in space or time (t1) and its limit as a function (d).
And it can represent any S=T duality, or more complex 5Ds=5Dt forms or simpler ones. We can also change the s and t coordinates according to the Galilean paradox, etc. Hence the ginormous numbers of applications, but essentially it will define a process of change in spacetime between the emergence of the phenomena at ST1 AND ITS DEATH mostly by scattering and entropic dissolution of form at d.
And in most cases will have a bell curved from of fast growth after emergence in its first age of maximal motion (youth, 1D) till a maximal point where it often will reproduce into a discontinuous parallel form (not shown in the graph at Max. S x Max. T; which will provoke its loss of energy and start its diminution till its extinction at point d.
Thus the best way to express quantitatively in terms of ST parameters (mostly information and energy), for any world cycle of any timespace super organism is a curve where we can find those key standing points in which a change of age, state or motion happens.
Of a special interest thus are the points of this graph whose abcissas are t1,2, 3, 4, 5.
At the point t0 the function f(t) is said to have a local maximum; by this we mean that at this point f(t) is greater than at neighboring points; more precisely for every t in a certain interval around the point x0.
A local minimum is defined analogously. For our function a local maximum occurs at the points t0 and t3, and a local minimum at the point t1.
At every maximum or minimum point, if it is inside the interval [a, b], i.e., if it does not coincide with one of the end points a or b, the derivative must be equal to zero.
This last statement, a very important one, follows immediately from the definition of the derivative as the limit of the ratio ΔS/ΔT. In fact, if we move a short distance from the maximum point, then ∆S≤0.
Thus for positive ΔT the ratio ΔS/ΔT is nonpositive, and for negative ΔT the ratio ΔS/ΔT is nonnegative. The limit of this ratio, which exists by hypothesis, can therefore be neither positive nor negative and there remains only the possibility that it is zero.
By inspection of the diagram it is seen that this means that at maximum or minimum points (it is customary to leave out the word “local,” although it is understood) the tangent to the graph is horizontal.
In the figure we should remark that at the points t2, and t4 also the tangent is horizontal, just as it is at the points t1, t3, although at these points the function has neither maximum nor minimum. In general, there may be more points at which the derivative of the function is equal to zero (stationary points) than there are maximum or minimum points.
Determination of the greatest and least values of a function.
In numerous technical questions it is necessary to find the point t at which a given function f(t) attains its greatest or its least value on a given interval.
In case we are interested in the greatest value, we must find x0 on the interval [a, b] for which among all x on [a, b] the inequality ƒ(to)≥ƒ(t) is fulfilled.
But now the fundamental question arises, whether in general there exists such a point. By the methods of modern analysis it is possible to prove the following existence theorem:
If the function f(t) is continuous on a finite interval, then there exists at least one point on the interval for which the function attains its maximum (minimum) value on the interval [a, b].
From what has been said already, it follows that these maximum or minimum points must be sought among the “stationary” points. This fact is the basis for the following wellknown method for finding maxima and minima.
First we find the derivative of, f(t) and then solve the equation obtained by setting it equal to zero.
If t1, t2, ···, tn, are the roots of this equation, we then compare the numbers f(t1, f(t2), ···, f(tn) with one another. Of course, it is necessary to take into account that the maximum or minimum of the function may be found not within the interval but at the end (as is the case with the minimum in figure) or at a point where the function has no derivative.
Thus to the points t1, t2, ···, tn, we must add the ends a and b of the interval and also those points, if they exist, at which there is no derivative. It only remains to compare the values of the function at all these points and to choose among them the greatest or the least.
With respect to the stated existence theorem, it is important to add that this theorem ceases, in general, to hold in the case that the function f(t) is continuous only on the interval (a, b); that is, on the set of points x satisfying the inequalities a <t < b.
It is then necessary to consider an initial time point and a final time point, birth and death, emergence and extinction to have a determined solution.
Derivatives of higher orders.
We have just seen how, for closer study of the graph of a function, we must examine the changes in its derivative f′(x). This derivative is a function of x, so that we may in turn find its derivative.
The derivative of the derivative is called the second derivative and is denoted by y”=ƒ”(x)
Analogously, we may calculate the third derivative y”‘=ƒ”‘(x) and more generally the nth derivative or, as it is also called, the derivative of nth order. But as there are not more than 3 ‘similar derivatives, with meaning’ in time (speed, acceleration, jerk) or space (distance, area and volume), beyond the 3rd derivative the use of derivatives is only as an approximation to polynomial equations, whose solvability itself is not possible by radicals beyond the 3rd power.
So it must be kept in mind that, for a certain value of x (or even for all values of x) this sequence may break off at the derivative of some order, say the kth; it may happen that f(k)(x) exists but not f(k + 1)(x). Derivatives of arbitrary order are therefore connected to the symmetry between power laws and ∫∂ operations in the 4th and inverse 5th Dimension, through the Taylor formula. For the moment we confine ourselves to the second and third derivatives for ‘real parameters’ of the 3 space volumes and time accelerations.
The second derivative has then as we have seen a simple significance in mechanics. Let s = f(t) be a law of motion along a straight line; then s′ is the velocity and s″ is the “velocity of the change in the velocity” or more simply the “acceleration” of the point at time t. For example, for a falling body under the force of gravity: That is, the acceleration of falling bodies is constant.
Significance of the second derivative; convexity and concavity.
The second derivative also has a simple geometric meaning. Just as the sign of the first derivative determines whether the function is increasing or decreasing, so the sign of the second derivative determines the side toward which the graph of the function will be curved.
“Suppose, for example, that on a given interval the second derivative is everywhere positive. Then the first derivative increases and therefore f′(x) = tan α increases and the angle a of inclination of the tangent line itself increases (figure 17). Thus as we move along the curve it keeps turning constantly to the same side, namely upward, and is thus, as they say, “convex downward.”
On the other hand, in a part of a curve where the second derivative is constantly negative the graph of the function is “convex upward.
Criteria for maxima and minima; study of the graphs of curves.
If throughout the whole interval over which x varies the curve is convex upward and if at a certain point x0 of this interval the derivative is equal to zero, then at this point the function necessarily attains its maximum; and its minimum in the case of convexity downward. This simple consideration often allows us, after finding a point at which the derivative is equal to zero, to decide thereupon whether at this point the function has a local maximum or minimum.
Now, the apparently equal nature on a first derivative of the minimal and maximal points of a being, have also deep philosophical implications, as it makes at ‘first sight’ indistinguishable often the processes of ‘reproductive expansion’ towards a maximal and explosive decay into death, the ‘two reversal’ points of the 5D (maximal) and 4D (minimal) states of a cycle of existence, for which we have to make a second assessment (second derivative) to know if we are in the point of maximal life (5D) or maximal death (4D) of a world cycle.
And further on to know if the cycle will cease in a continuous flat encephalogram or will restart a new upwards trend.
Or in other words is any scalar, e>cc>m bigbang both the death and the birth of matter?
Finitesimal Quanta, as the limit of populations in space and the minimal action in time.
So there is behind the duality between the concept of limits and differentials (Newton’s vs. Leibniz’s approach), the concept of a minimal quanta in space or in time, which has been hardly explored by classic mathematics in its experimental meaning but will be the key to understand ‘Planckton’ (Hplanck constants) and its role in the vital physics of atomic scales.
It is then essential to the workings of the Universe to fully grasp the relationship between scales and analysis. Both in the down direction of derivatives and the up dimension of integrals; in its parallelism with polynomials, which rise dimensional scales of a system in a different ‘more lineal social inter planar way’.
So polynomials and limits are what algebra is to calculus; space to time and lineal algebra to curved geometries.
The vital interpretation though of that amazing growth of polynomials is far more scary.
Power laws by the very fact of ‘being lineal’, and maximise the growth of a function ARE NOT REAL in the positive sense of infinite growth, a fantasy only taken seriously by our economists of greed and infinite usury debt interest… where the eª exponential function first appeared.
The fact is that in reality such exponentials only portrait the decay destruction of a mass of cellular/atomic beings ALREADY created by the much smaller processes of ‘re=production’ which is the second dimension mostly operated with multiplication (of scalars or anti commutative cross vectors).
So the third dimension of operandi is a backwards motion – a lineal motion into death, because it only reverses the growth of sums and multiplications polynomials makes sense of its properties.
Let us then see how the operations mimic the five dimensions, beyond the simplest ST, SS and TT steps, namely reproductive and 4D5D inverted arrows.
In general we can establish as the main parameter of the singularity, its time frequency, which will be synchronised to the rotary motion or angular momentum of the cyclical membrane. They will appear as the initial conditions and boundary conditions of a derivative/integral function, which often will be able to define the values of the vital energy within, as the law of superposition should work between the 3 elements, such as:
1D (singularity) + 2D (Holographic principle) = 3D (vital energy).
In practice this means the ‘synchronicity in time of the clocks of the 3 parts of the being’ and the superposition of the solutions that belong to each of the 3 elements of any T.œ
4th dimension: Entropy: S∂ polynomial death dimension of decay.
POLYNOMIALS DO NOT EVOLVE REALITY towards an impossible infinite growth. THEY ARE the inverse decay process; of exponential extinction, eˆx.
5th dimension: ∫T…
This is understood better observing that the inverse function does in fact model growth in the different models of biology and physics, limited by a carrying capacity straight flat line:
The logarithmic function has as derivative an infinitesimal, 1/x, which makes it interesting as it models better the curve of growth from o to 1 in the emergent fast explosive ∆1 seed state, while the inverse eˆx model the decay death process.
Integrals and derivatives which have a much slower growth, than polynomials on the other hand do model much better as they integrate the ‘indivisible’ finitesimal quanta of a system, its organic growth and ‘wholeness’ integrated in space.
Thus integrals do move up a social growth in new ∆+1 5D planes. And its graphs are a curved geometry, which takes each lineal step (differential) upwards, but as it creates a new whole, part of its energy growth sinks and curves to give birth to the mindsingularity @, the wholeness that warps the whole, and converts that energy into still, shrunk mindmappings of information, often within the 3D particlehead.
We will retake the analysis of the more complex steps on 3, 4 and 5D, since most of the complex process related to the 3rd dimension, as a mixture of S and T inner scales, will require a more complex double or triple derivative and integrals – only the 4D decay entropic explosion can be satisfied as the decay of the single ∆1 finitesimal with a single variable.
Let us now move to the inverse 5D function:
5D ∫∆1 INTEGRALS
5D: VORTEX OF INFORMATION: ∫T@. The culmination of the process of dimensional growth so far is the state of absolute stillness of the mind. It is the integral function of wholes, made of finitesimal 4D parts. And so as we integrate them we reconstruct the being in i=tself:
All what we have said changes though in 5D, where a force is exerted by a 5D SINGULARITY at the center of the vortex. How the Galilean paradox observes this ‘change’? Establishing a second dimension of ‘dynamic time motion’, which we call acceleration, or inversely establishing a new dimension of space as the angular momentum decelerates, creating a volume of space from a present flat spacetime sheet. And as both do imply change in ST volume, we find again relevant derivatives (finitesimal steps either in time frequency or space volume, and integrals (to bring the whole spatial static view of the phenomena to calculate that change.
Galileo discovered something essential to ∆ST: Relativity of motion, which is also a distance:
The state of rest and motionlessness is unknown in nature, but a construct of the mind (@view). The whole of nature, from the smallest particles up to the most massive bodies, is in a state of eternal creation and annihilation, in a perpetual flux, in unceasing motion and change. In the final analysis, every natural science studies some aspect of this temporal motion vs. spatial form duality. This is the ST question of analysis.
Next Newton and Leibniz studied the ∆question of analysis: how small instants of time or pieces of space gather into larger wholes and vice versa, how to extract the finitesimal quanta from the larger whole.
Both questions put together gave birth of analysis. Hence the classic book text definition:
“∆st Mathematical analysis is that branch of mathematics that provides methods for the quantitative investigation of various processes of change, motion, and dependence of one magnitude on another. ”
The name “infinitesimal analysis” says nothing about the subject matter under discussion but emphasizes the ∆method.
We are dealing here with the special mathematical method of infinitesimals, or in its modern form, of limits.
The error of CLASSIC science being to consider there are NO LIMITS to infintesimals. Yet ∆scales introduce a limit in the minimal quanta, or single frequency in space or time of the parameter study. So we shall talk of finitesimals and quanta, frequencies and bits≈minimal time cycles.
This was far more evident in the beginning, through the calculus of limits and Leibniz’s concept of an infinitesimal as the inverse, 1/n of a quantity, before the axiomatic method stubbornly decided to go ‘inflationary’ with the language of information (as all are) in its 3rd age and talk about bizarre infinities (Cantor).
In fact, analysis is just in its theory an inflationary extension (classic error of all languages from money to fiction words) of the method of limits.
Integrals
Thus integrals, tend to represent the growth of a space population, till it reaches a wholeness in a closed domain. So we can do ‘line integrals’, ‘surface integrals’, ‘volume integrals’, in simultaneous space.
Integrals though are also related to a world cycle, specially the motion if time, closer to the action of reproduction in space, as nature is ‘constantly building integrated wholes by the accumulation of single time actions of reproduction that become ‘clone’ cellsatomscitizens of an integrated supœrganism.
It is precisely in those more complex games of integration of a ‘flow’ of time actions of reproduction, and ‘constrains’ to time actions by an integral line membrane, where we find the more subtle use of both functions. Let us consider it in more detail.
But they can also portray the growth or diminution of populations of space, and then as space is symmetric, we can use inverse functions, normally related to ‘e’.
They can give birth unlike and when a system decreases, the space is dying when it grows it does so slower, so we find also different speeds on the two time arrows of space through the 5th dimension.
Space is symmetric; in its directions and they coexist together. Time is not symmetric and it is experienced as a sequential pattern of single Time cycles. So Time parameters are shorter in form, space is a more extended system. Of time we see only an instant, of space we integrate instants/cycles of time and sum them as frequencies which all play the same world cycle.
Time though often is just the reproduction of a new unit of space. Thus, time cycles become populations of a spatial herd due to its reproduction of a ‘seed’ form.
Space thus is the ‘mirror reproductive symmetry’ of ‘frequencies in time’, its tail of memories, by reproduction, expansion, and radiation along the path of the singular timeline of the wave.
So in broad strokes derivative and integrals cover a wide range of 5D themes: the infinitesimal units of time frequencies and complex herds of space populations.
Whereas given the simultaneity properties of space, integrals tend to be used to calculate space populations, and given the individual sequential structure of time frequencies, derivatives are most often used to calculate time motions.
HOLOGRAPHIC INTEGRALS: Area.
The area of a function has different meanings, but generally speaking is a measure of its vital energy between the singularity at 0 point or initial condition and its membrane, the function itself, between the limits of the domain. So it is an operation constantly performed by membrane and singularity as the initial point and boundary condition, by ‘exfoliating’ in layers the vital space and adding it up piece by piece, finitesimal by finitesimal.
The way mathematics treats integrals deals mostly with the obsession with perfect measure achieved by reducing the sections of the being to minimal infinitesimals. We have discussed the limits and irrelevance of such approach. It is of more interesting to discuss the different type of timelike, spacelike and combined spacetime dimensional functions integrated through this procedure.
And to consider the question of the ‘boundary conditions’, in which the membrane determines the volume which is integrated as the Spacetime area, surrounded by the being
Let us suppose that a curve above the xaxis forms the graph of the function y = f(x). We attempt to find the area S of the segment bounded by the line y = f(x), by the xaxis and by the straight lines drawn through the points x = a and x = b parallel to the yaxis.
To solve this problem we proceed as follows. We divide the interval [a, b] into n parts, not necessarily equal. We denote the length of the first part by Δx1, of the second by Δx2, and so forth up to the final part Δxn. In each segment we choose points ξ1, ξ2, ···, ξn and set up the sum:
The magnitude Sn is obviously equal to the sum of the areas of the rectangles shaded in figure:
The finer we make the subdivision of the segment [a, b], the closer Sn will be to the area S. If we carry out a sequence of such constructions, dividing the interval [a, b] into successively smaller and smaller parts, then the sums Sn will approach S.
The possibility of dividing [a, b] into unequal parts makes it necessary for us to define what we mean by “successively smaller” subdivisions. We assume not only that n increases beyond all bounds but also that the length of the greatest Δxi in the nth subdivision approaches zero. Thus the calculation of the desired area has in this way been reduced to finding the limit:
We note that when we first set up the problem, we had only an empirical idea of what we mean by the area of our curvilinear figure, but we had no precise definition. But now we have obtained an exact definition of the concept of area: It is the limit.:
We now have not only an intuitive notion of area but also a mathematical definition, on the basis of which we can calculate the area numerically .
We have assumed that: ƒ(x)≥0. If f(x) changes sign, then in figure, the limit will give us the algebraic sum of the areas of the segments lying between the curve y = f(x) and the xaxis, where the segments above the xaxis are taken with a plus sign and those below with a minus sign.
Definite integral.
The need to calculate the integral Sum limit arises in many other problems in which a new dimension is reached by the sum of finitesimal paths. For example, suppose that a point is moving along a straight line with variable velocity v = f(t). How are we to determine the distance s covered by the point in the time from t = a to t = b?
Let us assume that the function f(t) is continuous; that is, in small intervals of time the velocity changes only slightly. We divide the interval [a, b] into n parts, of length Δt1, Δt2, ···, Δtn. To calculate an approximate value for the distance covered in each interval Δti, we will suppose that the velocity in this period of time is constant, equal throughout to its actual value at some intermediate point ξ1. The whole distance covered will then be expressed approximately by the sum:
and the exact value of the distance s covered in the time from a to b, will be the limit of such sums for finer and finer subdivisions; that is, it will be the limit:
It would be easy to give many examples of practical problems leading to the calculation of such a limit. We will discuss some of them later, but for the moment the examples already given will sufficiently indicate the importance of this idea. The limit is called the definite integral of the function f(x) taken over the interval [a, b], and it is denoted by
The expression f(x)dx is called the integrand, a and b are the limits of integration; a is the lower limit, b is the upper limit.
The connection between differential and integral calculus.
The problem considered theN reduces to calculation of the definite integral:
Another example IS the problem of finding the area bounded by the parabola y = x².
Here the problem reduces to calculation of the integral:
We were able to calculate both these integrals directly, because we have simple formulas for the sum of the first n natural numbers and for the sum of their squares. But for an arbitrary function f(x), we are far from being able to add up the sum (that is, to express the result in a simple formula) if the points ξi, and the increments Δxi are given to suit some particular problem. Moreover, even when such a summation is possible, there is no general method for carrying it out; various methods, each of a quite special character, must be used in the various cases.
So we are confronted by the problem of finding a general method for the calculation of defiqite integrals. Historically this question interested mathematicians for a long period of time, since there were many practical aspects involved in a general method for finding the area of curvilinear figures, the volume of bodies bounded by a curved surface, and so forth.
We have already noted that Archimedes was able to calculate the area of a segment and of certain other figures. The number of special problems that could be solved, involving areas, volumes, centers of gravity of solids, and so forth, gradually increased, but progress in finding a general method was at first extremely slow. The general method could not be discovered until sufficient theoretical and computational material had been accumulated through the demands of practical life.
The work of gathering and generalizing this material proceeded very gradually until the end of the Middle Ages; and its subsequent energetic development was a direct consequence of the rapid growth in the productive powers of Europe resulting from the breakup of the former (feudal) methods of manufacturing and the creation of new ones (capitalistic).
The accumulation of facts connected with definite integrals proceeded alongside of the corresponding investigations of problems related to the derivative of a function. The reader already knows from that this immense preparatory labor was crowned with success in the 17th century by the work of Newton and Leibnitz. It is in this sense that Newton and Leibnitz are the creators of the differential and integral calculus.
One of the fundamental contributions of Newton and Leibnitz consists of the fact that they finally cleared up the profound connection between differential and integral calculus, which provides us, in particular, with a general method of calculating definite integrals for an extremely wide class of functions.
To explain this connection, we turn tq an example from mechanics.
We suppose that a material point is moving along a straight line with velocity v = f(t), where t is the time. We already know that the distance a covered by our point in the time between t = t1 and t = t2 is given by the definite integral:
Now let us assume that the law of motion of the point is known to us; that is, we know the function s = F(t) expressing the dependence on the time t of the distance s calculated from some initial point A on the straight line. The distance σ covered in the interval of time [t1, t2] is obviously equal to the difference: σ= F(t2) – F(t1)
In this way we are led by physical considerations to the equality:
which expresses the connection between the law of motion of our point and its velocity.
From a mathematical point of view the function F(t), may be defined as a function whose derivative for all values oft in the given interval is equal to f(t), that is:
F'(t)= ƒ(t). Such a function is called a primitive for f(t).
We must keep in mind that if the function f(t) has at least one primitive, then along with this one it will have an infinite number of others; for if F(t) is a primitive for f(t), then F(t) + C, where C is an arbitrary constant, is also a primitive. Moreover, in this way we exhaust the whole set of primitives for f(t), since if F1(t) and F2(t) are primitives for the same function f(t), then their difference ϕ(t) = F1(t) − F2(t) has a derivative ϕ(t) that is equal to zero at every point in a given interval so that ϕ(t) is a constant.*
From a physical point of view the various values of the constant C determine laws of motion which differ from one another only in the fact that they correspond to all possible choices for the initial point of the motion.
We are thus led to the result that for an extremely wide class of functions f(x), including all cases where the function f(x) may be considered as the velocity of a point at the time x, we have the following equality:
where F(x) is an arbitrary primitive for f(x).
This equality is the famous formula of Newton and Leibnitz, which reduces the problem of calculating the definite integral of a function to finding a primitive for the function and in this way forms a link between the differential and the integral calculus.
Many particular problems that were studied by the greatest mathematicians are automatically solved by this formula, stating that the definite integral of the function. f(x) on the interval [a, b] is equal to the difference between the values of any primitive at the left and right ends of the interval.‡ It is customary to write the difference (30) thus:
Example 1. The equality: (x³/3)’=x² shows that the function x³/3 is a primitive for the function x2. Thus, by the formula of Newton and Leibnitz:
Example 2. Let c and c′ be two electric charges, on a straight line at distance r from each other. The attraction F between them is directed along this straight line and is equal to:
F=a/r² (a = kcc′, where k is a constant). The work W done by this force, when the charge c remains fixed but c′ moves along the interval [R1, R2], may be calculated by dividing the interval [R1, R2] into parts Δri.
On each of these parts we may consider the force to be approximately constant, so that the work done on each part is equal to:. Making the parts smaller and smaller, we see that the work W is equal to the integral:
The value of this integral can be calculated at once, if we recall that:
So that:
In particular, the work done by a force F as the charge c′, initially at a distance R1, from c, moves out to infinity, is equal to:
From the arguments given above for the formula of Newton and Leibnitz, it is clear that this formula gives mathematical expression to an actual tie existing in the objective world. It is a beautiful and important example of how mathematics gives expression to objective laws.
We should remark that in his mathematical investigations, Newton always took a physical point of view. His work on the foundations of differential and integral calculus cannot be separated from his work on the foundations of mechanics.
The concepts of mathematical analysis, such as the derivative or the integral, as they presented themselves to Newton and his contemporaries, had not yet completely “broken away” from their physical and geometric origins, such as velocity and area. In fact, they were half mathematical in character and half physical. The conditions existing at that time were not yet suitable for producing a purely mathematical definition of these concepts. Consequently, the investigator could handle them correctly in complicated situations only if he remained in close contact with the practical aspects of his problem even during the intermediate (mathematical) stages of his argument.
From this point of view the creative work of Newton was different in character from that of Leibnitz.* Newton was guided at all stages by a physical way of looking at the problem. But the investigations of Leibnitz do not have such an immediate connection with physics, a fact that in the absence of clearcut mathematical definitions sometimes led him to mistaken conclusions. On the other hand, the most characteristic feature of the creative activity of Leibnitz was his striving for generality, his efforts to find the most general methods for the problems of mathematical analysis.
The greatest merit of Leibnitz was his creation of a mathematical symbolism expressing the essence of the matter. The notations for such fundamental concepts of mathematical analysis as the differential dx, the second differential d2x, the integral ∫y dx, and the derivative d/dx were proposed by Leibnitz. The fact that these notations are still used shows how well they were chosen.
One advantage of a wellchosen symbolism is that it makes our proofs and calculations shorter and easier; also, it sometimes protects us against mistaken conclusions. Leibnitz, who was well aware of this, paid especial attention in all his work to the choice of notation.
The evolution of the concepts of mathematical analysis (derivative, integral, and so forth) continued, of course, after Newton and Leibnitz and is still continuing in our day; but there is one stage in this evolution that should be mentioned especially. It took place at the beginning of the last century and is related particularly to the work of Cauchy.
Cauchy gave a clearcut formal definition of the concept of a limit and used it as the basis for his definitions of continuity, derivative, differential, and integral. These definitions have been introduced at the corresponding places in the present chapter. They are widely used in presentday analysis.
The great importance of these achievements lies in the fact that it is now possible to operate in a purely formal way not only in arithmetic, algebra, and elementary geometry, but also in this new and very extensive branch of mathematics, in mathematical analysis, and to obtain correct results in so doing.
Regarding practical application of the results of mathematical analysis, it is now possible to say: If the original data are verified in the actual world, then the results of our mathematical arguments will also be verified there. If we are properly assured of the accuracy of the original data, then there is no need to make a practical check of the correctness of the mathematical results; it is sufficient to check only the correctness of the formal arguments.
This statement naturally requires the following limitation. In mathematical arguments the original data, which we take from the actual world, are true only up to a certain accuracy. This means that at every step of our mathematical argument the results obtained will contain certain errors, which may accumulate as the number of steps in the argument increases.*
Returning now to the definite integral, let us consider a question of fundamental importance. For what functions f(x), defined on the interval [a, b], is it possible to guarantee the existence of the definite integral:Namely a number to which the sum:
Tends as limit as max Δxi, → 0? It must be kept in view that this number is to be the same for all subdivisions of the interval [a, b] and all choices of the points ξi.
Functions for which the definite integral, namely the limit (29), exists are said to be integrable on the interval [a, b]. Investigations carried out in the last century show that all continuous functions are integrable.
But there are also discontinuous functions which are integrable. Among them, for example, are those functions which are bounded and either increasing or decreasing on the interval [a, b].
The function that is equal to zero at the rational points in [a, b] and equal to unity at the irrational points, may serve as an example of a nonintegrable function, since for an arbitrary subdivision the integral sum sn, will be equal to zero or unity, depending on whether we choose the points ξi, as rational numbers or irrational.
Let us note that in many cases the formula of Newton and Leibnitz provides an answer to the practical question of calculating a definite integral. But here arises the problem of finding a primitive for a given function; that is, of finding a function that has the given function for its derivative. We now proceed to discuss this problem. Let us note by the way that the problem of finding a primitive has great importance in other branches of mathematics also, particularly in the solution of differential equations.
As we stated before integrals are mostly useful when we are studying a ‘defined’ full Spe<ST>Tiƒ system with a membrane or contour closing the surface. As integrals are more concerned with ‘space’ and ‘derivatives’ with time. And further on, those which integrate spacetime systems, or double and triple integrals.
Calculus of ALL type of vital spaces, enclosed by time functions, with a ‘scalar’ point of view, parameter that measured what the point of view extracted in symbiosis with the membrane, from the vital space it enclosed. Alas, this quantity absorbed and ab=used by the point of view, on the vital space would be called ‘Energy’, the vital space ‘field’, the membrane ‘frequency’, the finitesimal ‘quanta or Universal constant’, and the scalar point of view ‘active magnitude.
The fundamental language of physics are differential equations, which allow to measure the content of vital space of a system. The richness and varieties of ‘world species’ will define many variations on the theme. Sometimes there will not be a central point of view, and we talk of a liquid state’, where volumes will not have a ‘gradient’, but ‘Pressure’, the controlling parameter of the time membrane will be equal, or related to the gradient of the eternal world p.o.v. of the Earth (gravitational field).
Then we shall integrate along 3 parameters, the density that defines the liquid, the height that defines the gradient and the volume enclosed. Liquids, due to the simplicity of lacking an internal POV, would be the first physical application of Leibniz’s findings by his students, the Bernoulli family. Next a violin player would find the differential equation of waves – the essential equation of the membranes of present time of all systems. The 3^{rd} type of equations, those of the central point of view, will have to wait a mathematician, Poisson – latter refined by Einstein in his General Relativity.
This is the error of Newton. All cycles are finite, as they close into themselves. All worldcycles of life and death are finite as they end as they begun in the dissolution of death. All entropic motions stop. All time vortices once they have absorbed all the entropy of their territory become wrinkled, and die. Newton died, his ‘time duration’ did not extend to infinity.
But those minds measure from their selfcentered point of view, only a part of the Universe, and the rest remains obscure. So all of them display the paradox of the ego, as they confuse the whole Universe with their world, and see themselves larger than all what they don’t perceive. Hence as Descartes wittingly warned the reader in his first sentences ‘every human being thinks he is gifted with intelligence.
The ternary parts of a T.œ: its calculus.
We have already studied the process of integration for functions of one variable defined on a onedimensional region, namely an interval. But the analogous process may be extended to functions of two, three, or more variables, defined on corresponding regions.
For example, let us consider a surface:z= ƒ (x,y)
defined in a rectangular system of coordinates, and on the plane Oxy let there be given a region G bounded by a closed curve Γ. It is required to find the volume bounded by the surface, by the plane Oxy and by the cylindrical surface passing through the curve Γ with generators parallel to the Oz axis (figure 33). To solve this problem we divide the plane region G into subregions by a network of straight lines parallel to the axes Ox and Oy and denote by: G1, G2… Gn.
those subregions which consist of complete rectangles. If the net is sufficiently fine, then practically the whole of the region G will be covered by the enumerated rectangles. In each of them we choose at will a point:
and, assuming for simplicity that Gi denotes not only the rectangle but also its area, we set up the sum:
It is clear that, if the surface is continuous and the net is sufficiently fine, this sum may be brought as near as we like to the desired volume V. We will obtain the desired volume exactly if we take the limit of the sum (47) for finer and finer subdivisions (that is, for subdivisions such that the greatest of the diagonals of our rectangles approaches zero):
From the point of view of analysis it is therefore necessary, in order to determine the volume V, to carry out a certain mathematical operation on the function f(x, y) and its domain of definition G, an operation indicated by the left side of equality (48). This operation is called the integration of the function f over the region G, and its result is the integral of f over G. It is customary to denote this result in the following way:
Similarly, we may define the integral of a function of three variables over a threedimensional region G, representing a certain body in space. Again we divide the region G into parts, this time by planes parallel to the coordinate planes. Among these parts we choose the ones which represent complete parallelepipeds and enumerate them:G1, G2… Gn.
In each of these we choose an arbitrary point:
and set up the sum:
where Gi denotes the volume of the parallelepiped Gi. Finally we define the integral of f(x, y, z) over the region G as the limit:
to which the sum (50) tends when the greatest diagonal d(Gi) approaches zero.
Let us consider an example. We imagine the region G is filled with a nonhomogeneous mass whose density at each point in G is given by a known function ρ(x, y, z). The density ρ(x, y, z) of the mass at the point (x, y, z) is defined as the limit approached by the ratio of the mass of an arbitrary small region containing the point (x, y, z) to the volume of this region as its diameter approaches zero.* To determine the mass of the body G it is natural to proceed as follows. We divide the region G into parts by planes parallel to the coordinate planes and enumerate the complete parallelepipeds formed in this way: G1, G2, …, Gn
Assuming that the dividing planes are sufficiently close to one another, we will make only a small error if we neglect the irregular regions of the body and define the mass of each of the regular regions Gi (the complete parallelepipeds) as the product:
where (ξi, ηi, ζi) is an arbitrary point Gi. As a result the approximate value of the mass M will be expressed by the sum:
and its exact value will clearly be the limit of this sum as the greatest diagonal Gi approaches zero; that is:
The integrals, 49 and 51 are called double and triple integrals respectively.
Let us examine a problem which leads to a double integral. We imagine that water is flowing over a plane surface. Also, on this surface the underground water is seeping through (or soaking back into the ground) with an intensity f(x, y) which is different at different points. We consider a region G bounded by a closed contour (figure 34) and assume that at every point of G we know the intensity f(x, y), namely the amount of underground water seeping through per minute per cm2 of surface; we will have f(x, y) > 0 where the water is seeping through and f(x, y) < 0 where it is soaking into the ground. How much water will accumulate on the surface G per minute ?
If we divide G into small parts, consider the rate of seepage as approximately constant in each part and then pass to the limit for finer and finer subdivisions, we will obtain an expression for the whole amount of accumulated water in the form of an integral: Double (twofold) integrals were first introduced by Euler. Multiple integrals form an instrument which is used everyday in calculations and investigations of the most varied kind.
It would also be possible to show, though we will not do it here, that calculation of multiple integrals may be reduced, as a rule, to iterated calculation of ordinary onedimensional integrals.
Contour and surface integrals. Finally, we must mention that still other generalizations of the integral are possible. For example, the problem of defining the work done by a variable force applied to a material point, as the latter moves along a given curve, naturally leads to a socalled curvilinear integral, and the problem of finding the general charge on a surface on which electricity is continuously distributed with a given surface density leads to another new concept, an integral over a curved surface:
For example, suppose that a liquid is flowing through space ( and that the velocity of a particle of the liquid at the point (x, y)is given by a function P(x, y), not depending on z. If we wish to determine the amount of liquid flowing per minute through the contour Γ, we may reason in the following way. Let us divide Γ up into segments Δsi. The amount of water flowing through one segment Δsi is approximately equal to the column of liquid shaded in figure 35; this column may be considered as the amount of liquid forcing its way per minute through that segment of the contour. But the area of the shaded parallelogram is equal to:
P i (x,y) • ∆Si • cos α i where αi is the angle between the direction, ‾x of the xaxis and the outward normal of the surface bounded by the contour Γ; this normal is the perpendicular ñ to the tangent, which we may consider as defining the direction of the segment Δsi. By summing up the areas of such parallelograms and passing to the limit for finer and finer subdivisions of the contour Γ, we determine the amount of water flowing per minute through the contour Γ; it is denoted thus:
and is called a curvilinear integral. If the flow is not everywhere parallel, then its velocity at each point (x, y) will have a component P(x, y) along the xaxis and a component Q(x, y) along the yaxis. In this case we can show by an analogous argument that the quantity of water flowing through the contour will be equal to:
When we speak of an integral over a curved surface G for a function f(M) of its points M(x, y, z), we mean the limit of sums of the form:
for finer and finer subdivisions of the region G into segments whose areas are equal to Δσi.
General methods exist for transforming multiple, curvilinear, and surface integrals into other forms and for calculating their values, either exactly or approximately.
Ostrogradskiĭ Formula.
Several important and very general formulas relating an integral over a volume to an integral over its surface (and also an integral over a surface, curved or plane, to an integral around its boundary) have a very wide application, and are yet another striking proof on the constant transformations of S≈T DIMENSIONS, and interaction of the parts of the system, in this case between the membrane that encircles the vital space, whose parameters ARE ALWAYS CLOSELY RELATED, as we can consider the membrane, just the last ‘cover’ of maximal size of that inner 3D vital energy (unlike the quite distinct singularity, which ‘moves’ across ∆±i scales and tends to be quite different in form, parameters and substance)
Let us put an example: imagine, as we did before, that over a plane surface there is a horizontal flow of water that is also soaking into the ground or seeping out again from it. We mark off a region G, bounded by a curve Γ, and assume that for each point of the region we know the components P(x, y) and Q(x, y) of the velocity of the water in the direction of the xaxis and of the yaxis respectively.
Let us calculate the rate at which the water is seeping from the ground at a point with coordinates (x, y). For this purpose we consider a small rectangle with sides Δx and Δy situated at the point (x, y).
As a result of the velocity P(x, y) through the left vertical edge of this rectangle, there will flow approximately P(x, y)Δy units of water per minute into the rectangle, and through the right side in the same time will flow out approximately P(x + Δx, y)Δy units. In general, the net amount of water leaving a square unit of surface as a result of the flow through its left and right vertical sides will be approximately:
If we let Δx approach zero, we obtain in the limit: ∂P/∂x.
Correspondingly, the net rate of flow of water per unit area in the direction of the yaxis will be given by: ∂Q/∂y.
This means that the intensity of the seepage of ground water at the point with coordinates (x, y) will be equal to: ∂P/∂x + ∂Q/∂y
But in general, as we saw earlier, the quantity of water coming out from the ground will be given by the double integral of the function expressing the intensity of the seepage of ground water at each point, namely:
But, since the water is incompressible, this entire quantity must flow out during the same time through the boundaries of the contour Γ. The quantity of water flowing out through the contour Γ is expressed, as we saw earlier, by the curvilinear integral over Γ:
The equality of the magnitudes (52) and (53) gives in its simplest twodimensional case:
A key formula to mirror a widespread phenomenon in the external world, which in our example we interpreted in a readily visualized way as preservation of the volume of an incompressible fluid.
Which can be generalise to express the connection between an integral over a multidimensional volume and an integral over its surface. In particular, for a threedimensional body G, bounded by the surface Γ:
where dσ is the element of surface.
It is interesting to note that the fundamental formula of the integral calculus:
may be considered as a onedimensional case. The equation (54) connects the integral over an interval with the “integral” over its “nulldimensional” boundary, consisting of the two end points.
Formula (54) may be illustrated by the following analogy. Let us imagine that in a straight pipe with constant cross section s = 1 water is flowing with velocity F(x), which is different for different cross sections (figure 36). Through the porous walls of the pipe, water is seeping into it (or out of it) at a rate which is also different for different cross sections:
If we consider a segment of the pipe from x to x + Δx, the quantity of water seeping into it in unit time must be compensated by the difference F(x + Δx) – F(x) between the quantity flowing out of this segment and the quantity flowing into it along the pipe.
So the quantity seeping into the segment is equal to the difference F(x + Δx) – F(x), and consequently the rate of seepage per unit length of pipe (the ratio of the seepage over an infinitesimal segment to the length of the segment) will be equal to:
More generally, the quantity of water seeping into the pipe over the whole section [a, b] must be equal to the amount lost by flow through the ends of the pipe. But the amount seeping through the walls is equal to:and the amount lost by flow through the ends is F(b) – F(a). The equality of these two magnitudes produces formula.
GREEN’S THEOREM
Then there is of course the fact that a system in spacetime, in which there is a displacement in time, will be equivalent to a system in which this time motion is seen as fixed space. Such cases mean that we can integrate lines with motion into planes, and surfaces with motion into volumes.
The result is:
Green’s theorem, which gives the relationship between a line integral around a simple closed curve C and a double integral over the plane region D bounded by C. It is named after George Green[1] and is the twodimensional special case of the more general Kelvin–Stokes theorem.
Stoke’s theorem, which says that the integral of a differential form ω over the boundary of some orientable manifold Ω is equal to the integral of its exterior derivative dω over the whole of Ω: ∫∂Ω ω = ∫Ω d ω
In general such Integrals also follow the geometrical structure of a system built with an external membrane, which absorbs the information of the system, and internal 0point of scalar that focus it and a vital space between them. The result of these relationships allow to define the basic laws of integrals in time and space that relate line integrals to surface integrals to volume integrals, of which the best known example are the laws of electromagnetism, written not as Maxwell did, in terms of derivatives (curls and gradients) but of integrals.
So we can in this manner represent the laws of electromagnetism and fully grasp the meaning of magnetism, the external membrane and charge, the central point of view, with interactions between both, the electromagnetic waves and induced currents and magneto fluxes.
Thus the best examples in physics of this relationship are the 4 equations of Maxwell:
While the other 2 define the membrane of an electromagnetic field, the magnetic field and the central point of view or charge:
So we can consider that the Tƒ element of the electromagnetic field, the charge or opoint and the membrane, or closed outer path either in its integral or inverse differential equations, and the wave interaction between them, easily deduced from the stokes theorem or expressed inversely in differential form, give us the full description of an electromagnetic system in terms of the generator:
Spmembrane (magnetic fieldgauss’ law of magnetism) < ST (Faraday/Ampere’s laws of interaction between Sp and Tƒ) >Tƒ (Gauss Law of the central point).
And that those interactions are integrals of the quanta of the ∆1 field in which the electric charge and magnetic field that integrates them arouses.
Density integrals. The meaning of Tƒ/Sp and Sp/Tƒ: information and energy densities:
Some General Remarks on the Formation and Solution of Differential Equations
As we have expressed many times, equations do NOT have solutions till the whole information on its ternary TŒ are given. Which in time means to know an initial condition and an end, through which the function will run under the principle of completing its action in the least possible time: Max. S T (min. t), which for any combination of Space and Time dimensions implies to complete the action in the minimal possible time.
Conditions of this type arise in all equations of all stiences.
In the symmetry of space the boundary of the T.œ though must be expressed as lineal conditions of the 2D membrane and 1D singularity, which can be superposed 1D+2D to give the 3D solution of the vital space both enclose, normally through a product operator, 1 D x 2D=3D.
In any case each equation once determined by its space or time constrictions, can be found certain solutions, which form a sœT of possible frequencies or areas that are efficient parameters for the MIND equation to describe real T.œs (expressed here in the semantic inversion of sets and toes).
The key then to understand those solutions and its approximations is the fact that singularity and membrane conditions are expressed as scalars and lineal functions, while ternary vital energy solutions have cyclical form.
MORE COMPLEX DERIVATIVES: CURVATURE, TENSORS – ITS LIMIT OF UTILITY
Physical quantities may be of 3, s, t, st kinds.
NOW BEYOND 2 planes of existence, the utility of derivatives diminish, as organisms become invisible and do not organise further, and so because in the same plane we use a first derivative, in relationships between two any planes we use derivatives of the 2nd degree, the maximal use possible for derivatives are derivatives COMES from third degree derivatives, which give us the limit of information, in the form of a single parameter, 1/r², curvature.
Beyond that planes of pure space and pure time are not perceivable, so departing from the fact that all is S with some t (energy) or T with some S(information), we can still broadly talk of dominate spacelike parameters, timelike parameters and use the Spacetime parameters only for those in which S≈t (e=i) holds quite exactly.
Pure space and pure time.
Now the closest thing to the description of pure space as it emerges from ∆1 and influences a higher ∆ scale as a ‘field of forces’.
And the closest thing to pure time, is a process of ‘implosion’ that ‘forces down’ or ‘depresses’ (inverse of emergence), a system from an ∆+1, timelike implossive force. And that is the meaning of mass, the force down, in/form/ative force coming from the ∆+1 scale.
Since pure, implosive time and pure expansive, entropic space are not observable, the best way to ‘get it’, is when the implosive time process is felt by something which is smaller inside the vortex. So we feel mass, from the ∆+3 galactic scale as Mach and Einstein mused, because inward implosive informative forces affect mostly the internal parts, NOT the external ones. And we field inversely a field of expansive entropy, from smaller parts, exploding us from inside out.
Then we come to energylike (max. Se x min Ti) and informative like (max. Ti x min. Se) parameters.
Some are completely characterized by their numerical values, e.g., temperature, density, and the like, and are called scalars. Scalars are then to be considered parameters of closed informative functions. In the case of density is evident.
Temperature is not so clear a time parameter. But temperature, when properly constrain it to what and where we truly measure as temperature (Not the frequency of a wave), that is the vibrating motions of molecules of the ∆1 scale, in a closed space, hence a time parameter. So goes for massenergy, as energy becomes added to mass, always that we can measure it in an enclosed region of space, belonging therefore to a timeclosed world. So gluons of motionenergy enclosed in a nucleus do store most of the mass of atoms; as they are to be understood in terms of closed timeparameters from a potential point of view.
So goes from potential energy, which is stored in time cycles. So as a rule, regardless of how ‘distorted’ is conceptually current science and how unlikely will be a change of paradigm for centuries to come (we still drag, for example the – sign in electrons since Franklin chose it), the nondistorted truth can classify all parameters and physical quantities in terms of time and space.
On the other hand energylike parameters will have direction as vector quantities: velocity, acceleration, the strength of an electric field, etc. The simpler those parameters, with less ‘scalar’ elements the more spacelike, entropylike, fieldlike they will be. So again as a rule, the less dimensions we need to define a system the more spaceentropyfield like it will be.
Thus spacelike Vector quantities may be expressed just by the length of the vector and its direction or its spacedimensional “components” if we decompose it into the sum of three mutually perpendicular vectors parallel to the coordinate axes.
While a spacetime balanced process will have more ‘parameters’ to study than a simple vector, growing in dimensions till becoming 4vectors and finally a ‘tensor’, which will be the final ‘growth into a 5D ∆event
So it is easy just looking at a equation to know what kind of s, t, or st (e, i exi) process we study.
For example, a classic st process, which is, let us remember an S≈T process is one which tends to a dynamic balance between both parameters
So it is an oscillatory system in any ∆scale or ‘spacetime medium’. In such oscillations every point of the medium, occupying in equilibrium the position (x, y, z), will at time t be displaced along a vector u(x, y, z, t), depending on the initial position of the point (x, y, z) and on the time t.
In this case the process in question will be described by a vector field. But it is easy to see that knowledge of this vector field, namely the field of displacements of points of the medium, is not sufficient in itself for a full description of the oscillation.
It is also necessary to know, for example, the density ρ(x, y, z, t) at each point of the medium, the temperature T(x, y, z, t).
So we ad to the Spevector field, some TPARAMETERS, (closed Tvibration, density, and stress, which connects the system in a timenetwork like fashion to all other points of the whole).
∆events. Finally in processes which require the study of interactions between ∆scales, hence 5D processes, we need even more complex elements.
For example a classic ∆event is the internal stress, i.e., the forces exerted on an arbitrarily chosen volume of the body by the entire remaining part of it.
And so we arrive to systems defined by tensors, often with 6 dimensions (we forget the final r=evolution of thought that would make it all less distorted of working on bidimensional space and time, as to simplify understanding so the idea of a tensor is that the whole works dynamically into the ‘point’, described as a ‘cube’, with 6 faces, or ± elements on the pointparticlebeing from the 3 classic spacedimensions:
Examples of it are the mentioned stress, shown in the graph, or in the relativity description of how the ∆+3 scale of gravitational spacetime, influences the lower scales with the effect of implosive mass.
Thus in addition to Svector and Tscalar quantities, more complicated entities occur in SPACEtime events, often characterized everywhere by a set of functions of the same four independent variables; where the function is a description of the ∆1 scale, which come together into the upper new ‘function’ of functions, or ‘functional equation’.
And so we can classify in this manner according to the ∆•ST ternary method all the parameters of mathematical physics.
And they will reveal us the real structure and ∆ST symmetries they analyse according to the number of dimensions of complexity they use.
Yet beyond the range of ‘tensors’, which study relationships between 2 or at best a maximal of 3 scales of reality, there is nothing to be found. So happens when we consider the number of differential equations, we want to study. Nothing really goes on beyond the 3rd derivative of a system that scales down 2 scales (entropydeath events, dual emergence upwards from seed to world).
So one of the most fascinating events of the relationship of ilogic and the real world is to properly interpret what Einstein said he could not resolve:
‘I know when mathematics is truth but not when it is real’
By the fact that as the Soviet school of maths from Lobachevski to Aleksandrov rightly explained, we NEED an experimental method inserted in Mathematics to know when mathematics is both, LOGICALLY CONSISTENT as a language in its inner syntax but NOT fiction of beauty but REAL.
A bit on the ‘numbers’
Now the fundamental concept behind analysis is the ∂∫ duality of ‘derivatives’ and ‘integrals’, related at first sights to the concepts of ‘time’ and ‘space’ (you derivate in time, you integrate in space), and to the concept of scalar ‘evolution’ from parts into wholes (you derivate to obtain a higher scalar wholeness.
i.e. you derivate a past space into present speed, adding a timemotion, and then derivate into acceleration – future time – to obtain the ‘most thorough single parameter of the being in time’: its acceleration that encodes also its speed and distance.
On the other hand you integrate in space, and so it is also customary to consider the first and second integral, which will bring also the ternary scale of volume of a system.
And it is a tribute to the simplicity of the Universe that further ‘derivatives’ are not really needed, to perceive the system in its full time and space parameters. As further derivations and integrations are not needed (they happen in the search of curvature and tensions and jerks, rates of acceleration, which are really menial details and in some combined spacetime multiple systems).
Approximations.
We have though already commented on Algebra, that the third derivative, or higher derivatives however are used to improve the accuracy of an approximation to the function:
f(Xo+h)=f(xo)+f′(Xo)h+f″(Xo)h²/2!+f‴(ξ)h³/3!
Thus Taylor’s expansion of a function around a point involves higher order derivatives, and the more derivatives you consider, the higher the accuracy. This also translates to higher order finite difference methods when considering numerical approximations.
Now what this means is obvious: beyond the accuracy of the three derivatives canonical to an ∆º±1 supœrganism, information as it passes the potential barrier between scales of the 5th≈∆dimension, suffers a loss of precision so beyond the third derivative, we can only obtain approximations by using higher derivatives or in a likely less focused=exact procedure the equivalent polynomials, more clear expressions of ‘dimensional growth.
So their similitude first of all proves that both high derivatives and polynomials are representations of growth across planes and scales, albeit loosing accuracy.
However in the fifth dimensional correct perspective is more accurate the derivativeintegral game; as it ‘looks at the infinitesimal’ to integrate then the proper quanta.
Taylor’s Formula
The function: where the coefficients ak are constants, is a polynomial of degree=dimension n. In particular, y = ax + b is a polynomial of the first degree and y = ax² + bx + c is a polynomial of the second dimension. Dimensional polynomials have the particularity that are mostly 2manifolds symmetric in that x=y, that is both dimension square, D1 x D2.
To achieve this feat, S=t, tt or ss STEPS MUST BE CONSIDERED.
Polynomials may be considered IN THAT SENSE as the simplest of all poliDIMENSIONAL functions. In order to calculate their value for a given x, we require only the operations of addition, subtraction, and multiplication; not even division is needed. Polynomials are continuous for all x and have derivatives of arbitrary order. Also, the derivative of a polynomial is again a polynomial, of degree lower by one, and the derivatives of order n + 1 and higher of a polynomial of degree n are equal to zero. Yet the derivative diminishes slower than the simple square of the function; so if we consider the derivative the parts of the polynomial, the product of those parts would be more than the whole.
It is then when we can increase the complexity establishing for example, ratios of polynomials.
If to the polynomials we adjoin functions of the form:
for the calculation of which we also need division, and also the functions √x and ³√X, finally, arithmetical combinations of these functions, we obtain essentially all the functions whose values can be calculated by such methods.
But what a polynomial describes? All others are easier to get through approximations:
On an interval containing the point a, let there be given a function f(x) with derivatives of every order. The polynomial of first degree:
p1(x) = ƒ(a) + ƒ'(a) (xa) has the same value as f(x) at the point x = a and also, as is easily verified, has the same derivative as f(x) as this point. Its graph is a straight line, which is tangent to the graph of f(x) to the point a. It is possible to choose a polynomial of the second degree, namely: which at the point of x = a has with f(x) a common value and a common first and second derivative. Its graph at the point a will follow that of f(x) even more closely. It is natural to expect that if we construct a polynomial which at x = a has the same first n derivatives as f(x) at the same point, then this polynomial will be a still better approximation to f(x) at points x near a. Thus we obtain the following approximate equality, which is Taylor’s formula:The right side of this formula is a polynomial of degree n in (x − a). For each x the value of this polynomial can be calculated if we know the values of f(a), f′(a), ···, f(n)(a).
For functions which have an (n + 1)th derivative, the right side of this formula, as is easy to show, differs from the left side by a small quantity which approaches zero more rapidly than (x − a)n. Moreover, it is the only possible polynomial of degree n that differs from f(x), for x close to a, by a quantity that approaches zero, as x → a, more rapidly than (x − a)n. If f(x) itself is an algebraic polynomial of degree n, then the approximate equality (25) becomes an exact one.
Finally, and this is particularly important, we can give a simple expression for the difference between the right side of formula (25) and the actual value of f(x). To make the approximate equality (25) exact, we must add to the right side a further term, called the “remainder term”
has the peculiarity that the derivative appearing in it is to be calculated in each case not at the point a but at a suitably chosen point ξ, which is unknown but lies somewhere in the interval between a and x.
So we cab make use of the generalized meanvalue theorem quoted earlier:
Differentiating the functions ϕ(u) and ψ(u) with respect to u (it must be recalled that the value of x has been fixed) we find that: The equality of this last expression with the original quantity (27) gives Taylor’s formula in the form (26).
In the form (26) Taylor’s formula not only provides a means of approximate calculation of f(x) but also allows us to estimate the error.
And so with Taylor we close this introduction to derivatives and differentials, enlightened with the basic elements that relate them to the 5 dimensions of spacetime, specifically to the ∆1 finitesimals.
DETERMINISM IN SOLUTION TO ODEs.
Lineal vs cyclic; dis≈continuity; 1st vs 2nd order, partial vs ordinary,∂ vs ∫; 3 states of matter & its freedoms.
In the philosophy of science of mathematical physics some concepts come back once and again, based in dualities
And so the qualitative description of all those entropic, reproductive/moving and informative/Tiƒ timelike vortices became ‘only’ mathematical.
It is interesting at this stage, to consider that the whole world of ∫∂ mathematics has two approaches which humans as always being onedimensional did not fully find complementary but argue, the method of newton based in infinite series (arithmetic, temporal pov) and that of Leibniz using spatial, geometric concepts (tangents); which is the one, being more evident and simpler, that stood.
First, trivially speaking the existence of such 2 canonical, timespace ways to come to ∫∂ is a clear proof that both researchers found ∫∂ independently. Next, their argument about who was right and better shows how oneminded is humanity, and third, the dominance of Leibniz’s methods for being visual, geometrical, spatial tells a lot about the difficulty humans have to understand time, causality and the concepts of infinity, limit, discontinuity, continuity, and other key elements of ∆nalysis, which we shall argue in our mathematical posts on… ∆nalysis.
All this said, mathematical physics moved to the geometric, continental school with the help of Leibniz’s disciples, the Bernoulli.
And it is interesting to consider a diachronic concept to analyse the enormous flourishing of those equation…
PDE
INTRODUCTION: PHILOSOPHICAL CONSIDERATIONS
Physical equations are also equations related to the 3 elements of al the existential entities of the Universe, which we will develop in detail on a future post on physics, accompanying this one. It must then be understood that within the general ƒ(x)≈f(t) and y=S isomorphism between mathematical equations and STeps (not always te case as symmetric steps can repeat itself with the same parameters in SSS and TTT derivatives as we have seen in our intro to ODE), partial differential equations, will be combinations of analysis of systems in its ‘primary’ differential finitesimals of space and time then aggregated in more complex St SYSTEMS, giving as an enormous range of possible PDE studies, which we shall strive to order according to the concept that there is a geometric symmetry between algebra (s≈t symmetries) and geometry (Swholes sum of tdynamic points) and analysis (steps).
So it is a good guidance for all algebraic, analytic equations to make a comment of its significance in the vital ternary geometry of a T.œ or complex event between T.œs ACROSS different planes, ∆§ studied with those equations.
Partial Differential equations as ∆•stequations.
Physical events and processes occuring in a space time system always consist of the changes, during the passage of its finite time, of certain physical magnitudes related to its points of vital space.
This simple definition of spacetime processes is at the heart of the whole differential calculus, which with slight changes of interpretation apply to all GST.
Any of those ST processes can be described by functions with four ST, independent variables, S(x, y), and (z, ƒ), where x, y are the coordinates of a point of the space, and , and z and ƒ of time.
So ideally in a world in which humans had not distorted bidimensional time cycles, the way we work around mathematical equations would be slightly changed. As we are not reinventing the human mind of 7 billion people – we are not that arrogant, we just will feel happy trying to explain a few of those processes of bidimensional space and time here.
In the study of the phenomena of nature, partial differential equations are encountered just as often as ordinary ones. As a rule this happens in cases where an event is described by a function of several variables. From the study of nature there arose that class of partial differential equations that is at the present time the most thoroughly investigated and probably the most important in the general structure of human knowledge, namely the equations of mathematical physics.
Let us first consider oscillations in any kind of medium. In such oscillations every point of the medium, occupying in equilibrium the position (x, y, z), will at time t be displaced along a vector u(x, y, z, t), depending on the initial position of the point (x, y, z) and on the time t. In this case the process in question will be described by a vector field. But it is easy to see that knowledge of this vector field, namely the field of displacements of points of the medium, is not sufficient in itself for a full description of the oscillation. It is also necessary to know, for example, the density ρ(x, y, z, t) at each point of the medium, the temperature T(x, y, z, t), and the internal stress, i.e., the forces exerted on an arbitrarily chosen volume of the body by the entire remaining part of it.
Physical events and processes occuring in space and time always consist of the changes, during the passage of time, of certain physical magnitudes related to the points of the space. As we saw in Chapter II these quantities can be described by functions with four independent variables, x, y, z, and t, where x, y, and z are the coordinates of a point of the space, and and t is the time.
Physical quantities may be of different kinds. Some are completely characterized by their numerical values, e.g., temperature, density, and the like, and are called scalars. Others have direction and are therefore vector quantities: velocity, acceleration, the strength of an electric field, etc. Vector quantities may be expressed not only by the length of the vector and its direction but also by its “components” if we decompose it into the sum of three mutually perpendicular vectors, for example parallel to the coordinate axes.
In mathematical physics a scalar quantity or a scalar field is presented by one function of four independent variables, whereas a vector quantity defined on the whole space or, as it is called, a vector field is described by three functions of these variables. We can write such a quantity either in the form:
U (x,y,z,t) where the bold face type indicates the u is a vector, or in the form of three functions:Ux (x,y,z,t), U y(x,y,z,t), Uz (x,y,z,t)
where ux, uy, and uz denote the projections of the vector on the coordinate axes.
In addition to vector and scalar quantities, still more complicated entities occur in physics, for example the state of stress of a body at a given point. Such quantities are called tensors; after a fixed choice of coordinate axes, they may be characterized everywhere by a set of functions of the same four independent variables.
In this manner, the description of widely different kinds of physical phenomena is usually given by means of several functions of several variables. Of course, such a description cannot be absolutely exact.
For example, when we describe the density of a medium by means of one function of our independent variables, we ignore the fact that at a given point we cannot have any density whatsoever. The bodies we are investigating have a molecular structure, and the molecules are not contiguous but occur at finite distances from one another. The distances between molecules are for the most part considerably larger than the dimensions of the molecules themselves. Thus the density in question is the ratio of the mass contained in some small, but not extremely small, volume to this volume itself. The density at a point we usually think of as the limit of such ratios for decreasing volumes. A still greater simplification and idealization is introduced in the concept of the temperature of a medium. The heat in a body is due to the random motion of its molecules. The energy of the molecules differs, but if we consider a volume containing a large collection of molecules, then the average energy of their random motions will define what is called temperature.
Similarly, when we speak of the pressure of a gas or a liquid on the wall of a container, we should not think of the pressure as though a particle of the liquid or gas were actually pressing against the wall of the container. In fact, these particles, in their random motion, hit the wall of the container and bounce off it. So what we describe as pressure against the wall is actually made up of a very large number of impulses received by a section of the wall that is small from an everyday point of view but extremely large in comparison with the distances between the molecules of the liquid or gas. It would be easy to give dozens of examples of a similar nature. The majority of the quantities studied in physics have exactly the same character. Mathematical physics deals with idealized quantities, abstracting them from the concrete properties of the corresponding physical entities and considering only the average values of these quantities.
Such an idealization may appear somewhat coarse but, as we will see, it is very useful, since it enables us to make an excellent analysis of many complicated matters, in which we consider only the essential elements and omit those features which are secondary from our point of view.
The object of mathematical physics is to study the relations existing among these idealized elements, these relations being described by sets of functions of several independent variables.
The Simplest Equations of Mathematical Physics
The elementary connections and relations among physical quantities are expressed by the laws of mechanics and physics. Although these relations are extremely varied in character, they give rise to more complicated ones, which are derived from them by mathematical argument and are even more varied. The laws of mechanics and physics may be written in mathematical language in the form of partial differential equations, or perhaps integral equations, relating unknown functions to one another. To understand what is meant here, let us consider some examples of the equations of mathematical physics.
A partial differential equation (PDE) is a differential equation that contains unknown multivariable functions and their partial derivatives. (This is in contrast to ordinary differential equations, which deal with functions of a single variable and their derivatives.) PDEs are used to formulate problems involving functions of several variables, and are either solved by hand, or used to create a relevant computer model.
PDEs can be used to describe a wide variety of phenomena such as sound, heat, electrostatics, electrodynamics, fluid flow, elasticity, or quantum mechanics. These seemingly distinct physical phenomena can be formalised similarly in terms of PDEs. Just as ordinary differential equations often model onedimensional dynamical systems, partial differential equations often model multidimensional systems. PDEs find their generalisation in stochastic partial differential equations.
Both ordinary and partial differential equations are broadly classified as linear and nonlinear.
 A differential equation is linear if the unknown function and its derivatives appear to the power 1 (products of the unknown function and its derivatives are not allowed) and nonlinear. The characteristic property of linear equations is that their solutions form an affine subspace of an appropriate function space, which results in much more developed theory of linear differential equations. Homogeneous linear differential equations are a further subclass for which the space of solutions is a linear subspace i.e. the sum of any set of solutions or multiples of solutions is also a solution. The coefficients of the unknown function and its derivatives in a linear differential equation are allowed to be (known) functions of the independent variable or variables; if these coefficients are constants then one speaks of a constant coefficient linear differential equation.
 There are very few methods of solving nonlinear differential equations exactly; those that are known typically depend on the equation having particular symmetries. Nonlinear differential equations can exhibit very complicated behavior over extended time intervals, characteristic of chaos. Even the fundamental questions of existence, uniqueness, and extendability of solutions for nonlinear differential equations, and wellposedness of initial and boundary value problems for nonlinear PDEs are hard problems and their resolution in special cases is considered to be a significant advance in the mathematical theory (cf. Navier–Stokes existence and smoothness). However, if the differential equation is a correctly formulated representation of a meaningful physical process, then one expects it to have a solution.
Linear differential equations frequently appear as approximations to nonlinear equations. These approximations are only valid under restricted conditions. For example, the harmonic oscillator equation is an approximation to the nonlinear pendulum equation that is valid for small amplitude oscillations (see below).
Examples
In the first group of examples, let u be an unknown function of x, and c and ω are known constants.
 Inhomogeneous firstorder linear constant coefficient ordinary differential equation:
 Homogeneous secondorder linear ordinary differential equation:
 Homogeneous secondorder linear constant coefficient ordinary differential equation describing the harmonic oscillator:
 Inhomogeneous firstorder nonlinear ordinary differential equation:
 Secondorder nonlinear (due to sine function) ordinary differential equation describing the motion of a pendulum of length L:
In the next group of examples, the unknown function u depends on two variables x and t or x and y.
 Homogeneous firstorder linear partial differential equation:
 Homogeneous secondorder linear constant coefficient partial differential equation of elliptic type, the Laplace equation:
 Thirdorder nonlinear partial differential equation, the Korteweg–de Vries equation:
Existence of solutions
Solving differential equations is not like solving algebraic equations. Not only are their solutions often times unclear, but whether solutions are unique or exist at all are also notable subjects of interest.
For first order initial value problems, it is easy to tell whether a unique solution exists. Given any point in the xyplane, define some rectangular region , such that and is in . If we are given a differential equation and an initial condition , then there is a unique solution to this initial value problem if and are both continuous on . This unique solution exists on some interval with its center at .
However, this only helps us with first order initial value problems. Suppose we had a linear initial value problem of the nth order such that For any nonzero , if and are continuous on some interval containing , is unique and exists.
 A delay differential equation (DDE) is an equation for a function of a single variable, usually called time, in which the derivative of the function at a certain time is given in terms of the values of the function at earlier times.
 A stochastic differential equation (SDE) is an equation in which the unknown quantity is a stochastic process and the equation involves some known stochastic processes, for example, the Wiener process in the case of diffusion equations.
 A differential algebraic equation (DAE) is a differential equation comprising differential and algebraic terms, given in implicit form.
The theory of differential equations is closely related to the theory of difference equations, in which the coordinates assume only discrete values, and the relationship involves values of the unknown function or functions and values at nearby coordinates. Many methods to compute numerical solutions of differential equations or study the properties of differential equations involve approximation of the solution of a differential equation by the solution of a corresponding difference equation.
The study of differential equations is a wide field in pure and applied mathematics, physics, and engineering. All of these disciplines are concerned with the properties of differential equations of various types. Pure mathematics focuses on the existence and uniqueness of solutions, while applied mathematics emphasizes the rigorous justification of the methods for approximating solutions. Differential equations play an important role in modelling virtually every physical, technical, or biological process, from celestial motion, to bridge design, to interactions between neurons. Differential equations such as those used to solve reallife problems may not necessarily be directly solvable, i.e. do not have closed form solutions. Instead, solutions can be approximated using numerical methods.
Many fundamental laws of physics and chemistry can be formulated as differential equations. In biology and economics, differential equations are used to model the behavior of complex systems. The mathematical theory of differential equations first developed together with the sciences where the equations had originated and where the results found application. However, diverse problems, sometimes originating in quite distinct scientific fields, may give rise to identical differential equations. Whenever this happens, mathematical theory behind the equations can be viewed as a unifying principle behind diverse phenomena.
As an example, consider propagation of light and sound in the atmosphere, and of waves on the surface of a pond. All of them may be described by the same secondorder partial differential equation, the wave equation, which allows us to think of light and sound as forms of waves, much like familiar waves in the water. Conduction of heat, the theory of which was developed by Joseph Fourier, is governed by another secondorder partial differential equation, the heat equation. It turns out that many diffusion processes, while seemingly different, are described by the same equation; the Black–Scholes equation in finance is, for instance, related to the heat equation.
In physics:
Classical mechanics:
So long as the force acting on a particle is known, Newton’s second law is sufficient to describe the motion of a particle. Once independent relations for each force acting on a particle are available, they can be substituted into Newton’s second law to obtain an ordinary differential equation, which is called the equation of motion.
Electrodynamics:
Maxwell’s equations are a set of partial differential equations that, together with the Lorentz force law, form the foundation of classical electrodynamics, classical optics, and electric circuits. These fields in turn underlie modern electrical and communications technologies. Maxwell’s equations describe how electric and magnetic fields are generated and altered by each other and by charges and currents. They are named after the Scottish physicist and mathematician James Clerk Maxwell, who published an early form of those equations between 1861 and 1862.
General relativity:
The Einstein field equations (EFE; also known as “Einstein’s equations”) are a set of ten partial differential equations in Albert Einstein’s general theory of relativity which describe the fundamental interaction of gravitation as a result of spacetime being curved by matter and energy. First published by Einstein in 1915 as a tensor equation, the EFE equate local spacetime curvature (expressed by the Einstein tensor) with the local energy and momentum within that spacetime (expressed by the stress–energy tensor).
Quantum mechanics:
In quantum mechanics, the analogue of Newton’s law is Schrödinger’s equation (a partial differential equation) for a quantum system (usually atoms, molecules, and subatomic particles whether free, bound, or localized). It is not a simple algebraic equation, but in general a linear partial differential equation, describing the timeevolution of the system’s wave function (also called a “state function”).
Other important equations:
 Euler–Lagrange equation in classical mechanics
 Hamilton’s equations in classical mechanics
 Radioactive decay in nuclear physics
 Newton’s law of cooling in thermodynamics
 The wave equation
 The heat equation in thermodynamics
 Laplace’s equation, which defines harmonic functions
 Poisson’s equation
 The geodesic equation
 The Navier–Stokes equations in fluid dynamics
 The Diffusion equation in stochastic processes
 The Convection–diffusion equation in fluid dynamics
 The Cauchy–Riemann equations in complex analysis
 The Poisson–Boltzmann equation in molecular dynamics
 The shallow water equations
 Universal differential equation
The Lorenz equations whose solutions exhibit chaotic flow.
Simple examples.
In ∆st are therefore Mathematical statement containing one or more derivatives—that is, terms representing the rates of change of continuously varying quantities. Differential equations are very common in science and engineering, as well as in many other fields of quantitative study, because what can be directly observed and measured for systems undergoing changes are their rates of change. The solution of a differential equation is, in general, an equation expressing the functional dependence of one variable upon one or more others; it ordinarily contains constant terms that are not present in the original differential equation. Another way of saying this is that the solution of a differential equation produces a function that can be used to predict the behaviour of the original system, at least within certain constraints.
Differential equations are classified into several broad categories, and these are in turn further divided into many subcategories. The most important categories are ordinary differential equations and partial differential equations. When the function involved in the equation depends on only a single variable, its derivatives are ordinary derivatives and the differential equation is classed as an ordinary differential equation. On the other hand, if the function depends on several independent variables, so that its derivatives are partial derivatives, the differential equation is classed as a partial differential equation. The following are examples of ordinary differential equations:
In these, y stands for the function, and either t or x is the independent variable. The symbols k and m are used here to stand for specific constants.
Whichever the type may be, a differential equation is said to be of the nth order if it involves a derivative of the nth order but no derivative of an order higher than this.
The equation:
is an example of a partial differential equation of the second order. The theories of ordinary and partial differential equations are markedly different, and for this reason the two categories are treated separately.
Instead of a single differential equation, the object of study may be a simultaneous system of such equations. The formulation of the laws of dynamics frequently leads to such systems. In many cases, a single differential equation of the nth order is advantageously replaceable by a system of n simultaneous equations, each of which is of the first order, so that techniques from linear algebra can be applied.
An ordinary differential equation in which, for example, the function and the independent variable are denoted by y and x is in effect an implicit summary of the essential characteristics of y as a function of x.
These characteristics would presumably be more accessible to analysis if an explicit formula for y could be produced. Such a formula, or at least an equation in x and y (involving no derivatives) that is deducible from the differential equation, is called a solution of the differential equation. The process of deducing a solution from the equation by the applications of algebra and calculus is called solving or integrating the equation.
It should be noted, however, that the differential equations that can be explicitly solved form but a small minority. Thus, most functions must be studied by indirect methods. Even its existence must be proved when there is no possibility of producing it for inspection. In practice, methods from numerical analysis, involving computers, are employed to obtain useful approximate solutions.
Problems in the theory of differential equations.
We now give exact definitions. An ordinary differential equation of order n in one unknown function y is a relation of the form
between the independent variable x and the quantities
The order of a diflerential equation is the order of the highest derivative of the unknown function appearing in the differential equation. Thus the equation in example 1 is of the first order, and those in examples 2, 3, 4, 5, and 6, are of the second order.
A function ϕ(x) is called a solution of the differential equation (17) if substitution of ϕ(x) for y, ϕ′(x) for y′, · · ·, ϕ(n) (x) for y(n) produces an identity.
Problems in physics and technology often lead to a system of ordinary differential equations with several unknown functions, all depending on the same argument and on their derivatives with respect to that argument.
For greater concreteness, the explanations that follow will deal chiefly with one ordinary differential equation of order not higher than the second and with one unknown function. With this example one may explain the essential properties of all ordinary differential equations and of systems of such equations in which the number of unknown functions is equal to the number of equations.
We have spoken earlier of the fact that, as a rule, every differential equation has not one but an infinite set of solutions. Let us illustrate this first of all by intuitive considerations based on the examples given in equations (26). In each of these, the corresponding differential equation is already fully defined by the physical arrangement of the system. But in each of these systems there can be many different motions. For example, it is perfectly clear that the pendulum described by equation (8) may oscillate with many different amplitudes. To each of these different oscillations of the pendulum there corresponds a different solution of equation (8), so that infinitely many such solutions must exist. It may be shown that equation (8) is satisfied by any function of the form
where C1, and C2, are arbitrary constants.
It is also physically clear that the motion of the pendulum will be completely determined only in case we are given, at some instant t0, the (initial) value x0 of x (the initial displacement of the material point from the equilibrium position) and the initial rate of motion:
X’o=(dx/dt)t=0 These initial conditions determine the constants C1, and C2, in formula (18).
In exactly the same way, the differential equations we have found in other examples will have infinitely many solutions.
In general, it can be proved, under very broad assumptions concerning the given differential equation (17) of order n in one unknown function that it has infinitely many solutions. More precisely: If for some “initial value” of the argument, we assign an “initial value” to the unknown function and to all of its derivatives through order n – 1, then one can find a solution of equation (17) which takes on these preassigned initial values. It may also be shown that such initial conditions completely determine the solution, so that there exists only one solution satisfying the initial conditions given earlier. We will discuss this question later in more detail. For our present aims, it is essential to note that the initial values of the function and the first n – 1 derivatives may be given arbitrarily. We have the right to make any choice of n values which define an “initial state” for the desired solution.
If we wish to construct a formula that will if possible include all solutions of a differential equation of order n, then such a formula must contain n independent arbitrary constants, which will allow us to impose n initial conditions. Such solutions of a differential equation of order n, containing n independent arbitrary constants, are usually called general solutions of the equation. For example, a general solution of (8) is given by formula (18) containing two arbitrary constants; a general solution of equation (3) given by formula (5).
We will now try to formulate in very general outline the problems confronting the theory of differential equations. These are many and varied, and we will indicate only the most important ones.
If the differential equation is given together with its initial conditions, then its solution is completely determined. The construction of formulas giving the solution in explicit form is one of the first problems of the theory. Such formulas may be constructed only in simple cases, but if they are found, they are of great help in the computation and investigation of the solution.
The theory should provide a way to obtain some notion of the behavior of a solution: whether it is monotonic or oscillatory, whether it is periodic or approaches a periodic function, and so forth.
Suppose we change the initial values for the unknown function and its derivatives; that is, we change the initial state of the physical system. Then we will also change the solution, since the whole physical process will now run differently. The theory should provide the possibility of judging what this change will be. In particular, for small changes in the initial values will the solution also change by a small amount and will it therefore be stable in this respect, or may it be that small changes in the initial conditions will give rise to large changes in the solution so that the latter will be unstable ?
We must also be able to set up a qualitative, and where possible, quantitative picture of the behavior not only of the separate solutions of an equation, but also of all of the solutions taken together.
In machine construction there often arises the question of making a choice of parameters characterizing an apparatus or machine that will guarantee satisfactory operation. The parameters of an apparatus appear in the form of certain magnitudes in the corresponding differential equation. The theory must help us make clear what will happen to the solutions of the equation (to the working of the apparatus) if we change the differential equation (change the parameters of the apparatus).
Finally, when it is necessary to carry out a computation, we will need to find the solution of an equation numerically. and here the theory will be obliged to provide the engineer and the physicist with the most rapid and economical methods for calculating. the solutions.
Partial differential equations
In mathematics, equation relating a function of several variables to its partial derivatives. A partial derivative of a function of several variables expresses how fast the function changes when one of its variables is changed, the others being held constant (compare ordinary differential equation). The partial derivative of a function is again a function, and, if f(x, y) denotes the original function of the variables x and y, the partial derivative with respect to x—i.e., when only x is allowed to vary—is typically written as f_{x}(x, y) or ∂f/∂x. The operation of finding a partial derivative can be applied to a function that is itself a partial derivative of another function to get what is called a secondorder partial derivative. For example, taking the partial derivative of f_{x}(x, y) with respect to y produces a new function f_{xy}(x, y), or ∂^{2}f/∂y∂x. The order and degree of partial differential equations are defined the same as for ordinary differential equations.
In general, partial differential equations are difficult to solve, but techniques have been developed for simpler classes of equations called linear, and for classes known loosely as “almost” linear, in which all derivatives of an order higher than one occur to the first power and their coefficients involve only the independent variables.
Functions of Several Variables. Geometrical view.
Up to now we have spoken only of functions of one variable, but in practice it is often necessary to deal also with functions depending on two, three, or in general many variables. For example, the area of a rectangle is a function S=xy of its base x and its height y. The volume of a rectangular parallelepiped is a function V=xyz of its three dimensions. The distance between two points A and B is a function:
of the six coordinates of these points. The wellknown formula: pv = nRT expresses the dependence of the volume v of a definite amount of gas on the pressure p and absolute temperature T.
Functions of several variables, like functions of one variable, are in many cases defined only on a certain region of values of the variables themselves. For example, the function
U = ln (1x²y²z²) is defined only for values of x, y and z that satisfy the condition x²+y²+z²=1
(For other x, y, z its values are not real numbers.) The set of points of space whose coordinates satisfy the inequality (35) obviously fills up a sphere of unit radius with its center at the origin of coordinates. The points on the boundary are not included in this sphere; the surface of the sphere has been so to speak “peeled off.” Such a sphere is said to be open. The function (34) is defined only for such sets of three numbers (x, y, z) as are coordinates of points in the open sphere G. It is customary to state this fact concisely by saying that the function (34) is defined on the sphere G.
Let us give another example. The temperature of a nonuniformly heated body V is a function of the coordinates x, y, z of the points of the body. This function is not defined for all sets of three numbers x, y, z but only for such sets as are coordinates of points of the body V.
Finally, as a third example, let us consider the function:where ϕ is a function of one variable defined on the interval [0, 1]. Obviously the function u is defined only for sets of three numbers (x, y, z) which are coordinates of points in the cube: 0≤x≤1,0≤y≤1,0≤z≤1.
We now give a formal definition of a function of three variables. Suppose that we are given a set E of triples of numbers (x, y, z) (points of space). If to each of these triples of numbers (points) of E there corresponds a definite number u in accordance with some law, then u is said to be a function of x, y, z (of the point), defined on the set of triples of numbers (on the points) E, a fact which is written thus: u= F(x,y,z)
In place of F we may also write other letters: f, ϕ, ψ.
In practice the set E will usually be a set of points, filling out some geometrical body or surface: sphere, cube, annulus, and so forth, and then we simply say that the function is defined on this body or surface. Functions of two, four, and so forth, variables are defined analogously.
Implicit definition of a function.
Let us note that functions of two variables is a useful means for the definition of functions of one variable. Given a function F(x, y) of two variables let us set up the equation: F(s,t)=0
In general, this equation will define a certain set of points (s,t) of the surface on which our function is equal to zero. Such sets of points usually represent curves that may be considered as the graphs of one or several onevalued functions y = ϕ(s) or s = ψ(t) of one variable. In such a case these onevalued functions are said to be defined implicitly by the equation (36). For example, the equation:
s²+t²=r²=0 gives an implicit definition of two functions of one variable:
s=+√r²t² and s= – √r²t²
But it is necessary to keep in mind that an equation of the form (36) may fail to define any function at all. For example, the equation: t²+s²+1=0 obviously does not define any real function, since no pair of real numbers satisfies it.
Geometric representation. Functions of two variables may always be visualized as surfaces by means of a system of space coordinates. Thus the function: z=ƒ(s,t)
is represented in a threedimensional rectangular coordinate system by a surface, which is the geometric locus of points M whose coordinates s, t, z satisfy the equation:
There is another, extremely useful method, of representing the function (37), which has found wide application in practice. Let us choose a sequence of numbers z1, z2, ···, and then draw on one and the same plane Ost the curves: z1=ƒ(s,t); z2=ƒ(s,t)
which are the socalled level lines of the function f(s, t). From a set of level lines, if they correspond to values of z that are sufficiently close to one another, it is possible to form a very good image of the variation of the function f(s,t), just as from the level lines of a topographical map one may judge the variation in altitude of the locality.
Figure shows a map of the level lines of the function z = s2 + t2, the diagram at the right indicating how the function is built up from its level lines. In Chapter III, figure 50, a similar map is drawn for the level lines of the function z = st.
Partial derivatives and differential.
Let us make some remarks about the differentiation of the functions of several variables. As an example we take the arbitrary function of two variables: z=ƒ(x,y)
If we fix the value of y, that is if we consider it as not varying, then our function of two variables becomes a function of the one variable x. The derivative of this function with respect to x, if it exists, is called the partial derivative with respect to x and is denoted thus: ∂z/∂x or ƒx/∂x or ƒ’x(x,y)
The last of these three notations indicates clearly that the partial derivative with respect to x is in general a function of x and y. The partial derivative with respect to y is defined similarly.
The general case for space change through any volume.
When we generalise the case to any combination of space or time dimensions, the same method can be used to obtain the ginormous quantity of possible changes in multiple Dimensional analysis.
Thus, in order to determine the function that represents a given physical process, we try first of all to set up an equation that connects this function in some definite way with its derivatives of change of various orders and dimensions.
The method of obtaining such an equation, which is called a differential equation, often amounts to replacing increments of the desired functions by their corresponding differentials.
As an example let us solve a classic problem of change in 3 pure dimensions of euclidean space, which by convention we shall call Sxyz.
In a rectangular system of coordinates Oxyz, then we consider the surface obtained by rotation of the parabola whose equation (in the Oyz plane) is z = y2. This surface is called a paraboloid of revolution). Let v denote the volume of the body bounded by the paraboloid and the plane parallel to the Oxy plane at a distance z from it. It is evident that v is a function of z (z >0).
To determine the function v, we attempt to find its differential dv. The increment Δv of the function v at the point z is equal to the volume bounded by the paraboloid and by two planes parallel to the Oxy plane at distances z and z + Δz from it.
It is easy to see that the magnitude of Δv is greater than the volume of the circular cylinder of radius √z and height Δz but less than that of the circular cylinder with radius √z+∆z and height Δz. Thus:
πz ∆z < ∆v ≤ π (z +∆z) ∆z. And so:
where θ is some number depending on Δz and satisfying the inequality 0 < θ < 1.
So we have succeeded in representing the increment Δv in the form of a sum, the first summand of which is proportional to Δz, while the second is an infinitesimal of higher order than Δz (as Δz → 0). It follows that the first summand is the differential of the function v:
dv=πz ∆z or dv=πz dz
since Δz = dz for the independent variable z. The equation so obtained relates the differentials dv and dz (of the variables v and z) to each other and thus is called a differential equation. If we take into account that:
dv/dz =v’ where v′ is the derivative of v with respect to the variable z, our differential equation may also be written in the form: v’=π z
To solve this very simple differential equation we must find a function of z whose derivative is equal to πz.
A solution of our equation is given by v = πz²/2 + C, where for C we may choose an arbitrary number. In our case the volume of the body is obviously zero for z = 0 (see figure 22), so that C = 0. Thus our function is given by v = πz²/2.
Geometrically the function f(x, y) represents a surface in a rectangular threedimensional system of coordinates. The corresponding function of x for fixed y represents a plane curve (figure) obtained from the intersection of the surface with a plane parallel to the plane Oxz and at a distance y from it. The partial derivative ∂z/∂x is obviously equal to the trigonometric tangent of the angle between the tangent. to the curve at the point (x, y) and the positive direction of the xaxis.
More generally, if we consider a function z = f(x1, x2, . . ., xn) of the n variables x1, x2, . . ., xn, the partial derivative ∂z/∂x, is defined as the derivative of this function with respect to xi, calculated for fixed values of the other variables.
We may say that the partial derivative of a function with respect to the variable xi is the rate of change of this function in the direction of the change in xi. It would also be possible to define a derivative in an arbitrary assigned direction, not necessarily coinciding with any of the coordinate axis, but we will not take the time to do this.
It is sometimes necessary to form the partial derivatives of these partial derivatives; that is; the socalled partial derivatives of second order. For functions of two variables there are four of them: However, if these derivatives are continuous, then it is not hard to prove that the second and third of these four (the socalled mixed derivatives) coincide:
For example, in the case of first function considered:
the two mixed derivatives are seen to coincide.
For functions of several variables, just as was done for functions of one variable, we may introduce the concept of a differential.
For definiteness let us consider a function:
z = ƒ (x,y) of two variables. If it has continuous partial derivatives, we can prove that its increment: corresponding to the increments Δx and Δy of its arguments, may be put in the form:where ∂f/∂x and ∂f/∂y are the partial derivatives of the function at the point (x, y) and the magnitude a depends on Δx and Δy in such a way that α → 0 as Δx → 0 and Δy → 0.
The sum of the first two components:is linearly dependent on Δx and Δy and is called the differential of the function. The third summand, because of the presence of the factor α, tending to zero with Δx and Δy, is an infinitesimal of higher order than the magnitude:describing the change in x and y.
Let us give an application of the concept of differential. The period of oscillation of a pendulum is calculated from the formula:
where l is its length and g is the acceleration of gravity. Let us suppose that l and g are known with errors respectively equal to Δl and Δg. Then the error in the calculation of T will be equal to the increment ΔT corresponding to the increments of the arguments Δl and Δg. Replacing ΔT approximately by dT, we will have:
The signs of Δl and Δg are unknown, but we may obviously estimate ΔT by the inequality:
Thus we may consider in practice that the relative error for T is equal to the sum of the relative errors for l and g.
For symmetry of notation, the increments of the independent variables Δx and Δy are usually denoted by the symbols dx and dy and are also called differentials. With this notation the differential of the function u = f(x, y, z) may be written thus:Partial derivatives play a large role whenever we have to do with functions of several variables, as happens in many of the applications of analysis to technology and physics.
THE GEOMETRIV VIEW: Multiple Integrals.
Integrals in topology.
As the Universe is a kaleidoscopic mirror of symmetries between all its elements, this dominant of analysis on ∆scaling must ad also the use of analysis on a single plane, in fact the most common, whereas the essential consideration is the ∆§ocial decametric and eπ ternary scaling, with minimal distortion (which happens in the Lorentzian limits between scales).
This key distinction on GST (∆§ wellbehaved scaling versus ∆±i ‘distorted emerging and dissolution, which does change the form of the system) does have special relevance in analysis as for very long it was necessary the ‘continuity’ without such distortions of the function studied, and so analysis was restricted to ∆(±1 – 0) intervals and ‘broke’ when jumping two scales as in processes of entropy (deathfeeding). But with improved approximation techniques, functionals and operators (which assume a whole scale of ∞ parts as a function of functions in the operator of the larger scale) and renormalisation in double and triple integrals and derivatives by praxis, without understanding the scalar theory behind it, this hurdle today…
And it has always amused me that humans can get so far in all disciplines by trial and error, when a ‘little bit of thought on first principles’ could make thinks much easier. It seems thoughthought beings are scarce in our species and highly disregarded, as the site’§ight§how (allow me, a bit of cacophony and repetition the trade mark of this blog and the Universe 🙂 As usual I shall also repeat, I welcome comments, and offers of serious help from specialists and Universities, since nothing would make me happier than unloading tons of nowconfusing analysis not only of analysis, before I get another health crisis and all goes to waste in the eternal entropic arrow of two derivatives, aka death.
Existence and Uniqueness of the Solution of a Differential Equation; Approximate Solution of Equations
The question of existence and uniqueness of the solution of a differential equation. We return to the differential equation (17) of arbitrary order n. Generally, it has infinitely many solutions and in order that we may pick from all the possible solutions some one specific one, it is necessary to attach to the equation some supplementary conditions, the number of which should be equal to the order n of the equation. Such conditions may be of extremely varied character, depending on the physical, mechanical, or other significance of the original problem.
For example, if we have to investigate the motion of a mechanical system beginning with some specific initial state, the supplementary conditions will refer to a specific (initial) value of the independent variable and will be called initial conditions of the problem. But if we want to define the curve of a cable in a suspension bridge, or of a loaded beam resting on supports at each end, we encounter conditions corresponding to different values of the independent variable, at the ends of the cable or at the points of support of the beam. We could give many other examples showing the variety of conditions to be fulfilled in connection with differential equations.
We will assume that the supplementary conditions have been defined and that we are required to find a solution of equation:
that satisfies them.
The first question we must consider is whether any such solution exists at all. It often happens that we cannot be sure of this in advance. Assume, say, that equation (17) is a description of the operation of some physical apparatus and suppose we want to determine whether periodic motion occurs in this apparatus. The supplementary conditions will then be conditions for the periodic repetition of the initial state in the apparatus, and we cannot say ahead of time whether or not there will exist a solution which satisfies them.
In any case the investigation of problems of existence and uniqueness of a solution makes clear just which conditions can be fulfilled for a given differential equation and which of these conditions will define the solution in a unique manner.
But the determination of such conditions and the proof of existence and uniqueness of the solution for a differential equation corresponding to some physical problem also has great value for the physical theory itself. It shows that the assumptions adopted in setting up the mathematical description of the physical event are on the one hand mutually consistent and on the other constitute a complete description of the event.
The methods of investigating the existence problem are manifold, but among them an especially important role is played by what are called direct methods. The proof of the existence of the required solution is provided by the construction of approximate solutions, which are proved to converge to the exact solution of the problem. These methods not only establish the existence of an exact solution, but also provide a way, in fact the principal one, of approximating it to any desired degree of accuracy.
For the rest of this section we will consider, for the sake of definiteness, a problem with initial data, for which we will illustrate the ideas of Euler’s method and the method of successive approximations.
Euler’s method of broken lines.
Consider in some domain G of the (x, y) plane the differential equation: dy/dx = ƒ (x,y)
As we have already noted, equation (34) defines in G a field of tangents. We choose any point (x0, y0) of G. Through it there will pass a straight line L0 with slope f(x0, y0). On the straight line L0 we choose a point (x1, y1), sufficiently close to (x0, y0); in figure 9 this point is indicated by the number 1.
We draw the straight line L1, through the point (x1, y1) with slope f(x1, y1) and on it mark the point (x2, y2); in the figure this point is denoted by the number 2. Then on the straight line L2, corresponding to the point (x2, y2) we mark the point (x3, y3), and continue in the same manner with x0, < x1, < x2, < x3, < · · ·. It is assumed, of course, that all the points (x0, y0), (x1, y1), (x2, y2), · · · are in the domain G. The broken line joining these points is called an Euler broken line.
One may also construct an Euler broken line in the direction of decreasing x; the corresponding vertices on our figure are denoted by –1, –2, –3.
It is reasonable to expect that every Euler broken line through the point (x0, y0) with sufficiently short segments gives a representation of an integral curve l passing through the point (x0, y0), and that with decrease in the length of the links, i.e., when the length of the longest link tends to zero, the Euler broken line will approximate this integral curve.
Here, of course, it is assumed that the integral curve exists. In fact it is not hard to prove that if the function f(x, y) is continuous in the domain G, one may find an infinite sequence of Euler broken lines, the length of the largest links tending to zero, which converges to an integral curve l. However, one usually cannot prove uniqueness: there may exist different sequences of Euler broken lines that converge to different integral curves passing through one and the same point (x0, y0). M. A. Lavrent’ev has constructed an example of a differential equation of the form (29) with a continuous function, f(x, y), such that in any neighborhood of any point P of the domain G there passes not one but at least two integral curves.
In order that through every point of the domain G there pass only one integral curve, it is necessary to impose on the function f(x, y) certain conditions beyond that of continuity. It is sufficient, for example, to assume that the function f(x, y) is contitiuous and has a bounded derivative with respect to y on the whole domain G. In this case it may be proved that through each point of G there passes one and only one integral curve and that every sequence of Euler broken lines passing through the point (x0, y0) converges uniformly to this unique integral curve, as the length of the longest link of the broken lines tends to zero. Thus for sufficiently small links the Euler broken line may be taken as an approximation to the integral curve of equation (34).
From the preceding it can be seen that the Euler broken lines are so constituted that small pieces of the integral curves are replaced by line segments tangent to these integral curves. In practice, many approximations to integral curves of the differential equation (34) consist not of straightline segments tangent to the integral curves, but of parabolic segments that have a higher order of tangency with the integral curve. In this way it is possible to find an approximate solution with the same degree of accuracy in a smaller number of steps (with a smaller number of links in the approximating curve).
The method of successive approximations.
We now describe another method of successive approximation, which is as widely used as the method of the Euler broken lines. We assume again that we are required to find a solution y(x) of the differential equation (34) satisfying the initial condition: y (xo) = yo
For the initial approximation to the function y(x), we take an arbitrary function y0(x). For simplicity we will assume that it also satisfies the initial condition, although this is not necessary. We substitute it into the right side f(x, y) of the equation for the unknown function y and construct a first approximation y1, to the solution y from the following requirements:Since there is a known function on the right side of the first of these equations the function y1(x) may be found by integration:
It may be expected that y1(x) will differ from the solution y(x) by less than y0(x) does, since in the construction of y1(x) we made use of the differential equation itself, which should probably introduce a correction into the original approximation. One would also think that if we improve the first approximation y1(x) in the same way, then the second approximation:
will be still closer to the desired solution.
Let us assume that this process of improvement has been continued indefinitely and that we have constructed the sequence of approximations: yo(x), y1(x),…yn(x)….
Will this sequence converge to the solution y(x)?
More detailed investigations show that if f(x, y) is continuous and ƒ’y is bounded in the domain G, the functions yn(x) will in fact converge to the exact solution y(x) at least for all x sufficiently close to x0 and that if we break off the computation after a sufficient number of steps, we will be able to find the solution y(x) to any desired degree of accuracy.
Exactly in the same way as for the integral curves of equation (34), we may also find approximations to integral curves of a system of two or more differential equations of the first order. Essentially the necessary condition here is to be able to solve these equations for the derivatives of the unknown functions. For example, suppose we are given the system:Assuming that the right sides of these equations are continuous and have bounded derivatives with respect to y and z in some domain G in space, it may be shown under these conditions that through each point (x0, y0, z0) of the domain G, in which the right sides of the equations in (37) are defined, there passes one and only one integral curve:
y = Φ (x), z = ψ (x) of the system (37). The functions f1(x, y, z) and f2(x, y, z) give the direction numbers at the point (x, y, z), of the tangent to the integral curve passing through this point. To find the functions ϕ(x) and ψ(x) approximately, we may apply the Euler broken line method or other methods similar to the ones applied to the equation (34).
The process of approximate computation of the solution of ordinary differential equations with initial conditions may be carried out on computing machines. There are electronic machines that work so rapidly that if, for example, the machine is programmed to compute the trajectory of a projectile, this trajectory can be found in a shorter space time than it takes for the projectile to hit its target (cf. Chapter XIV).
The connection between differential equations of various orders and a system of a large number of equations of first order. A system of ordinary differential equations, when solved for the derivative of highest order of each of the unknown functions, may in general be reduced, by the introduction of new unknown functions, to a system of equations of the first order, which is solved for all the derivatives. For example, consider the differential equation: d²y/dx²= ƒ (x, y, dy/dx). We set dy/dx = z. Then equation (38) may be written in the form: dz/dx = ƒ (x, y, z)
Hence, to every solution of equation (38) there corresponds a solution of the system consisting of equations (39) and (40). It is easy to show that to every solution of the system of equations (39) and (40) there corresponds a solution of equation (38).
Equations not explicitly containing the independent variable. The problems of the pendulum, of the Helmholtz acoustic resonator, of a simple electric circuit, or of an electrontube generator considered in §1 lead to differential equations in which the independent variable (time) does not explicitly appear. We mention equations of this type here, because the corresponding differential equations of the second order may be reduced in each case to a single differential equation of the first order rather than to a system of firstorder equations as in the paragraph above for the general equation of the second order. This reduction greatly simplifies their study.
Let us then consider a differential equation of the second order, not containing the argument t in explicit form:F (x, dx,dt, d²x.dt²)=0. We set dx/dt=y and consider y as a function of x, so that:
Then equation (41) may be rewritten in the form: F (x, y, y dy/dx)=0
In this manner, to every solution of equation (41) there corresponds a unique solution of equation (43). Also to each of the solutions y = ϕ(x) of equation (43) there correspond infinitely many solutions of equation (41). These solutions may be found by integrating the equation: dx/dt=Φ (x) where x is considered as a function of t.
It is clear that if this equation is satisfied by a function x = x(t), then it will also be satisfied by any function of the form x(t + t0), where t0 is an arbitrary constant.
It may happen that not every integral curve of equation (43) is the graph of a single function of x. This will happen, for example, if the curve is closed. In this case the integral curve of equation (43) must be split up into a number of pieces, each of which is the graph of a function of x. For every one of these pieces, we have to find an integral of equation (44).
The values of x and dx/dt which at each instant characterize the state of the physical system corresponding to equation (41) are called the phases of the system, and the (x, y) plane is correspondingly called the phase plane for equation (41). To every solution x = x(t) of this equation there corresponds the curve: y = x'(t)
in the (x, y) plane; t here is considered as a parameter. Conversely, to every integral curve y = ϕ(x) of equation (43) in the (x, y) plane there corresponds an infinite set of solutions of the form x = x(t + t0) for equation (41); here t0 is an arbitrary constant. Information about the behavior of the integral curves of equation (43) in the plane is easily transformed into information about the character of the possible solutions of equation (41). Every closed integral curve of equation (43) corresponds, for example, to a periodic solution of equation (41).
If we subject equation (6) to the transformation (42), we obtain: dy/dx = ay – bx/my.
Setting ν = x and dv/dt = y in equation (16), in like manner we get:
Just as the state at every instant of the physical system corresponding to the secondorder equation (41) is characterized by the two magnitudes* (phases) x and y = dx/dt, the state of a physical system described by equations of higher order or by a system of differential equations is characterized by a larger number of magnitudes (phases). Instead of a phase plane, we then speak of a phase space.
DUALITIES: The behavior of integral curves in the large DOMAIN selfcentred in the small singularity.
the behavior of the integral curves “in the large”; that is, in the entire domain of the given system of differential equations, without attempting to preserve the scale. We will consider a space in which this system defines a field of directions as the phase space of some physical process. Then the general scheme of the integral curves, corresponding to the system of differential equations, will give us an idea of the character of all processes (motions) which can possibly occur in this system:
In figures we have constructed approximate schematized representations of the behavior of the integral curves in the neighborhood of an isolated singular point.
Why those matter obviously because singularities @ matter. We can divide those curves which are canonical of extensive families that exhaust the 3 possibilities:
∑=∏: 3D communication. What first calls attention is the symmetry of the upper fig. 12, when the singularity merely acts as in a tetraktys configuration as the identity neutral element that communicates all the flows that touch the T.œ system, entering and leaving symmetrically the opoint (having hence a 0 line of symmetry diagonal to the point).
It is also noticeable that the paths are ‘fast’, as the points of those paths know they will not be changed by the identity element.
ð•: 1D predation. But in the case the 0point acts as a predator that won’t let the pointprey go, the form is a spiralled, slow motion.
$: 2D flows. Finally as usual we have a ternary case in which the curves do NOT touch the singularity, which curiously start with the points going straight, perpendicular to it, hence this case tends to apply to spatial points of vital energy with a certain ‘discerning’ view, which makes them feed first on the field established by the singularity to escape it when being aware of what lies ahead. The 2 last cases can be compared in vital terms – not trajectories, the behaviour of smallish ‘blind comets’ spiralling into stars that will feed on them as opposed to symbiotic planets that herd gravitational quanta together with the star but will NOT fall in the gravitational trap.
Mathematically the drawing of those curves, is one of the most fundamental problems in the theory of differential equations: finding as simple a method as possible for constructing such a scheme for the behavior of the family of integral curves of a given system of differential equations in the entire domain of definition, in order to study the behavior of the integral curves of this system of differential equations “in the large.”
And since we exist in a bidimensional Universe, this problem remains almost untouched for spaces of dimension higher than 2 (a recurrent fact of all mathematical mirrors from Fermat’s last theorem to the proof of almost all geometrical theorems in a plane).
But the problem is still very far from being solved for the single equation of the form: dy/dx = M (x,y)/N (x.y) even when M(x, y) and N(x, y) are polynomials, which shows how so many times the whys of ∆st are truly synoptic and simple, even if the detailed paths of 1D motions, the obsession of onedimensional humans are ignored.
In fact the only solution quite resolved is… yes you guess it, that in which the particle has no ‘freedom of thought’ so to speak and falls down the spiral path of entropic death and informative capture by the singularity.
THIS WILL again be a rule of ∆st, the simplest solutions are those related with death, dissolution, entropy and onedimensional ‘fall’.
In what follows, we will assume that the functions M(x, y) and N(x, y) have continuous partial derivatives of the first order.
If all the points of a simply connected domain G, in which the right side of the differential equation is defined, are ordinary points, then the family of integral curves may be represented schematically as a family of segments of parallel straight lines; since in this case one integral curve will pass through each point, and no two integral curves can intersect. For an equation of more general form, which may have singular points, the structure of the integral curves may be much more complicated. The case in which the previous equation has an infinite set of singular points (i.e., points where the numerator and the denominator both vanish) may be excluded, at least when M(x, y) and N(x, y) are polynomials.
Thus we restrict our consideration to those cases in which the previous equation has a finite number of isolated singular points. The behavior of the integral curves that are near to one of these singular points forms the essential element in setting up a schematized representation of the behavior of all the integral curves of the equation.
A very typical element in such a scheme for the behavior of all the integral curves of the previous equation is formed by the socalled limit cycles. Let us consider the equation 64: dρ/dΦ = ρ1 where ρ and ϕ are polar coordinates in the (x, y) plane.
The collection of all integral curves of the equation is given by the formula (65):where C is an arbitrary constant, different for different integral curves. In order that ρ be nonnegative, it is necessary that ϕ have values no larger than – In  C , C < 0. The family of integral curves will consist of
1. the circle ρ = 1 (C = 0);
2. the spirals issuing from the origin, which approach this circle from the inside as ϕ → – ∞(C < 0);
3. the spirals, which approach the circle ρ = 1 from the outside as ϕ → – ∞ (C > 0)
The circle ρ = 1 is called a limit cycle for its equation (65). In general a closed integral curve l is called a limit cycle, if it can be enclosed ina disc all points of which are ordinary for equation (64) and which is entirely filled by nonclosed integral curves.
From equation (65) it can be seen that all points of the circle are ordinary. This means that a small piece of a limit cycle is not different from a small piece of any other integral curve.
Every closed integral curve in the (x, y) plane gives a periodic solution [x(t), y(t)] of the system:
dx/dt =N (x.y), dy/dt=M (x,y) describing the law of change of some physical system. Those integral curves in the phase plane that as t → + ∞ approximate a limit cycle are motions that as t → ∞ approximate periodic motions.
Let us suppose that for every point (x0, y0) sufficiently close to a limit cycle l, we have the following situation: If (x0, y0) is taken as initial point (i.e., for t = t0) for the solution of the system (67), then the corresponding integral curve traced out by the point [x(t), y(t)], as t → + ∞ approximates the limit cycle l in the (x, y) plane. (This means that the motion in question is approximately periodic.) In this case the corresponding limit cycle is called stable. Oscillations that act in this way with respect to a limit cycle correspond physically to selfoscillations. In some selfoscillatory systems, there may exist several stable oscillatory processes with different amplitudes, one or another of which will be established by the initial conditions. In the phase plane for such “selfoscillatory systems,” there will exist corresponding limit cycles if the processes occuring in these systems are described by an equation of the form (67).
The problem of finding, even if only approximately, the limit cycles of a given differential equation has not yet been satisfactorily solved. The most widely used method for solving this problem is the one suggested by Poincaré of constructing “cycles without contact.” It is based on the following theorem. We assume that on the (x, y) plane we can find two closed curves L1 and L2 (cycles) which have the following properties:
1. The curve L2 lies in the region enclosed by L1.
2. In the annulus Ω, between L1 and L2, there are no singular points of equation (64).
3. L1 and L2 have tangents everywhere, and the directions of these tangents are nowhere idertical with the direction of the field of directions for the given equation (64).
4. For all points of L1 and L2 the cosine of the angle between the interior normals to the boundary of the domain Ω and the vector with components [N(x, y), M(x, y)] never changes sign.
Then between L1 and L2 there is at least one limit cycle of equation (64).
Poincaré called the curves L1 and L2 cycles without contact.
The proof of this theorem is based on the following rather obvious fact.
We assume that for decreasing t (or for increasing t) all the integral curves: x = x(t), y = y (t), of equation (64) (or, what amounts to the same thing, of equations (67), where t is a parameter), which intersect L1 or L2 enter the annulus Ω between L1 and L2. Then they must necessarily tend to some closed curve l lying between L1 and L2, since none of the integral curves lying in the annulus can leave it, and there are no singular points there.
Singular Points.
Now when considering the singular points in relationship to the vital energy mapped out in its cyclical trajectories by those curves, we observe there are 3 cases, the absorption, the crossing and the isolated point, which in abstract math are studied as follows.
Let the point P(x, y) be in the interior of the domain G in which we consider the differential equation: dy/dx = M (x,y)/N (x.y).
If there exists a neighborhood R of the point P through each point of which passes one and only one integral curve (47), then the point P is called an ordinary point of equation (47). But if such a neighborhood does not exist, then the point P is called a singular point of this equation. The study of singular points is very important in the qualitative theory of differential equations, which we will consider in the next section.
Particularly important are the socalled isolated singular points, i.e., singular points in some neighborhood of each of which there are no other singular points. In applications one often encounters them in investigating equations of the form(47), where M(x, y) and N(x, y) are functions with continuous derivatives of high orders with respect to x and y. For such equations, all the interior points of the domain at which M(x, y) ≠ 0 or N(x, y) ≠ 0 are ordinary points.
Let us now consider any interior point (x0, y0) where M(x, y) = N(x, y) = 0. To simplify the notation we will assume that x0 = 0 and y0 = 0. This can always be arranged by translating the original origin of coordinates to the point (x0, y0). Expanding M(x, y) and N(x, y) by Taylor’s formula into powers of x and y and restricting ourselves to terms of the first order, we have, in a neighborhood of the point (0, 0):Equations (45) and (46) are of this form. Equation (45) does not define either dy/dx or dx/dy for x = 0 and y = 0. If the determinant:then, whatever value we assign to dy/dx at the origin, the origin will be a point of discontinuity for the values dy/dx and dx/dy, since they tend to different limits depending on the manner of approach to the origin. The origin is a singular point for our differential equation.
It has been shown that the character of the behavior of the integral curves near an isolated singular point (here the origin) is not influenced by the behavior of the terms ϕ1(x, y) and ϕ2(x, y) in the numerator and denominator, provided only that the real part of both roots of the equation:is different from zero. Thus, in order to form some idea of this behavior, we study the behavior near the origin of the integral curves of the equation:We note that the arrangement of the integral curves in the neighborhood of a singular point of a differential equation has great interest for many problems of mechanics, for example in the investigation of the trajectories of motions near the equilibrium position.
It has been shown that everywhere in the plane it is possible to choose coordinates ξ, η, connected with x, y by the equations:
where the kij are real numbers such that equation (50) is tranformed into one of the the following three types:
If these roots are real and different, then equation (50) is transformed into the form (52). If these roots are equal, then equation (50) is transformed either into the form (52) or into the form (53), depending on whether a2 + d2 = 0 or a2 + d2 ≠ 0. If the roots of equation (55) are complex, λ = α ± βi, then equation (51) is transformed into the form (54).
We will consider each of the equations (52), (53), (54). To begin with, we note the following.
Even though the axes Ox and Oy were mutually perpendicular, the axes Oξ and Oη need not, in general, be so. But to simplify the diagrams, we will assume they are perpendicular. Further, in the transformation (51) the scales on the Oξ and Oη axes may be changed; they may not be the same as the ones originally chosen on the axes Ox and Oy. But again, for the sake of simplicity, we assume that the scales are not changed. Thus, for example, in place of the concentric circles, as in figure 8, there could in general occur a family of similar and similarly placed ellipses with common center at the origin.
All integral curves of equation 1 are given by:
where a and b are arbitrary constants.
The integral curves of equation (52) are graphed in figure 10; here we we have assumed that k > 1. In this case all integral curves except one, the axis Oη, are tangent at the origin to the axis Oξ. The case 0 < k < 1 is the same as the case k > 1 with interchange of ξ and η, i.e., we have only to interchange the roles of the axes ξ and η. For k = 1, equation (52) becomes equation (30). whose integral curves were illustrated in figure 7.
An illustration of the integral curves of equation (52) for k < 0 is given in figure 11. In this case we have only two integral curves that pass through the point O: these are the axis Oξ and the axis Oη. All other integral curves, after approaching the origin no closer than to some minimal distance, recede again from the origin. In this case we say that the point O is a saddle point because the integral curves are similar to the contours on a map representing the summit of a mountain pass (saddle).
All integral curves of equation (53) are given by the equation:where a and b are arbitrary constants. These are illustrated schematically in figure 12; all of them are tangent to the axis Oη at the origin.
If every integral curve entering some neighborhood of the singular point O passes through this point and has a definite direction there, i.e., has a definite tangent at the origin, as is illustrated in figures 10 and 12, then we say that the point O is a node.
Equation (54) is most easily integrated, if we change to polar coordinates ρ and ϕ, putting:
If k > 0 then all the integral curves approach the point O, winding infinitely often around this point as ϕ → – ∞ (figure 13). If k < 0,
then this happens for ϕ → + ∞. In these cases, the point O is called a focus. If, however, k = 0, then the collection of integral curves of (56) consists of curves with center at the point O. Generally, if some neighborhood of the point O is completely filled by closed integral curves, surrounding the point O itself, then such a point is called a center.
A center may easily be transformed into a focus, if in the numerator and the denominator of the right side of equation (54) we add a term of arbitrarily high order; consequently, in this case the behavior of integral curves near a singular point is not given by terms of the first order.
Equation (55), corresponding to equation (45), is identical with the characteristic equation (19). Thus figures 10 and 12 schematically represent the behavior in the phase plane (x, y) of the curves:
x=x(t), y = x'(t) corresponding to the solutions of equation (6) for real λ1, and λ2, of the same sign; Figure 11 corresponds to real λ1, and λ2, of opposite signs, and figures 13 and 8 (the case of a center) correspond to complex λ1, and λ2. If the real parts of λ1, and λ2, are negative, then the point (x(t), y(t)) approaches 0 for t → + ∞; in this case the point x = 0, y = 0 corresponds to stable equilibrium. If, however, the real part of either of the numbers λ1, and λ2, is positive, then at the point x = 0, y = 0, there is no stable equilibrium.
There are not many differential equations with the property that all their solutions can be expressed explicitly in terms of simple functions, as is the case for linear equations with constant coefficients. It is possible to give simple examples of differential equations whose general solution cannot be expressed by a finite number of integral of known functions, or as one says, in quadratures.
An equation of the form dy/dx + ay² = x², for a > 0, cannot be expressed as a finite combination of integrals of elementary functions.
So it becomes important to develop methods of approximation to the solutions of differential equations, which will be applicable to wide classes of equations.
The fact that in such cases we find not exact solutions but only approximations should not bother us. First of all, these approximate solutions may be calculated, at least in principle, to any desired degree of accuracy. Second, it must be emphasized that in most cases the differential equations describing a physical process are themselves not altogether exact, as can be seen in all the examples discussed in §1.
An especially good example is provided by the equation (12) for the acoustic resonator. In deriving this equation, we ignored the compressibility of the air in the neck of the container and the motion of the air in the container itself. As a matter of fact, the motion of the air in the neck sets into motion the mass of the air in the vessel, but these two motions have different velocities and displacements. In the neck the displacement of the particles of air is considerably greater than in the container. Thus we ignored the motion of the air in the container, and took account only of its compression. For the air in the neck, however, we ignored the energy of its compression and took account only of the kinetic energy of its motion.
To derive the differential equation for a physical pendulum, we ignored the mass of the string on which it hangs. To derive equation (14) for electric oscillations in a circuit, we ignored the selfinductance of the wiring and the resistance of the coils. In general, to obtain a differential equation for any physical process, we must always ignore certain factors and idealize others.
For physical investigations we are especially interested in those differential equations whose solutions do not change much for arbitrary small changes, in some sense or another, in the equations themselves. Such differential equations are called “intensitive.” These equations deserve particularly complete study.
It should be stated that in physical investigations not only are the differential equations that describe the laws of change of the physical quantities themselves inexactly defined but even the number of these quantities is defined only approximately. Strictly speaking, there are no such things as rigid bodies. So to study the oscillations of a pendulum, we ought to take into account the deformation of the string from which it hangs and the deformation of the rigid body itself, which we approximated by taking it as a material point. In exactly the same way, to study the oscillations of a load attached to springs, we ought to consider the masses of the separate coils of the springs.
But in these examples it is easy to show that the character of the motion of the different particles, which make up the pendulum and its load together with the springs, has little influence on the character of the oscillation. If we wished to take this influence into account, the problem would become so complicated that we would be unable to solve it to any suitable approximation. Our solution would then bear no closer relation to physical reality than the solution given in §1 without consideration of these influences. Intelligent idealisation of a problem is always unavoidable.
To describe a process, it is necessary to take into account the essential features of the process but by no means to consider every feature without exception. This would not only complicate the problem a great deal but in most cases would result in the impossibility of calculating a solution.
The fundamental problem of physics or mechanics, in the investigation of any phenomenon, is to find the smallest number of quantities, which with sufficient exactness describe the state of the phenomenon at any given moment, and then to set up the simplest differential equations that are good descriptions of the laws governing the changes in these quantities. This problem is often very difficult. Which features are the essential ones and which are nonessential is a question that in the final analysis can be decided only by long experience. Only by comparing the answers provided by an idealized argument with the results of experiment can we judge whether the idealization was a valid one.
The mathematical problem of the possibility of decreasing the number of quantities may be formulated in one of the simplest and most characteristic cases, as follows.
Suppose that to begin with we characterize the state of a physical system at time t by the two magnitudes x1(t) and x2(t). Let the differential equations expressing their rates of change have the form: In the second equation the coefficient of the derivative is a small constant parameter ε . If we put ε= 0, the second equation will cease to be a differential equation. It then takes the form: ƒ2(t, x1,x2)=0
From this equation, we define x2, as a function of t and x1, and we substitute it into the first equation. We then have the differential equation: dx1/dt = F (t, x1) for the single variable x1. In this way the number of parameters entering into the situation is reduced to one. We now ask, under what conditions will the error introduced by taking ε = 0 be small. Of course, it may happen that as ε → 0 the value dx2/dt grows beyond all bounds, so that the right side of the second of equations (28) does not tend to zero as ε→ 0.
Generalized Solutions
The range of problems in which a physical process is described by continuous, differentiable functions satisfying differential equations may be extended in an essential way by introducing into the discussion discontinuous solutions of these equations.
In a number of cases it is clear from the beginning that the problem under consideration cannot have solutions that are twice continuously differentiable; in other words, from the point of view of the classical statement of the problem given in the preceding section, such a problem has no solution. Nevertheless the corresponding physical process does occur, although we cannot find functions describing it in the preassigned class of twicedifferentiable functions. Let us consider some simple examples.
 If a string consists of two pieces of different density, then in the equation:
the coefficient will be equal to a different constant on each of the corresponding pieces, and so equation (24) will not, in general, have classical (twice continuously differentiable) solutions.
 Let the coefficient a be a constant, but in the initial position let the string have the form of a broken line given by the equation ui = 0 = ϕ(x). At the vertex of the broken line, the function ϕ(x) obviously cannot have a first derivative. It may be shown that there exists no classical solution of equation (24) satisfying the initial conditions:
(here and in what follows ut denotes ∂u/∂t).
 If a sharp blow is given to any small piece of the string, the resulting oscillations are described by the equation:
where f(x, t) corresponds to the effect produced and is a discontinuous function, differing from zero only on the small piece of the string and during a short interval of time. Such an equation also, as can be easily established, cannot have classical solutions.
These examples show that requiring continuous derivatives for the desired solution strongly restricts the range of the problems we can solve. The search for a wider range of solvable problems proceeded first of all in the direction of allowing discontinuities of the first kind in the derivatives of highest order, for the functions serving as solutions to the problems, where these functions must satisfy the equations except at the points of discontinuity. It turns out that the solutions of an equation of the type Δu = 0 or ∂u/∂t − Δu = 0 cannot have such (socalled weak) discontinuities inside the domain of definition.
Solutions of the wave equation can have weak discontinuities in the space variables x, y, z, and in t only on surfaces of a special form, which are called characteristic surfaces. If a solution u(x, y, z, t) of the wave equation is considered as a function defining, for t = t1, a scalar field in the x, y, z space at the instant t1, then the surfaces of discontinuity for the second derivatives of u(x, y, z, t) will travel through the (x, y, z) space with a velocity equal to the square root of the coefficient of the Laplacian in the wave equation.
The second example for the string shows that it is also necessary to consider solutions in which there may be discontinuous first derivatives; and in the case of sound and light waves, we must even consider solutions that themselves have discontinuities.
The first question that comes up in investigating the introduction of discontinuous solutions consists in making clear exactly which discontinuous functions can be considered as physically admissible solutions of an equation or of the corresponding physical problem. We might, for example, assume that an arbitrary piecewise constant function is “a single solution” of the Laplace equation or the wave equation, since it satisfies the equation outside of the lines of discontinuity.
In order to clarify this question, the first thing that must be guaranteed is that in the wider class of functions, to which the admissible solutions must belong, we must have a uniqueness theorem. It is perfectly clear that if, for example, we allow arbitrary piecewise smooth functions, then this requirement will not be satisfied.
Historically, the first principle for selection of admissible functions was that they should be the limits (in some sense or other) of classical solutions of the same equation. Thus, in example 2, a solution of equation (24) corresponding to the function ϕ(x), which does not have a derivative at an angular point may be found as the uniform limit of classical solutions un(x, t) of the same equation corresponding to the initial conditions un  t = 0 = ϕn(x), unt  t = 0 = 0 where the ϕ(x) are twice continuously differentiable functions converging uniformly to ϕ(x) for n → ∞.
In what follows, instead of this principle we will adopt the following: An admissible solution u must satisfy, instead of the equation Lu = f, an integral identity containing an arbitrary function Ф.
This identity is found as follows: We multiply both sides of the equation Lu = f by an arbitrary function Ф, which has continuous derivatives with respect to all its arguments of orders up through the order of the equation and vanishes outside of the finite domain D in which the equation is defined. The equation thus found is integrated over D and then transformed by integration by parts so that it does not contain any derivatives of u. As a result we get the identity desired. For equation (24), for example, it has the form:
For equations with constant coefficients these two principles for the selection of admissible (or as they are now usually called, generalized) solutions, are equivalent to each other. But for equations with variable coefficients, the first principle may turn out to be inapplicable, since these equations may in general have no classical solutions (cf. example 1). The second of these principles provides the possibility of selecting generalized solutions with very broad assumptions on the differentiability properties of the coefficients of the equations. It is true that this principle seems at first sight to be overly formal and to have a purely mathematical character, which does not directly indicate how the problems ought to be formulated in a manner similar to the classical problems.
In order that a larger number of problems may be solvable, we must seek the solutions among functions belonging to the widest possible class of functions for which uniqueness theorems still hold. Frequently such a class is dictated by the physical nature of the problem. Thus, in quantum mechanics it is not the state function ψ(x), defined as a solution of the Schrödinger equation, that has physical meaning but rather the integral av = ∫E ψ(x) ψv(x)dx, where the ψv are certain functions for which:. Thus the solution ψ is to be sought not among the twice continuously differentiable functions but among the ones with integrable square. In the problems of quantum electrodynamics, it is still an open question which classes of functions are the ones in which we ought to seek solutions for the equations considered in that theory.
Progress in mathematical physics during the last thirty years has been closely connected with this new formulation of the problems and with the creation of the mathematical apparatus necessary for their solution.
Particularly convenient methods of finding generalized solutions in one or another of these classes of functions are: the method of finite differences, the direct methods in the calculus of variations and functionaloperator methods. These latter methods basically depend on a study of transformations generated by these problems. Here we will explain the basic ideas of the direct methods of the calculus of variations.
Let us consider the problem of defining the position of a uniformly stretched membrane with fixed boundary. From the principle of minimum potential energy in a state of stable equilibrium the function u(x, y) must give the least value of the integral:
in comparison with all other continuously differentiable functions υ(x, y) satisfying the same condition on the boundary, u S = ϕ, as the function u does. With some restrictions on ϕ and on the boundary S it can be shown that such a minimum exists and is attained by a harmonic function, so that the desired function u IS a solution of the Dirichlet problem Δu = 0, uS = ϕ. The converse is also true: The solution of the Dirichlet problem gives a minimum to the integral J with respect to all υ satisfying the boundary condition.
The proof of the existence of the function u, for which J attains its minimum, and its computation to any desired degree of accuracy may be carried out, for example, in the following manner (Ritz method). We choose an infinite family of twice continuously differentiable functions {υn(x, y)}, n = 0, 1, 2, …, equal to zero on the boundary for n > 0 and equal to ϕ for n = 0. We consider J for functions of the form:
where n is fixed and the Ck are arbitrary numbers. Then J(υ) will be a polynomial of second degree in the n independent variables C1, C2, …, Cn. We determine the Ck from the condition that this polynomial should assume its minimum. This leads to a system of n linear algebraic equations in n unknowns, the determinant of which is different from zero. Thus the numbers Ck are uniquely defined. We denote the corresponding υ by vn(x, y). It can be shown that if the system {υn) satisfies a certain condition of “completeness” the functions υn will converge, as n → ∞, to a function which will be the desired solution of the problem.
In conclusion, we note that in this chapter we have given a description of only the simplest linear problem of mechanics and have ignored many further questions, still far from completely worked out, which are connected with more general partial differential equations.
Methods of Constructing Solutions
On the possibility of decomposing any solution into simpler solutions. Solutions of the problems of mathematical physics formulated previously may be derived by various devices, which are different specific problems. But at the basis of these methods there is one general idea. As we have seen, all the equations of mathematical physics are, for small values of the unknown functions, linear with respect to the functions and their derivatives. The boundary conditions and initial conditions are also linear.
If we form the difference between any two solutions of the same equation, this difference will also be a solution of the equation with the righthand terms equal to zero. Such an equation is called the corresponding homogeneous equation. For example, for the Poisson equation Δu = − 4πρ, the corresponding homogeneous equation is the Laplace equation Δu = 0.
If two solutions of the same equation also satisfy the same boundary conditions, then their difference will satisfy the corresponding homogeneous condition: The values of the corresponding expression on the boundary will be equal to zero.
Hence the entire manifold of the solutions of such an equation, for given boundary conditions, may be found by taking any particular solution that satisfies the given nonhomogeneous condition together with all possible solutions of the homogeneous equation satisfying homogeneous boundary conditions (but not, in general, satisfying the initial conditions).
Solutions of homogeneous equations, satisfying homogeneous boundary conditions may be added, or multiplied by constants, without ceasing to be solutions.
If a solution of a homogeneous equation with homogeneous conditions is a function of some parameter, then integrating with respect to this parameter will also give us such a solution. These facts form the basis of the most important method of solving linear problems of all kinds for the equations of mathematical physics, the method of superposition.
The solution of the problem is sought in the form:
where u, is a particular solution of the equation satisfying the boundary conditions but not satisfying the initial conditions, and the u, are solutions of the corresponding homogeneous equation satisfying the corresponding homogeneous boundary conditions. If the equation and the boundary conditions were originally homogeneous, then the solution of the problem may be sought in the form: U = ∑ Uk.
In order to be able to satisfy arbitrary initial conditions by the choice of particular solutions uk of the homogeneous equation, we must have available a sufficiently large arsenal of such solutions.
The method of separation of variables.
For the construction of the necessary arsenal of solutions there exists a method called separation of variables or Fourier’s method.
Let us examine this method, for example, for solving the problem:
In looking for any particular solution of the equation, we first of all assume that the desired function u and satisfies the boundary condition u  S = 0 and can be expressed as the product of two functions, one of which depends only on the time t and the other only on the space variables: u (x,y,z,t) = U (x,y,z) T (t). Substituting this assumed solution into our equation, we have: T (t) ∆ U = T” (t) U.
Dividing both sides by TU gives: T”/T = ∆U/U.
The right side of this equation is a function of the space variables only and the left is independent of the space coordinates. Hence it follows that the given equation can be true only if the left and right sides have the same constant value. We are led to a system of two equations:
The constant quantity on the right is denoted here by:in order to emphasize that it is negative (as may be rigorously proved). The subscript k is used here to note that there exist infinitely many possible values of:Where the solutions corresponding to them form a system of functions complete in a wellknown sense.
Crossmultiplying in both equations, we get:
The first of these equations has, as we know, the simple solution:
where Ak and Bk are arbitrary constants. This solution may be further simplified by introducing the auxiliary angle ϕ. We have:
Then:
The function T represents a harmonic oscillation with frequency λk, shifted in phase by the angle ϕk.
More difficult and more interesting is the problem of finding a solution of the equation:
for given homogeneous boundary conditions; for example, for the conditions: Us=0
(where S is the boundary of the volume Ω under consideration), or for any other homogeneous condition. The solution of this problem is not always easy to construct as a finite combination of known functions, although it always exists and can be found to any desired degree of accuracy.
The equation: for the condition U  S = 0 has first of all the obvious solution U ≡ 0. This solution is trivial and completely useless for our purposes. If the λk are any randomly chosen numbers, then in general there will not be any other solution to our problem. However, there usually exist values of λk for which the equation does have a nontrivial solution.
All possible values of the constant: are determined by the requirement that equation (19) have a nontrivial solution, i.e., distinct from the identically vanishing function, which satisfies the condition U  S = 0. From this it also follows that the numbers denoted by:must be negative.
For each of the possible values of λk in equation (19), we can find at least one function Uk. This allows us to construct a particular solution of the wave equation (18) in the form:
Such a solution is called a characteristic oscillation (or eigenvibration) of the volume under consideration. The constant λk is the frequency of the characteristic oscillation, and the function Uk(x, y, z) gives us its form. This function is usually called an eigenfunction (characteristic function). For all instants of time, the function uk, considered as a function of the variables x, y, and z, will differ from the function Uk(x, y, z) only in scale.
We do not have space here for a detailed proof of the many remarkable properties of characteristic oscillations and of eigenfunctions; therefore we will restrict ourselves merely to listing some of them.
The first property of the characteristic oscillations consists of the fact that for any given volume there exists a countable set of characteristic frequencies. These frequencies tend to infinity with increasing k.
Another property of the characteristic oscillations is called orthogonality. It consists of the fact that the integral over the domain Ω of the product of eigenfunctions corresponding to different values of λk, is equal to zero: For j = k we will assume:
This can always be arranged by multiplying the functions Uk(x, y, z) by an appropriate constant, the choice of which does not change the fact that the function satisfies equation (19) and the condition U  S = 0.
Finally, a third property of the characteristic oscillations consists of the fact that, if we do not omit any value of λk, then by means of the eigenfunctions Uk(x, y, z), we can represent with any desired degree of exactness a completely arbitrary function f(x, y, z), provided only that it satisfies the boundary condition f  S = 0 and has continuous first and second derivatives. Any such function f(x, y, z) may be represented by the convergent series:
The third property of the eigenfunctions provides us in principle with the possibility of representing any function f(x, y, z) in a series of eigenfunctions of our problem, and from the second property we can find all the coefficients of this series. In fact, if we multiply both sides of equation (20) by Uj(x, y, z) and integrate over the domain Ω, we get:
In the sum on the right, all the terms in which k ≡ j disappear because of the orthogonality, and the coefficient of Cj is equal to one. Consequently we have:
These properties of the characteristic oscillations now allow us to solve the general problem of oscillation for any initial conditions.
For this we assume that we have a solution of the problem in the form:
and try to choose the constants Ak and Bk so that we have:
Putting t = 0 in the right side of (21), we see that the sine terms disappear and cos λkt becomes equal to one, so that we will have:
From the third property, the characteristic oscillations can be used for such a representation, and from the second property, we have: