Submitted to Nature, 17 November 2002, and rejected the following day without refereeing with the following email: Dear xxx (sic!), The Editor thanks you for your communication but regrets that he is unable to publish it. He regrets also that he cannot enter into further correspondence on this matter. Yours sincerely, Nature Administration; but published in: Journal of Mathematics and Design (Buenos Aires) 2 (2002) 63-73 (Editor: Prof. Dr. Vera W. De Spinadel), ISSN 1515-7881

The aesthetic appeal of the golden mean seems to be a consequence of the kind of information coding by the brain. Since decades psychologists have claimed memory span to be the missing link between psychometric intelligence and cognition. By applying Bose-Einstein-statistics to learning experiments, Pascual-Leone obtained a fit between predicted and tested span. Multiplying span by mental speed (bits processed per unit time) and using the entropy formula for bosons, we obtain the same result. If we understand span as the quantum number n of a harmonic oscillator, we get this result from the EEG. The metric of brain waves can always be understood as a superposition of n harmonics times 2 F, where half of the fundamental is the golden mean F (1.618 ) as the point of resonance. Such wave packets scaled in powers of the golden mean have to be understood as numbers with directions, where bifurcations occur at the edge of chaos, i.e. 2 F.

It bothers me that, according to the laws as we understand them today, it takes ... an infinite number of logical operations to figure out what goes on in no matter how tiny a region of space, and no matter how tiny a region of time. How can all that be going on in that tiny space? Why should it take an infinite amount of logic to figure out what a tiny piece of space-time is going to do? So I have often made the hypothesis that ultimately physics will not require a mathematical statement, that in the end the machinery will be revealed and the laws will turn out to be simple, like the checker board with all its apparent complexities, wrote FEYNMAN in 1965 (cited from ILACHINSKI 2002. p. 638). WOLFRAM (2002), too, believes that there are quite simple mechanisms that underlie human reasoning. He asserts that the use of memory is what in fact underlies almost every aspect of human thinking. Capabilities like generalization, analogy and intuition immediately seem very closely related to the ability to retrieve data from memory on the basis of similarity.

Already in 1966, KAC had put
forward the question: Can one hear the shape of a drum? In order to find an
answer, Kac asks for the energy in the frequency interval df. To this
end, he calculates the number of harmonics which lie between the frequencies f
und df and multiplies this number by the energy which belongs to the
frequency f, and which according to the theory of quantum mechanics is the same for all frequencies. By solving the eigenvalue problem of the wave equation, Kac is able to state that one can not only hear the area of a reflecting surface, its volume and circumference, but also the connectivity of paths of an irregular shaped network. If the brain waves had the possibility to measure and hence to know the eigenvalues of a spatially distributed information amount, they would have nearly perfect access to information and - in terms of communication theory - perform nearly perfect bandlimited processing. As we know, the eigenvalues are proportional to the squares (i.e. variances) of resonant frequencies (FOGLEMAN 1987).

The question whether brain waves reflect underlying information processing is as old as EEG research itself. Therefore, relationships between well-confirmed psychometric and psychophysiological empirical facts (EYSENCK 1986) and EEG spectral density are very interesting.

Memory span as the quantum of action of thought

Ever since attention became the object of scientific study, psychologists have recognised that it possesses a quantitative dimension in terms of the maximum number of items to which a person can attend at one time. It now seems almost universally accepted (KAWAI and MATSUZAWA 2000) that short-term memory has a capacity limit of seven plus or minus two (MILLER 1956). The possibility that such quantitative limits on attention span might be related to qualitative differences in thought and reasoning was recognised by PIAGET (1971). Beginning with PASCUAL-LEONE (1970), the prediction of childrens reasoning from estimates of their memory span has been a major goal of neo-Piagetian theories of cognitive development.

In a typical Piagetian class inclusion task, children are shown a collection of objects (e.g., wooden beads), most of which are of one colour (e.g., red) and the rest of another colour (e.g., white). Children are asked if there are more red beads or more wooden beads and are credited with class inclusion if they indicate that there are more wooden beads because the red beads are included in the total class of wooden beads. Under the assumption that each simultaneous value assignment requires a unit of capacity, the operation of class inclusion would require a minimum of 3 such units that means a memory span of 3. It was shown by HUMPHREYS et al. (1985) that a total score on 27 Piagetian tasks was very highly correlated (r = .88) with the 14-item Wechsler IQ test. From only 13 Piagetian tasks Humphreys et. al. could form a test that is an excellent measure of general cognitive ability in its own right but can also add to the information furnished by Wechsler Verbal and Performance IQs and academic achievement. Piagetian tasks and ordinary IQ test item differ only that in Piagetian tasks this minimum of memory span to solve the task is known, in ordinary tests not or not explicitly.

Pascual-Leone understands memory span as the maximum of discrete and equal energy units (i.e. quanta) which every subject has at his disposal. In the first step of Pascual-Leone`s experimental procedure all subjects learned a small repertoire of stimulus-response units. The responses were overlearned motor behaviours such as: raise-the-hand, hit-the-basket, clap-hands, etc. If a subject has a memory span of 5 and it has to keep in mind a memory set of 5 elements, he cannot arrange element 1 corresponding to span or attention space 1, element 2 to span 2 and so on. This will be impossible. Because access to chunks in working memory is random, the available energy quanta are not distinguishable and have to be defined as bosons (i.e. indistinguishable quanta). By applying the Bose-Einstein occupancy model of combinatorics to his learning experiments with children of different age, Pascual-Leone obtained a very good agreement between empirical probabilities and Bose-Einstein predicted theoretical probabilities. Weiss (see 1992 for detailed statistics) calculated from Pascual-Leones sample of 11.8-year-olds a mean information entropy H of 86.4 bits. A mean IQ of 119 for 11.8 year-olds corresponds in performance to an adult IQ of 102 for about 40-year-olds. In tables of IQ test results edited by LEHRL et al. (1991)and based on concepts of information theory (see below), we read for this age and IQ 102 a short-term memory storage capacity of 84 bits. Two approaches with seemingly completely differing theoretical starting points lead on the absolute scale of information entropy to practically the same result. For Pascual-Leones data the latter result was even obtained after applying quantum mechanics twice in series, for calculating Bose-Einstein statistics and information entropy.

The variance of the Bose-Einstein distribution equals m2 + m, where m reflects the granularity of the energy due to Einsteins photons (cited from SCHROEDER 1991, p.189]). If we set the variance 1 and m = x, we get x2 + x = 1. The solution of this equation is, the golden mean F, also called the golden ratio, the golden number, the golden section or the divine proportion. It has the property that 1 + F = F2. Therefore the double geometric F-series:

..., 1/F2, 1/F, 1, F, F2, F3, ... .

has the properties,

..., 1/F2 + 1/F = 1, 1/F + 1 = F, 1 + F = F2, ... (1)

and is thus a Fibonacci series. It is the only geometric series that is also a Fibonacci series. Essential is the fact that the fractional parts .618033... of all powers of F are identical. The title chosen by us refers to this golden mean in the broader sense.

Forces are now recognised as resulting from the exchange of huge numbers of discrete particles, or information patterns called vector bosons, which are exchanged between two or more particle information patterns. The absorption of a vector boson information pattern changes the internal oscillation state of a particle, and causes an impulse of motion to occur along a particular direction. This turns out to be the quantum origin of all forces. Therefore, forces can be thought of being digital rather than analogue.

In 2001 BIANCONI and BARÁBASI discovered that not only
neural networks but all evolving networks, including the World Wide Web and
business networks, can be mapped into an equilibrium Bose gas, where nodes
correspond to energy levels and links represent particles. Still unaware of the
research by Pascual-Leone, for these network researchers this correspondence
between network dynamics and a Bose gas was highly unexpected (ALBERT and BARÁBASI 2002).

The information entropy of working memory capacity

Shannon’s information entropy H is the logarithm of the number of microstates or patterns
consistent with our information. To reduce the information to this scalar
measure (evaluated in bits of accuracy gained) is to reduce the form to its
topological complexity. Each global state of a network can be assigned a single
number, called the energy of that state, i.e. the differences in the log
probabilities, and hence information entropy of two global states (thoughts) is
just their energy difference. By extension of Shannon’s concept of channel capacity, in 1959
FRANK had claimed that cognitive performance to be limited by the channel
capacity of short-term memory. He argued that the capacity H of short-term memory (measured in bits of information) is the
product of the processing speed S of
information flow (in bits per second) and the duration time D (in seconds) of information in short-term memory absent rehearsal.

The first experimental approach to
determine mental processing speed in bits per second was accomplished by NAYLOR
(1968). His method of testing enabled the subjects to present to themselves a
stimulus which remained as long as they kept a finely balanced switch
depressed. The stimuli were digits between 1 and 9 or numbers between 1 and 32
presented singly or in groups of two, three, four, or five. By this procedure
the time was measured until the signs were perceived by the subjects. The
information content of one digit of the repertoire of nine possibilities was 2 3.17
= 9. That is, 3.17 bits. Recognition of one of the 32 possibilities (= 25) was equal to 5 bits. Thus, Naylor measured not only the time between stimulus and reaction but also the amount of stimulus information. This is the prerequisite for the more striking observation by LEHRL and FISCHER (1990), that the results (in bits/s) are numerically equal although the repertoires if signs differ. The measurement of stimuli and reaction in terms of the information unit (the bit) und physical time will only reveal properties of the subject if the information content of the objective repertoire agrees with that of the subjective repertoire. When a repertoire of signs (such as letters, digits or chunks) is overlearned, independently presented signs, whether of sense or nonsense in common usage, have the same objective as subjective information.

Instead of applying one of the elementary
cognitive tasks already mentioned, Lehrl et al. operationalised Frank`s concept
of short-term memory storage capacity (in bits) by testing memory span and
reading rate. The subject is simply asked to read a series of mixed up letters
in an undertone as quickly as possible. As soon as the subject begins to speak,
the stopwatch is started. The time from the first to the last spoken letter is
measured. It should be documented in tenths of a second, e.g., 7.3 s. When evaluating the raw scores it must be
remembered that a subject can only perform full binary decisions. Therefore,
the recognition of a letter out of the repertoire of 27 letters, which
theoretically has an information content of 4.7 bits (27 = 24.7)
needs five binary decisions. Since each letter contains 5 bits of information,
the 20 letters contain 100 bits. This is divided by the time of reading to
obtain the amount of information processed in a second S (bits/s). For example, if the best time of a subject is 7.3 s,
then S = 100/7.3 (bits/s) = 13.7 bits/s. By standardising letter reading on adults, normative data are available (see Table 1; column mental speed).

Forward memory span D can be predicted on the basis of the number of simple words which
the subject can read out in 1.8 seconds. Regardless of the number of syllables,
any subject in an empirical investigation by BADDELEY et al. (1975) was able to
recall as many words as he could read in 1.8 s. This result can easily be
confirmed by the normative data from Lehrl et al. For example, for IQ 100
holds: The 20 letters of their reading task are read in 6.6s; D (memory span) corresponds to 5.4. Now we can calculate x = 6.6s x 5.4 / 20 = 1.8s. Hence, span and processing rate are both measures of the same working memory system (WEISS 1995). The greater the memory span, the faster the processing rate. The time required to process a full memory load is a constant, independent of the type of material stored.

The overall importance of reading speed in everyday
life and as an indicator of processing speed is obvious. With increasing age,
children name familiar objects more rapidly, and these naming times are related
to reading ability. Greater memory capacity is associated with greater reading
recognition skill, and the same comprehension processes underlie both reading
and auding. The fastest rate that individuals can successfully operate their
reading and auding rate is limited by their thinking rate. Consequently, there
is an inverse relationship between the length of words and their frequencies of
usage. Because words are stored in neural networks, the discovery by BIANCONI
and BARÁBASI (2001) reveals the deeper meaning of ZIPF’s (1949) and Pareto’s
power law by which the size of the vocabulary of a given individual can be
understood as a function of his memory span n.

.

Memory span and EEG

During the last decades a number of authors
have claimed not only correlations between memory span and mental speed, but
also with electrophysiological variables of the EEG. In 1935, GIBBS et al. had
already documented that patients (sample size was 55) with petit mal epilepsy
show, in all cases during seizures, an outburst of spike waves of great
amplitude at a frequency of about 3/s. The fact that such seizures can be
aborted using brief stimuli is very suggestive of an underlying multistable
dynamical system. This finding is part of our confirmed knowledge and can be
read in every textbook on EEG or epilepsy. From this LIBERSON (1985) had drawn
the conclusion that all significant channels in EEG could be n multiples of one fundamental frequency of about 3.3 Hz. According to his empirical data the number of these multiples (harmonics) is nine as the maximum of memory span (see Table 1). Assuming these numbers one to nine to be quanta of action (as Pascual-Leone did), we again obtain a relationship between the classical formulae of quantum statistics and empirical results of both EEG and psychometric research.

Table 1 Memory span
(corresponding to the number of an EEG harmonic), frequency of EEG harmonics and mental speed and their relationships
with information entropy, power density of short-term memory storage
capacity, and IQ

Assuming the numbers 1 to 9 of memory span
to be equivalent of harmonics in the sense of wave theory, the power spectral
density E is given by the eigenstate energy-frequency relationship

E = nf (kT x ln2), (3)

where f is frequency. According to
thermodynamics, the measurement of 1 bit of information entropy (SZILARD 1929)
a minimum energy of 1 kT x ln2, where
k is Boltzmann’s constant and T is absolute temperature. During the
duration of 1 perceptual moment 1 bit of information is processed per harmonic.
That means that 1 break of symmetry and 1 phase reversal after each
zero-crossing of an EEG wave corresponds with a possible 1 bit decision between
two alternatives. Consequently, each degree of freedom and of translation
corresponds to an energy of kT/2 or its macroscopic analogon.

Because the frequency of EEG harmonics can be expressed as n x 2F Hz, for the expected latencies of harmonics follows 1000 ms/n x 2F and for power density follows

E = Sum n x 2F (kT x ln2). (4)

The physical term power is appropriate because it is a measure of the ability of waves at frequency f to do work. The power spectrum to the EEG describes the total variance in the amplitudes due to the integer multiples of the fundamental frequency (i.e. the first harmonic 1 x 2F). In order to calculate power density in this way, the waveform must be squared and then integrated for the duration of its impulse response, i.e. the duration of the transient of 1 complete wave packet containing all the harmonics of the memory span of a given subject.

The relationships in Table 1 are further
supported by data from BENNETT (1974), who reanalysed the ERTL and SCHAFER
(1969) findings of a correlation between IQ and latencies of EEG evoked
potential components. Bennett (confirmed by FLINN 1977 accomplished a Fourier transformation of the original data and found that high IQ subjects (IQ above 123) go through 20 or more perceptual moments per second, low IQ subjects (IQ below 75) only through 8 moments or even less (compare Table 1, columns b and d). This striking parallelism between EEG results and channel capacity, measured with mental tests, is emphasised by results from HARWOOD and NAYLOR (1969). 42 young university students had a mean channel capacity of 21.4 bit/s; 105 "average normal" adults who were 60-69 years old performed 14.2 bit/s; the age group of 70-79 years (sample size was 67) achieved 12.9 bit/s; and 13 subjects being 80 years and older 10.2 bit/s, thus reflecting the usual decline of mental performance of old aged people. Pure coincidence in this parallelism of channel capacity and EEG frequencies (compare Table 1) seems impossible: neither Liberson nor Lehrl, neither Bennett nor Naylor nor Pascual-Leone knew anything about the results and theories of the others.

Higher IQ subjects have not only a higher
memory span, but consequently also more complex waveforms of EEG than lower IQ
subjects. The most extreme compression of information is represented by the
eigenvalues of the power spectrum. There are as much eigenvalues of a spectrum
as are harmonics. Already in 1959 BURCH (cited from SALTZBERT and BURCH 1971)
had found that "the parameters ... of the power spectral density ... can
be estimated in a completely adequate way without the necessity of performing
squaring and integrating operations but simply by counting the zero
crossings." The number of zero-crosses up to the P300 of evoked potentials
is the upper bound of the memory span of an individual.

In such a way memory span has to be understood as the quantum of action of thought. In fact, these quanta of action represent macroscopic ordered states in the sense of quantum mechanics. Empirical analysis shows that Libersons fundamental is lower than 3.3 Hz and in the range between 3.1 and 3.3 Hz. The reliability of the empirical data allows no more precise calculation. Nevertheless, it could be imagined that a numerical constant underlies the harmonics of the EEG, enabling brain waves to process information in the most efficient way. From technical applications we know that an array consists of equally spaced sensors making measurements at discrete intervals (BATH 1974). Only under this condition frequency bands and wavenumber can be detected in the spatiotemporal domain. If a travelling wave is spatially sampled using such a discrete array of sensors, an estimate of the wave is obtained by appropriately delaying or advancing the signals on each of the channels and summing the results. Therefore the idea that brain architecture and neural networks, respectively, should be understood in terms of sequences of delaying chains and matched filters facilitating run-length coding is not a new one.

The golden mean as resonant frequency

It is a psychoacoustic fact, known as octave equivalence (GLASSMAN 1999) that all known musical cultures consider a tone twice the frequency of another to be, in some sense, the same tone as the other (only higher). The point of resonance, corresponding to the eigenvalues and zero-crossings of a wave packet (wavelet), is not the frequency of its fundamental, but half of its frequency. If we assume the fundamental to be twice the golden mean (DE SPINADEL 1998) F, that means 2 x 1,618 = 3,236 Hz, a point of resonance at F = 1,618 Hz follows. DATTA (2002) showed how a sense of time and evolution is intrinsically defined by the infinite continued fraction of the golden mean and its inverse. The real number set gets replaced by an extended physical set, each element of which is endowed with an equivalence class of infinitesimally separated neighbours in the form of random fluctuations. Time thereby undergoes random inversions generating well defined random scales, thus allowing a dynamical system to evolve self similarly over the set of multiple scales. This insight opens an astounding variety of possibilities to encode and decode information in the most efficient way. With this property the brain can use simultaneously the powers of the golden mean and the infinite Fibonacci word (FROUGNY and SAKAROVICH 1999) (synonymously called the golden string, the golden sequence, or the rabbit sequence) for coding and classifying. Every positive integer can be a sum of Fibonacci numbers; it can also be understood as a finite sum of positive and negative powers of the golden mean. A binomial graph of a memory span n has n distinct eigenvalues and these are powers of the golden mean. The number of closed walks of length k in the binomial graph is equal to the nth power of the (k+1)-st Fibonacci number (KENNEDY and CHRISTOPHER 1997). The total number of closed walks of length k within memory is the nth power of the kth Lucas number.

Lifeforms maximise their adaptive capacities by entering the region of complexity on the edge of chaos. From the period doubling route of chaos it turns out that when R = 2 F = 2 times 1.618 = 3.236 one gets a super-stable period with two orbits, producing the first island of stability. Thus, the quasi periodic F toroid geometry is the most stable under perturbation. The orbit is of the lowest period possible (being two) and therefore crucially, consumes the least energy to maintain. Bands of order in the Feigenbaum diagram occur at a fixed scaling mean, where all bifurcations, representing the length w1 = (F  1)/2, are positioned at a = 2 F [42]. This is how F is embedded within dynamical systems, as a universal binary shift operator, or primary eigenfunction. All constants so derived have to be eigenvalues of this operator (think of resonances and harmonics).

The existence of a Fibonacci series and the convergence of the ratio of the winding numbers of an orbit towards (Ö5  1)/2 = 1/F in a Hamiltonian system is a numerically well-know phenomenon of physics. The mathematical foundation and proof of this phenomenon is the essence of the theorem of Kolmogorov, Arnold and Moser (KAM). From this theorem follows, too, that the golden mean, which is the most irrational number, must give the most stable orbit. Irrational values of the winding number correspond to an uncountable set of zero measures of values  in other words the irrationals are squeezed into a Cantor dust (EL NASCHIE 1999).

A slide-rule computes
products because the marks on the sliding ruler correspond to logarithms, and
adding two logarithms is equivalent to multiplying the corresponding pair of
numbers. Also the Fibonacci and Lucas numbers can be understood like the
markings on a ruler that is recursively divided into golden mean pieces. By
using powers of the golden mean any multiplication can be reduced to an addition.
The golden mean is the mean of the sides of a rectangle circumscribed about a
logarithmic spiral, too. Logarithmic spirals are, like fractals, self-similar
at all scales. Therefore our brain performs visual computation at several
scales (demagnifications of the image) and compares the results. With a
sampling algorithm, based on Fibonaccis and phyllotaxis, even coloured images can be quantized and processed (MOJSILOVIC and SOLJANIN 2001).

If we draw a line y = F x on a graph (i.e. a line whose gradient is F) there we can see directly the binary expression of the Fibonacci sequence known as the infinite Fibonacci word. Where the F line crosses a horizontal grid line (imagine the discrete columns of the brain) we write 1 by the line and where the F line crosses a vertical line we write a 0. As we travel along the F line from the origin we meet a sequence of 1s and 0s. The 1s in the Fibonacci string 1011010110 occur at positions given by the spectrum of F and only at those positions. Trajectories of dynamic systems whose phase spaces have a negative curvature everywhere can be completely characterised by such a discrete sequence of 0s and 1s. The self-similar Fibonacci string reproduces itself upon reverse mapping or decimation, both fundamental properties from the point of information storage and retrieval. After decimation by a factor of the golden mean every unit in the original lattice coincides precisely with a unit in the compressed lattice. From the point of view of renormalization theories of physics, the decimation process is the complement of deflation or block renaming. Any 1 in the Fibonacci string forces an infinite number of symbols in a characteristic quasi periodic pattern. For any such Sturmian sequence the topological structure completely determines all the Markov approximations. It means that only one ergodic measure is compatible with the topological structure.

For computer science the Fibonacci string is no newcomer. Processing of strings of symbols and string rewriting is the most fundamental and the most common form of computer processing: every computer instruction is a string, and every piece of data processed by these instructions is a string. A repetition in a string is a word of the form 11 or 00, called a square. The frequency of such squares is a function of the logarithm of the golden mean (MIGNOSI et al 1998).

Since the fabrication of semiconductor
superlattices arranged according to the Fibonacci and other sequences, there
has been a growing interest in their electronic properties. When a homogeneous
electric field is applied perpendicular to the layer plan, electronic states
become localised and the energy spectrum consists of a Wannier-Stark-ladder,
characterised by a sequence of metastable states of resonance separated by equal
energy intervals. An initial Gaussian wave packet is filtered selectively when
passing through the superlattice. This means that only those components of the
wave packet whose wave number belong to the allowed harmonics of the
fractal-like energy spectrum can propagate over the lattice. DIEZ et al. (1996)
discuss therefore, aside from the possibility of building filterlike devices, designed with Fibonacci or a binary quasi periodic sequence according to the desired application, the possibility that such a kind of system can be used in processing information. Surely, the insight that our brain uses very similar physical and mathematical properties will accelerate technical progress in this area. CUESTA and SATIJA (1999), studying empirically Fibonacci lattices with defects, even found: Novel result of our studies is the relationship between the resonant states and the states where the energy bands cross. We show that the resonant states are fully transmitting states in the quasi periodic limit and are described by the wave functions that are related to the harmonics of the sine wave with fundamental Bloch number equal to the golden mean. Bloch Waves are known as the most important effect due to the discrete lattice translational symmetry. This arises because the Hamiltonian must commute with the translational operator for any discrete integer lattice translation. The wave function can be represented as the product of a plane wave with a periodic function. The translational invariance of the wave function is of utmost importance. This basically indicates that all information about the system is stored within an excited subset of the system; the rest of the non-resonant information is redundant at this very moment. If we stress the analogy between waves in quasi periodic lattices and the phenomenon of memory span in our brain, this seems to be an especially important point. We confess to have the vision of multilayer hierarchical binary or Fibonacci semiconductor superlattices simulating the calculating and classifying capabilities of our brain, far surpassing the brain by the higher speed of the technical application.

It is already well-known among electrical engineers (SRINIVASAN 1992) that the characteristic impedance of an electrical ladder network, which is needed for an error-free connection, has to be a function of the golden mean. Even the sound by any stereo system depends on the purity of the audio signal it produces. Each strand in a cable has its own beat. When the cable linking all components together imparts its own sound, the audio signal is corrupted. George Cardas received U.S. Patent Number 4,628,151 for creating Golden Mean Stranding Audio Cable. Individual strands are arranged so each strand is coupled to another, whose note or beat is irrational with its own, thus nulling interstrand resonance.

The universe as a world of numbers

EL NASCHIE (2000) and others developed for the fundamental question of time reversibility the notion of a Cantorian space-time. What is really remarkable of this Cantorian space-time is that applying all the probabilistic necessary laws, the values of the Hausdorff dimension are intrinsically linked to the golden mean and its successive powers. The correlated fluctuations of the fractal space-time are analogous to the Bose-Einstein condensation phenomenon. The polynomial roots of higher order Fibonaccis, scaling a quasi periodic hierarchy, are based on golden mean powers.

There can be no doubt that our brain uses for computing inherent and inborn properties of the physical universe. We have or learn into the neural network of our brains the relationships between external stimuli, the integer powers of the golden mean, the Fibonacci word and Lucas numbers, and we are probably able to use the relationships between the Beatty sequences of e, p and F, and we use hundreds of similar relationships (many of them may still be undiscovered by contemporary mathematics, see, e.g., ATANASSOV et al. 2002) between numbers for encoding and decoding information simultaneously and unconsciously by wavelets. A genius like Ramanajun gave us some closed fraction formulae which contain p, e and F all together in a single equation. Together with Eulers famous formula eip + 1 = 0 for the unit circle we all understand in our subconsciousness these irrational numbers as rules for superposition and time reversal by folding, symmetry breaking and compactification. By raising F = 1 + f to the third power , we get the Hausdorff dimension of Cantorian spacetime

(1 + f)3 = 2 + Ö5 = 4 + f3 = 1 + f/1 - f = 1/f3 = 4.236 (5)

Quantum mechanics seems to require the quantization of all physical quantities on the small scale, yet space and time are still treated in most cases as a classical space-time continuum, where there are an infinite number of space points between any two given locations, no matter how close. Therefore many physicists agree that the current set of fundamental physical laws is incomplete. Because Hz, oscillations per second, is superficially seen only a man-made measure, this seems to be the weakest point of our line of reasoning. Behind the definition of the second is the velocity of light (c = 299792458 m/s), which is the constant on which size all other physical constants depend upon and hence represents the inherent speed limit that any particle information pattern is able to achieve. In this system of present-day constants the Planck length has the value 1,6160 x 10-35 m (standard uncertainty 0,0012 x 10-35 m). If we fix instead the Planck length at the value of F at 1,6180 x 10-35 m and recalculate consequently all other physical variables, this means for the numerical size of the second only a trivial correction not relevant for our argument.

Indeed, there is a growing minority of scholars who understand the world as something like a cellular automaton running with and counting numbers. The numerical state of all the cells, everywhere, changes at a regular synchronised interval called a clock cycle. The universal cellular automaton seems to be capable of updating its entire memory in a single clock cycle, which according to Occams razor could be nothing else than the Planck time as the relation between the velocity of light and the Planck length, the latter fixed at the value of F. If we, for example, look into Wolframs A New Kind of Science we see that this Model 3 and Model 4 automata are full of Pascal triangles. Behind such a triangle are always the Fibonaccis and hence the golden mean. That means that by encoding and decoding the information of such and any automaton or system no other wave could be more optimal than a wavelet containing the golden mean itself. The quantization of time simply represents the number of regular clock cycles elapsed between two events and all changes that occur must occur as localised changes. At the lowest level our brain seems to be utterly simple, deterministic and mathematical in nature. Despite this, we can never read out the numeric state of any brain in a foreseeable future. We can only infer this type of information by observing larger scale patterns as it is the phenomenon of memory span.

Frougny C, Sakarovitch J. Automatic conversion from Fibonacci representation to representation in base phi, and a generalization. International Journal of Algebra and Computing, 1999; 9:351-384.

Gibbs FA, Davis H, Lennox WG. The electroencephalogram in epilepsy and in conditions of impaired consciousness. Archivs of Neurology and Psychiatry, 1935; 34:1133-1148.

Glassman RB. Hypothesized neural dynamics of working memory: Several chunks might be marked simultaneously by harmonic frequencies within an octave band of brain waves. Brain Research Bulletin, 1999; 50:77-94.