The Grossone Methodology Perspective On Turing Machines
The Grossone Methodology Perspective On Turing Machines
Turing machines
Abstract This chapter discusses how the mathematical language used to describe
and to observe automatic computations influences the accuracy of the obtained
results. The chapter presents results obtained by describing and observing dif-
ferent kinds of Turing machines (single and multi-tape, deterministic and non-
deterministic) through the lens of a new mathematical language named Grossone.
This emerging language is strongly based on three methodological ideas borrowed
from Physics and applied to Mathematics: the distinction between the object (indeed
mathematical object) of an observation and the instrument used for this observation;
interrelations holding between the object and the tool used for the observation; the
accuracy of the observation determined by the tool. In the chapter, the new results
are compared to those achievable by using traditional languages. It is shown that
both languages do not contradict each other but observe and describe the same ob-
ject (Turing machines) but with different accuracies.
1 Introduction
Turing machines represent one of the simple abstract computational devices that can
be used to investigate the limits of computability . In this chapter, they are consid-
ered from several points of view that emphasize the importance and the relativity of
Yaroslav D. Sergeyev
Dipartimento di Ingegneria Informatica, Modellistica, Elettronica e Sistemistica (DIMES), Uni-
versità della Calabria, Rende (CS), Italy.
N.I. Lobatchevsky State University, Nizhni Novgorod, Russia.
Istituto di Calcolo e Reti ad Alte Prestazioni, C.N.R., Rende (CS), Italy.
e-mail: [email protected]
Alfredo Garro
Dipartimento di Ingegneria Informatica, Modellistica, Elettronica e Sistemistica (DIMES), Uni-
versità della Calabria, Rende (CS), Italy.
e-mail: [email protected]
1
2 Yaroslav D. Sergeyev and Alfredo Garro
1 We are reminded that a numeral is a symbol or group of symbols that represents a number. The
difference between numerals and numbers is the same as the difference between words and the
things they refer to. A number is a concept that a numeral expresses. The same number can be
represented by different numerals. For example, the symbols ‘7’, ‘seven’, and ‘VII’ are different
numerals, but they all represent the same number.
A new perspective on Turing machines 3
2 Turing machines
The Turing machine is one of the simple abstract computational devices that can be
used to model computational processes and investigate the limits of computability.
In the following subsections, deterministic Single and Multi-tape Turing machines
are described along with important classical results concerning their computational
power and related equivalences (see Section 2.1 and 2.2 respectively); finally, non-
deterministic Turing machines are introduced (see Section 2.3).
where Q is a finite and not empty set of states; Γ is a finite set of symbols; b̄ ∈ Γ is
a symbol called blank; Σ ⊆ {Γ − b̄} is the set of input/output symbols; q0 ∈ Q is the
initial state; F ⊆ Q is the set of final states; δ : {Q − F} × Γ 7→ Q × Γ × {R, L, N} is
a partial function called the transition function, where L means left, R means right,
and N means no move .
Specifically, the machine is supplied with: (i) a tape running through it which is
divided into cells each capable of containing a symbol γ ∈ Γ, where Γ is called the
tape alphabet, and b̄ ∈ Γ is the only symbol allowed to occur on the tape infinitely
often; (ii) a head that can read and write symbols on the tape and move the tape left
and right one and only one cell at a time. The behavior of the machine is specified
by its transition function δ and consists of a sequence of computational steps ; in
each step the machine reads the symbol under the head and applies the transition
function that, given the current state of the machine and the symbol it is reading on
the tape, specifies (if it is defined for these inputs): (i) the symbol γ ∈ Γ to write on
the cell of the tape under the head; (ii) the move of the tape (L for one cell left, R for
one cell right, N for no move); (iii) the next state q ∈ Q of the machine.
Starting from the definition of Turing machine introduced above, classical results
(see, e.g., [1]) aim at showing that different machines in terms of provided tape and
alphabet have the same computational power, i.e., they are able to execute the same
computations. In particular, two main results are reported below in an informal way.
Given a Turing machine M = {Q, Γ, b̄, Σ, q0 , F, δ}, which is supplied with an infi-
nite tape, it is always possible to define a Turing machine M ′ = {Q′ , Γ′ , b̄, Σ′ , q′0 , F ′ , δ′ }
which is supplied with a semi-infinite tape (e.g., a tape with a left boundary) and is
equivalent to M , i.e., is able to execute all the computations of M .
4 Yaroslav D. Sergeyev and Alfredo Garro
Let us consider a variant of the Turing machine defined in (1) where a machine
is equipped with multiple tapes that can be simultaneously accessed and updated
through multiple heads (one per tape). These machines can be used for a more direct
and intuitive resolution of different kind of computational problems. As an example,
in checking if a string is palindrome it can be useful to have two tapes on which
represent the input string so that the verification can be efficiently performed by
reading a tape from left to right and the other one from right to left.
Moving towards a more formal definition, a k-tapes, k ≥ 2, Turing machine
(see [12]) can be defined (cf. (1)) as a 7-tuple
D E
M K = Q, Γ, b̄, Σ, q0 , F, δ(k) , (2)
S
where Σ = ki=1 Σi is given by the union of the symbols in the k input/output al-
phabets Σ1 , . . . , Σk ; Γ = Σ ∪ {b̄} where b̄ is a symbol called blank; Q is a finite and
not empty set of states; q0 ∈ Q is the initial state; F ⊆ Q is the set of final states;
δ(k) : {Q − F} × Γ1 × · · · × Γk 7→ Q × Γ1 × · · · × Γk × {R, L, N}k is a partial func-
tion called the transition function, where Γi = Σi ∪ {b̄}, i = 1, . . . , k, L means left, R
means right, and N means no move .
This definition of δ(k) means that the machine executes a transition starting from
an internal state qi and with the k heads (one for each tape) above the characters
ai1 , . . . , aik , i.e., if δ(k) (q1 , ai1 , . . . , aik ) = (q j , a j 1 , . . . , a j k , z j 1 , . . . , z j k ) the machine
goes in the new state q j , write on the k tapes the characters a j 1 , . . . , a j k respec-
tively, and moves each of its k heads left, right or no move, as specified by the
z j l ∈ {R, L, N}, l = 1, . . . , k.
A machine can adopt for each tape a different alphabet, in any case, for each tape,
as for the Single-tape Turing machines, the minimum portion containing characters
distinct from b̄ is usually represented. In general, a typical configuration of a Multi-
tape machine consists of a read-only input tape, several read and write work tapes,
and a write-only output tape, with the input and output tapes accessible only in one
direction. In the case of a k-tapes machine, the instant configuration of the machine,
A new perspective on Turing machines 5
as for the Single-tape case, must describe the internal state, the contents of the tapes
and the positions of the heads of the machine. D E
More formally, for a k-tapes Turing machine M K = Q, Γ, b̄, Σ, q0 , F, δ(k) with
Sk
Σ= i=1 Σi (see 2) a configuration of the machine is given by:
It is worth noting that, although the k-tapes Turing machine can be used for a more
direct resolution of different kind of computational problems, in the classical frame-
work it has the same computational power of the Single-tape Turing machine. More
formally, given a Multi-tape Turing machine it is always possible to define a Single-
tape Turing machine which is able to fully simulate its behavior and therefore to
completely execute its computations. In particular, the Single-tape Turing machines
adopted for the simulation use a particular kind of the tape which is divided into
tracks (multi-track tape). In this way, if the tape has m tracks, the head is able to
access (for reading and/or writing) all the m characters on the tracks during a sin-
gle operation. If for the m tracks the alphabets Γ1 , . . . Γm are adopted respectively,
the machine alphabet Γ is such that |Γ| = |Γ1 × · · · × Γm | and can be defined by an
injective function from the set Γ1 × · · · × Γm to the set Γ; this function will asso-
ciate the symbol b̄ in Γ to the tuple (b̄, b̄, . . . , b̄) in Γ1 × · · · × Γm . In general, the
6 Yaroslav D. Sergeyev and Alfredo Garro
A non-deterministic Turing machine (see [12]) can be defined (cf. (1)) as a 7-tuple
M N = Q, Γ, b̄, Σ, q0 , F, δN , (5)
where Q is a finite and not empty set of states; Γ is a finite set of symbols; b̄ ∈ Γ is
a symbol called blank; Σ ⊆ {Γ − b̄} is the set of input/output symbols; q0 ∈ Q is the
initial state; F ⊆ Q is the set of final states; δN : {Q−F}×Γ 7→ P (Q×Γ×{R, L, N})
is a partial function called the transition function, where L means left, R means right,
and N means no move .
A new perspective on Turing machines 7
An important result for the classic theory on Turing machines (see e.g., [1]) is that
for any non-deterministic Turing machine M N there exists an equivalent determinis-
tic Turing machine M D . Moreover, if the depth of the computational tree generated
by M N is equal to k, then for simulating M N , the deterministic machine M D will
execute at most
k
KM D = ∑ jd j = O(kd k )
j=0
computational steps.
Intuitively, for simulating M N , the deterministic Turing machine M D executes
a breadth-first visit of the computational tree of M N . If we consider the example
from Fig. 1 with k = 3, then the computational tree of M N has d k = 27 leaf nodes
and d k = 27 computational paths consisting of k = 3 branches (i.e., computational
steps) . Then, the tree contains d k−1 = 9 computational paths consisting of k − 1 = 2
branches and d k−2 = 3 computational paths consisting of k − 2 = 1 branches . Thus,
for simulating all the possible computations of M N , i.e., for complete visiting the
8 Yaroslav D. Sergeyev and Alfredo Garro
Fig. 1 The computational tree of a non-deterministic Turing machine M N having the non-
deterministic degree d = 3
In this section, we give just a brief introduction to the methodology of the new
approach [30, 32] dwelling only on the issues directly related to the subject of the
chapter. This methodology will be used in Section 4 to study Turing machines and
to obtain some more accurate results with respect to those obtainable by using the
traditional framework [4, 44] .
In order to start, let us remind that numerous trials have been done during
the centuries to evolve existing numeral systems in such a way that numerals
representing infinite and infinitesimal numbers could be included in them (see
[2, 3, 5, 17, 18, 25, 28, 46]). Since new numeral systems appear very rarely, in each
concrete historical period their significance for Mathematics is very often underes-
timated (especially by pure mathematicians). In order to illustrate their importance,
let us remind the Roman numeral system that does not allow one to express zero and
negative numbers. In this system, the expression III-X is an indeterminate form. As
a result, before appearing the positional numeral system and inventing zero math-
ematicians were not able to create theorems involving zero and negative numbers
and to execute computations with them.
There exist numeral systems that are even weaker than the Roman one. They se-
riously limit their users in executing computations. Let us recall a study published
recently in Science (see [11]). It describes a primitive tribe living in Amazonia (Pi-
rahã). These people use a very simple numeral system for counting: one, two, many.
For Pirahã, all quantities larger than two are just ‘many’ and such operations as 2+2
and 2+1 give the same result, i.e., ‘many’. Using their weak numeral system Pirahã
are not able to see, for instance, numbers 3, 4, 5, and 6, to execute arithmetical op-
erations with them, and, in general, to say anything about these numbers because in
their language there are neither words nor concepts for that.
In the context of the present chapter, it is very important that the weakness of
Pirahã’s numeral system leads them to such results as
which are very familiar to us in the context of views on infinity used in the traditional
calculus
∞ + 1 = ∞, ∞ + 2 = ∞. (7)
The arithmetic of Pirahã involving the numeral ‘many’ has also a clear similarity
with the arithmetic proposed by Cantor for his Alephs2 :
ℵ0 + 1 = ℵ0 , ℵ0 + 2 = ℵ0 , ℵ1 + 1 = ℵ1 , ℵ1 + 2 = ℵ1 . (8)
2 This similarity becomes even more pronounced if one considers another Amazonian tribe –
Mundurukú (see [26]) – who fail in exact arithmetic with numbers larger than 5 but are able to
compare and add large approximate numbers that are far beyond their naming range. Particularly,
they use the words ‘some, not many’ and ‘many, really many’ to distinguish two types of large
numbers using the rules that are very similar to ones used by Cantor to operate with ℵ0 and ℵ1 ,
respectively.
10 Yaroslav D. Sergeyev and Alfredo Garro
①
0 · ① = ① · 0 = 0, ① − ① = 0, = 1, ①0 = 1, 1① = 1, 0① = 0, (9)
①
3 Analogously, when the switch from Roman numerals to the Arabic ones has been done, numerals
ones can be studied. For all of them it becomes possible to introduce a new notion
of continuity that is closer to our modern physical knowledge. Functions assuming
finite and infinite values can be differentiated and integrated.
By using the new numeral system it becomes possible to measure certain infinite
sets and to see, e.g., that the sets of even and odd numbers have ①/2 elements
each. The set, Z, of integers has 2①+1 elements (① positive elements, ① negative
elements, and zero). Within the countable sets and sets having cardinality of the
continuum (see [20, 36, 37]) it becomes possible to distinguish infinite sets having
different number of elements expressible in the numeral system using grossone and
to see that, for instance,
①
< ① − 1 < ① < ① + 1 < 2① + 1 < 2①2 − 1 < 2①2 < 2①2 + 1 <
2
2①2 + 2 < 2① − 1 < 2① < 2① + 1 < 10① < ①① − 1 < ①① < ①① + 1. (12)
Another key notion for our study of Turing machines is that of infinite sequence.
Thus, before considering the notion of the Turing machine from the point of view
of the new methodology, let us explain how the notion of the infinite sequence can
be viewed from the new positions.
4 The Postulate 3 states: The part is less than the whole is applied to all numbers (finite, infinite,
and infinitesimal) and to all sets and processes (finite and infinite), see[30].
12 Yaroslav D. Sergeyev and Alfredo Garro
① ① ①
{bn : − 1} = {4, 8, . . . 4( − 2), 4( − 1)}; (14)
2 2 2
2① 2① 2①
{cn : } = {4, 8, . . . 4( − 1), 4 }. (15)
3 3 3
The three sequences have an = bn = cn = 4n but they are different because they
have different number of members. Sequence {an } has ① elements and, therefore,
is complete, {bn } has ① ①
2 − 1, and {cn } has 2 3 elements. 2
Let us consider now infinite sequences as one of the instruments used by math-
ematicians to study the world around us and other mathematical objects and pro-
cesses. The first immediate consequence of Theorem 1 is that any sequential pro-
cess can have at maximum ① elements. This means that a process of sequential
observations of any object cannot contain more than ① steps5 . We are not able to
5 It is worthy to notice a deep relation of this observation to the Axiom of Choice. Since Theorem 1
states that any sequence can have at maximum ① elements, so this fact holds for the process of a
sequential choice, as well. As a consequence, it is not possible to choose sequentially more than
① elements from a set. This observation also emphasizes the fact that the parallel computational
paradigm is significantly different with respect to the sequential one because p parallel processes
can choose p · ① elements from a set.
A new perspective on Turing machines 13
execute any infinite process physically but we assume the existence of such a pro-
cess; moreover, only a finite number of observations of elements of the considered
infinite sequence can be executed by a human who is limited by the numeral system
used for the observation. Indeed, we can observe only those members of a sequence
for which there exist the corresponding numerals in the chosen numeral system; to
better clarify this point the following example is discussed.
Example 2. Let us consider the numeral system, P , of Pirahã able to express only
numbers 1 and 2. If we add to P the new numeral ①, we obtain a new numeral
system (we call it Pb ). Let us consider now a sequence of natural numbers {n : ①}.
It goes from 1 to ① (note that both numbers, 1 and ①, can be expressed by numerals
from Pb ). However, the numeral system Pb is very weak and it allows us to observe
only ten numbers from the sequence {n : ①} represented by the following numerals
① ① ① ① ①
1, 2 , ... − 2, − 1, , + 1, + 2, ... ① − 2, ① − 1, ① . (16)
|{z} |2 2 2 2
{z 2 } | {z }
f inite in f inite
in f inite
The first two numerals in (16) represent finite numbers, the remaining eight numer-
als express infinite numbers, and dots represent members of the sequence of natural
numbers that are not expressible in Pb and, therefore, cannot be observed if one uses
only this numeral system for this purpose. 2
In the light of the limitations concerning the process of sequential observations,
the researcher can choose how to organize the required sequence of observations
and which numeral system to use for it, defining so which elements of the object
he/she can observe. This situation is exactly the same as in natural sciences: before
starting to study a physical object, a scientist chooses an instrument and its accuracy
for the study.
Example 3. Let us consider the set A={1, 2, 3, . . . , 2①-1,2①} as an object of our
observation. Suppose that we want to organize the process of the sequential counting
of its elements. Then, due to Theorem 1, starting from the number 1 this process can
arrive at maximum to ①. If we consider the complete counting sequence {n : ①},
then we obtain
1, 2, 3, 4, . . . ①− 2, ①− 1, ①, ①+ 1, ①+ 2, ①+ 3, . . . , 2①− 1, 2①
x xxx x xx
| {z } (17)
① steps
Analogously, if we start the process of the sequential counting from 5, the process
arrives at maximum to ① + 4:
1, 2, 3, 4, 5 . . . ①− 1, ①, ①+ 1, ①+ 2, ①+ 3, ①+ 4, ①+ 5, . . . , 2①− 1, 2①
x x xx x x x
| {z } (18)
① steps
14 Yaroslav D. Sergeyev and Alfredo Garro
The corresponding complete sequence used in this case is {n + 4 : ①}. We can also
change the length of the step in the counting sequence and consider, for instance,
the complete sequence {2n − 1 : ①}:
If we use again the numeral system Pb , then among finite numbers it allows us to see
only number 1 because already the next number in the sequence, 3, is not expressible
in Pb . The last element of the sequence is 2① − 1 and Pb allows us to observe it. 2
The introduced definition of the sequence allows us to work not only with the
first but with any element of any sequence if the element of our interest is express-
ible in the chosen numeral system independently whether the sequence under our
study has a finite or an infinite number of elements. Let us use this new definition
for studying infinite sets of numerals, in particular, for calculating the number of
points at the interval [0, 1) (see [30, 32]). To do this we need a definition of the term
‘point’ and mathematical tools to indicate a point. If we accept (as is usually done
in modern Mathematics) that a point A belonging to the interval [0, 1) is determined
by a numeral x, x ∈ S, called coordinate of the point A where S is a set of numerals,
then we can indicate the point A by its coordinate x and we are able to execute the
required calculations.
It is worthwhile to emphasize that giving this definition we have not used the
usual formulation “x belongs to the set, R, of real numbers”. This has been done be-
cause we can express coordinates only by numerals and different choices of numeral
systems lead to different sets of numerals and, as a result, to different sets of num-
bers observable through the chosen numerals. In fact, we can express coordinates
only after we have fixed a numeral system (our instrument of the observation) and
this choice defines which points we can observe, namely, points having coordinates
expressible by the chosen numerals. This situation is typical for natural sciences
where it is well known that instruments influence the results of observations. Re-
mind the work with a microscope: we decide the level of the precision we need and
obtain a result which is dependent on the chosen level of accuracy. If we need a
more precise or a more rough answer, we change the lens of our microscope.
We should decide now which numerals we shall use to express coordinates of
the points. After this choice we can calculate the number of numerals expressible
in the chosen numeral system and, as a result, we obtain the number of points at
the interval [0, 1). Different variants (see [30, 32]) can be chosen depending on the
precision level we want to obtain. For instance, we can choose a positional numeral
system with a finite radix b that allows us to work with numerals
Then, the number of numerals (20) gives us the number of points within the inter-
val [0, 1) that can be expressed by these numerals. Note that a number using the
positional numeral system (20) cannot have more than grossone digits (contrarily
to sets discussed in Example 3) because a numeral having g > ① digits would not
be observable in a sequence. In this case (g > ①) such a record becomes useless in
sequential computations because it does not allow one to identify numbers entirely
since g − ① numerals remain non observed.
expressing integers in the positional system with a finite radix b in the alphabet
{0, 1, . . . b − 2, b − 1} is equal to b① .
Proof. The proof is a straightforward consequence of Theorem 2 and is so omit-
ted. 2
In this Section the different types of Turing machines introduced in Section 2 are
analyzed and observed by using as instruments of observation the Grossone lan-
guage and methodology presented in Section 3 . In particular, after introducing a
16 Yaroslav D. Sergeyev and Alfredo Garro
distiction between physical and ideal Turing machine (see Section 4.1), some re-
sults for Single-tape and Multi-tape Turing machines are summarized (see Sections
4.2 and 4.3 respectively), then a discussion about the equivalence between Single
and Multi-tape Turing machine is reported in Section 4.4. Finally, a comparison be-
tween deterministic and non-deterministic Turing machines through the lens of the
Grossone methodology is presented in Section 4.5.
Let us consider first a physical machine T P and discuss about the number of
computational steps it can execute and how the obtained results then can be inter-
preted by a human observer (e.g. the researcher) . We suppose that its output is
written on the tape using an alphabet Σ containing b symbols {0, 1, . . . b − 2, b − 1}
where b is a finite number (Turing uses b = 10).Thus, the output consists of a se-
quence of digits that can be viewed as a number in a positional system B with the
radix b. By definition, T P should stop after a finite number of iterations. The mag-
nitude of this value depends on the physical construction of the machine, the way
the notion ‘iteration’ has been defined, etc., but in any case this number is finite.
A physical machine T P stops in two cases: (i) it has finished the execution of its
program and stops; (ii) it stops because its breakage. In both cases the output se-
quence
not clear whether the obtained output is a solution or just a result of the depletion of
its computational resources.
In particular, with respect to the halting problem it follows that all algorithms stop
on T P .
In order to be able to read and to understand the output, the researcher (the user)
should know a positional numeral system U with an alphabet {0, 1, . . . u − 2, u − 1}
where u ≥ b. Otherwise, the output cannot be decoded by the user. Moreover, the
researcher must be able to observe a number of symbols at least equal to the maximal
length of the output sequence that can be computed by machine (i.e., KU ≥ KT P ).
If the situation KU < KT P holds, then this means that the user is not able to inter-
pret the obtained result. Thus, the number K ∗ = min{KU , KT P } defines the length
of the outputs that can be computed and interpreted by the user.
As a consequence, algorithms producing outputs having more than K ∗ positions be-
come less interesting from the practical point of view.
After having introduced the distinction between physical and ideal Turing ma-
chines, let us analyze and observe them through the lens of the Grossone Method-
ology. Specifically, the results obtained and discussed in [42] for deterministic and
non-deterministic Single-tape Turing machines are summarized in Section 4.2 and
4.4 respectively; whereas, Section 4.3 reports additional results for Multi-tape Tur-
ing machines (see [43]).
As stated in Section 4.1, single-tape ideal Turing machines M I (see Section 2.1)
can produce outputs with an infinite number of symbols k. However, in order to be
observable in a sequence, an output should have k ≤ ① (see Section 3). Starting
from these considerations the following theorem can be introduced.
Proof. This result follows from the definitions of the complete sequence and the
form of numerals
Corollary 3. Let us consider an ideal Turing machine M 1I working with the alpha-
bet {0, 1, 2} and computing the following complete computable sequence
18 Yaroslav D. Sergeyev and Alfredo Garro
0, 1, 2, 0, 1, 2, 0, 1, 2, . . . 0, 1, 2, 0, 1, 2 . (22)
| {z }
① positions
Then ideal Turing machines working with the output alphabet {0, 1} cannot produce
observable in a sequence outputs computing (22).
Since the numeral 2 does not belong to the alphabet {0, 1} it should be coded by
more than one symbol. One of codifications using the minimal number of symbols
in the alphabet {0, 1} necessary to code numbers 0, 1, 2 is {00, 01, 10}. Then the
output corresponding to (22) and computed in this codification should be
00, 01, 10, 00, 01, 10, 00, 01, 10, . . . 00, 01, 10, 00, 01, 10. (23)
Since the output (22) contains grossone positions, the output (23) should contain
2① positions. However, in order to be observable in a sequence, (23) should not
have more than grossone positions. This fact completes the proof. 2
The mathematical language used by Turing did not allow one to distinguish these
two machines. Now we are able to distinguish a machine from another also when
we consider infinite sequences. Turing’s results and the new ones do not contradict
each other. Both languages observe and describe the same object (computable se-
quences) but with different accuracies.
It is not possible to describe a Turing machine (the object of the study) without
the usage of a numeral system (the instrument of the study). As a result, it becomes
not possible to speak about an absolute number of all possible Turing machines T I .
It is always necessary to speak about the number of all possible Turing machines
T I expressible in a fixed numeral system (or in a group of them).
Theorem 4. The maximal number of complete computable sequences produced by
ideal Turing machines that can be enumerated in a sequence is equal to ①.
We have established that the number of complete computable sequences that can
be computed using a fixed radix b is less or equal b① . However, we do not know
how many of them can be results of computations of a Turing machine. Turing es-
tablishes that their number is enumerable. In order to obtain this result, he used the
mathematical language developed by Cantor and this language did not allow him
to distinguish sets having different infinite numbers of elements. The introduction
of grossone gives a possibility to execute a more precise analysis and to distinguish
within enumerable sets infinite sets having different numbers of elements. For in-
stance, the set of even numbers has ① 2 elements and the set of integer numbers has
2① + 1 elements. If the number of complete computable sequences, MT I , is larger
than ①, then there can be differen sequential processes that enumerate different se-
quences of complete computable sequences. In any case, each of these enumerating
sequential processes cannot contain more than grossone members.
A new perspective on Turing machines 19
Proof. Due to the definition of the complete sequence, on each tape at maximum
① symbols can be produced and observed and thus by considering a complete se-
quence on each of the k tapes of the machine the complete output of the machine
will result in k① symbols. 2
Having proved that a complete output that can be produced by a k-tapes Turing
machine results in k① symbols, it is interesting to investigate what part of the com-
plete output produced by the machine can be observed in a sequence taking into
account that it is not possible to observe in a sequence more than ① symbols (see
Section 3). As examples, we can decide to make in a sequence one of the following
observations: (i) ① symbols on one among the k-tapes of the machine, (ii) ① k sym-
bols on each of the k-tapes of the machine; (iii) ① symbols on 2 among the k-tapes
2
of the machine, an so on.
D E
Theorem 6. Let M KI = Q, Γ, b̄, Σ, q0 , F, δ(k) be an ideal k-tapes, k ≥ 2, Turing
machine. Let M be the number of all possible complete outputs that can be produced
by M KI . Then it follows M = ∏ki=1 (bi + 1)① .
Proof. Due to the definition of the complete sequence, on each tape i, with 1 ≤ i ≤
k, at maximum ① symbols can be produced and observed by using the bi symbols
of the alphabet Σi of the tape plus the blank symbol (b̄); as a consequence, the
number of all the possible complete sequences that can be produced and observed
on a tape i is (bi + 1)① . A complete output of the machine is obtained by considering
a complete sequence on each of the the k-tapes of the machine, thus by considering
all the possible complete sequences that can be produced and observed on each of
the k tapes of the machine, the number M of all the possible complete outputs will
results in ∏ki=1 (bi + 1)① . 2
As the number M = ∏ki=1 (bi + 1)① of complete outputs that can be produced
by M K is larger than grossone, then there can be different sequential enumerating
processes that enumerate complete outputs in different ways, in any case, each of
these enumerating sequential processes cannot contain more than grossone members
(see Section 3).
In the classical framework ideal k-tape Turing machines have the same computa-
tional power of Single-tape Turing machines and given a Multi-tape Turing ma-
chine M KI it is always possible to define a Single-tape Turing machine which is able
to fully simulate its behavior and therefore to completely execute its computations.
As showed for Single-tape Turing machine (see [42]), the Grossone methodology
allows us to give a more accurate definition of the equivalence among different ma-
chines as it provides the possibility not only to separate different classes of infinite
sets with respect to their cardinalities but also to measure the number of elements
A new perspective on Turing machines 21
Let us discuss the traditional and new results regarding the computational power of
deterministic and non-deterministic Turing machines.
Classical results show an exponential growth of the time required for reaching
a final configuration by the deterministic Turing machine M D with respect to the
time required by the non-deterministic Turing machine M N , assuming that the time
required for both machines for a single step is the same. However, in the classic
theory on Turing machines it is not known if there is a more efficient simulation of
M N . In other words, it is an important and open problem of Computer Science theory
to demonstrate that it is not possible to simulate a non-deterministic Turing machine
by a deterministic Turing machine with a sub-exponential numbers of steps.
Let us now return to the new mathematical language. Since the main interest to
non-deterministic Turing machines (5) is related to their theoretical properties, here-
inafter we start by a comparison of ideal deterministic Turing machines, T I , with
ideal non-deterministic Turing machines T I N . Physical machines T P and T P N are
considered at the end of this section. By taking into account the results of Section
4.4, the proposed approach can be applied both to single and multi-tape machines,
however, single-tape machines are considered in the following.
Due to the analysis made in Section 4.3, we should choose the accuracy (gran-
ularity) of processes of observation of both machines, T I and T I N . In order to be
close as much as possible to the traditional results, we consider again an applica-
tion of the transition function of the machine as our observation granularity. With
respect to T I N this means that the nodes of the computational tree are observed.
With respect to T I we consider sequences of such nodes. For both cases the ini-
tial configuration is not observed, i.e., we start our observations from level 1 of the
computational tree.
This choice of the observation granularity is particularly attractive due to its ac-
cordance with the traditional definitions of Turing machines (see definitions (1) and
(5)). A more fine granularity of observations allowing us to follow internal oper-
ations of the machines can be also chosen but is not so convenient. In fact, such
an accuracy would mix internal operations of the machines with operations of the
algorithm that is executed. A coarser granularity could be considered, as well. For
instance, we could define as a computational step two consecutive applications of
the transition function of the machine. However, in this case we do not observe all
the nodes of the computational tree. As a consequence, we could miss some results
of the computation as the machine could reach a final configuration before complet-
ing an observed computational step and we are not able to observe when and on
which configuration the machine stopped. Then, fixed the chosen level of granular-
ity the following result holds immediately.
Theorem 8. (i) With the chosen level of granularity no more than ① computational
steps of the machine T I can be observed in a sequence. (ii) In order to give possi-
bility to observe at least one computational path of the computational tree of T I N
A new perspective on Turing machines 23
Fig. 2 The maximum number of computational steps of the machine T I that can be observed in a
sequence
from the level 1 to the level k, the depth, k ≥ 1, of the computational tree cannot be
larger than grossone, i.e., k ≤ ①.
Proof. Both results follow from the analysis made in Section 3.1 and Theorem 1.
2
In Figure 2 the first result of Theorem 8 concerning the maximum number of
computational steps of the machine T I that can be observed in a sequence is ex-
emplified with reference to the computational tree of the machine introduced in
Section 2.3.
Similarly, the second result of Theorem 8 concerning the depth of the computational
tree of T I N is exemplified in Figure 3.
Corollary 4. Suppose that d is the non-deterministic degree of T I N and S is the
number of leaf nodes of the computational tree with a depth k representing the pos-
sible results of the computation of T I N . Then it is not possible to observe all S
possible results of the computation of T I N if the computational tree of T I N is
complete and d k >①.
Proof. For the number of leaf nodes of the tree, S, of a generic non-deterministic
Turing machine T I N the estimate S ≤ d k holds. In particular, S = d k if the computa-
tional tree is complete, that is our case. On the other hand, it follows from Theorem 1
24 Yaroslav D. Sergeyev and Alfredo Garro
that any sequence of observations cannot have more than grossone elements. As a
consequence, the same limitation holds for the sequence of observations of the leaf
nodes of the computational tree. This means that we are not able to observe all the
possible results of the computation of our non-deterministic Turing machine T I N
if d k >①. 2
In Figure 4 the result of Corollary 4 concerning the maximum number of compu-
tational results of the machine T I that can be observed in a sequence is exemplified
with reference to the computational tree of the machine introduced in Section 2.3 .
steps (note that if the computational tree of T I N is not complete, T I will execute
less than KT I ). Due to Theorems 1 and 8, and Corollary 5, it follows that in order
26 Yaroslav D. Sergeyev and Alfredo Garro
to prove the theorem it is sufficient to show that under conditions of the theorem it
follows that
KT I 6 ①. (27)
To do this let us use the well known formula
k
d k+1 − 1
∑ dj = d −1
, (28)
j=0
and derive both parts of (28) with respect to d. As the result we obtain
k
kd k+1 − (k + 1)d k + 1
∑ jd j−1 = (d − 1)2
. (29)
j=1
Notice now that by using (26) it becomes possible to represent the number KT I as
k k
KT I = ∑ jd j = d ∑ jd j−1 .
j=1 j=1
Due to assumption (25), it follows that (27) holds. This fact concludes the proof of
the theorem. 2
Corollary 6. Suppose that the length of the input sequence of symbols of a non-
deterministic Turing machine T I N is equal to a number n and T I N has a complete
computational tree with the depth k such that k = nl , i.e., polynomially depends on
the length n. Then, if the values d, n, and l satisfy the following condition
l l
d(nl d n +1 − (nl + 1)d n + 1)
6 ①, (31)
(d − 1)2
then: (i) there exists a deterministic Turing machine T I that can be observed and
able to simulate T I N ; (ii) the number, KT I , of computational steps required to a
deterministic Turing machine T I to simulate T I N for reaching a final configuration
exponentially depends on n.
Proof. The first assertion follows immediately from theorem 9. Let us prove the
second assertion. Since the computational tree of T I N is complete and has the depth
k, the corresponding deterministic Turing machine T I for simulating T I N will ex-
ecute KT I steps where KT I is from (27). Since condition (31) is satisfied for T I N ,
we can substitute k = nl in (30). As the result of this substitution and (31) we obtain
that
A new perspective on Turing machines 27
l l
d(nl d n +1 − (nl + 1)d n + 1)
KT I = 6 ①, (32)
(d − 1)2
i.e., the number of computational steps required to the deterministic Turing machine
T I to simulate the non-deterministic Turing machine T I N for reaching a final con-
figuration is KT I 6 ① and this number exponentially depends on the length of the
sequence of symbols provided as input to T I N . 2
Results described in this section show that the introduction of the new mathemat-
ical language including grossone allows us to perform a more subtle analysis with
respect to traditional languages and to introduce in the process of this analysis the
figure of the researcher using this language (more precisely, to emphasize the pres-
ence of the researcher in the process of the description of automatic computations).
These results show that there exist limitations for simulating non-deterministic Tur-
ing machines by deterministic ones. These limitations can be viewed now thanks to
the possibility (given because of the introduction of the new numeral ①) to observe
final points of sequential processes for both cases of finite and infinite processes.
Theorems 8, 9, and their corollaries show that the discovered limitations and
relations between deterministic and non-deterministic Turing machines have strong
links with our mathematical abilities to describe automatic computations and to con-
struct models for such descriptions. Again, as it was in the previous cases studied
in this chapter, there is no contradiction with the traditional results because both
approaches give results that are correct with respect to the languages used for the
respective descriptions of automatic computations.
We conclude this section by the note that analogous results can be obtained for
physical machines T P and T P N , as well. In the case of ideal machines, the pos-
sibility of observations was limited by the mathematical languages. In the case of
physical machines they are limited also by technical factors (we remind again the
analogy: the possibilities of observations of physicists are limited by their instru-
ments). In any given moment of time the maximal number of iterations, Kmax , that
can be executed by physical Turing machines can be determined. It depends on the
speed of the fastest machine T P available at the current level of development of
the humanity, on the capacity of its memory, on the time available for simulating
a non-deterministic machine, on the numeral systems known to human beings, etc.
Together with the development of technology this number will increase but it will
remain finite and fixed in any given moment of time. As a result, theorems presented
in this section can be re-written for T P and T P N by substituting grossone with Kmax
in them.
5 Concluding Remarks
Since the beginning of the last century, the fundamental nature of the concept of
automatic computations attracted a great attention of mathematicians and computer
scientists (see [4, 14, 15, 16, 23, 24, 27, 44]). The first studies had as their ref-
28 Yaroslav D. Sergeyev and Alfredo Garro
erence context the David Hilbert programme, and as their reference language that
introduced by Georg Cantor [3]. These approaches lead to different mathematical
models of computing machines (see [1, 6, 9]) that, surprisingly, were discovered
to be equivalent (e.g., anything computable in the λ-calculus is computable by a
Turing machine). Moreover, these results, and expecially those obtained by Alonzo
Church, Alan Turing [4, 10, 44] and Kurt Gödel, gave fundamental contributions to
demonstrate that David Hilbert programme, which was based on the idea that all of
the Mathematics could be precisely axiomatized, cannot be realized.
In spite of this fact, the idea of finding an adequate set of axioms for one or
another field of Mathematics continues to be among the most attractive goals for
contemporary mathematicians. Usually, when it is necessary to define a concept or
an object, logicians try to introduce a number of axioms describing the object in
the absolutely best way. However, it is not clear how to reach this absoluteness;
indeed, when we describe a mathematical object or a concept we are limited by
the expressive capacity of the language we use to make this description. A richer
language allows us to say more about the object and a weaker language – less.
Thus, the continuous development of the mathematical (and not only mathematical)
languages leads to a continuous necessity of a transcription and specification of
axiomatic systems. Second, there is no guarantee that the chosen axiomatic system
defines ‘sufficiently well’ the required concept and a continuous comparison with
practice is required in order to check the goodness of the accepted set of axioms.
However, there cannot be again any guarantee that the new version will be the last
and definitive one. Finally, the third limitation already mentioned above has been
discovered by Gödel in his two famous incompleteness theorems (see [10]).
Starting from these considerations, in the chapter, Single and Multi-tape Turing
machines have been described and observed through the lens of the Grossone lan-
guage and methodology . This new language, differently from the traditional one,
makes it possible to distinguish among infinite sequences of different length so en-
abling a more accurate description of Single and Multi-tape Turing machines. The
possibility to express the length of an infinite sequence explicitly gives the pos-
sibility to establish more accurate results regarding the equivalence of machines in
comparison with the observations that can be done by using the traditional language.
It is worth noting that the traditional results and those presented in the chapter
do not contradict one another. They are just written by using different mathematical
languages having different accuracies. Both mathematical languages observe and
describe the same objects – Turing machines – but with different accuracies. As a
result, both traditional and new results are correct with respect to the mathematical
languages used to express them and correspond to different accuracies of the obser-
vation. This fact is one of the manifestations of the relativity of mathematical results
formulated by using different mathematical languages in the same way as the usage
of a stronger lens in a microscope gives a possibility to distinguish more objects
within an object that seems to be unique when viewed by a weaker lens.
Specifically, the Grossone language has allowed us to give the definition of com-
plete output of a Turing machine, to establish when and how the output of a ma-
chine can be observed, and to establish a more accurate relationship between Multi-
A new perspective on Turing machines 29
tape and Single-tape Turing machines as well as between deterministic and non-
deterministic ones. Future research efforts will be geared to apply the Grossone
language and methodology to the description and observation of new and emerging
computational paradigms.
References
26. P. Pica, C. Lemer, V. Izard, and S. Dehaene. Exact and approximate arithmetic in an amazonian
indigene group. Science, 306(15 October):499–503, 2004.
27. E. Post. Finite combinatory processes – formulation 1. Journal of Symbolic Logic, 1:103–105,
1936.
28. A. Robinson. Non-standard Analysis. Princeton Univ. Press, Princeton, 1996.
29. E.E. Rosinger. Microscopes and telescopes for theoretical physics: How rich locally and large
globally is the geometric straight line? Prespacetime Journal, 2(4):601–624, 2011.
30. Ya.D. Sergeyev. Arithmetic of Infinity. Edizioni Orizzonti Meridionali, CS, 2003.
31. Ya.D. Sergeyev. Blinking fractals and their quantitative analysis using infinite and infinitesi-
mal numbers. Chaos, Solitons & Fractals, 33(1):50–75, 2007.
32. Ya.D. Sergeyev. A new applied approach for executing computations with infinite and in-
finitesimal quantities. Informatica, 19(4):567–596, 2008.
33. Ya.D. Sergeyev. Evaluating the exact infinitesimal values of area of Sierpinski’s carpet and
volume of Menger’s sponge. Chaos, Solitons & Fractals, 42(5):3042–3046, 2009.
34. Ya.D. Sergeyev. Numerical computations and mathematical modelling with infinite and in-
finitesimal numbers. Journal of Applied Mathematics and Computing, 29:177–195, 2009.
35. Ya.D. Sergeyev. Numerical point of view on Calculus for functions assuming finite, infinite,
and infinitesimal values over finite, infinite, and infinitesimal domains. Nonlinear Analysis
Series A: Theory, Methods & Applications, 71(12):e1688–e1707, 2009.
36. Ya.D. Sergeyev. Counting systems and the First Hilbert problem. Nonlinear Analysis Series
A: Theory, Methods & Applications, 72(3-4):1701–1708, 2010.
37. Ya.D. Sergeyev. Lagrange Lecture: Methodology of numerical computations with infinities
and infinitesimals. Rendiconti del Seminario Matematico dell’Università e del Politecnico di
Torino, 68(2):95–113, 2010.
38. Ya.D. Sergeyev. Higher order numerical differentiation on the infinity computer. Optimization
Letters, 5(4):575–585, 2011.
39. Ya.D. Sergeyev. On accuracy of mathematical languages used to deal with the Riemann zeta
function and the Dirichlet eta function. p-Adic Numbers, Ultrametric Analysis and Applica-
tions, 3(2):129–148, 2011.
40. Ya.D. Sergeyev. Using blinking fractals for mathematical modelling of processes of growth in
biological systems. Informatica, 22(4):559–576, 2011.
41. Ya.D. Sergeyev. Solving ordinary differential equations by working with infinitesimals nu-
merically on the infinity computer. Applied Mathematics and Computation, 219(22):10668–
10681, 2013.
42. Ya.D. Sergeyev and A. Garro. Observability of Turing machines: A refinement of the theory
of computation. Informatica, 21(3):425–454, 2010.
43. Ya.D. Sergeyev and A. Garro. Single-tape and Multi-tape Turing Machines through the lens
of the Grossone methodology. The Journal of Supercomputing, 65(2):645–663, 2013.
44. A.M. Turing. On computable numbers, with an application to the entscheidungsproblem.
Proceedings of London Mathematical Society, series 2, 42:230–265, 1936-1937.
45. M.C. Vita, S. De Bartolo, C. Fallico, and M. Veltri. Usage of infinitesimals in the Menger’s
Sponge model of porosity. Applied Mathematics and Computation, 218(16):8187–8196, 2012.
46. J. Wallis. Arithmetica infinitorum. 1656.
47. A.A. Zhigljavsky. Computing sums of conditionally convergent and divergent series using the
concept of Grossone. Applied Mathematics and Computation, 218(16):8064–8076, 2012.
48. A. Žilinskas. On strong homogeneity of two global optimization algorithms based on sta-
tistical models of multimodal objective functions. Applied Mathematics and Computation,
218(16):8131–8136, 2012.