Thermodynamics
Thermodynamics
Faculty of physics
Notes of thermodynamics
Author: LPT
Professor: Marco Aurelio Diaz
March 2021
Contents
1 Appendix 4
1.1 Language of set theory . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.1.1 Other logical conectives . . . . . . . . . . . . . . . . . . . . . 5
1.2 Axioms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.2.1 Axiom of extensionality . . . . . . . . . . . . . . . . . . . . . 6
1.2.2 Axiom of comprehension . . . . . . . . . . . . . . . . . . . . . 6
1.2.3 Sets that can defined by the comprehension axiom . . . . . . 8
1.2.4 Axiom of pairs . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.2.5 Axiom of union . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.2.6 Axiom of the parts . . . . . . . . . . . . . . . . . . . . . . . . 9
1.2.7 Classic sets defined in the ZF language . . . . . . . . . . . . . 9
1.2.8 Axiom of infinity . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.2.9 Principle of induction . . . . . . . . . . . . . . . . . . . . . . 14
2
pag. 3
Appendix
In this part of the document we will analyze in an introductory way some useful
concepts of axiomatic set theory (assuming that the student has knowledge of the
intuitive theory and first-order logical) that will help to understand concepts of
combinatorics and probabilistic theory.
This way, it is ensure the formation of a rigorous theory that does not have
circular reasoning and is based at all times. Any axiomatic theory also presents a
language with logical and non-logical symbols, in particular, the set theory works
with a first-order language, which is formed by:
1) Logical simbols:
-Individual variables commonly represented with the letters x,y,z, among others
-Logical conectives (basics): ¬, ∨, ∃
-Binary equality predicate:=
∃∧ ∈=
4
pag. 5
The latter generates the need to think of a rule that allows to obtain valid
expressions (in the fundamental logic).
With this, we can easily differentiate between formulas and expressions without
logical sense, for example, in the following question:
∃x (x ∈ y) ∨ (z = x)
Answer: By 1, x ∈ y and z = x are formulas, we will call φ to x ∈ y and γ to
z = x, thus, it is easy to see that φ ∨ γ is formula by 2, let’s call it , finally (∃x)
is a formula by 3, therefore, the latter expression is a formula.
1) x 6= y ≡ ¬(x = y)
2) φ ∧ γ ≡ ¬(¬φ ∨ ¬y)
3) φ → γ ≡ ¬φ ∨ γ
4) φ ↔ γ ≡ φ → γ ∧ γ → φ
5) (∀x)φ ≡ ¬((∃x)¬φ)
6) ∃x ∈ y(φ) ≡ ∃x(x ∈ y ∧ φ)
7) x ∈
/ y ≡ ¬(x ∈ y)
A common notation for formulas with free variables is φ(x1 , x2 , ...xn ), where
x1 ,x2 ,..., and xn represent these variables. Another commonly used terminology is
that of open or closed formula, if a formula has no quantifiers it is said to be open,
and if it has no free variables it is called closed.
pag. 6
1.2 Axioms
The axioms that will be presented below are those used in zermelo frankel’s theory
(ZF), these will be taught chronologically trying to present the motivation behind
each one.
Identity:
∀x(x = x)
Sustitution:
Axiom extensionality
∀a∀b [∀x(x ∈ a ↔ x ∈ b) → a = b]
R = {x : x ∈
/ x}
With the definition given at the time, asking the question ”R ∈ R?” would lead
to an obvious contradiction:
R∈R↔R∈
/R
Therefore, mathematical objects such as these cannot be allowed to be sets, so
the idea of creating the concept of class to designate them was born and the axiom
of comprehension was created to differentiate them from sets. (It should be noted
that since every element of the theory represents a set, then classes of the form
A = {x : φ(x)} turn out to be just a notation to refer to a grouping of sets with a
property in common).
pag. 7
Axiom of comprehension
∀x∃y∀z (z ∈ y ↔ z ∈ x ∧ φ(z))
Thus, any class of type A = {x : φ(x)} is a set if and only if there is a set
containing it, this way a set is write (by the axiom) with the notation y = {z ∈ x :
φ(z)}
1) A = B ≡ ∀x(φ(x) ↔ γ(x))
2) A ⊂ B ≡ ∀x(φ(x) → γ(x))
3) A ∩ B ≡ {x : x ∈ A ∧ x ∈ B}
4) A ∪ B ≡ {x : x ∈ A ∨ x ∈ B}
5) A − B ≡ {x : x ∈ A ∧ x ∈
/ B}
The classes that do not belong to a set are called proper classes, it is clear that
russell’s class is evidently a proper class, however there is another class that must be
taken into account as proper, this is the universe, that is, the class that categorizes
all the mathematical elements, defined by:
U ≡ {x : x = x}
This is a proper class, because if it were a set, then russell’s class would belong
to it and therefore it would also be a set, which would generate the contradiction
again.
Another consequence of the axiom of comprehension, is that it allows us to define
the empty set, since given any x, the empty set belong x, therefore is a set, but a
set of the form:
∅ ≡ {z ∈ x : z 6= z}
Exercise: Prove that the empty set is unique with the extensionality and com-
prehension axioms.
Answer: Let a and b be any two sets that are different from each other, such
that:
a = {x ∈ z : x 6= x}
b = {x ∈ y : x 6= x}
by comprehension axiom:
x∈a x∈b
↔ x ∈ z ∧ x 6= x ↔ x ∈ y ∧ x 6= x
↔x∈z∧F ↔x∈y∧F
↔F ↔F
pag. 8
Finally:
F ↔F
↔x∈a↔x∈b
→ a = b (by extensionality)
a ∩ b ≡ {x ∈ a : x ∈ b}
Another set that can be defined by choosing in this case φ(x) = x ∈
/ b, is the
difference set:
a − b ≡ {x ∈ a : x ∈
/ b}
Finally, we can defined a arbitrary intersection of sets. If F 6= ∅ and x ∈ F ,
then:
\
F ≡ {z ∈ x : ∀y ∈ F (z ∈ y)}
(∀x, y)(∃z)(∀a)[a ∈ z ↔ a = x ∨ a = y]
With this axiom can defined the pairs:
{x, y} ≡ {a : a = x ∨ a = y}
{x, x} ≡ {x}
That help to designate the definition of ordered pair:
(a, b) = (c, d) ↔ a = c ∧ b = d
pag. 9
Axiom of union
∀x∃y∀z(z ∈ y ↔ z ⊂ x)
z ⊂ x ≡ ∀u(u ∈ z → u ∈ x)
By the axiom, can be denotated the parts sets y by:
P (x) ≡ {z : z ⊂ x}
a × b = {(x, y) : x ∈ a ∧ y ∈ b}
Notation:
(x, y) ∈ R ≡ xRy
R⊂a×b≡R:a→b
.
Dom(R) = {x ∈ a : ∃y ∈ b(xRy)}
Rang(R) = {y ∈ b : ∃x ∈ a(xRy)}
∀x ∈ a, ∃!y ∈ b, xf y
Notation:
xf y ≡ y = f (x)
Application properties
∀x ∈ a, ∀y ∈ b (f −1 (y) = x ↔ f (x) = y)
Order relations
Let a homogeneous relation R : a → a, then:
1) R is a preorder ↔ R is reflexive and transitive.
2) R is a partial order ↔ R is a preorder and antisymmetric.
3) R is a total order ↔ R is a partial order and total.
Distinguished elements
Let a partial order (a, ≤) it is say that:
1) M is maximun ↔ M ∈ a ∧ ∀x ∈ a (x ≤ M )
2) m in minimum ↔ m ∈ a ∧ ∀x ∈ a (m ≤ x)
Notation:
≤−1 ≡ ≥
3) p is maximal ↔ p ∈ a ∧ ∀x ∈ a (p ≤ a → p = a)
4) q is minimal ↔ q ∈ a ∧ ∀x ∈ a (q ≥ x → q = x)
5) b ⊂ a → [c is upper bound of b ↔ c ∈ a ∧ ∀x ∈ b (x ≤ c)]
6) b ⊂ a → [c is lower bound of b ↔ c ∈ a ∧ ∀x ∈ b (x ≥ c)]
Set of bounds
Let a partial order (a, ≤) and b ⊂ a, then:
1) U B = {c ∈ a : ∀x ∈ b (x ≤ c)}
2) LB = {c ∈ a : ∀x ∈ b (x ≥ c)}
Strict orders
A order (a, <) is strict if:
- < is irreflexive
- < is antisymmetric
- < is transitive
pag. 13
S(X) = X ∪ {X}
∅≡0
Then, the successor of 0 is equal to a set with one element:
S(0) = {0} ≡ 1
And the successor of 1 is equal to a set with two elements:
S(1) = {0, 1} ≡ 2
and so on...
∅ ∈ X ∧ [∀y (y ∈ x → S(y) ∈ x]
Axiom of infinity:
∃X (Ind(X))
With this we can find an important observation, which is that, any inductive set
should have the numbers 0,1,2,... etc, since by definition, 0 will always belong to
them, and therefore 1, 2,... and so on. This allows us to define the naturals as the
following set:
N = {n ∈ J : ∀X (Ind(X) → n ∈ X)}
pag. 14
Proof 1:
In first place, by definition of inductive set we can obtain that:
∀X (Ind(X) → 0 ∈ X)
∴0∈N
And in second place, by definition of natural set, given a n arbitrary:
n∈N
↔ ∀X (Ind(X) → n ∈ X)
→ ∀X (Ind(X) → S(n) ∈ X)
→ S(n) ∈ N
∴ ∀n (n ∈ N → S(n) ∈ N)
Finally, putting 1 and 2 together we get that Ind(N) is a true proposition.
Proof 2:
By definition of natural set, for any inductive set X, the elements n on N also
belong to X, that is:
∀X (Ind(X) → ∀n ∈ N (n ∈ X))
∴ ∀X (Ind(X) → N ⊂ X)
A = {n ∈ N : φ(n)}
This set is a subset of N whose elements comply with the property φ(n), now,
What would happen if A were also an inductive set?, then, the naturals should also
be subsets of A, therefore:
A ⊂ N ∧ Ind(A) → A = N
pag. 15
A = {n ∈ N : φ(n)}
Now, by definition, if A is a inductive set, then:
0 ∈ A ∧ ∀n (n ∈ A → S(n) ∈ A)
Implicitly understanding n as a natural number, then, the above proposition
becomes equivalent to:
∀n (n ∈ A ↔ n ∈ N)
But since n is a natural number, the second proposition is trivially true, obtain
thus:
∀n (φ(n))
Therefore:
In this chapter, the basic concepts of combinatorics and probability theory will be
presented, assuming that the student has knowledge of finite sets (definition and
cardinality), this way they can start with a solid foundation in the course.
Combinatorics is the branch of discrete mathematics that studies the number
of elements in finite sets with different orderings. This is based on two principles,
(demonstrable in set theory, but that in most documents, including this one, are
practically axioms), these are the additive and multiplicative principle.
A ∩ B = ∅ → |A ∪ B| = |A| + |B|
|A × B| = |A| ∗ |B|
16
pag. 17
X n X
Y n
Y
n
|( ) | = | |= 26 = 26n
i=1 i=1
But we need the number of words at most n figures, this is, the number of words
with
Sn P n figures, or n-1 figures, or ..., or one figure, that is to say, the cardinality of
i
i=1 , so that, applying the additive principle, we obtain:
n X n n n n
[ i X Xi X
i
X
i
X 26n+1 − 26
| |= | |= 26 = 26 ± 1 = 26i − 1 =
25
i=1 i=1 i=1 i=1 i=0
Answer
P 2: For this case wePneed the number of n-tuples (xi )ni=1 such that
x1 ∈ −{b} and ∀i ∈ [2; n], xi ∈ , in other words:
X X X X
n n
| −{b} × Xi=2 |=| −{b}| ∗ |Xi=2 | = 25 ∗ 26n−1
Definition 1.2 (Set of functions): It defines the set B A as the set whose
elements are functions from A to B, that is:
B A = {f ⊂ A × B : f is f uncion}
from the definition of a finite set it follows that, let A and B finite set:
φ : B A → B |A| Ψ : B |A| → B A
Where ai are the elements of A. We can prove that φ = Ψ−1 , since knowing
φ ◦ Ψ : B |A| → B |A| and Ψ ◦ φ : B A → B A , it is enough to verify that:
f is surjective
↔ ∀i ∈ [1; m], f −1 ({bi }) 6= ∅
→ ∀i ∈ [1; m], |f −1 ({bi })| =
6 0
↔ ∀i ∈ [1; m], |f −1 ({bi })| ≥ 1
→ m −1 ({b })| ≥ m
P
i=1 |f i
Sn −1
↔ | i=1 f ({bi })| ≥ m
↔ |f −1 ( ni=1 {bi })| ≥ m
S
↔ |f −1 (B)| ≥ m
↔ |A| ≥ m
↔ |A| ≥ |B|
This type of relations helps to solve different types of problems, some of them are
very well known, as for example, the problem of calculating the minimal number of
moves to solve the game of the towers of Hanoi with n washers. The game consist in
its traditional form of three vertical poles, where they are stacked a indeterminate
number of discs with decreasing radius and one perforation in the center of more
diameter that the poles (so that they can be traversed by the latter). At the begin-
ning, all the disks are stacked on one of the poles in a decreasing way from the base
up, this in order to move the tower to another pole following two simple rules:
pag. 20
an = 2an−1 + 1
This is a recurrence, where the initial condition a1 is equal 1. The sucession
generated with this formule is a1 = 1 ; a2 = 3 ; a3 = 7 ; a4 = 15 ; a5 = 31 ; ... etc,
if sum one to any number then, a1 + 1 = 2 ; a2 + 2 = 4 ; a3 + 1 = 8 ; a4 + 1 = 16 ;
a5 + 1 = 32 ; ... etc, so that it is no unreasonable to conjecture that
an = 2n − 1
To check this we will use mathematical induction:
B.S: a1 = 1 = 21 − 1
I.H: an−1 = 2n−1 − 1
I.S:
an−1 = 2n−1 − 1
↔ 2an−1 = 2n − 2
↔ 2an−1 + 1 = 2n − 1
↔ an = 2n − 1
Therefore, the minimal number of moves for solve the problem is 2n − 1.
2.3 Variations
If B is a finite set with elements bi and i ∈ [1; n], it is called variation of n objects
taken of r in r, to all orderly sequence of elements bi with length r, this also can be
understood as all r-tuples formed with elements of B. Sometimes, it is possible to
encounter problems in which it is necessary to know the number of variations of a
set, however, this problem can be solved, both for the case in which the elements of
B can be repeated and for the case in which they cannot.
V Rnr = |B r | = nr
This is equal to the cardinality of the set B A when |A| = r (see page 5), therefore,
a variation with repeticion is also represented as a function f : A → B.
r−1 r−1
Y Y (n − r)! n!
Vnr = n−i= n−i∗ =
(n − r)! (n − r)!
i=0 i=0
Exercise: How many ways can you have 2 rooks attacking each other in chess?
Answer: As the towers move in a straight line, then, a tower attacks another if
it is positioned in the same column or row, therefore, once the position of the first
tower is chosen (64 options), we only have seven options with the same column or
seven with the same row, so that, the number n of ways is :
n = 64 ∗ 7 + 64 ∗ 7 = 896
Proposed exercise: how many ways can you have n towers that not attacking
each other on an nxn board?
pag. 22
2.4 Combinations
The combinations is a selection of r elements of a set A = {ai }ni=1 where no matter
the order, this can be understood as a subset with cardinal equal to r in the case
that elements are no repeated, this way, the number of combinations is equal the
cardinality of existing subsets on A with length r. To know this number one must
take into account that any variation (without repetition) can be obtained by order-
ing the elements of the combinations in all possible ways, then, by multiplicative
principle, since there are nr subsets in A with cardinal r and r! ways of ordering
The previous expression is called the Newton binomial in honor of its discoverer,
and can be proved by induction with the help of pascal’s identity, which indicates
that:
n n n+1
+ =
k k+1 k+1
Proposed exercises:
1) Prove the following identity:
n
X n
= 2n
k
k=0
2) Prove that:
n n
=
k n−k
3) Prove the following identity and find its relation with 1):
|P ([1; n])| = 2n
pag. 24
For example, the multiset {(a, 3), (b, 1), (c, 4)} can be denoted simply as {a3 , b, c4 }.
Thus a combination with the elements of A that have repetition, can be understood
as a multiset m : A → N0 . It can be observed that the cardinality of the set m is
not the same as the one represented by the multiset, therefore:
But this cardinal is equal to the number of ways that k elements can be placed
in n different containers xi :
However, it is not necessary that the compartments are boxes as in the previous
image, the containers can also be defined as a divisions of the objects throught
vertical lines as in the following image:
pag. 25
Where {•k , |n−1 }n+k−1 is the set of sequences with k points (•) and n-1 vertical
lines (|) (i.e, an n+k-1-tuple). The sequence is defined simply by choosing the coor-
dinates of the n+k-1-tuple that have a vertical line, if the coordinates are considered
as elements of a set, the number of sequence is equal to the number of selections of
n-1 coordinates for a total of n+k-1-tuples, this is:
n+k−1 n+k−1 n
= =
n−1 k k
Exercise: Calculate the number of solutions to the following inequation:
x1 + x2 + x3 ≤ 7 : x1 , x2 , x3 ∈ N0
Answer: knowing that:
7
_
x1 + x2 + x3 ≤ 7 ↔ x1 + x2 + x3 = k
k=0
Then, the number of solutions (x1 , x2 , x3 ) is equal to the cardinality of the next
set:
7
[
{(xi )3i=1 ∈ N0 : x1 + x2 + x3 = k}
k=0
These sets are disjoint, therefore, aplying the additive principle, the cardinality
is equal to:
7 7 7
X X 3 X 3+k−1
{(xi )3i=1 ∈ N0 : x1 + x2 + x3 = k} = = = 120
k k
k=0 k=0 k=0
Proposed excercise:
Prove the following property:
k
X n+i n+k+1
=
i k
i=0
pag. 26
2.6 Permutations
The permutation is a specific case of the variation without repetition when n=r,
therefore, the cardinal of the set with all permutations is:
Pn = Vnn = n!
This can also be interpreted as the cardinal of the set Bij(A, B) = {f ∈ B A :
f is bijective} since in this case, the selection is with all the elements of B, so that
any element of B corresponds to a unique f (ai ) (being unique due to the restriction
that indicates the non-repetition of the elements).
With I = {Ij }rj=1 a partition of [1; n] such that |Ij | = m(adj ), this repetitions
number of the value adj we called nj for simplify the notation. With all this, is
possible to note that the number of variations without repetition is equal to the
number of combinations for the partitions I. For calculate this value we cancaltulate
the number of combinations for the elements Ij , where if for I1 exists nn1 possible
results, then for I2 will exist n−n 1
n2 possible results (the indexes already chosen for
n−n1 −n2
I1 cannot be chosen), and for I3 n3 posibble results, ... and so on, in general
we can see that the number of permutation with repetition for a multiset with n
elements and r distinguishable elements adj that are repeated nj times is:
r r r
n − j−1 nk (n − j−1 (n − j−1
P P P
k=1 nk )! nk )!
Y Y Y
k=1 = Pj−1 = Pj k=1
nj
j=1 j=1 (n − k=1 nk − nj )!(nj )! j=1 (n − k=1 nk )!(nj )!
pag. 27
r
(n − j−1
P
1 Y nk )!
= Qr ∗ Pk=1
j
j=1 (nj )! j=1 (n − k=1 nk )!
Q
r aj−1 a1 Pr
But, by telescopic property j=1 aj = ar and recalling that k=1 nk = n,
we finally obtain:
!
1 n! n!
= Qr Pr = Qr
j=1 nj ! n − k=1 nk )! j=1 nj !
Notation:
n n!
= Qr
n1 , n2 , ..., nr j=1 nj !
Exercise: if there are 52! different permutations of the cards in a deck, given
two identical decks, how many permutations are there?
Answer: if there are 52! different permutations in a deck, then the number of
cards is 52, therefore, the number of cards in two decks is 104, in addition, as the
decks are identicals, the cards are repeated 2 by 2 a total of 52 times, thus, the
number of permutations is:
104!
252
(Ω, Σ, P )
Where it is called Ω to the sample space, which is an arbitrary set interpreted as
one that contains all possible outcomes of a random experiment; on the other hand
Σ is the called σ − algebra this is a family of subsets of Ω defined by the following
rules:
pag. 28
1- Ω ∈ Σ
2- A ∈ Σ → Ac ∈ Σ
S
3- ∀i ∈ N, Ai ∈ Σ → i∈N Ai ∈Σ
These rules guarantee that sets generated by the basic operations also belong to
Σ:
Exercise: Prove the next proposition:
\
∀i ∈ N, Ai ∈ Σ → Ai ∈ Σ
i∈N
Proof:
∀i ∈ N, Ai ∈ Σ
→ ∀i ∈ N, Aci ∈ Σ
→ i∈N Aci ∈ Σ
S
T c
→ i∈N Ai ∈Σ
T
→ i∈N Ai ∈ Σ
\
∴ ∀i ∈ N, Ai ∈ Σ → Ai ∈ Σ
i∈N
This mathematical structure also allows us to separate from the sample space
those sets that interest us to calculate their probability, these sets are called events,
and when a result w of the experiment is measured, it is said that the event A
happened if and only if w ∈ A, so the probability theory is intimately related to the
set theory.
Finally, P is known as the probability measure, and is a function governed by
kolmorov’s axioms, formally:
P :Σ→R:
1) 0 ≤ P (S)
2) P (Ω) = 1
3) ∀i 6= j ∈ N, Ai ∩ Aj = ∅ → P ( ∞
S P∞
i=1 Ai ) = i=1 P (Ai )
With all this in mind, we can move on to some important properties and defini-
tions that will be constantly addressed in the course of this document.
pag. 29
3) P (Ac ) = 1 − P (A)
4) A ⊂ B → P (B − A) = P (B) − P (A)
5) A ⊂ B → P (A) ≤ P (B)
6) 0 ≤ P (A) ≤ 1
7) P (A ∪ B) = P (A) + P (B) − P (A ∩ B)
8) P (A ∪ B) ≤ P (A) + P (B)
Proofs:
1) It is possible to write the empty set as a succession of sets of the following
form:
∅ = ∅ ∪ ∅ ∪ ....
we can observe that all the elements of the sequence are disjoint, therefore, using
the second axiom of Kolmorov we can obtain that:
∞
X
P (∅) = P (∅)
i=1
But by definition of probabilitie this serie should converge, which will occur only
when P (∅) = 0
therefore:
∞ ∞ n ∞ n n
! !
[ X X X X [
P Ai = P (Ai ) = P (Ai ) + P (Ai ) = P (Ai ) = P Ai
i=1 i=1 i=1 i=n+1 i=1 i=1
3) Since A and Ac are disjoints, using the second propertie, we can obtain that:
P (A ∪ Ac ) = P (A) + P (Ac )
But A ∪ Ac = Ω and P (Ω) = 1 by the first axiom of Kolmorov, therefore:
4)
P (B) = P ((B − A) ∪ A) = P (B − A) + P (A)
↔ P (B − A) = P (B) − P (A)
0 ≤ P (B − A)
Adding P(A):
P (A) ≤ P (B − A) + P (A)
Using the propertie 4:
P (A) ≤ P (B)
0 ≤ P (A)
And by definition of Σ, ∀A ∈ Σ, A ⊂ Ω, then, using the propertie 5:
P (A) ≤ P (Ω)
↔ P (A) ≤ 1
1) C = A ∩ B c
2) D = A ∩ B
3) E = Ac ∩ B
We can be noted that:
C ∩D =∅
C ∩E =∅
D∩E =∅
In addition:
C ∪ D ∪ E = (A ∩ B c ) ∪ (A ∩ B) ∪ (Ac ∩ B)
= (A ∪ (A ∩ B)) ∩ (B c ∪ (A ∩ B)) ∪ (Ac ∩ B)
= (A ∩ ((B c ∪ A) ∩ Ω) ∪ (Ac ∩ B)
= (A ∩ (B c ∪ A)) ∪ (Ac ∩ B)
= (A ∪ (Ac ∩ B))
=A∪B
pag. 31
So that:
C ∪D =A
D∪E =B
Therefore:
P (A ∩ B) ≥ 0
Adding P (A ∪ B):
P (A ∩ B) + P (A ∪ B) ≥ P (A ∪ B)
Using the propertie 7:
P (A) + P (B) ≥ P (A ∪ B)
2) ni=1 P ({wi }) = 1
P
P
3) P (A) = w∈A P ({w})
Proof:
1) Kolmorov’s axiom 1 is fulfilled by property 1.
3) Since no element of the sample space belongs to ∅, then by the properties two
and three, we can state that P (∅) = 0, thus, we can defined the sucession {Ai } such
that ∀i > n, Ai = ∅ and ∀i ≤ n, Ai are disjoint elements in 2Ω , then, if we calculate
the probability of the union of the elements belonging to the image of the sequence
we can notice that:
∞
[ X ∞ X
X ∞
X
P( Ai ) = P ({w}) = P ({w}) = P (Ai )
i=1 w∈ ∞ i=1 w∈Ai i=1
S
i=1 Ai
∗B
f (B) =
n
∗(B ∩ A)
f (B ∩ A) =
n
Where ∗ is used by represents the number of times the result was obtained. On
the other hand, the frequency of A given that B occurred should be the number
of times that A and B occurred among all the times that B occurred, this can be
understood intuitively as if upon the occurrence of B, the sample space is restricted
to only the values that can intersect with it, so it is only necessary to focus on the
repetition of these values among all the times that event B occurred, therefore:
∗B ∩ A ∗B ∩ A/n f (B ∩ A)
f (A|B) = = =
∗B ∗B/n f (B)
Finally, since the probability is the frequency to which an event tends when the
experiment has been repeated a large number of times (P (A) = lim ∗A/n), then
n→∞
the probability of A happening given that B is:
P (B ∩ A)
P (A|B) =
P (B)
Exercise: Given a probabilistic space (Ω, Σ, P ), prove that the 3-tuple (Ω, Σ, P (X|B)),
where X, B ∈ Σ (for all X and P (B) 6= 0), is a probabilistic space.
Answer:
1) ∀X ∈ Σ, P (B ∩ X) ≥ 0 ∧ P (B) > 0
P (B ∩ X)
→ ∀X ∈ Σ, = P (X|B) ≥ 0
P (B)
P (B ∩ Ω) P (B)
2) P (Ω|B) = = =1
P (B) P (B)
pag. 34
∞ S∞
P( ∞
S P∞
[ P( i=1 Xi ∩ B) i=1 (Xi ∩ B)) P (Xi ∩ B)
P( Xi |B) = = = i=1
P (B) P (B) P (B)
i=1
∞ ∞
X P (Xi ∩ B) X
= = P (Xi |B)
P (B)
i=1 i=1
As a corollary of the above formula, it follows that:
P (A|B) = P (A)
P (A ∩ Ω)
P (A|Ω) = = P (A)
P (Ω)
Therefore:
∞ ∞
! S∞ !
P( i=1 Ci ∩ A)
[ [
P (A) = P (A|Ω) = P A| Ci = =P (Ci ∩ A)
P (Ω)
i=1 i=1
pag. 35
∞ ∞ ∞
X X P (Ci ∩ A) X
= P (Ci ∩ A) = ∗ P (Ci ) = P (A|Ci ) ∗ P (Ci )
P (Ci )
i=1 i=1 i=1
Bayes’ Theorem
Given a probability space (Ω, Σ, P ) and a partition C = {Ci }∞
i=1 of Ω (with
P (Ci ) > 0 for all i), then, for all B ∈ Σ such that P (B) > 0:
P (B|Ci ) ∗ P (Ci )
∀i ∈ [1; n], P (Ci |B) = P∞
j=1 P (B|Cj ) ∗ P (Cj )
P (B ∩ Ci )
P (Ci |B) =
P (B)
But remembering the total probability theorem, we have that:
∞
X
P (B) = P (B|Cj ) ∗ P (Cj )
j=1
P (B ∩ Ci ) = P (B|Ci ) ∗ P (Ci )
Then, replacing:
P (B|Ci ) ∗ P (Ci )
P (Ci |B) = P∞
j=1 P (B|Cj ) ∗ P (Cj )
Proof:
Suppose we have any set of indices I such that ∀i ∈ I, Σi is a σ − algebra of Ω,
if F is the following class:
F = {Σi : i ∈ I}
Then, the arbitrary intersection of the elements of F is:
\
F = {A : ∀i ∈ I, A ∈ Σi }
1) ∀i ∈ I, Σi is a σ − algebra of Ω
→ ∀i ∈ I, Ω ∈ Σi
T
↔Ω∈ F
pag. 36
T
2) A ∈ F
↔ ∀i ∈ I, A ∈ Σi
→ ∀i ∈ I, Ac ∈ Σi
↔ Ac ∈ F
T
T
3) ∀j ∈ N, Aj ∈ F
↔ ∀j ∈ N, ∀i ∈ I, Aj ∈ Σi
↔ ∀i ∈ I, ∀j ∈ N, Aj ∈ Σi
S
→ ∀i ∈ I, j∈N Aj ∈ Σi
S T
↔ j∈N Aj ∈ F
With this property already demonstrated, we can define the following set:
∴ σ({A}) = {A, Ac , Ω, ∅}
pag. 37
1- (x, ∞) 6- (x, y]
2- (−∞, x) 7- (x, y)
3- [x, ∞) 8- {x}
4- [x, y] 9- N, Z, Q, Qc
5- [x, y) 10- R
1) (−∞, x] ∈ B(R)
→ (−∞, x]c ∈ B(R)
↔ (x, ∞) ∈ B(R)
In other hand, we know that lim x − 1/n = x, then, ∃n0 such that for arbitrary
n→∞
y ∈ (−∞, x), y < x − 1/n0 , so:
y ∈ (−∞, x − 1/n0 ]
∞
[
→y∈ (−∞, x − 1/n]
i=1
∞
[
∴ (−∞, x − 1/n) = (−∞, x) ∈ B(R)
i=1
3) By the property 2:
(−∞, x) ∈ B(R)
→ (−∞, x)c ∈ B(R)
↔ [x, ∞) ∈ B(R)
pag. 38
8) If x=y, by property 4:
[x, y] ∈ B(R)
↔ [x, x] ∈ B(R)
↔ {x} ∈ B(R)
9) By the property 8:
S∞
i=1 {i} ∈ B(R)
↔ N ∈ B(R)
On the other hand, we can define the index sets E = {2, 4, ...} and O = {1, 3, ...},
such that:
[ ∞
[
F = zi ∈ B(R)
i=1
pag. 39
But:
∞
[
zi = Z
i=1
Therefore, Z ∈ B(R).
∞
[
→ {mi /j} ∈ B(R)
j=1
∞ [
[ ∞
→ {mi /j} ∈ B(R)
i=1 j=1
But:
∞ [
[ ∞
{mi /j} = Q
i=1 j=1
10) Finally, the R = Ω, then, R ∈ B(R) by the first property of the σ − algebra.
PX (B) = P (X −1 (B))
X −1 (B) = {w ∈ Ω : X(w) ∈ B}
pag. 40
Then, the latter can be denoted as X ∈ B, and say that the probability that
event X −1 (B) occurs is equal to the probability that X ∈ B, since knowing that for
a arbitrary result w, w ∈ Ω, it follows that:
X −1 (B) ∈ Σ ↔ ∀x ∈ R, (X ≤ x) ∈ Σ
Proof:
⇒) As {(−∞, x] : x ∈ R} ⊂ B(R), if for any B, X −1 (B) ∈ Σ, then in specific,
for all x ∈ R, X ≤ x ∈ Σ.
β = {B ∈ B(R) : X −1 (B) ∈ Σ}
γ = {(−∞, x] : x ∈ R}
Because of the hypothesis, for any x ∈ R, (−∞, x] ∈ Σ, then, the intervals are
also contained in β, therefore γ ⊂ β. In other hand by definition of β, β ⊂ B(R) =
σ(γ). Thus, in order to prove what has been proposed, we must make sure that
β = B(R), but with the above deduction this is equivalent to saying that B(R) ⊂ β,
however, we know that γ ⊂ β, therefore, if we prove that β is a sigma algebra, then
we prove that the sigma algebra generated from γ ( i.e. B(R)) is a subset of β, since
σ(γ) is minimal by definition.
2) Given a arbitrary B:
B∈β
↔ B ∈ B(R) ∧ X −1 (B) ∈ Σ
→ B c ∈ B(R) ∧ (X −1 (B))c ∈ Σ
pag. 41
1) Z = C 4) XY
2) CX 5) X/Y : Y 6= 0
3) X + Y 6) |X|
Proofs:
1) Since Z is a function and Z=C, then the preimage of any Borelian containing
C must be the domain of the function, that is, given a B ∈ B(R):
C ∈ B → Z −1 (B) = Ω
On the other hand, if C does not belong to the Borelian then no domain value
will be associated with it, therefore:
/ B → Z −1 (B) = ∅
C∈
pag. 42
Finally, since these are the only values that the preimages can take and belong
to sigma, then Z is a random variable.
∀x ∈ R, (CX ≤ x) ∈ Σ
(or their equivalent expressions) Given a arbitrary real x, we have 3 cases:
1- C > 0:
CX ≤ x
x
↔ X≤ ∈ Σ (because X is a random variable)
C
2- C < 0
CX ≤ x
x
= (X ≥ )∈Σ
C
3- C = 0
→ CX = 0
Therefore CX is a random variable (By property 1).
Since X and Y are random variables then X > r and Y > x − r are events, then,
their intersection is also, and finally their numerable union (since the rationals are
a numerable set as seen on page 38) also, so it suffices to demonstrate this identity
to prove the proposed property.
⊂): Given a arbitrary w ∈ Ω
X(w) + Y (w) > x
↔ X(w) > x − Y (w)
→ X(w) > r0 ∧ r0 > x − Y (w) (by the density of rationals)
↔ X(w) > r0 ∧ Y (w) > x − r0
→ ∃r ∈ Q, X(w) > r ∧ Y (w) > x − r
S
↔ w ∈ r∈Q (X > r) ∩ (Y > x − r)
(X + Y )2 − (X − Y )2
XY =
4
And this latter is a random variable.
1
= ≤x∩Y <0
Y
1
=Y ≥ ∩Y <0
x
1
= Y ∈ [ , 0) ∈ Σ
x
3) Case x=0:
1
≤0
Y
1 1
= ≤0∩Y >0 ∪ ≤0∩Y <0
Y Y
1
= ≤0∩Y <0
Y
= (Y < 0) ∈ Σ
6) Case x ≥ 0:
|X| ≤ x
= (−x ≤ X ≤ x) ∈ Σ
Case x < 0:
As for all w ∈ Ω, |X(w)| ≥ 0, then X −1 (−∞, 0) = ∅, therefore is a event.
The distribution function can also be defined on the basis of their characteristic
properties, which are as follows:
1) lim F (x) = 1
x→+∞
2) lim F (x) = 0
x→−∞
3) x1 ≤ x2 → F (x1 ) ≤ F (x2 )
4) ∀x0 ∈ R, lim F (x) = F (x0 )
x→x+
0
pag. 45
Proofs:
1) To test this we will use the following property:
Given a increasing sequence {An }, then:
∞
[
P( An ) = lim P (An )
n→∞
n=1
Proof 1: There are two cases of trichotomy, j < i or i < j, in the first case it
follows that Aj−1 ⊂ Aj ⊂ Ai−1 ⊂ Ai since the sequence is increasing, then:
Bi ∩ Bj
= Ai ∩ Aci−1 ∩ Aj ∩ Acj−1
= Ai ∩ Acj−1 ∩ Aci−1 ∩ Ai−1
=∅
On the other hand, the case i < j implies Ai−1 ⊂ Ai ⊂ Aj−1 ⊂ Aj obtaining
analogously the same result.
Proof 2:
B.S:
[ [
An = Bn = B1 = A1
n∈{1} n∈{1}
I.H:
[ [
An = Bn
n∈[1;m] n∈[1;m]
I.S:
[ [ [
Bn = Bn ∪ Bm+1 = An ∪ Bm+1
n∈[1;m+1] n∈[1;m] n∈[1;m]
[ [ [
= An ∪ (Am+1 ∩ Acm ) = An ∩ An ∪ Acm
n∈[1;m] n∈[1;m+1] n∈[1;m]
[ [
= An ∩ Ω = An
n∈[1;m+1] n∈[1;m+1]
pag. 46
Then:
[ [
∀m ∈ N, Bn = An
n∈[1;m] n∈[1;m]
[ [
∴ An = Bn
n∈N n∈N
Thus:
∞
[ ∞
[ ∞
X ∞
X
P( An ) = P (B1 ∪ Bn ) = P (B1 ) + P (Bn ) = P (A1 ) + P (An − An−1 )
n=1 n=2 n=2 n=2
m
X
= P (A1 ) + lim P (An ) − P (An−1 ) = P (A1 ) + lim P (Am ) − P (A1 )
m→∞ m→∞
n=2
= lim P (Am )
m→∞
Finally, Given a sequence of real numbers {xn } that is positively divergent, then:
∞
[
lim F (xn ) = lim P (X ≤ xn ) = lim P (X −1 (−∞, xn )) = P ( X −1 (−∞, xn ))
n→∞ n→∞ n→∞
n=1
P (Ω) = 1
∴ lim F (x) = 1
x→∞
With this, given a sequence of real numbers {xn } that is negatively divergent,
then:
∞
\
lim F (xn ) = lim P (X ≤ xn ) = P ( X −1 (−∞, x) = P (∅) = 0
n→∞ n→∞
n=1
∴ lim F (x) = 0
x→−∞
3) x1 ≤ x2
→ X ≤ x1 ⊂ X ≤ x2
→ P (X ≤ x1 ) ≤ P (X ≤ x2 )
↔ F (x1 ) ≤ F (x2 )
Given a sequence of real numbers {xn } decreasing to 0, then it can be seen that:
F (x0 + xn ) = P (X ≤ x0 + xn )
= P (X ≤ x0 ∪ x0 < X ≤ x0 + xn )
= P (X ≤ x0 ) + P (x0 < X ≤ x0 + xn )
= F (x0 ) + P (x0 < X ≤ x0 + xn )
Then:
∞
\
= F (x0 ) + P ( x0 < X ≤ x0 + xn ) = F (x0 ) + P (∅) = F (x0 )
n=1
Proofs:
1) Given a real sequence {xn } decreasing to 0, then {a − xn } is a real sequence
increasing to a a, therefore:
∞
[
lim F (x) = lim F (a−xn ) = lim P (X ≤ a−xn ) = P ( X ≤ a−xn ) = P (X < a)
x→a− n→∞ n→∞
n=1
2) By the property 1:
3) Since a < b:
P (a < X ≤ b) = P (X ≤ b ∩ (X ≤ a)c ) = P (X ≤ b − X ≤ a)
= P (X ≤ b) − P (X ≤ a) = F (b) − F (a)
4) By the property 1:
5)
P (a < X < b) = P (X < b − X ≤ a) = P (X < b) − P (X ≤ a)
6)
P (a ≤ X < b) = P (X < b − X < a) = P (X < b) − P (X − a)
Note: First of all, since the function is positive, the probability of an event can
be interpreted as the area under the curve given by the distribution function, on the
other hand, since it is continuous, the properties of F described above are drastically
simplified, obtaining that:
∀x, y ∈ R, P (X ≤ x ∩ Y ≤ y) = P (X ≤ x) ∗ P (Y ≤ y)
On other hand in the discrete case, we have that:
∀x, y ∈ R, P (X = x ∩ Y = y) = P (X = x) ∗ P (Y = y)
As it will be noticed, the probability measure takes values of two random vari-
ables X and Y, in such a case a distribution function can be associated which will
depend on the values taken by these two variables, that is:
FXY : R2 → R
FXY (x, y) = P (X ≤ x ∩ Y ≤ y)
Thus having for the discrete and continuous case distribution functions of the
form:
P (X = x ∩ Y = y) : ∀(x, y) ∈ Rang(X) × Rang(Y )
f (x, y) =
0 : in other case
XX
→ F (x, y) = f (t, r)
t≤x r≤y
Z x Z y
F (x, y) = f (t, r) dt dr
−∞ −∞
Note: Generally, as with multiparticle mechanics, theorem proving will be done
for the discrete case only since the proofs are analogous in the continuous case only
with integrals.
Notation:
P (X = x ∩ Y = y) ≡ P (X = x, Y = y)
Proof:
[ X
P (X = x) = P (X = x, Ω) = P (X = x, Y = y) = P (X = x) ∗ P (Y = y)
y y
X X
P (X = x)P (Y = y) = P (X = x, Y = y)
y y
pag. 52
(X = x)∗
f (x) = P (X = x) = lim
n→∞ n
Then, the expectation E(X) is (in discrete case):
x ∗ (X = x)∗ ∗ (X = x)∗
P
xx
X X X
E(X) = xf (x) = xP (X = x) = lim = lim
x x x
n→∞ n n→∞ n
And since x is the value that the random variable will take, (X = x)∗ the
frequency with which it will be repeated, and n the number of runs of the experiment,
then the value that E will deliver will be a representative of the random variable
associated with the value that will be repeated the most times.
Knowing that:
g −1 ({y}) = {x : g(x) = y}
pag. 54
Then:
X ∈ g −1 ({y}) ↔ Y = y
Thus:
E(Y ) = y yP (X ∈ g −1 ({y}))
P
P P
= y y x∈g−1 ({y}) P (X = x)
P P
= y x∈g−1 ({y}) y P (X = x)
P P
= y x∈g−1 ({y}) g(x) P (X = x)
P
= x g(x) P (X = x)
In the case of having a function of two random variables, these results can be
extended as follows:
X
E(g(X, Y )) = g(x, y) f (x, y)
x,y
Z Z
E(g(X, Y )) = g(x, y) f (x, y) dy dx
R R
Proofs
P
1) E(C) = xx ∗ P (C = x)
= C ∗ P (C = C)
=C
P
2) E(C ∗ X) = x C ∗ xP (X = x, C = C)
P
= C x P (X = x)
= C ∗ E[X]
P
3) E(X + Y ) = x,y (x + y) P (X = x, Y = y)
P P
= x,y x P (X = x, Y = y) + x,y y P (X = x, Y = y)
P P P P
= x x y P (X = x, Y = y) + y y x P (X = x, Y = y)
pag. 55
P P
= x x P (X = x) + y y P (Y = y)
= E(X) + E(Y )
P
4) E(X ∗ Y ) = x,y (x ∗ y) P (X = x, Y = y)
P
= x,y x P (X = x) ∗ y P (Y = y)
P P
= x x P (X = x) ∗ y y P (Y = y)
= E(X) ∗ E(Y )
2.8.16 Variance
The variance represents a measure of dispersion of the values taken by the random
variable with respect to the mean and is defined as:
Proofs:
1) V (C) = E(C 2 ) − E(C)2
= C2 − C2
=0
= C 2 ∗ V (X)
V (X + Y ) = E((X + Y )2 ) − E(X + Y )2
= E(X 2 + 2XY + Y 2 ) − (E(X) + E(Y ))2
= E(X 2 ) + 2E(XY ) + E(Y 2 ) − (E(X)2 + 2E(X) ∗ E(Y ) + E(Y )2 )
= E(X 2 ) + 2E(X)E(Y ) + E(Y 2 ) − (E(X)2 + 2E(X) ∗ E(Y ) + E(Y )2 )
= E(X 2 ) − E(X)2 + E(Y 2 ) − E(Y )2
= V (X) + V (Y )
Note: It is important to mention that squaring the deviation values will result
in a measure of dispersion with units also squared, so that to return to the respective
units of interest it is proposed to calculate the square root of the variance, which is
known as the standard deviation:
p
s= V (X)
pag. 57
: ∀x ∈ {xi }ni=1
1/n
f (x) =
0 : in other case
In this way it can also be noted that:
n
!
X 1 1 X
E[X] = x = xi = A.M ({xi }ni=1 )
x
n n
i=1
n
X 1 1X
V (X) = E[(X − E[X])2 ] = (x − A.M ({xi }ni=1 ))2 = (xi − A.M ({xi }ni=1 ))2
x
n n
i=1
In addition, if the set {xi }ni=1 is ordered in the form ∀i < j, xi < xj , then, for
i ∈ [1; n − 1] :
0 : ∀x < x1
F (x) = i/n : ∀x ∈ [xi , xi+1 )
1 : ∀x ≥ xn
px (1 − p)1−x : ∀x ∈ {0, 1}
f (x) =
0 : in other case
So that, P (X = 1) = p and P (X = 0) = (1−p). As for the distribution function,
it can be deduced that:
0 : ∀x < 0
F (x) = 1−p : ∀x ∈ [0, 1)
1 : ∀x ≥ 1
In addtion:
E(X) = 0 ∗ (1 − p) + 1(p) = p
X : {S, F }n → R
Such that X indicates the number of times the trial was successful. That said, it
can be seen that Rang(X) = [0; n], on other hand, to know the probability P(X=k)
with k ∈ [0; n] we will perform the following analysis. First we should note that the
events are mutually independent, so the probability of the intersection will be the
product of the probabilities, so we can calculate the probability of obtaining first k
consecutive successes and then n-k failures, which will give a total of :
pk (1 − p)n−k
However, since only the number of successes matters, it is necessary to multiply
by the number of combinations in which a sequence with k successes can be obtained,
therefore:
n k
P (X = k) = p (1 − p)n−k
k
thus obtaining the density function:
n x
x p (1 − p)n−x : ∀x ∈ [0; n]
f (x) =
0 : in other case
And with the latter, its distribution function (with i ∈ [0; n − 1]):
0 : ∀x < 0
P
i n
k n−k
F (x) =
k=0 k p (1 − p) : ∀x ∈ [i, i + 1)
: ∀x ≥ n
1
X ∼ Bin(n, p)
And as for the expectation and variance of this variable, they can be calculated
as follows:
n n
X n x n−x
X n x
E(X) = x p (1 − p) = x p (1 − p)n−x
x x
x=0 x=1
n n
X n! X n!
= x px (1 − p)n−x = px (1 − p)n−x
(n − x)!(x)! (n − x)!(x − 1)!
x=1 x=1
n n
X (n − 1)! X (n − 1)!
= n∗ px (1−p)n−x = n∗ px (1−p)n−x
(n − x)!(x − 1)! ((n − 1) − (x − 1))!(x − 1)!
x=1 x=1
n n−1
X
X n−1 x n−x n − 1 x+1
=n∗ p (1 − p) =n∗ p (1 − p)n−(x+1)
x−1 x
x=1 x=0
n−1
X
n−1 x
= np ∗ p (1 − p)(n−1)−x = np ∗ (p + (1 − p))n−1 = np
x
x=0
Then:
V (X) = (np)2 − np2 + np − (np)2
= np − np2
= np(1 − p)
pag. 61
(1 − p)x p : ∀x ∈ N0
f (x) =
0 : in other case
As for its distribution function, for its construction we will first perform the
deduction with F (x) for x ∈ N0 and then we will complement it with other sections,
then:
x x
1 − (1 − p)x+1
X X X
t t
F (x) = f (t) = (1 − p) p = p ∗ (1 − p) = p ∗
1 − (1 − p)
t≤x t=0 t=0
= 1 − (1 − p)x+1
Therefore, for i ∈ N0 :
0 : ∀x < 0
F (x) =
1 − (1 − p)i+1 : ∀x ∈ [i, i + 1)
The expectation and variance of the random variable can be calculated as follows:
E(X) = ∞ x
P
x=0 x (1 − p) p
= (p)(1 − p) ∗ ∞ x−1
P
x=0 x(1 − p)
P∞ d
= −(p)(1 − p) ∗ x=0 (1 − p)x
dp
pag. 62
d P∞
= −(p)(1 − p) ∗ (1 − p)x
dp x=0
d 1
= −(p)(1 − p) ∗
dp (1 − p) − 1
d 1
= −(p)(1 − p) ∗
dp p
1−p
=
p
d2 P∞
= (p)(1 − p)2 ∗ (1 − p)x
dp2 x=0
d2 1
= (p)(1 − p)2 ∗ 2
dp p
2
= (p)(1 − p)2
p3
(1 − p)2
=2∗
p2
Then:
1 − p (1 − p)2
V (X) = +
p p2
1−p
=
p2
Note: It should be noted that the above distribution can also be described as a
function of the naturals, for which the random variable X must be determined as
the number of attempts necessary to obtain a success, so that:
(1 − p)x−1 p
: ∀x ∈ N
f (x) =
0 : in other case
And under the same reasoning as above, it can also be shown that:
1 1−p
E(X) = ; V (X) =
p p2
pag. 63
0 : ∀x < 1
F (x) =
1 − (1 − p)i : ∀x ∈ [i, i + 1)
With i ∈ N.
Notation: X ∼ Geo(p).
x−1
!
λ x λ n−x
n! 1 Y
f (x) = n→∞
lim px (1−p)n−x = lim n−k 1−
p→0
(n − x)!(x)! x! n→∞
p→0
n n
k=0
x−1
!
λx λ n λ −x λx e−λ
Y n−k
= lim lim 1 − lim 1 − =
x! n→∞
p→0
nx n→∞
p→0
n n→∞
p→0
n x!
k=0
Then, under such conditions, the random variable is said to have a Poisson
distribution with parameter lambda, denoted as X ∼ P oi(λ) (with λ ∈ R), on the
other hand, complementing, its density function can be described by:
x −λ
λ e
: ∀x ∈ N0
f (x) = x!
0 : in other case
And the distribution function (with i ∈ N0 ) by:
0 : ∀x < 0
F (x) =
Pi λk e−λ
k=0 : ∀x ∈ [i, i + 1)
k!
pag. 64
But:
P∞ e−λ λx
E(X(X − 1)) = x=0 x(x − 1)
x!
P∞ λx
= e−λ x=0 x(x − 1)
x!
λx−2
= e−λ λ2 ∞
P
x=0 x(x − 1)
x!
2
x
d λ
= e−λ λ2 ∞
P
x=0 2
dλ x!
pag. 65
d2 P∞ λx
= e−λ λ2
dλ2 x=0 x!
d2
= e−λ λ2 2 eλ
dλ
=λ 2
Then:
V (X) = λ
Chapter 3
Basic definitions of
thermodynamics
66
pag. 67
In this example, the system is usually the volume of the gas, so the environment
can be reduced to the walls of the cylinder and the piston. Finally, the environment
would be the regions that delimit the gas and allow it not to escape, i.e. the base
of the piston and the inner walls of the cylinder.
The systems are mainly classified into 3 types, open, closed and isolated. On the
one hand, the open system is one that exchanges energy and matter with the envi-
ronment, an example of this could be a pot of boiling water, as energy is exchanged
in the form of heat and matter in the form of water molecules that evaporate and
escape to the environment.
On the other hand, a closed system only exchanges energy, for example, if we
cover the pot with boiling water then it would not allow the gas to escape, however
heat would still enter the system and through the walls of the pot.
Finally, the isolated system is one that exchanges neither matter nor energy,
although these systems do not exist, there are some that significantly reduce the
transfer of energy, such as thermos flasks.