0% found this document useful (0 votes)
171 views124 pages

Intro To Space Technology

This document is a comprehensive introduction to space technology authored by Ignacio Chechile, covering various topics such as artificial satellites, semiconductors, spacecraft configuration, and machine learning applications in space. It includes detailed sections on the design, development, and operation of satellites, as well as frequently asked questions about space. The book serves as a resource for understanding the complexities and innovations within the field of space technology.

Uploaded by

Arijit Mitra
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
171 views124 pages

Intro To Space Technology

This document is a comprehensive introduction to space technology authored by Ignacio Chechile, covering various topics such as artificial satellites, semiconductors, spacecraft configuration, and machine learning applications in space. It includes detailed sections on the design, development, and operation of satellites, as well as frequently asked questions about space. The book serves as a resource for understanding the complexities and innovations within the field of space technology.

Uploaded by

Arijit Mitra
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 124

Ignacio Chechile

Space
Technology
A Short Introduction
Space Technology
Ignacio Chechile

Space Technology
A Short Introduction
Ignacio Chechile
ReOrbit
Helsinki, Finland

ISBN 978-3-031-34817-4 ISBN 978-3-031-34818-1 (eBook)


https://doi.org/10.1007/978-3-031-34818-1

© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Switzerland AG 2023

This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.

This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Contents

1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Space and Startups, a.k.a “NewSpace” . . . . . . . . . . . . . . . . . . . . . . . 2
2 Artificial Satellites; The Shortest Introduction Ever . . . . . . . . . . . . . . . . 7
3 Semiconductors in Space: From Sand to Satellites . . . . . . . . . . . . . . . . . . 11
3.1 Let’s Meet at the Junction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.2 The Transistor Drama . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.3 The Space Environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.3.1 Unpacking Single Event Effects (SEEs) . . . . . . . . . . . . . . . 27
4 The Hectic Ride to Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4.1 Rideshares, Dispensers, and Orbital Transfer Vehicles . . . . . . . . 33
5 Configuring Spacecraft . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
6 A Peek Under the Hood . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
6.1 The Skeleton: Structures and Mechanisms . . . . . . . . . . . . . . . . . . . . 39
6.2 The Data Links: From Sparks to Mobile Networks,
Lasers, and In-Orbit Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
6.2.1 Mobile Networks and Satellites . . . . . . . . . . . . . . . . . . . . . . . 47
6.2.2 Lasers in Orbit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
6.2.3 Connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
6.3 The Software: Hello World in Space . . . . . . . . . . . . . . . . . . . . . . . . . 55
6.3.1 What Does It Take to Run Software
on a Spacecraft? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
6.3.2 How Is Software Updated or Changed in Orbit? . . . . . . . 58
6.3.3 What Type of Languages Are Used for Coding
Flight Software? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
6.3.4 Can I Run Linux on a Satellite? What About
Windows? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
6.3.5 Can I Host a Website on a Satellite? . . . . . . . . . . . . . . . . . . 61
6.3.6 What Kind of Skills Are Required for Doing
Flight Software? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
6.3.7 How Is Flight Software Designed? . . . . . . . . . . . . . . . . . . . . 61

v
vi Contents

6.3.8 What Does “Software-Defined Satellites” Mean? . . . . . . 62


6.3.9 Bugs and Glitches in Orbit . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
6.4 The Orientation: Attitude Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
6.5 The Space Sauna: Thermal Control . . . . . . . . . . . . . . . . . . . . . . . . . . 74
6.6 The Avionics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
6.7 The Payload . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
6.8 Putting It Together: Assembly, Integration and Test . . . . . . . . . . . 82
6.8.1 Mechanical Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
6.8.2 Thermal Vacuum Test (TVAC) . . . . . . . . . . . . . . . . . . . . . . . . 85
6.8.3 Software Verification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
6.8.4 Concluding and Shipping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
7 Satellites and Machine Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
7.1 Can There Be Too Much Data? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
8 Operating Distant Machines Floating in Space . . . . . . . . . . . . . . . . . . . . . 99
9 Making Reliable and Dependable Spacecraft . . . . . . . . . . . . . . . . . . . . . . . 103
10 TL:DR; Frequently Asked Questions About Space . . . . . . . . . . . . . . . . . 109
10.1 Q0: Why Launch a Metal Box into Space? . . . . . . . . . . . . . . . . . . . 109
10.2 Q1: What Are the Rules for Launching Something
into Space? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
10.3 Q2: How Are Satellites Designed and Developed? Also,
Is It Done Differently in NewSpace Versus Classic Space? . . . 110
10.4 Q3: What’s Typically Under the Hood of a Satellite? . . . . . . . . . 110
10.5 Q4: How Are Satellites Launched? . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
10.6 Q5: Ok the Thing Is up in Space, Now What? . . . . . . . . . . . . . . . . 111
10.7 Q6: How Do Satellites Orient Themselves in Space? . . . . . . . . . 111
10.8 Q7: How Are Satellites Operated? . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
10.9 Q8: What Does the Software on Board of a Satellite Do
Exactly? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
10.10 Q9: How Do Satellites Generate Power? . . . . . . . . . . . . . . . . . . . . . 113
10.11 Q10: How Big and Heavy Are Current Commercial
Satellites? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
10.12 Q11: What Is the Lifetime of a Satellite? . . . . . . . . . . . . . . . . . . . . . 114
10.13 Q12: What Can Affect the Lifetime of a Satellite? . . . . . . . . . . . . 115
10.14 Q13: What Is an Orbit Exactly? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
10.15 Q14: Why Are Orbits Crowded and How Is This an Issue? . . . 116
10.16 Q15: Why Are Satellites Assembled in Clean Rooms? . . . . . . . . 116
10.17 Q16: How Are Satellites Currently Distributed Across
Different Orbits? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
List of Figures

Fig. 2.1 Newton’s cannon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8


Fig. 3.1 SiO2 structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Fig. 3.2 A bipolar transistor with one junction in forward-bias
and another one in reverse-bias . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
Fig. 3.3 NOT gate with BJT transistor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
Fig. 3.4 NAND gate with BJT transistor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
Fig. 3.5 Van Allen radiation belts; cross them is not the nicest ride
for a satellite going somewhere (public domain) . . . . . . . . . . . . . . 24
Fig. 3.6 Magnetic field strength at Earth’s surface (Creative
Commons) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
Fig. 3.7 Applicability of SEE to different device types . . . . . . . . . . . . . . . . 27
Fig. 6.1 Junkers J 1 (public domain) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Fig. 6.2 Subsystem faults and types of faults. Source Fault-Tolerant
Attitude Control of Spacecraft (Qinglei Hu, Bing Xiao,
Bo Li, Youmin Zhang) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
Fig. 6.3 Types of faults in the attitude control subsystem. Source
Fault-Tolerant Attitude Control of Spacecraft (Qinglei Hu,
Bing Xiao, Bo Li, Youmin Zhang) . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
Fig. 6.4 Funny little configuration to have by default . . . . . . . . . . . . . . . . . . 65
Fig. 6.5 Good luck running for the tram in Helsinki
without friction forces (Creative Commons) . . . . . . . . . . . . . . . . . . 68
Fig. 6.6 FA and FB are forces applied at different distances
from the center O, creating different torques. Credit
public domain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
Fig. 6.7 Building blocks of a computer-based control system . . . . . . . . . . 74
Fig. 6.8 A generic avionics block diagram . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
Fig. 6.9 AOCS functional chain as a member of the avionics
architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
Fig. 6.10 Subsystem federated architecture with a star topology . . . . . . . . . 79
Fig. 6.11 A backplane connecting 1 CPU unit and 2 peripheral
boards . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
Fig. 7.1 An unannotated graph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
Fig. 7.2 Google searches over time about something unspecified . . . . . . . 88

vii
viii List of Figures

Fig. 7.3 Plot is about people searching about dogs . . . . . . . . . . . . . . . . . . . . 89


Fig. 7.4 Google searches about Christmas are obviously seasonal . . . . . . 90
Fig. 7.5 A bit of less obvious seasonal spikes in data . . . . . . . . . . . . . . . . . . 91
Fig. 7.6 Pythagorean theorem searches versus time . . . . . . . . . . . . . . . . . . . . 92
Fig. 7.7 A probable correlation: Pythagorean theorem searches
and school season (note the interesting noise
during the COVID-19 pandemic) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
Fig. 7.8 Life of a Turkey: all is great, and nothing indicates
the trend will change; until it does. Source The Black
Swan, by Nicholas Nassim Taleb; Wikimedia Commons . . . . . . . 94
Fig. 8.1 Fuel pump state machine (you probably don’t think of this
while you’re topping your car, but it’s what the pump
needs to deal with) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
Fig. 10.1 A ground antenna (photo by Donald Giannatti
on Unsplash) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
Fig. 10.2 A sketch illustrating deployable solar panels . . . . . . . . . . . . . . . . . . 114
Fig. 10.3 A team in action in a cleanroom (photo by Laurel
and Michael Evans on Unsplash) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
Fig. 10.4 Distribution of satellites for altitudes between 0
and 50,000 km altitude . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
Fig. 10.5 Distribution of satellites for altitudes between 0 and 2000
km (LEO) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
Fig. 10.6 Distribution of satellites for altitudes between 400
and 700 km (LEO) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
Introduction
1

The idea of this text is not about conveying tedious, encyclopedic information
about space in the condescending tone books seem to have adopted lately, but to
provide a quick, lightweight introduction to the fundamentals you need to know
about space technology in a colloquial tone.
Ideally, you are reading this text during a short flight back home from a job
interview, a conference or similar. During short flights, there’s nothing much to do
really, infotainment is nothing to write home about; so why not read a bit about
all things space and get up to speed?
You might be a newcomer to the industry; a marketing manager, a legal counsel,
an investor, a software engineer, or a project manager joining a space tech company
after a gig in a different industry. How much do you know about space? Unless
you are an enthusiast who did some research on your own, chances are that you
know little or nothing about it, other than the fact that satellites somehow seem to
go to space. But there is of course more than that, for example:

. What are the physics laws behind an object orbiting a celestial body?
. What kind of environments do satellites face when in orbit?
. What are satellites made of?
. What does it entail to ride a rocket into orbit?
. What is the process to design and build satellites?
. How do people on the ground keep track of satellites as they fly?
. What happens if a satellite fails?
. What kind of data do satellites deal with?

This short guide got your back. When you finish reading these lines, you will be
equipped with a good dose of the fundamentals about the peculiar endeavor of
creating artificial satellites. Moreover, you will also get an idea about the tech-
nologies that have enabled—and keep enabling—space activities, like materials,

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 1


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_1
2 1 Introduction

radio, telecommunications, optics, and software, with some brief historical back-
ground provided whenever possible. But beware: this is not a handbook, nor a
doctoral thesis. The depth of the topics overall is at the introductory level. I tried
on purpose not to bloat this text with distracting figures and focus on content and
fundamentals. This is nothing but a primer. If any section in this text interests you
more than any other, I trust you will do a deeper research using the links provided.
Google is your friend. Of course, once you land if you happen to be flying while
reading.

1.1 Space and Startups, a.k.a “NewSpace”

Lately, space technology has been happening at the dungeons of very small and
dynamic companies: space startups. Life in startups is quite well documented, per-
haps somewhat over documented and a tad romanticized. There are books, blogs,
and popular TV series. Startup platitudes are all over the place in social media—
everyone seems to know how to run them, how to scale them, adding their own
bit of advice, their own experiences of what works, what doesn’t. Although most
startups more or less go through similar phases, in reality each one of them is
unique. No surprise there: the same applies for us people; we all go through
the same life stages, from infancy to adolescence and adulthood, although we
all—luckily—experience each of these life cycle stages very differently.
Space startups are a subset of the startup universe, yet a peculiar type. What’s so
special about them? Think about a space startup for a moment. A handful of people
trying to build spaceships, on hair-thin budgets, short runways (bankruptcy is con-
stantly lurking around the corner), and often without having a customer in sight.
There isn’t perhaps another startup type with so many odds against, just by looking
at the challenges they face. In a way, space startups are like salmon. Salmon swim
against the river’s current the whole way—sometimes up to thousands of kilome-
ters, leaping over obstacles, waterfalls, rapids, and dams. These amazing fish can
jump two meters high or even higher. And all while hungry predators like bears
and eagles wait around every river bend to catch them when they jump out of the
water. Just like salmon, many space startups will perish along the way. But a fair
lot will make it up the river, in the process becoming what Nicolas Nassim Taleb
calls “antifragile”—what doesn’t kill them makes them stronger—and eventually
reaching orbit. Antifragility is a property of systems in which they increase their
probability of survival as a result of shocks, volatility, mistakes, faults, attacks,
or failures. For antifragile space startups, an extra day they exist, the higher the
chances of continuing existing.
When analyzing space startups, survivorship bias1 is a trap we repeatedly fall
into. For each startup that makes it big, there are thousands of others who didn’t

1Survivorship bias is the logical error of concentrating on entities that passed a selection process
while overlooking those that did not. This can lead to incorrect conclusions because of incomplete
data.
1.1 Space and Startups, a.k.a “NewSpace” 3

make it and have volumes to speak but die in silence. In terms of survivability,
there is more insight from the salmon who did not make it up the river than
from those that did, for the former knows what didn’t work—where the bears are
lurking—whereas the latter could have been just lucky. Granted, a dead salmon
can’t talk. Also live salmon can’t talk, but you get the gist of the analogy.
What’s inside a space startup if you crack it open? How can a small bunch
of people get to launch something into space? Wasn’t space supposed to be done
by governmental agencies, with their billion-dollars budgets, their thousands of
employees, their bureaucracy? No. You can get to fly something in space with, say,
less than 10 people and a modest budget. How? The magic tends to revolve around
vision, motivation, industrial loads of hard work, commitment, and an obsessively
systemic mindset. Yes, this sounds like yet another of those platitudes out of a
vanity-published management book sold in airport bookstores. But there is no
magic and it all boils down to common sense and few things to pay attention
whenever possible.
The first one is complexity. Designing a satellite is a complex task. If you open
the hood of any satellite, what you will see is an intricate network of computers,
each one performing a specialized job—command and data handling, attitude con-
trol, radio communications, payload control, data processing, etc. Each computer
is a world on its own, running lots of software. Making sure those “worlds” com-
bine seamlessly in order to give a spacecraft its functional integrity in a harmonic
way requires a good deal of cross-functional and system level analysis such as
architecture design, thermal analysis, structural analysis, power generation, phys-
ical configuration, and a very long et cetera. What’s more, all those things are
heavily intertwined. Such interdisciplinary is nothing else than Systems Engineer-
ing,2 which more a craft than a profession. The good thing is that it does not
require you hiring a veteran Systems Engineering wizard you couldn’t possibly
dream to afford. In fact, Systems Engineering is a glorified term for a combina-
tion of good knowledge of the technical fundamentals, critical thinking, problem
solving and common sense. Although these factors do improve with experience,
there’s plenty of young people with a good dose of them.
The second one is wheel reinvention (the avoidance thereof). One of the most
important factors in small space startups is to minimize rediscovering said round
artifact used to help things move from A to B. This also means, space startups
must stay extremely focused on what their métier is. Mind you, the métier may
change along the way and the space startups that prevail are those who identify in
time when the focal point is wrong and are able to swerve before the iceberg gets
too close.

2 Systems engineering is an interdisciplinary field of engineering and engineering management that


focuses on how to design, integrate, and manage complex systems over their life cycles. At its
core, systems engineering utilizes systems thinking principles to organize its body of knowledge.
The individual outcome of such efforts, an engineered system, can be defined as a combination of
components that work in synergy to collectively perform a useful function.
4 1 Introduction

The third one. The not-so-glamorous side of building things: supply chain. The
grocery list. Supply chain management is an art. It deals with uncertainty, change,
prices, inflation, secrecy, proprietary data, volatility, convoluted technical specifica-
tions, variants, and a ridiculous amount of foresight to predict what a company will
manufacture years from now. Supply chain is a challenging endeavor when you are
small, young, and the new kid on the block. Suppliers tend to pay attention to the
big guns—their established customers—and rightly so; who could blame them?
Then, the small guys must elbow their way to source themselves parts and com-
ponents, at times closing partnerships with other fellow young startups in need. In
NewSpace, John Does attract Jane Does.
In the process, some space startups may choose to maverick their make vs buy
strategy and go vertical or in-house, supposedly to shield themselves from supplier
uncertainty, only to continue being locked in with suppliers because they realize
they cannot produce up to the last bolt. On the flip side, horizontal integration
may create uncomfortable situations if a critically important, complex subsystem
is provided by a third party in which the startup has zero control (and sometimes
zero trust).
Fourth one, the almost literally million-dollar question: what on earth to sell.
Next time you walk past a pizza place, think about how clear the product is for
them. They make pizzas, that’s what they sell. Simple as that. Zero ambiguity.
They define flavors, toppings, variants. They choose names for the variants and
print menus. They make people happy by selling warm flat tasty discs with cheese,
tomato sauce and stuff. It’s so clear that if you go and ask anyone working there
what it is that they sell, they will all say the same: we sell pizza—what a silly
question to ask! Now, for a visitor entering the office of a space startup and pick-
ing up someone from the team and asking: what the hell is this startup selling, the
answer may vary depending on who the mysterious visitor should ask. That’s the
situation usually at the early—and not so early—stages: an amorphous idea involv-
ing space technology does not always automatically map to a product. It can be
data, can be insight from the data, can be the spacecraft, can be a subsystem, can
be a service on top of all that. Products must be discovered, and such discovery
process may take long and be exhausting. Ideally, the product shall be discovered
before the money runs out.
And fifth, last and perhaps the most important thing: everyday life. A space
startup is not just a romantic adventure about reaching the stars. Or, it might be,
but reaching the stars comes as the culmination of disciplined work and sound
day-to-day company operations as people share many hours a week, shoulder to
shoulder, overcoming hurdles and finding their way through the job. In short, a
space startup is—no surprise there—an actual company which needs to be run.
There are operational matters such as talent capture, facility management, frozen
pizza, coffee, and of course, team matters. Building healthy teams where learning
and making mistakes is part of the job and, more fundamentally, where it is fun
to do all that is a bigger accomplishment than flinging shiny boxes beyond the
1.1 Space and Startups, a.k.a “NewSpace” 5

Kármán line.3 All this is what NewSpace startups are about. Satellites, at the end
of the day, are by-products. The main ‘product’ of a space startup is the network
of brains behind the technology.
So, let’s dive into this. The chapters of this text are reasonably self-contained,
although there might be suggestions in certain parts to jump here and there for
elaboration. I do not expect you to read this from cover to cover, but to selectively
sift through the pages as the topics that resonate on you and your curiosity will
capture your attention. Some chapters go a bit more technical than others, and if
the content in those makes absolutely no sense, jump back to the safety of the
less technical sections. If you are really, really busy, there is a TL; DR (too long,
didn’t read) chapter at the end (Chap. 10) which summarizes the text in a set of
frequently asked questions.
As a CTO at a space startup like ReOrbit, I am responsible for ensuring that
the technology roadmap comes together and aligns well with the business model.
But my job is, as I see it, more than that. Fundamentally, as a CTO, my role is
to ensure the team of engineers I lead enjoy developing space technology and feel
safe trying things out and screwing up in the process, learning from the mistakes
and charging back stronger than before. There is no innovation possible without
experimentation, and space technology moves forward thanks to those who ven-
ture themselves into the unknown, for most of the ‘knowns’ today in space were
unknowns yesterday.
Last but definitely not least, a mention of ReOrbit. ReOrbit is a space company
based in Helsinki, Finland, and with offices in Stockholm and Argentina. Founded
in 2019, ReOrbit designs and develops satellites for a variety of different payloads
and applications. At ReOrbit, satellites are designed as network routers and thus
equipped with the capabilities to ensure secure and reliable data transport from
satellite to satellite or satellite to ground. Find more information at www.reorbit.
space.
With all this being said, here we go.
Ignacio Chechile, Chief Technology Officer, ReOrbit. April 2023.
Helsinki, Finland.

3 The Kármán line is a proposed conventional boundary between Earth’s atmosphere and outer
space set by the international record-keeping body FAI (Fédération Aéronautique Internationale)
at an altitude of 100 km. However, such definition of the edge of space is not universally adopted.
Artificial Satellites; The Shortest
Introduction Ever 2

No one here is alone. Satellites in every home.


—Blur, “The Universal”

Abstract

Three and a half years after the launch of the first artificial satellite, Sputnik 1,
there were already 115 artificial satellites orbiting the Earth. Things escalated
quickly. What is the story behind the first artificial satellites? What are the
physics laws involved? This chapter presents the shortest introduction ever to
the topic.

The first published mathematical study of the possibility of an artificial satellite


was the now famous Newton’s cannonball, a thought experiment by Isaac Newton
to explain the motion of natural satellites, published in his Philosophiæ Naturalis
Principia Mathematica (1687). In it, Newton thought of a cannon situated at the
summit of a mountain and being fired. Now, depending on the velocity imprinted
by the cannon, the ball would fall at different distances from the muzzle. See the
image below: a certain initial velocity would cause the ball to fall at the point D.
A slightly higher velocity would bring the ball up to point E, F and G. Now, if we
increased the velocity consistently in few more steps, there would be a velocity
for which the ball just does not fall back to the surface of the planet anymore but
keeps on falling “eternally” (provided no friction), which is the closed curve in
the illustration, and what rockets basically do to satellites: imprint them the right
velocity and letting them achieve closed paths (yes, this is a bit oversimplistic and
there’s more than that, as we will see). Mind that if we kept increasing the velocity
after this point, the ball will eventually escape the planet orbit and start wandering
in interplanetary space. But that’s out of the scope for this text (Fig. 2.1).

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 7


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_2
8 2 Artificial Satellites; The Shortest Introduction Ever

Fig. 2.1 Newton’s cannon

The first fictional depiction of a satellite being launched into orbit was a short
story by Edward Everett Hale, “The Brick Moon” (1869). The idea appeared again
in Jules Verne’s The Begum’s Fortune (1879).
In 1903, Konstantin Tsiolkovsky published Exploring Space Using Jet Propul-
sion Devices, which is the first academic treatise on the use of rocketry to launch
spacecraft.
Herman Potočnik entertained the idea of using orbiting spacecraft for detailed
peaceful and military observation of the ground in his 1928 book, The Problem of
Space Travel. He described how the special conditions of space could be useful
for scientific experiments. The book described geostationary satellites (first put for-
ward by Tsiolkovsky) and discussed communication between them and the ground
using radio but fell short of the idea of using satellites for mass broadcasting and
as telecommunications relays.
In a 1945 Wireless World article, the English science fiction writer Arthur C.
Clarke described in detail the possible use of communications satellites for mass
communications. He suggested that three geostationary satellites would provide
coverage over the entire planet.
In May 1946, the United States Air Force’s Project RAND released the Pre-
liminary Design of an Experimental World-Circling Spaceship, which stated “A
satellite vehicle with appropriate instrumentation can be expected to be one of the
most potent scientific tools of the Twentieth Century”. The United States had been
considering launching orbital satellites since 1945 under the Bureau of Aeronautics
of the United States Navy.
2 Artificial Satellites; The Shortest Introduction Ever 9

In 1946, American theoretical astrophysicist Lyman Spitzer proposed an


orbiting space telescope.
In February 1954, Project RAND released “Scientific Uses for a Satellite Vehi-
cle”, by R. R. Carhart. This expanded on potential scientific uses for satellite
vehicles and was followed in June 1955 with “The Scientific Use of an Artificial
Satellite”, by H. K. Kallmann and W. W. Kellogg.
In the context of activities planned for the International Geophysical Year
(1957–1958), the White House announced on 29 July 1955 that the U.S. intended
to launch satellites by the spring of 1958. This became known as Project Vanguard.
On 31 July, the Soviet Union announced its intention to launch a satellite by the
fall of 1957. The game was on.
The first real artificial satellite would end up being Sputnik 1, launched by the
Soviet Union on 4 October 1957 under the Sputnik program. The 84 kg spacecraft
worked for roughly 2 weeks, and it reentered the atmosphere a few months after. Its
architecture was rather rudimentary: its batteries weighed 51 kg, it was equipped
with a 1Watt transmitter which encoded telemetry in low frequency pulses which
would be broadcast and heard on AM radio, and it was pressurized with nitrogen.
Sputnik 1 helped to identify the density of high atmospheric layers through
measurement of its orbital change and provided data on radio signal distribution
in the ionosphere. The unanticipated announcement of Sputnik 1’s success precip-
itated the Sputnik crisis in the United States and ignited the so-called Space Race
within the Cold War.
Sputnik 2 was launched on 3 November 1957 and carried the first living
passenger into orbit, a dog named Laika.
Explorer 1 became the United States’ first artificial satellite, launched on 31
January 1958. The information sent back from its radiation detector led to the dis-
covery of the Earth’s Van Allen radiation belts. The TIROS-1 spacecraft, launched
on April 1, 1960, as part of NASA’s Television Infrared Observation Satellite
(TIROS) program, sent back the first television footage of weather patterns to
be taken from space.
In June 1961, three and a half years after the launch of Sputnik 1, the United
States Space Surveillance Network had already cataloged 115 Earth-orbiting
satellites. Things escalated really quick.
Expectedly, early satellites were built to unique designs. With advancements in
technology, multiple satellite missions began to be built on single model platforms
called satellite buses. The first standardized satellite bus design was the HS-333
geosynchronous (GEO) communication satellite made by Hughes and launched in
1972. Oddly enough, many satellites are still designed and built as one offs—in
other words, the 70s way—although multi-mission buses are growing in popularity.
We will talk about this in due time.
As of today, there are more than 5000 operative satellites orbiting our planet.
If we count both operative and inoperative spacecraft, forgotten stages of rockets
and whatnot, we need to talk about 10,000 objects flying over our heads.
Since Sputnik 1, satellite architecture and design methods have evolved con-
sistently. Satellites’ capabilities have improved fast thanks to the progress certain
10 2 Artificial Satellites; The Shortest Introduction Ever

enabling technologies have made on their own. One of those foundational tech-
nologies stand out from the rest: semiconductors. Let’s talk about that in the next
chapter.
Semiconductors in Space: From Sand
to Satellites 3

I don’t like sand. It’s coarse and rough and irritating.


—Anakin Skywalker, Star Wars: Episode II, Attack of the
Clones

Abstract
In space, microprocessors and solid-state devices are ubiquitous because satel-
lites need software, storage, and digital logic in order to process information
on-board, and operate. Systems on Chip (SoCs), FPGAs and logic gates are
heavily used. The software and machine code that spacecraft run on-board to
manage their resources, their orientation or to control a payload sensor executes
on these types of devices, and the space environment is not precisely nice with
their underlying microscopic structure. In this chapter, we delve into how sand
is converted into electronic devices and how those devices survive in orbit.

The sand we find on the beaches is mostly composed of silica, which is another
name for silicon dioxide, or SiO2 . Silica is one of the most complex and abundant
families of materials, existing as a compound of several minerals. Silica is a crys-
talline material, which means that its atoms are linked in an orderly spatial lattice
of silicon-oxygen tetrahedra, with each oxygen being shared between two adjacent
tetrahedra.
Sand is abundant of silica and many other things, including macro particles
such as plastic and other stuff, so SiO2 must be cleaned to be industrially used.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 11


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_3
12 3 Semiconductors in Space: From Sand to Satellites

Fig. 3.1 SiO2 structure

Once all the macro impurities are removed, silica is melted in a furnace at high
temperature and is reacted with carbon to produce silicon of a relative purity.1
Somewhere in 1915 a Polish scientist called Jan Czochralski woke up one morn-
ing on the wrong side of the bed and made a mistake: instead of dipping his pen
into his inkwell, he dipped it in molten tin—why our Jan had molten tin on his
desk is beyond me—and drew a tin filament, which later proved to be a single
crystal. He had invented by accident a method2 which remains in use in most semi-
conductor industries around the world to grow silicon monocrystalline structures,
manufactured as ingots3 that are then sliced into ultra-thin wafers that companies
use to etch their integrated circuits layouts on.4 The process provides an almost
pure, monocrystalline silicon chip makers can work with.
Crystals and their orderly structure have fascinated scientists for ages, perhaps
due to the fact they provide an illusion of order and for that reason offer a rel-
atively easier grasp of the underlying physics: condensed matter is a complex
matter—heh—but when it’s arranged in a more or less symmetrical way in three
dimensions, it may give the impression to be a tad simpler to comprehend.
In a silicon crystal, each silicon atom forms four covalent bonds with four
oxygen atoms, that is, each silicon atom sharing electrons with four oxygen atoms
(see Fig. 3.1).
As we know, temperature is the quantitative measure of the kinetic energy of
all particles that form a substance or material. In crystals, atoms do not really

1 More about the process here: https://aip.scitation.org/doi/pdf/10.1063/5.0046150.


2 Expectedly, the process is called after him: https://en.wikipedia.org/wiki/Czochralski_method.
3 More about the production of ingots here: https://www.microchemicals.com/technical_inform

ation/czochralski_floatzone_silicon_ingot_production.pdf.
4 See more about the etching process here: https://www.kth.se/social/upload/510f795cf276544e1d

dda13f/Lecture7Etching.pdf.
3 Semiconductors in Space: From Sand to Satellites 13

go anywhere but they vibrate in their fixed positions. Temperature in crystalline


structures indicates how violently atoms shake at their spots. Valence electrons,5
in thermal equilibrium with the crystal they belong to, share the kinetic energy
with the rest of the material. But temperature tends to describe the average energy
across the lattice. Momentary differences in local temperature may cause an elec-
tron to muster the guts to break its covalent bond and go free.6 A bond without
its precious electron is a broken bond, and as such will try to recover from this
absence, so the affinity with neighboring electrons intensifies. If the broken bond
manages to capture an electron from a neighboring bond, the problem is only
passed to the neighbor, which will also soon pass it to the next one, and so on.
The “hole” left behind by the initial emancipated electron spreads across the lat-
tice. What happens with the initial fugitive electron? It travels across the structure,
emotionally disengaged from the problem it caused. Worth noting is that a bro-
ken bond creates two phenomena: wandering holes and wandering free electrons.
Another way of calling such free electrons is conduction electrons.
Undisputed kings of negative charge, electrons leave positively charged zones
behind them. Therefore, in the vicinity of holes, the charge is now more positive,
and such positivity travels as the hole travels. Therefore, we can say holes have
positive charge.
A wafer of pure monocrystalline silicon or germanium does not do much in
and of itself. It is just an ‘intrinsic’ material with electrons and holes moving
around because of bonds constantly being broken due to thermal agitation. Intrinsic
materials create electron–hole pairs in exact numbers because one exists because of
the other (along with some other particles existing inside intrinsic silicon as well,
like photons). Intrinsic materials would be of little practical use if we couldn’t
break the balance between electrons and holes. How to break that harmony? By
opportunistically sprinkling our crystals with more electrons (or more holes) by
means of adding impurities. Didn’t we say impurities were bad? Yes, but these
are more sophisticated, controlled impurities, unlike the microplastic that washes
ashore on beaches as a product of our pointless mass consumption urges.
But here’s the catch: we cannot just add loose electrons like we add pepper
to salad—the Coulomb forces would be insane due to the sudden electric charge
imbalance. All we can do is to add atoms that can contribute with electrons, called
donors. Examples of donors are phosphorus or arsenic. Typical proportions of
impurity atoms is one of these guys for every million silicon atoms.
When a donor atom is implanted in the lattice, it mimics the Si atom quite well;
it completes the four covalent bonds the same way as Si atoms do. But arsenic hap-
pens to have 5 valence electrons, so one electron does not belong to any bond, and
because it’s not trapped in any potential barrier, it has a higher energy than their

5 A valence electron is an electron in the outer shell associated with an atom, and that can partici-
pate in the formation of a chemical bond if the outer shell is not closed.
6 This is due to thermionic emission. Thermionic emission (also known as Edison effect) is the

liberation of electrons by virtue of its temperature. This occurs because the thermal energy given
to the charge carrier overcomes the work function of the material.
14 3 Semiconductors in Space: From Sand to Satellites

other 4 cousins, and thus it has high chances of leaving the atom behind, leaving
it positively charged as a gift. An ion is born, fixed in the crystalline structure.
The material remains electrically neutral at the macro level, but it’s now populated
with positively charged spots, all balanced by the free electrons wandering around.
Conversely, acceptor impurities do the opposite. Aluminum, Indium and Gal-
lium, for instance, are good examples of acceptor elements. Adding acceptors is a
way of adding holes to a lattice, without breaking the macro electric neutrality. An
Indium atom fits comfortably in the lattice, impersonating a Silicon atom, but it
has only 3 valence electrons. You get the score. A hole is now there, because one
covalent bond is missing. This vacant bond is open for business, and eventually
it will get filled by an electron, breaking the impurity atom neutrality, and thus
creating a negative ion.
In summary: impurities, whether donors or acceptors, will end up all being
ionized. Donors will quickly lose an electron, and acceptors will quickly lose a
hole (or gain an electron) because the energy to allow such ionization is quite low.
Thermal agitation will make sure that practically all impurities will be ionized,
therefore we can consider that all donors will lose their extra electron. This sim-
plifies the math: we can estimate that the density of conduction electrons will be
more or less equal to the density of donor atoms. The same goes for conduction
holes. This is important: a piece of silicon crystal with more donor impurities than
acceptor impurities will be called type n. Similarly, if more acceptors than donors
are added to the silicon, the material will be called type p. Conduction electrons
and holes will not have it easy while traveling inside the lattice. Multiple things
will alter their trajectories: repulsion forces coming from fellow moving carriers,
un-ionized impurity atoms, ionized impurity atoms, and whatnot. Life of a charge
carrier is not simple.

3.1 Let’s Meet at the Junction

The magic starts to unfold when we sandwich type-n and type-p materials together.
This is called a junction, and its properties are worth mentioning, because it sets
the foundations of all solid-state devices out there.
Junctions are not perfect; it is impossible to define an ideally abrupt boundary
between a material partially doped with donors and another part partially doped
with acceptors. Junctions must be gradual, and this does not affect the physics
behind them. It is very important to note that junctions are not made by welding
one type-n crystal with a type-p crystal. A junction must still be made of a sin-
gle crystal; there is no practical means of attaching together two bars of silicon
with different impurities dosage and expect that it will work. The crystal lattice
perfection is a key factor when it comes to junction’s performance.
In equilibrium (that is, with the piece of silicon that hosts the junction at some
nonzero temperature, with no electric field applied), the concentration of acceptors
will be maximum on the p-side, then decrease to zero as we approach the junc-
tion, and the same for donors on the n-side. With carriers moving due to thermal
3.1 Let’s Meet at the Junction 15

agitation, they cross the boundary thrusted by the gradient of impurities concen-
trations at the far ends. Holes come across the chasm and reach to the n-side,
where they recombine easily because of the high density of electrons there. Equiv-
alently, electrons cross the boundary to the p side, and recombine. Then, a zone
starts to appear around the border, a zone without carriers. A no man’s land of
sorts, where all ions are complete. Because acceptor and donor ions are fixed to
the lattice, the area around the boundary will be charged slightly negative on the
p side (because electrons have found their spots in acceptors) and slightly positive
on the n side, because electrons have fled the scene. These non-zero charge levels
stemming from the fixed ions create an electric field, which causes the diffusion
process to settle when such electric field is intense enough to create displacement
currents that cancel further currents from the doping concentration gradient.
In all our analyses thus far, we have only considered the piece of material to
be only interacting with its surroundings by thermal energy. But that is only one
part of the story. There are several other ways equilibrium in a silicon bar can
be disrupted: electric fields, magnetic fields, and light. In a n-type material, holes
are the minority carriers. Equivalent, in a p-type, electrons are minority carriers.
Minority carriers are many, many orders of magnitude less numerous than majority
carriers. Now if we put the silicon bar under uniform light, the photons of the light
beam will break bonds all across the lattice, creating pairs of electron-holes. Light
photons have created carriers of both signs in equal amounts, but the minority
carriers are the ones noticed here. Imagine that an extra number of electrons on
the n-side will not move the needle; at the end of the day there were a myriad
of other electrons there, so they are nothing special. But an increasing number
of holes on the n-side will be comparatively noticed. The injection of minority
carriers is an important effect which will also play a part in the discovery of the
bipolar transistor. You start to see the tendency of semiconductors to easily become
a mess just by being beamed with some harmless light.
Now, to break the equilibrium in the junction, we must apply a voltage to the
junction. In forward bias, the p-type is connected with the positive terminal and
the n-type is connected with the negative terminal of a voltage source.
Only majority carriers (electrons in n-type material or holes in p-type) can
flow through a semiconductor for a macroscopic length. The forward bias causes
a force on the electrons pushing them from the n side toward the p side. With
forward bias, the depletion region is narrow enough that electrons can cross the
junction and inject into the p-type material. However, they do not continue to
flow through the p-type material indefinitely, because it is favorable for them to
recombine with holes. The average length an electron travels through the p-type
material before recombining is called the diffusion length, and it is typically on
the order of micrometers.
Although the electrons penetrate only a short distance into the p-type material,
the electric current continues uninterrupted, because holes (the majority carriers
on that side) begin to flow in the opposite direction. The total current (the sum
of the electron and hole currents) is constant, in spatial terms. The flow of holes
from the p-type region into the n-type region is exactly analogous to the flow
16 3 Semiconductors in Space: From Sand to Satellites

of electrons from n to p. Therefore, the macroscopic picture of the current flow


through this device involves electrons flowing through the n-type region toward the
junction, holes flowing through the p-type region in the opposite direction toward
the junction, and the two species of carriers constantly recombining in the vicinity
of the junction. The electrons and holes travel in opposite directions, but they also
have opposite charges, so the overall current is in the same direction on both sides
of the material, as required.
Now we do the opposite. Connecting the p-type region to the negative terminal
of the voltage source and the n-type region to the positive terminal corresponds
to reverse bias. Because the p-type material is now connected to the negative ter-
minal of the power supply, the holes in the p-type material are pulled away from
the junction, leaving behind charged ions. Likewise, because the n-type region is
connected to the positive terminal, the electrons are pulled away from the junction,
with similar effect. This increases the voltage barrier causing a high resistance to
the flow of charge carriers, thus allowing minimal electric current to cross the
boundary. But some current—a leakage current—does flow. Leakage current is
caused by the movement of minority carriers (electrons in p-type and holes in n-
type) across the depletion region of the junction. As the depletion region widens,
the potential barrier at the junction increases. However, even though the potential
barrier is high, a small number of minority carriers can still cross the junction
by thermionic emission7 or tunneling. The amount of leakage current depends on
several factors, including the doping concentration of the semiconductor material,
the temperature, and the voltage applied across the diode. Higher doping concen-
trations and higher temperatures can increase the number of minority carriers and
therefore increase the leakage current.
The increase in resistance of the p–n junction results in the junction behaving
as an insulator. The strength of the depletion zone electric field increases as the
reverse-bias voltage increases.
But everything has a limit. Once the electric field intensity increases beyond a
critical level, the p–n junction depletion zone may break down and current shall
begin to flow even when reverse-biased, usually by what is called the avalanche
breakdown8 processes. When the electric field is strong enough, the mobile elec-
trons or holes may be accelerated to high enough speeds to knock other bound
electrons free, creating more free charge carriers, increasing the current and lead-
ing to further “knocking out” processes and creating an avalanche. In this way,
large portions of a normally insulating crystal can begin to conduct.
This breakdown process is non-destructive and is reversible, as long as the
amount of current flowing does not reach levels that cause the semiconductor
material to overheat and cause thermal damage.

7 Thermionic emission is the process by which electrons escape from the surface of a material due
to their thermal energy. When a material is heated to a sufficiently high temperature, the thermal
energy of the electrons increases, and some of them gain enough energy to overcome the potential
barrier that holds them in the material.
8 https://en.wikipedia.org/wiki/avalanche_breakdown.
3.2 The Transistor Drama 17

It is important to say that the hectic scene inside a semiconductor described in


this section can be noticed from the outside. All these electrons and holes knocking
about the junction create a good deal of noise which can affect external circuits.
For instance, shot noise, also known as Schottky noise, is a type of electrical
noise that arises from the random nature of the flow of electric charge carriers in
the material. In semiconductors, shot noise occurs when the electrons and holes
cross the junction, and is caused by the discrete nature of charge carriers and their
motion. Because of the discrete nature of charge carriers, current in a junction does
not flow smoothly but rather in bursts or “shots” of current. These bursts occur
when electrons or holes overcome the potential barrier and move from one side to
the other. The size and frequency of these bursts depend on several factors, includ-
ing the voltage applied, the temperature of the material, and the concentration of
charge carriers. At the beginning of this section, we commented that thermal agi-
tation caused electrons to break loose from their atoms in the lattice and go wild,
creating electron–hole pairs. This process causes a noise called Johnson-Nyquist
noise, also known as thermal noise, and is a type of electrical noise that arises
from the random thermal motion of charge carriers in the presence of thermal
energy, which means that it obviously increases with temperature. Thermal noise
is present in all electric circuits, and in radio receivers it may affect weak signals.
There is also flicker noise, which although not fully understood, it is believed to
be related to the trapping and release of charge carriers by defects or impurities in
the semiconductor material.
All these noises can affect the performance of the external circuits—and more
importantly, low-noise circuits—using the semiconductors, and the relevance of
these noises may change depending on the application, the current levels and
frequencies involved.
Overall, what we have described in this section is nothing by the inner workings
of a diode. A diode is a solid-state device which conducts current primarily in one
direction. As we will see, being able to control the direction of flow of electrons
and holes would prove to be of importance. Why stop with only one junction?

3.2 The Transistor Drama

A drama you didn’t expect: the transistor drama. After Bardeen and Brattain’s
December 1947 invention of the point-contact transistor,9 William Shockley dis-
sociated himself from many of his colleagues at Bell Labs, and eventually became
disenchanted with the institution itself. Some hint that this was the result of
jealousy at not being fully involved in the final, crucial point-contact transistor

9 https://en.wikipedia.org/wiki/Point-contact_transistor.
18 3 Semiconductors in Space: From Sand to Satellites

Fig. 3.2 A bipolar transistor with one junction in forward-bias and another one in reverse-bias

experiments and frustration at not progressing rapidly up the laboratory man-


agement ladder. Mr. Shockley had, in the words of his employees, an unusual
management style.10
Shockley recognized that the point-contact transistor delicate mechanical con-
figuration would be difficult to manufacture in high volume with sufficient
reliability. He also disagreed with Bardeen’s explanation of how their transis-
tor worked. Shockley claimed that positively charged holes could also penetrate
through the bulk germanium material, not only trickle along a surface layer. And
he was right. On February 16, 1948, physicist John Shive achieved transistor action
in a sliver of germanium with point contacts on opposite sides, not next to each
other, demonstrating that holes were indeed flowing through the thickest part of
the crystal.
All we have said before about the p–n junction before applies to transistors. But
transistors have three distinctive areas, with two boundaries or junctions: n–p–n,
or p–n–p, typically called emitter, base and collector. Emitters are heavily doped
with impurities, and for that it is usually called n++ or p++. The base is weakly
doped, and for the collector this is not so important, and its doping depends on
the manufacturing process. The most important constructive factor is the based
width, or W. The junction separating emitter from base is called, no wonder, emit-
ter junction, whereas the junction separating base from collector is called—drum
roll—collector junction. Naming at least is not complicated (Fig. 3.2).
To understand the inner workings of a transistor of this kind, let’s assume a p–
n–p arrangement where we forward-bias the emitter junction, that is, the positive
terminal of the voltage source connected to the emitter, and the negative terminal
to the base (see figure above). Conversely, we reverse-bias the collector junction:

10 https://www.nature.com/articles/442631a. (Eventually, William Shockley would also be famous


for being a notorious racist)
3.2 The Transistor Drama 19

negative terminal of a power source to the collector, positive terminal to the base.
This way, the emitter to base current is large because the junction is forward-
biased—with the current value being governed by the diode equation.11 Given that
this junction is highly asymmetric (the doping of the emitter p-region is orders of
magnitude higher than the doping of the base n-region), the emitter current will
be largely composed of holes going from the p-side to the n-side (current 1 in
the figure). If the base width (W) is narrow enough, and because the base area is
electrically neutral, the holes traversing through the emitter junction will find their
way to the collector junction where the electric field will capture and inject them
into the collector area (currents 3 and 4 in the figure). Some holes will recombine
in the base (current 6), creating a base current which is very small due to the
low doping of the base section and the small width of the base. With all this, the
emitter current is passing almost unaltered to the collector. The collector current is
almost independent of the collector–base voltage, as long as this voltage remains
negative. Otherwise, the collector would also inject holes into the base, altering
the overall functioning of the device. This is an important mode (saturation mode)
we will talk about.
The electric field at the collector junction injects the holes into the collector
area, and the magnitude of this electric field does not affect the number of holes
arriving to that place. It is the base and the diffusion that happens there which
defines the number of holes that will make it to the collector. Even zero volts
between collector and base would keep that current flowing.
Thus far, we have been analyzing the behavior of the transistor mostly from
its direct-current (DC) biasing perspective. The analysis to follow should be about
observing how the transistor behaves while in the active region and when fed with
small—and not so small—AC signals superimposed to base voltages, causing the
device’s biasing to fluctuate around certain points, and how the input and output
signals should match each other, minimizing alterations (i.e., distortion). Although
understanding this is of great importance and a topic in itself which finds applica-
tions in a myriad of fields such as analog circuits, radiofrequency, communications,
hi-fi audio, and whatnot, for this discussion we shall focus on the device in switch-
ing mode, that is, moving between defined, discrete conduction states: from cut-off
to saturation, and swinging between them as fast as possible. In this mode, the tran-
sistor acts as a switch, evolving from one extreme state (cutoff, or open switch) to
the other (saturation, closed switch) as fast as possible.
A transistor operating in the cutoff region has its two junctions working in
reverse bias mode. In this situation, only leakage current flows from collector
to emitter. Conversely, in saturation, the device has both junctions in forward-
bias mode, allowing a small depletion layer and allowing the maximum current to
flow through it. By controlling the biasing of the emitter–base junction, we can
make the transistor transition between these two modes; full current conduction or
practically zero.

11 https://en.wikipedia.org/wiki/shockley_diode_equation.
20 3 Semiconductors in Space: From Sand to Satellites

Fig. 3.3 NOT gate with BJT transistor

Table 3.1 NOT gate truth


A Output
table
0 1
1 0

The transistor in switching mode sets the foundation of the underlying behavior
of practically all digital electronics and computer systems out there.
So, all this hassle with electrons, holes, donors, acceptors, minority carriers and
gossip at Bell Labs only to create a switch?
Really? Yes.
A very special kind of switch, one that would go down history to spark a
revolution. The junctions we described above, in the form of diodes and transistors,
would become the basic building blocks of our modern digital toolbox. A toolbox
that supports today’s machine learning, artificial intelligence, cloud computing, but
also Instagram, TikTok and the metaverse. How?
Combining transistors in switching mode can form logic gates. For instance, a
simple bipolar junction transistor (BJT, the one whose inner working we described
in Sect. 3.2) can form a NOT gate, which basically takes an input and inverts it
(Fig. 3.3; Table 3.1).12
Similarly, a BJT can form a NAND gate (Fig. 3.4; Table 3.2).

12Note that “1” and “0” states here are just logic states and do not represent specific voltages.
Eventually, the semiconductor industry would standardize the voltage thresholds in logic among
other specs, giving way to logic families.
3.2 The Transistor Drama 21

Fig. 3.4 NAND gate with BJT transistor

Table 3.2 NAND gate truth


A B Output
table
0 0 1
0 1 1
1 0 1
1 1 0

Eventually, logic gates would form flip-flops.13 Flip-flops would form registers,
decoders, multiplexers, demultiplexers, but also adders, subtractors and multipli-
ers, which in turn would form arithmetic units (ALUs). As integration technology
and processes would mature, designers would start packing several logic blocks
such as memories, ALUs and buses inside smaller and smaller silicon dies. Then,
engineers would create a clever digital machine whose behavior could be slightly
modified—this means, it would perform different arithmetic operations and data

13 A flip-flop is a digital circuit that can be thought of as a single bit of memory that can store either
a 0 or a 1. It has two stable states, which are typically referred to as “SET” and “RESET”. The flip-
flop can be set to either of these states using appropriate input signals, and it will remain in that
state until it is reset or set again.
22 3 Semiconductors in Space: From Sand to Satellites

movements between parts of its architecture—by means of binary words called


instructions stored in a memory, giving way to machine code and CPU architec-
tures. Corrado Böhm in his Ph.D. thesis14 would conceive the foundations for the
first compiler—which still lacked the name as he called it “automatic program-
ming”, with Böhm being one of the first computer science doctorates awarded
anywhere in the world—an invention that would appear as a way of coping with
the natural lack of human readability of machine code. The word ‘compiler’ would
eventually be coined by Grace Hopper, who would go and implement the first
compiler ever. Compilers would accelerate the process of development run time
behavior in CPUs, what we now call software. Not without creating some crisis in
the process.15
In our eternal quest for more and more abstraction, and as different CPU archi-
tectures would proliferate, porting software from architecture to architecture would
become more problematic, so we would sort this by packing layers of standardized
software libraries and services that would dramatically ease our way of program-
ming application software on top of dissimilar hardware, giving way to what
we now call operating systems that would, in the process, make some people
obnoxiously rich.
And as bipolar integrated circuits would pass the baton to more efficient fabrica-
tion processes,16 and as the physical lengths of integrated transistors would shrink
and their density would double roughly every 2 years,17 their switching speed
from cutoff to saturation would continue decreasing and with better integration
technologies, more complex architectures became possible, making System-On-
Chips, CPLDs and later FPGAs feasible devices and products. Combined with a
new breadth of spectrum-efficient digital modulation and signal processing tech-
niques, mobile devices would materialize, maturing with them important related

14 Translated version of the thesis: http://www.itu.dk/~sestoft/boehmthesis/boehm.pdf.


15 Software crisis is a term used in the early days of computing science to describe the difficulty of
writing efficient computer programs in the required time. The software crisis was due to the rapid
increases in computer power and the complexity of the problems that could not be tackled. With
the increase in the complexity of the software, many software problems arose because existing
methods were inadequate.
16 CMOS is actually based on metal–oxide–semiconductor field-effect transistors (MOSFET)

which is a different kind of transistor compared to the Bipolar Junction Transistor (BJT). MOS-
FETs have an insulated gate, the voltage of which determines the existence and conductivity of a
conduction channel used for amplifying or switching electronic signals.
17 The doubling period is often misquoted as 18 months because of a prediction by Moore’s col-

league, Intel executive David House. In 1975, House noted that Moore’s revised law of doubling
transistor count every 2 years in turn implied that computer chip performance would roughly dou-
ble every 18 months (with no increase in power consumption). Mathematically, Moore’s Law
predicted that transistor count would double every 2 years due to shrinking transistor dimensions
and other improvements. As a consequence of shrinking dimensions, Dennard scaling predicted
that power consumption per unit area would remain constant. Combining these effects, David
House deduced that computer chip performance would roughly double every 18 months.
3.3 The Space Environment 23

domains and technologies like displays, allowing us to create arbitrary arrange-


ments of pixels in screens whose colorful photons would hit our retinas, creating
appealing user human–machine interfaces in applications that would allow us to,
for example, send an emoji to a friend for comedic purposes.
How does space technology relate to all these happenings?
In space, microprocessors and solid-state devices are ubiquitous because satel-
lites need software, storage, and digital logic in order to process information
on-board and act accordingly. Systems on Chip (SoCs), FPGAs and logic gates
are heavily used. The software and machine code that spacecraft run on-board
to manage their resources, orientation or to control a payload sensor executes on
these types of devices, and the space environment is not precisely nice with the
microscopic structure that we have just described above. Let’s see why.18

3.3 The Space Environment

Although we all are technically in space as we travel across interstellar regions


while riding on this geoid we call earth,19 we tend to live in a sort of crystal
bubble in terms of the coziness of this blue dot we live in. Space is a harsh place
to be, at least compared to life here at the surface of the ground. We happen to
be protected by two huge shields: the magnetosphere, which captures and deflects
particles of different energies that otherwise would be harmful for us, and by a
thick layer of gas we call atmosphere which captures and neutralizes space debris
wanting to hit us in the head. And both shields complement each other well.
Unlike Mercury, Venus, and Mars, Earth is surrounded by an immense magnetic
field called the magnetosphere. The Earth has a magnetic field because it has a
molten outer core of iron and nickel that is constantly in motion. The motion of the
liquid outer core creates electrical currents, which in turn generate a magnetic field,
as André-Marie Ampère stated in his eponymous circuital law. Our magnetosphere
shields us from erosion of our atmosphere by the solar wind (charged particles
the Sun continually spews at us), erosion and particle radiation from coronal mass
ejections (massive clouds of energetic and magnetized solar plasma and radiation),
and cosmic rays from deep space. The magnetosphere plays the role of gatekeeper,
repelling this unwanted energy that’s harmful to life on Earth, trapping most of it a
safe distance from Earth’s surface in doughnut-shaped zones called the Van Allen
Belts.

18 Content for this section has been adapted and illustration has been taken from the book “Elec-
trónica del Estado Sólido” by Ángel Tremosa.
19 The name Earth is an English/German name which simply means the ground. It comes from

the Old English words ‘eor(th)e’ and ‘ertha’. In German it is ‘erde’. The name Earth is at least
1000 years old.
24 3 Semiconductors in Space: From Sand to Satellites

Fig. 3.5 Van Allen radiation belts; cross them is not the nicest ride for a satellite going somewhere
(public domain)

The inner Van Allen belt is located typically between 6000 and 12,000 km
(1–2 Earth radii20 ) above Earth’s surface, although it dips much closer over the
South Atlantic Ocean. The outer radiation belt covers altitudes of approximately
25,000–45,000 km (4–7 Earth radii).
As you may imagine, any semiconductor on-board of a satellite crossing these
regions will not have the best time ever. Geostationary satellites must pierce
through the inner belt on their way to their final orbits (Fig. 3.5).
Hardware exposed to space must be ready to withstand all aspects of the
environment. This includes vacuum, thermal cycling, charged particle radiation,
ultraviolet radiation, plasma effects and atomic oxygen.
Radiation is generally classified as being either ionizing or non-ionizing. The
basic dividing line between the two is the energy levels involved. Ionizing radia-
tion has sufficient energy to strip electrons from atoms, thus creating ions, that is,
atoms with charge. Remember the previous section about holes, electrons, impu-
rities, and the delicate mechanism inside the silicon lattice? Now imagine such a
fragile microscopic scenery being bombarded by highly energetic particles impact-
ing the crystal structure and knocking electrons out of place and disrupting the
charge balance across the place. This is what electronics on board of every satellite
flying over our heads is experiencing as we speak. Examples of ionizing radiation
are alpha and beta particles, protons, X-rays, and gamma rays. Neutrons are not
directly ionizing, but the resulting radiation from their collisions with nuclei is
ionizing.

20The Earth is almost, but not quite, a perfect sphere. Its equatorial radius is 6378 km, whereas its
polar radius is 6357 km. A radius value of 6371 km is usually adopted.
3.3 The Space Environment 25

In contrast, non-ionizing radiation only has sufficient energy to change the


energy state of electrons. Examples of non-ionizing radiation are visible and
infrared light, microwaves, and radio waves. Non-Ionizing radiation cannot induce
upsets in electronic devices but can still create undesired effects. Additionally,
Galactic cosmic rays (GCR) comprised of high-energy particles, overwhelmingly
protons, impact the Earth’s atmosphere constantly. These particles, when they col-
lide with molecules in the Earth’s atmosphere, produce a wide range (and a high
number) of particles, primarily neutrons and protons. Neutrons are particularly
troublesome because they can penetrate most man-made construction.21
The hard vacuum of space with its pressures below 10E−4 Pa (0.0010 Pa)
causes some materials to outgas, which in turn affects any spacecraft compo-
nent with a line-of-sight to the emitting material, principally optics sensitive to
impurities in their lenses.
Another effect to suffer in space is thermal cycling. Thermal cycling occurs as
the spacecraft moves through sunlight and shadow while in orbit or while maneu-
vering. Thermal cycling temperatures are dependent on the spacecraft component
thermo-optical properties, i.e., solar absorptance, or how much solar energy the
material absorbs, and infrared emittance, or how much thermal energy can be
emitted to space. The lower the ratio of absorptance to emittance, the cooler the
temperature of the spacecraft surface. Thermal cycling can cause cracking, delam-
ination, and other mechanical problems, particularly in assemblies where there is
mismatch in the coefficient of thermal expansion between materials.
Radiation can also affect materials. Charged particle radiation, along with ultra-
violet radiation can cause cross-linking (hardening) and chain scission (weakening)
of polymers, darkening and color center formation in windows and optics.
If all that was not enough, micrometeoroids and space debris particles may
impact at high velocities. All of these may have significant effects on material
properties (Fig. 3.6).
Satellites are designed to incorporate mitigation measures for the undesired
effects from radiation mentioned: Ionization Dose, which refers to the cumulative
effect of the energy deposited in matter by ionizing radiation per unit mass (known
as Total Ionization Dose, or TID), and Single Event Effects (SEE), which are
related to single, highly energetic particles interacting with the atomic structure of
semiconductors and altering its behavior, both destructive and non-destructively.
TID affects semiconductors in several ways, for example by modifying thresh-
old voltages. The mechanics of this is as follows: the trapping of holes in the
material may cause a charge buildup and it occurs in the bulk of the semiconduc-
tor oxide. These charges will increase the gate oxide electric fields, leading to a
change in the current–voltage (I–V) characteristic of the device. The most promi-
nent change is the shift of the power-ON (threshold) voltage which is negative for
NMOS and positive for PMOS. As a result, a device might become unresponsive
to some commands as it might get “stuck” in a specific state.

21 https://www.microsemi.com/document-portal/doc_view/130760-neutron-seu-faq.
26 3 Semiconductors in Space: From Sand to Satellites

Fig. 3.6 Magnetic field strength at Earth’s surface (Creative Commons)

Devices might also see an increase in their leakage current (remember the con-
cept of leakage current at the beginning of this chapter). In NMOS transistors,
charges might draw an image charge in the semiconductor which can reverse the
interface and free leakage paths. These parasitic leakage currents cause degraded
timings and increase power consumption.
In general, BJT transistors are more robust against radiation compared to MOS
(Metal–Oxide–Semiconductor) transistors. This is because the operation of a BJT
transistor is based on the physical movement of minority carriers (see Sect. 3.2),
which are not as susceptible to radiation-induced damage as the oxide layer in
MOS transistors. In contrast, MOS transistors rely on the formation of a thin oxide
layer, which can be disrupted by ionizing radiation. Furthermore, MOS transistors
are more prone to Single Event Effects (SEEs, see next section), which occur when
a charged particle strikes the gate oxide and alters the state of the transistor. BJT
transistors are less sensitive to SEEs since the minority carrier transport is not
affected by the radiation.
TID can also cause amplifier gain degradation. This usually manifests as a
reduction in gain with increasing total dose exposure. To compensate for it, more
power needs to be supplied to the device.
And TID can also cause dark signals in camera sensors as a direct effect of
the charging of gate oxides. This is manifested as an increased noise background
and is observed in both CCD and CMOS technologies. As a consequence, the
dynamical range of the imager is compromised. This is a major problem with Star
Trackers that could fail to locate reference stars.
For TID mitigation, for the on-board electronics to maintain its electric perfor-
mance (timing, current consumption) throughout the mission lifetime while subject
to ionizing radiation energy deposition, a typical measure is to add shielding,
which basically consists in adding barriers of certain materials. The effectiveness
3.3 The Space Environment 27

Fig. 3.7 Applicability of SEE to different device types

of a material in shielding radiation is determined by its half-value thicknesses,


that is, the thickness of material that reduces the radiation by half. This value is a
function of the material itself and of the type and energy of ionizing radiation.
As for single event effects, the on-board electronics and on-board software shall
be designed in a way that the SEEs will disrupt the nominal operations of the
subsystem the least. Given that SEEs can be of destructive and non-destructive
nature, different strategies will be defined for each case. Let’s unpack SEEs in the
next section.

3.3.1 Unpacking Single Event Effects (SEEs)

There are different kinds of single event effects, and different types of electronic
devices are susceptible to SEEs in different ways. The table below summa-
rizes how SEEs impact different types of devices, both for non-destructive and
destructive single event effects (Fig. 3.7).22
It can be seen that, for instance, analog and mixed signal circuits tend to be
more robust (immune to different types of SEEs), as opposed to memories and
Field Gate Programmable Arrays (FPGA) which tend to be highly susceptible to
several kinds of SEEs.
The SEEs listed above are split in two halves: non-destructive (that is, unable to
cause permanent damage) or destructive (able to create permanent damage). Let’s
unpack the acronyms.

22FAA, “Single Event Effects Mitigation Techniques Report” https://www.faa.gov/aircraft/air_


cert/design_approvals/air_software/media/TC-15-62.pdf.
28 3 Semiconductors in Space: From Sand to Satellites

3.3.1.1 Non-destructive Effects

. SEU: Single event upset


. MBU: Multiple bit upset
. MCU: Multiple cell upset
. SEFI: Single event functional interrupt
. SET: Single event transient
. SEC-DED: Single error correction and double error detection
. SED: Single event disturb.

Single Event Upset (SEU)


An SEU causes a change of state in a storage cell. The SEU affects memory
devices, latches, registers, and sequential logic. Depending on the size of the depo-
sition region and the amount of charge deposited, a single event can upset more
than one storage cell in which case the effect is called a multiple cell upset (MCU).
Multiple Bit Upset (MBU)
An MBU is defined as a single event that causes more than one bit to be upset
during a single measurement. During an MBU, multiple bit errors in a single word
can be introduced, as well as single bit errors in multiple adjacent words.
Single Event Functional Interrupt (SEFI)
The loss of functionality (or interruption of normal operation) in complex inte-
grated circuits due to perturbation of control registers or clocks is called a single
event functional interrupt (SEFI). An SEFI can generate a burst of errors or long
duration loss of functionality (e.g., lockup). The functionality may be recovered
either by cycling the power, resetting, or reloading a configuration register.
Single Event Transient (SET)
A single event transient (SET) is a short impulse generated in a gate resulting in
the wrong logic state at the combinatorial logic output. The wrong logic state will
propagate if it appeared during the active clock edge.
Single Event Disturb (SED)
The transient unstable state of a static random-access memory (SRAM) cell is
described as resulting from a single event disturb (SED). This unstable SRAM state
will eventually reach a stable state and the characterization will fall under SEU.
Because the unstable state of the cell can be long enough that read instructions
can be performed and soft errors generated, SEDs are identified separately.
3.3 The Space Environment 29

3.3.1.2 Destructive Effects


Destructive Single Event Effects are:

. SHE: Single Event Hard Error


. SEL: Single event latch-up
. SESB: Single Event Snap-Back
. SEB: Single Event Burnout
. SEGR: Single event gate rupture
. SEDR: Single event dielectric rupture.

Single Event Hard Error (SHE)


A single event hard error (SHE) is used to highlight the fact that a neutron-induced
upset (e.g., SEU, MBU) is not recoverable. For example, when a particle hit causes
damage to the device substrate in addition to the flipping bit, an SHE is declared
in lieu of an SEU.
Single Event Latch up (SEL)
In a four-layer semiconductor device, an SEL occurs when the energized particle
activates one of a pair of the parasitic transistors, which combines into a circuit
with large positive feedback. As a result, the circuit turns fully on and causes
a short across the device until it burns up or the power is cycled. The effect of
an electric short is potentially destructive when it results in overheating of the
structure and localized metal fusion.
Single Event Snap-Back (SESB)
SESBs are a subtype of SEL and, like SEL, they exhibit a high current consuming
condition in the affected device. When the energized particle hits near the drain,
an avalanche multiplication of the charge carriers is created. The transistor is open
and remains so (hence, the reference to a latch-up condition) until the power is
cycled (the device snaps back).
Single Event Burnout (SEB)
A single event burnout (SEB) is a condition that can cause device destruction due
to a high current state in a power transistor, and the resulting failure is permanent.
An SEB susceptibility has been shown to decrease with increasing temperature.
SEBs include burnout of power metal oxide–semiconductor field effect transistors
(MOSFET), gate rupture, frozen bits, and noise in charge-coupled devices.
Single Event Gate Rupture (SEGR)
An SEGR is caused by particle bombardment that creates a damaging ionization
column between the gate oxide and drain in power components. It typically results
in leakage currents at the gate and drain that exceed the normal leakage current on
a non-exposed device. SEGRs may have destructive consequences.
30 3 Semiconductors in Space: From Sand to Satellites

Single Event Dielectric Rupture (SEDR)


The single event dielectric rupture (SEDR) has been observed in testing but not in
space-flight data. Therefore, it is currently considered mostly an academic curios-
ity. An SEDR is identified from a small permanent jump in the core power supply
current.

3.3.1.3 Mitigation
For the typical mitigation techniques against SEEs, two distinctive approaches are
frequently used: internal and external. Internal here means, intra-integrated circuit
(inside the chip). For example, certain mitigation techniques require increasing dis-
tances in the semiconductors or adding capacitive hardening of dynamic random
access memories by inserting trench capacitors23 and transmission gates. All these
methods need to be implemented by the chip designers and carry the penalty of
requiring larger semiconductor area, with all costs related. Hardened chips may be
manufactured on insulating substrates instead of the usual semiconductor wafers.
Silicon on insulator (SOI) and silicon on sapphire (SOS) are commonly used.
Another approach is about shielding the package against radiation to reduce expo-
sure of the die or shielding the chips themselves (from neutrons) by use of special
elements like depleted boron consisting of isotope boron-11 in the glass passivation
layer protecting the chips.
External mitigation techniques are those which can be added at the board or
subsystem level, i.e., outside the chip. Current limiters, scheduled power cycling,
memory scrubbing, generous design margins; these are all techniques which do not
require tampering semiconductors and can be added by the system designers, with
the penalty of increasing the overall complexity of the design, especially mass and
power.

23A deep trench capacitor (DTC) is a three-dimensional vertical capacitor formed by etching a
deep trench into a silicon substrate.
The Hectic Ride to Space
4

It can’t be that hard, it’s just lift versus drag and rotation.
—Jamaal (Me, Myself and Irene, 2000) while they try to
figure out how to fly a helicopter

Abstract
Satellites are given the necessary velocity to achieve orbit by means of a launch
vehicle that extracts chemical energy stored in the fuel and transforms it into
heat and work to increase its own positional and motion energy, and with that
the energy of the occasional passengers. Rockets are way more than just heat
engines: they need structures, instrumentation, data links, piping, guidance, nav-
igation, control, computers, and software. The road to orbit is a wild ride, full
of shocks, accelerations and vibrations spacecraft must survive before they even
have a chance to provide a service in orbit. This chapter describes the hectic
journey and how it impacts the satellite’s design process.

The most typical use case of launchers is a vertical rocket in a launchpad at ground
level in which vertical speed gained by its engines thrust ensures departure from
the thickest atmosphere layers, only to then maneuver to imprint the right tan-
gential velocity in the higher atmosphere to achieve the required orbital velocity.
It sounds simple, but it is extremely complicated: rockets have hectic dynamics,
engines are complex, so an incredible variety of situations are experienced by a
spacecraft while attached to a rocket on its way to space. Other launcher technolo-
gies besides vertical launchers exist, which are less popular and less cost effective.
Aircraft based, balloon based, and even slingshot-based launchers exist to disparate
levels of maturity. We will refer to vertical launch as the baseline for this text, due
to its popularity among all available options as of today.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 31


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_4
32 4 The Hectic Ride to Space

A launch generates high stresses on spacecraft structures. From a satellite per-


spective, launch starts when the boosters’ engines ignite (the launcher lifts off from
the pad) and ends with the spacecraft separation, when it is finally allowed to float
alone in space. Mind you, it is a wild ride.
Launchers are generally designed in stages which discretizes a typical flight in
a set of events where different parts of the rocket are separated and jettisoned in
a complex sequence which involves stopping and starting rocket engines, creating
a number of accelerations and shocks in the process. Rockets undergo different
aerodynamic scenarios and can suffer from turbulence just as commercial airliners
do, adding to the overall hecticness of the ride.
In vehicles such as the Ariane 5 launcher, peak axial load factors of 4.2 g are
expected (a load factor is a dimensionless multiple of g that represent the inertia
force acting on the spacecraft or unit).
There are many loads present during rocket flight, both in the lateral and lon-
gitudinal (i.e., axial, or along the rocket long body) directions. Some loads are
predicted as a function of time and others can only be estimated statistically as
random loads. In the mechanical environment, we find:

. Steady state accelerations


. Low frequency vibrations
. Broadband vibrations, such as:
– Random vibrations
– Acoustic vibrations
. Shock vibrations.

Loads are transmitted from the launch vehicle to the payload—the satellite—
through the mechanical interface which bonds the rocket and spacecraft together,
which is usually a ring with bolts.
During lift-off and early phases of the launch, a high level of vibrations transfer
from the vehicle to the payload through their common interface. The principal
sources of vibrations are the rocket engines thrusting, and aerodynamic turbulence
and acoustic noise as pressure waves impinging on light weight appendages which
can be excited beyond safe margins. Everything is so hectic that bolts may simply
come off, as well as cables, boards, or brackets on board the satellite if not properly
secured. As you may imagine, assembling a spacecraft requires adding a good
amount of glue.
In summary, the design process must consider all these factors and include the
necessary measures to ensure launch survivability, considering that the satellite will
be violently shaken and hit several times before it can even get to start thinking of
providing its service.
4.1 Rideshares, Dispensers, and Orbital Transfer Vehicles 33

Tip: Launch vehicle manuals are great sources of information about loads,
sequences, and cool infographics.1 ,2

4.1 Rideshares, Dispensers, and Orbital Transfer Vehicles

NewSpace companies, in general, cannot afford first class tickets on rockets. That
is, having a comfy space for them alone to stretch the legs and sleep like a baby.
Such tickets are worth tens of millions, and because small satellites coming from
NewSpace in general are made on a budget, they can afford what is called a
‘rideshare’ in space jargon. Rideshares are ways the launch providers can make
a better use of their available launch volume and make a profit in the process.
For this, rocket companies provide adapters such as the popular EELV Secondary
Payload Adapter (ESPA3 ) which can accommodate several secondary payloads
attached to the available rings. Launch vehicles can stack several ESPA rings
together. Once they deliver the primary payload, launch providers begin to deliver
the different secondary payloads sequentially. One rocket may contain hundreds
of secondary payloads.4
Broker companies can also make a buck or two by attaching dispensers into
ESPA ports and exploit the space for dispensing smaller satellites. Even further,
companies may attach Orbital Transfer Vehicles (OTVs) which are satellite dis-
pensers with their own propulsion, computers, and attitude control—motherships
of sorts—which can carry several payloads and insert them into customer-specific
target orbits that rocket launchers cannot (or do not want to) go for. A breed of
new actors in the market offering these services have bloomed in the last 5 years
or so, and many of them are still maturing their technologies and trying to reach
commercial capacity, not without problems. The fundamental challenge for OTVs
is to convince customers that it is worth the risk of riding on-board of another full
blown spacecraft in order to reach to the destination orbit. Constellation operators
face a dilemma: acquiescing to the non-ideal orbits rocket launchers drop them
to with high reliability and proven heritage, versus reaching a customized orbit
(within certain limits) on board of a low-maturity, comparatively low-reliability
mothership? At the time of writing—April 2023—the market appears to prefer the
former, but it is true that OTVs are still new, and the trend may eventually change if
they prove to be a reliable option, not only cost-wise but also technically speaking.

1 https://www.arianespace.com/wp-content/uploads/2011/07/Ariane5_Users-Manual_Octobe
r2016.pdf.
2 https://www.spacex.com/media/falcon-users-guide-2021-09.pdf.
3 The EELV Secondary Payload Adapter (ESPA) is an adapter for launching secondary payloads

on orbital launch vehicles. Originally developed for US launch vehicles in the 2000s to launch
secondary payloads on space missions of the United States Department of Defense that used the
Atlas V and Delta IV, the adapter design has become a de facto standard and is now also used for
spaceflight missions on non-governmental private spacecraft missions as well.
4 Transporter-6 mission from SpaceX launched 114 secondary payloads in early January 2023.
Configuring Spacecraft
5

Make it simple, but significant.


—Don Draper, fictional character on Mad Men

Abstract
Configuring a spacecraft means finding an optimal geometric and physical shape
that can satisfy requirements within cost and schedule. And it also means defin-
ing the spatial and functional arrangement of the on-board equipment. Satellites
are like Russian dolls: boxes inside boxes. Some of those ‘boxes’ are required
to be at specific locations and looking in specific directions. This chapter dives
into the complex, multidisciplinary challenge of configuring spacecraft where
the main goal is to make everybody as happy as possible, only to end up making
almost everybody unhappy.

Satellites are highly intertwined collections of subsystems (see next chapter for
more details about these subsystems), and the main outcome of the configuration
analysis phase is to define a geometric baseline of the placement and orientation of
the most relevant domains (namely structure, power, attitude control and payload,
which concentrate a high percentage of the spacecraft mass and complexity) while
observing the high level requirements, including the envelope constraints, which is
the volume available to occupy inside a selected launch vehicle. If the spacecraft
flies in a rideshare (see Sect. 4.1), the envelope is typically a tight squeeze, which
makes the configuration process more interesting, to use a polite term.
The configuration process starts at the beginning of the project, and it is peri-
odically reviewed throughout the system lifecycle. In short, configuration analysis

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 35


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_5
36 5 Configuring Spacecraft

is practically a permanent, on-going activity. During configuration, each subsys-


tem (including payload) contributes to a tree-like decomposition called the product
breakdown structure (PBS), including details such as:

. List of units, with unique designators


. Environmental limits (allowable flight temperatures, accelerations, radiation)
. Power, size and mass of subsystems
. Interfaces.

Configuring spacecraft requires analyzing the information provided by the selected


launch vehicle thoroughly to understand the allowed envelope inside the fairing,
the mechanical interfaces, and relevant requirements in terms of structural fre-
quencies and loads as we saw in the previous chapter. Configuring spacecraft must
also include a good amount of technical bookkeeping and budgeting, which entails
analyzing the stocks and flows of on-board resources such as power, pointing, data
links, propellant, thermal and data, in order to understand and ensure their gener-
ation, consumption, and storage. In an iterative capacity, and as more information
flows in, the configuration analysis shall:

. Refine the detailed Bill of Materials (also called Required Equipment List, or
REL)
. Refine mass of each unit + contingency
. Refine power consumption of each unit + contingency
. Refine mass properties along the body axes for control purposes.

While configuring spacecraft, the feedback loops are everywhere. For instance,
the power budget feeds solar arrays sizing which feeds the mass budget which
impacts the propulsion design which feeds the power budget (if electric propulsion
is chosen). The efficiencies of the power subsystem design process feeds battery
dimensioning activities, which is also shaped by orbital period and the eclipse
time (during eclipses, there is no power generation, so satellites survive with the
power batteries source). Propulsion analysis feeds the propellant budget which in
turn feeds thrusters’ definition (thrust, orientation, type, and number of thrusters)
which also feeds the mass budget and the power budget because thrusters are
electrically actuated.
Attitude control configuration activities require analyzing star-tracker orienta-
tion to avoid sun intrusion, which may require to move equipment out of their
way, but as more elements come into the satellite body, there’s less and less space
for playing this “Tetris” of sorts. Analysis of shadows must be performed to avoid
a deployable appendage affecting optics or solar panels, same as plume impinge-
ment analysis, which is a phenomenon that occurs in propulsion subsystems where
the exhaust plume impinges on a surface such as a spacecraft structure, and causes
damage or degradation.
In time, the mass budget starts to provide an inertia matrix—which indicates
how mass is distributed along the spacecraft axes—the actuators must be able to
5 Configuring Spacecraft 37

slew for maneuvers, which feeds the reaction wheel sizing, which in turn depends
on orbital disturbances, and so on.
There is an unwritten axiom that silently rules space engineers: you move a
screw in a satellite, and you unleash hell.
For other subsystems, things don’t get any easier. For the payload: size, mass,
mounting requirements, deployment sequence (if any), fields of view, power
required, electrical interfaces, data transfer requirements, electromagnetic inter-
ference requirements. All with the added complexity that payloads are usually
supplied by a different organization compared to the bus designer, so add on top
of that all the complexities involved with two dissimilar companies with different
cultures, possibly different languages, and different processes and methodologies
interacting together.
For communications, the process includes antennas dimensioning and orienta-
tion, uplink/downlink requirements in terms of signal to noise ratio (or its more
specialized derivatives like Eb/N01 ), power output of transmitters, sensitivity for
receivers, antenna apertures, attenuators, weather attenuation, and so forth. For
data links, ground stations are rarely owned by the spacecraft operators, so bus
providers must ensure compatibility with equipment of third-party ground stations
which may vary from provider to provider.
As subsystem design matures, light appears at the end of the tunnel, and
the configuration starts to feed from better and better data, taking more detailed
information in stemming from all the subdomains and allowing for certain modi-
fications to be made. The earlier the changes are done, the cheaper and simpler to
execute them.
Configuration analysis is a collaborative activity between systems engineers,
domain experts and project managers. Change is, ironically, a constant in space
design, and a fluid communication between the team is key to catch the changes
as they inevitably happen. Big changes in space missions happening too late in the
schedule is usually bad news: the bigger the change, the bigger the ripple waves,
and the broader the analysis that must be performed to assess that everything is
still compliant with requirements.
Once the configuration process gains enough maturity, it provides the confi-
dence to procure or manufacture the different elements and components, and soon
the design starts to be ready for Assembly, Integration and Test (AIT, see Sect. 6.8
for more details). This involves handling each component—either in-house or sup-
plied by a third party—and assembling them into the final configuration, without
forgetting the glue of course (see Chap. 4).
With the spacecraft assembled, the next step is to test and verify it. This involves
a series of comprehensive tests to ensure that the spacecraft can withstand the
stresses of the launch and operate effectively in the indented environment. This

1 In digital communications, Eb/N0 (energy per bit to noise power spectral density ratio) is a
normalized signal-to-noise ratio (SNR) measure, also known as the “SNR per bit”. It is espe-
cially useful when comparing the bit error rate (BER) performance of different digital modulation
schemes without taking bandwidth into account.
38 5 Configuring Spacecraft

stage involves verifying that each subsystem is functioning correctly, and that the
spacecraft design is, a priori, able to meet the mission objectives, which largely
drive the design process. We’ve talked about verification thus far. In terms of
validation, the mission objectives will only be validated with the satellite placed
in the target environment. Bibliography loves to confuse these two terms.
As we hinted few paragraphs ago, some spacecraft designs are conceived as
“buses” or “platforms”, that is, they are generic designs which are not strictly
mapped to one particular mission but aim to be “multi-mission”. Multi-mission
design requires a different, more modular approach where the architecture must
be broken down into building blocks that can be reused from mission to mission.
This is far beyond the scope of this chapter, but we will say here that multi-mission
spacecraft design requires applying product variant management techniques to
ensure that the “core” design and the specific mission designs map consistently
without jeopardizing reliability while at the same time maximizing reuse and com-
monality. I wrote extensively about modularity in spacecraft in some other text,2
feel free to go and take a deeper look.

2 https://link.springer.com/chapter/10.1007/978-3-030-66898-3_5.
A Peek Under the Hood
6

A fist is more than the sum of its fingers.


—Margaret Atwood

Abstract
To have a working satellite in orbit, you need a set of “building blocks” in place.
As a bare minimum, all satellites need housing, need a way of generating and
storing power, need radio links, orientation control, computers, and industrial
amounts of software, and glue. This chapter describes these building blocks and
how their interconnection creates working spacecraft.

6.1 The Skeleton: Structures and Mechanisms

The history of aerospace is also the history of materials.


In 1915, just 12 years after the Wright Brothers had made the first ever pow-
ered flight, this was a hot topic for discussion among aviation experts: how could
metal fly? A century later, we are flying rockets made of sophisticated composite
materials carrying satellites made of aluminum alloys, carbon fiber, titanium, and
many other materials, including some 3D printed parts.
But, a hundred years ago, things were a tad different. Aircraft were made to be
as light as possible, often using wood, steel wires and canvas; the idea of a plane
made entirely of metal seemed technically infeasible.
One man knew differently, though. German pioneer Hugo Junkers saw the
future of aviation not only in aerial battles and flying competitions but in large-
scale transport of goods and passengers. That would require a major change to the
way aircraft were made. His revolutionary J 1 (the space character between the

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 39


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_6
40 6 A Peek Under the Hood

Fig. 6.1 Junkers J 1 (public domain)

J and the 1 is intentional) was the world’s first all-metal aircraft, as well as the
first to use a single monoplane wing. The J 1 was nicknamed the Blechesel (“Tin
Donkey” or “Sheet Metal Donkey”).
Junkers found that steel made the J 1 tough and durable, but heavy and sluggish
to handle. He turned his attention to aluminum, which had emerged at the start of
the twentieth century as a viable manufacturing material. Lightweight and strong,
it is a third the weight of steel, making it ideal for aircraft.
He used it to develop the world’s first civil airliners, such as the F13 and G24.
Junkers’ work was noticed by Henry Ford, who borrowed heavily from it (too
heavily, said Junkers, and decided to take him to court1 ) to make the Ford Trimotor
in 1925. These aircraft welcomed the age of long-distance passenger aviation,
although it wasn’t until the early 1930s that metal aircraft could be manufactured
cost-effectively (Fig. 6.1).
In space, materials are incredibly important, although not from an aerodynamic
perspective. Aerodynamics does not play such a relevant part even though there is,
to be technically accurate, aerodynamic drag in space. Such small drag—due to the
thin remnants of the atmosphere’s upper layers—manifests as a disturbance force
that alters satellites’ orbits and as a torque acting in the spacecraft body which
affects the satellite’s orientation. Spacecraft actuators must be sized to account for
this disturbance, and other disturbances such as gravity, radiation pressure, etc.
(see Sect. 6.4 for more details on disturbances).
Mass reduction is a strong driver for material research in space because launch
cost is related to the amount of mass to be launched. But mass cannot be opti-
mized or minimized in an isolated manner, because of the complex tradeoff mass
reduction may cause on stiffness, thermal conductivity, and structural response.
Schematically, spacecraft metallic structures can be modeled as a distributed
network of springs, masses, and dampers. The “spring” part in a structure comes

1 http://www.blancorincon.com/Fragatas/The%20Story%20of%20the%20Ford%20Trimotor.htm.
6.1 The Skeleton: Structures and Mechanisms 41

from the incapacity of the materials of being perfectly rigid, as all materials dis-
tort under load. Up to certain loads, materials show a linear relationship between
load and displacement (just like a linear spring). Within this region, the material
or component will return to its original shape should the load be removed. If we
continue loading it, the material will not be linear anymore, and it will show per-
manent deformation and rupture will be reached. The stiffness and strength depend
on cross-sectional area, length, and material. Stiffness is defined as a measure of
the load required to cause a certain amount of deflection in the material.
Strength-to-weight ratio is a critical factor in choosing structural materials for
spacecraft. Static and dynamic loads must be considered, along with thermal per-
formance, corrosion protection, manufacturability, reparability, and cost. Static
loads are those loads which remain constant in time whereas dynamic loads vary
with time. Examples of static loads are the weight of computers and units on board
(load is present when a steady acceleration is applied). Examples of dynamic loads
are engine thrust, sound pressure and gusts of wind during launch.
A structure’s stiffness and mass are distributed as a result of the material’s
elasticity and density. When subjected to time-varying forces, a structure’s total
response is the sum of the responses of its mode of vibration. Fortunately, only
a small number of modes (typically those with lowest frequency) are of interest.
The displacement associated with higher order modes isn’t enough to produce
significant stress.
The frequency band over which vibration is a concern depends on the structure’s
parameters, such as size, material, geometry and on the excitation environment.
Every structure has an infinite number of natural frequencies corresponding to dif-
ferent mode shapes of vibration. For a long highway bridge, for example, excited
by wind gusts, only the modes with frequencies less than one or two Hz might be
significant. For a spacecraft’s primary structure during launch, one might need to
consider modes up to 50 Hz. For smaller structures like unit boxes and electronic
components, their frequency responses can be at several hundred Hertz.
One of the main requirements driving spacecraft structures is to withstand the
wild ride on top of the rockets (see Chap. 4) without degradation to ensure a
correct performance of the mission once in orbit, because, well, that’s when the
actual work starts. Structures provide housing to spacecraft’s key components in
specific locations and orientations, considering thermal management, field of view,
orientation, etc. But structures also:

. Protect the spacecraft’s components from the environment during ground


operations, launch, deployment, and mission operations.
. Ensure that structural vibration will not interfere with the launch vehicle’s
control system. Similarly, the spacecraft’s structural vibration in its deployed
configuration must not interfere with its own control system.
. Ensure on-board vibrations will not affect payload performance.
. Ensure that the materials used will survive ground, launch and on-orbit
environments (pressure, humidity, radiation, contamination, thermal cycling,
42 6 A Peek Under the Hood

and atomic particles) without rupturing, collapsing, excessively distorting or


contaminating critical components.

The primary structure is the skeleton of the spacecraft, and the main vessel of
the loads transferred between the spacecraft’s components and the launch vehicle.
The primary structure provides the mechanical interface not only with the launch
vehicle but also with all the mechanical ground support equipment (MGSE) needed
during the design and development phase, notably during AIT activities.
One of the most important design decisions in space projects is the defi-
nition of the primary structure. But secondary structures are crucial as well.
Secondary structures include support beams, booms, trusses, and solar panels.
Smaller structures, such as boxes and brackets that support harnesses are called
tertiary structures. The primary structure is designed for stiffness (or natural fre-
quency) and to survive steady-state accelerations and transient loading. Secondary
structures are also designed for stiffness, but they are also affected by on-orbit
thermal cycling, acoustic pressure, and the high frequency environment.
Structure design complexity is constrained by cost and by the requirement that
all its elements should be producible, and testable. A producible design is one
that the engineers can build using affordable raw materials, using established and
simple processes and available equipment and tooling. A testable product is one
that can be handled and verified on ground without the need of overly complex
infrastructure and procedures, and with measurable success criteria.
Most materials expand when heated and contract when cooled. In space, the Sun
ensures that an orbiting spacecraft’s temperatures are neither uniform nor constant.
As a result, spacecraft structures distort. The various materials that make up a
spacecraft expand or contract different amounts as temperatures change. Thus, they
push and pull on each other, resulting in stresses that can cause them to yield or
rupture. Most spacecraft requires accurate predictions of thermal deformations to
verify pointing and alignment requirement for sensors or communication antennas.
Thermo-elastic forces are usually the ones which drive the joints design in structure
with dissimilar materials regarding CTE (coefficient of thermal expansion) thus
causing high shear loads in fasteners that joint dissimilar materials.
There are guidelines and practices to help to reduce the effects of thermal
deformation and stresses:

. One shall contain critical members and assemblies in multilayer insulation


(MLI) to keep temperature changes and gradients as low as possible.
. When possible, one must design structures and their interfaces with isostatic
mount designs in order to avoid thermally induced boundary loads.
. Design structural members from materials with low CTE to minimize thermal
deformations.
. For structures with large temperature excursions, it’s recommended to use mate-
rials with similar CTEs. When attaching members of different materials, one
must design the joints to allow the expected differences in thermal expansion.
6.2 The Data Links: From Sparks to Mobile Networks, Lasers, and In-Orbit Networks 43

The structure is an essential subsystem in any spacecraft. There is little point in


having the most sophisticated computers, the fanciest software-defined avionics if
the whole thing will fall apart into pieces as soon as the rocket lifts off.

6.2 The Data Links: From Sparks to Mobile Networks, Lasers,


and In-Orbit Networks

German physicist Heinrich Hertz was among the first to experiment with the gen-
eration of radio waves by creating high voltage sparks between two conductors,
pioneering with the investigations of electromagnetics and propagation.
In reality, there were vast numbers of radio waves in the universe long before
Hertz performed his famous experiments. Electromagnetic waves are generated
naturally whenever electric charges are accelerated or decelerated. All hot objects,
in which charged particles are in rapid random motion, radiate electromagnetic
energy in various frequencies. The stars are potent sources of electromagnetic
energy, which is the basis of radio astronomy. On our own planet, atmospheric
events such as lightning strikes produce showers of radio energy, noticeable as
the background crashes and crackles heard on broadcast receivers during thunder-
storms. In most of these cases of natural generation, the radio energy is incoherent,
characterized by a jumble of photons of disparate energies. The same is true
of many human-made sources of electromagnetic disturbances, such as electrical
machinery.2
Italian inventor and electrical engineer Guglielmo Marconi had successfully
transmitted radio energy for a bit more than a kilometer, generating his own “in-
visible waves” in Pontecchio, Italy. After failing to impress the Italian government,
Marconi traveled to England at the age of 22, where he found financial backers
for his work. By 1897, Marconi was broadcasting radio waves up to 20 km away
and was commissioned to set up a wireless station on the Isle of Wight so that
Queen Victoria could send Morse code messages to her son while he was aboard
his yacht.
In 1899, Marconi began work on a transatlantic broadcast, believing (unlike the
leading physicists of the day) that the signals would follow the curvature of the
earth. In 1901, he achieved a 3430 km transmission of three Ss (or three sets of
three dots in Morse code) from Poldhu, Cornwall to Signal Hill in St. John’s.
Needless to say, his work led to a communications revolution. But, why exactly?
Marconi will not be remembered for generating precise radio waves—he was,
technically speaking, spouting noise in a more or less controlled manner—but he
will be remembered for leading the work to convey information wirelessly by

2Adapted from Radio Antennas and Propagation: Radio Engineering Fundamentals (1st Edition),
William Gosling.
44 6 A Peek Under the Hood

manipulating electrical signals with the help of electric circuitry. In more modern
terms, he fathered radio access technology.3
In the late 1890s, Canadian American inventor Reginald Fessenden started to
realize that spark transmitters were not as elegant, and he realized he could develop
a far more efficient transmitter and coherent receiver combination. To this end,
he worked on developing a high-speed alternator that generated pure sine waves
and produced a continuous train of propagating waves of substantially uniform
nature, or, in modern terminology, a continuous-wave (CW) transmitter, which
still consisted in turning on and off a sine wave signal of fixed characteristics in a
rather sloppy way.
In time, the idea of doing such a thing—switching a signal on and off—would
start to give way to a more subtle approach: keeping the carrier on and altering the
attributes of said carrier at the rhythm of another signal containing the information
to be transmitted. In other words, modulation. Fessenden would at some point
introduce the concept of signal heterodyning, where a receiver would have a local
oscillator producing a radio signal adjusted to be close in frequency to the signal
being received. When the two signals would be mixed, a “beat” frequency equal
to the difference between the two frequencies was created. Adjusting the local
oscillator frequency would correctly put the beat frequency in the audio range,
where it would be heard as a tone in a receiver’s speaker whenever the transmitter
signal is present. Now we could hear tones, and thus the Morse code “dots” and
“dashes” would become audible as beeping sounds. Nice.
Soon, we would realize that nothing was preventing us from opportunistically
altering a carrier’s amplitude with something more than mere tones (a tone is a
signal composed of a single frequency) but using spectrally richer signals, such
as audio.4 Consequently, voice communication and broadcasting would become a
very popular thing.
With all this, it soon became clear that:

. The bandwidth of the modulating signal (its spectral richness of frequencies)


would turn out to be of great importance since it would start requiring cer-
tain capabilities from transmitters and receivers for protecting such bandwidth
across their signal chains if one pretended to ensure audio quality.
. The rapid proliferation of radio transmitters would make evident the need for
spectrum coordination. After all, the experimental early phase of radio commu-
nications was starting to be over, and we had to start regulating transmitters’
activity to reduce unwanted interference and aim for a more responsible use of
a now key natural resource like the radio spectrum.

3 A Radio Access Technology or (RAT) is the underlying physical connection method for a radio
based communication network.
4 Audio is the conversion of sounds waves into electric signals by means of a transducer such as a

microphone.
6.2 The Data Links: From Sparks to Mobile Networks, Lasers, and In-Orbit Networks 45

Additionally, and as other technologies like transistors, integrated circuits and com-
puters were maturing fast, we would realize that the modulating signals did not
need to be only of analog nature. Digital signals could also be used to alter a car-
rier’s attribute. In the meantime, we also would figure out ways of altering other
attributes of sinusoidal signals, namely frequency, and phase, which showed better
bandwidth efficiencies, where a change in the carrier’s phase could be coded to
represent several data bits.
Digital modulation would make a glorious appearance, giving us the possibil-
ity of sending digital data between a transmitter and a receiver. A small practical
problem showed up: Fourier analysis5 taught us that perfectly squared waves are
equivalent to a sine wave at the fundamental frequency summed with an infinite
series of odd-multiple harmonics at decreasing amplitudes (describing a sinc func-
tion in the frequency domain). This meant a quasi-infinite bandwidth to handle!
So, inevitably, a breadth of techniques for limiting the bandwidth of digital signals
prior to transmission materialized for ensuring that practical bandwidths could be
achieved while guaranteeing the ones and zeros present in the data stream could
still be discerned at the receiving end.
A metric started to surface as important: data rate; that is, the number of bits of
information that could be transmitted or received per unit of time (usually per sec-
ond). We wanted it to be higher and higher (and we still do). Shannon and Hartley
with their homonymous theorem6 would come to tell us to hold our horses because
the theoretical upper bound on the information rate of data that could be commu-
nicated at an arbitrarily low error rate for a given received signal power through
an analog communication channel subject to noise would be constrained by the
bandwidth of the channel, the average received signal power over such bandwidth
and the average power of the noise and interference over said bandwidth. Said
theorem would make it very visible that such rates would be painfully governed
by the signal-to-noise ratio (SNR).
In time, our engineering predecessors would burn their brains devising tech-
niques to squeeze more and more from a now contested and noisy radio spectrum.
This included techniques to improve said SNR—pumping out power is not always
a practical solution—with the aim to achieve lower and lower bit error rates (BER).
Antenna technology improvements and techniques combined with the appearance
of a multitude of coding techniques would allow radio engineers to reach the same
BERs at lower SNRs, making their designs simpler.
Spectrum use sophistication wildly kicked in. It became understood that we
could also partition the spectrum of a communications channel in a series of
non-overlapping bands called subcarriers (a technique called frequency-division
multiplexing, or FDM). And now the concept of orthogonality would find a great
application: remember we said before that a digital signal’s spectrum is described

5 In mathematics, Fourier analysis is the study of the way general functions may be represented or
approximated by sums of simpler trigonometric functions.
6 https://en.wikipedia.org/wiki/Shannon%E2%80%93Hartley_theorem.
46 6 A Peek Under the Hood

by a sinc function? Well, if we chose the frequencies of contiguous sub-bands


wisely, the individual peaks of the subcarriers’ sincs will line up with the nulls
of the other subcarriers. A clever, constructive use of interference. Such over-
lap of spectral energy still gives us the ability to recover the original signal,
which is the foundation of Orthogonal Frequency-Division Multiplexing (OFDM),
a method to convey data using a set of orthogonal subcarriers at lower data rates
as opposed as to modulating a single carrier at a very high data rate. OFDM would
become the foundation of Wi-Fi (IEEE 802.11a/g/n/ac), 4G LTE and 5G mobile
communication technologies.
Fast-forwarding to these days, wireless networks are ubiquitous. The mobile
phones we carry in our pockets are continuously exchanging data with cell
phone towers at high data rates, with antenna towers operating arrays of anten-
nas equipped with techniques to track users spatially, serving a big number of
moving targets such as smartphones or connected devices as they move around
cities, across the globe.
Note that, thus far, we have not talked at all about what’s in the data being
transmitted. And this is intentional. Wireless links can be thought of as onions of
sorts, in the sense that they are designed in layers. So far, we have been strictly
talking about the lowermost layer: the physical layer. This is about electric signals
traversing across a transmitter, through an antenna that emits them as a shower of
radio quanta which propagates towards a receiver, without caring about what those
bits being transmitted mean.
Engineers would spend a long time defining different formats and protocols for
the data to follow. Such protocols would address important needs such as address-
ing, error handling, automatic retransmission requests, forward-error-correction,
fragmentation, virtual channels multiplexing, and many other features. Still, at the
end of the day, the fact is that only bits are communicated. We will expand on this
in Sect. 6.2.3.
But how does space fits in all this? Space hasn’t grown unaware of the tech-
nological progress of radio engineering in the last century. Radio links have been,
for obvious reasons, the weapons of choice for transferring data from satellites
to the ground. But the space industry seems to be showing, in terms of the
state-of-the-art, two different faces: SATCOM and Earth Observation. On the
SATCOM face, the spectrum exploitation level of sophistication is very high,
using advanced frequency-division multiplexing or time-division multiplexing
techniques to squeeze as many customers as possible into a single communica-
tions satellite. Called multiple access, it allows several carriers from several earth
stations to access a SATCOM’s antenna. To serve its customers, the payload of a
SATCOM satellite incorporates a repeater that consists of one or more channels,
called transponders, operating at all times in parallel on different sub-bands of the
total bandwidth used. SATCOM networks can be based on a single-beam antenna
payload or multibeam antenna payload. In the context of a single-beam payload,
the carriers transmitted by earth stations access the same satellite receiving antenna
beam, and these same earth stations can receive all the carriers retransmitted by
6.2 The Data Links: From Sparks to Mobile Networks, Lasers, and In-Orbit Networks 47

the same antenna. The number of users that can be served with a single SATCOM
satellite and with only one parabolic antenna is mesmerizing.
On the Earth Observation face, the level of sophistication in terms of spectrum
use decreases a bit. Earth observation satellites are still using predominantly sin-
gle carrier channels and techniques, somewhat “classic” modulation schemes and
protocols, and limited or no beam forming or beam scanning capabilities. In Earth
Observation missions in LEO, the wireless links are mechanically steered with the
satellite moving its body-mounted antenna to track a ground-based, big antenna
and the ground station antenna, in turn, tracking the satellite.
Satellites, irrespective of SATCOM, EO, pure science, or any other application,
are about data. No satellite exists for the sake of being up there isolated, floating
alone in space. Satellites are always part of a mission, and missions are about
acquiring, transporting, forwarding, or downlinking various kinds of digital data,
be it health status telemetry, internet traffic, IoT transactions, deep space astron-
omy, or a hyperspectral image of some location on the ground. Such bits and bytes
need to move fast because someone, somewhere is eagerly waiting to make sense
out of them.
Augmented with other technologies like lasers, satellites are becoming nodes in
a global-area wireless network. Nothing is really stopping us from being able to
integrate satellites to our ground-based wireless networks, including our modern
mobile networks.

6.2.1 Mobile Networks and Satellites

Mobile communications have evolved from generation to generation, adding better


and better capabilities, and the trend is far from being over.
Initially, the first two mobile comms generations aimed to ensure the efficient
transmission of voice information. Newer generations added more digital technolo-
gies, bandwidth-efficient modulation schemes and a smarter use of an increasingly
contested spectrum, allowing faster data rates, more secure radio access technology
and a breadth of protocol layers ready to carry internet protocol (IP) datagrams.
This, in turn, unlocked the possibility of merging mobile and global networks
seamlessly, giving the end users the possibility of enjoying the same services and
applications they used on their computers now from their smartphones, without
compromising on quality.
The last generation of deployed mobile communications—the fifth generation,
or 5G—incorporates capabilities that were unheard of in previous generations. To
exploit the potential, a new radio-access technology known as NR (New Radio)
48 6 A Peek Under the Hood

was devised. The key innovative factor of defining a brand-new radio access tech-
nology meant that NR, unlike previous evolutions, was not restricted by a need to
retain backwards compatibility. This permitted rolling out a set of key use cases:

. Enhanced mobile broadband (eMBB), which appears as the most straightfor-


ward evolution from previous generations, enabling larger data volumes and
further enhanced user experience.
. Massive machine-type communication (mMTC), providing services that are
characterized by a massive number of devices, such as remote sensors, actu-
ators, and monitoring equipment. Key requirements for such services include
ultra-low device cost and ultra-low power consumption, allowing for extended
device battery life of up to at least several years. Typically, each device con-
sumes and generates only a relatively small amount of data, thus, support for
high data rates is of less importance.
. Ultra-reliable and low-latency communication (URLLC), with services targeted
to ensure very low latency and extremely high reliability. Examples hereof are
traffic safety, smart cities, automatic control, power grid and factory automation.

Although 5G was deployed several years ago (Release 157 ), it keeps evolving.
The newest evolution of 5G (Release 18>), is called 5G Advanced and it will add
support for new applications and use cases. 5G Advanced is expected to bring
significant enhancements around smarter network management by incorporating
artificial intelligence and machine learning techniques for beam management,
load balancing, channel state information feedback enhancement, improvements
in positioning accuracy and user equipment network slicing. 5G Advanced is
envisioning incorporating low-latency audio and video streaming services aimed
for Extended reality (XR), along with a more energy-efficient use of network
resources, and Deterministic Networking (DetNet8 ) capabilities to ensure deter-
ministic data paths for real-time applications with extremely low data loss rates
and packet delay variation.
More importantly, recent releases of 5G have been progressing on integrating
satellite communications with 5G NR techniques called “non-terrestrial networks”,
or NTN. The study of non-terrestrial networks includes identifying NTN scenarios,
architectures and use cases by considering the integration of satellite access in the
5G network including roaming, broadcast/multicast, secure private networks, etc.

7 3GPP uses a system of “Releases” which provide developers with a stable platform for the imple-
mentation of features at a given point and then allow for the addition of new functionality in
subsequent Releases.
8 Deterministic Networking (DetNet) is an effort by the IETF DetNet Working Group to study

implementation of deterministic data paths for real-time applications with extremely low data loss
rates, packet delay variation (jitter), and bounded latency, such as audio and video streaming,
industrial automation, and vehicle control.
6.2 The Data Links: From Sparks to Mobile Networks, Lasers, and In-Orbit Networks 49

The synergy between 5G and satellites is neither mere speculation nor hype,
but tangible technical traction towards integrating space technology and mobile
communications together.
While 5G Advanced is about adapting the already established generation for
new incremental use cases, 6G (Release 20>) is natively designed for the human
digital needs of the next decade. The sixth generation is already in the mak-
ing, coordinated by the 3rd Generation Partnership Project (3GPP9 ), the standards
development organization behind the 6G initial research of enabling technologies,
the definition of the requirements, the technical steering, and identification of use
cases. An activity which is already ongoing and will span for the next half decade
or so, refining the architecture and commencing implementation.
The core driving factors for 6G will revolve around enhancing human commu-
nication, including immersive experience, telepresence, and multimodal collabora-
tion and interaction. 6G will also aim to enhance machine communication, with
the focus on autonomous machines and vehicles capable of sensing their surround-
ing environment in real time (network as a sensor). 6G will provide key enabling
services, such as hyper-precise positioning, mapping, and smart health.
Satellites and humans carrying smartphones have more in common than what
meets the eye. Both are moving nodes in adaptive, time-variant networks. Unlike
us humans, who roam around cities following rather complex adaptive patterns,
satellites describe more deterministic paths in orbit, and by choosing their orbiting
geometry carefully, connected constellations in Low Earth Orbit (LEO) can be
deployed to achieve global coverage with low latency and smaller propagation
losses.
Today, almost half of the world’s population still lives in areas that do not have
basic connectivity services. Non-terrestrial networks can provide affordable and
reliable broadband services for areas where mobile operators do not find commer-
cial feasibility to build terrestrial networks. By integrating different non-terrestrial
network systems together, such as LEO satellites, unmanned aerial vehicles, and
high-altitude platforms, non-terrestrial networks can be easily rolled out, connect-
ing people through various devices such as smartphones and laptops, and helping
sense and monitor critical infrastructure in a secure and power-efficient manner.

6.2.2 Lasers in Orbit

A reality which most Earth Observation companies face every day is: data bot-
tlenecks. Satellites generate lots of data—a single acquisition from a camera or
a radar can account for several gigabytes of raw data, creating terabytes of pay-
load data per day. According to the European Space Agency (ESA), the Sentinel-1
radar satellite produces approximately 10 terabytes (TB) of data per day, while

9 See more here: https://www.3gpp.org/.


50 6 A Peek Under the Hood

the Sentinel-2 multispectral instrument generates around 20 TB of data per day.10


Another example is Landsat 8, a joint mission of NASA and the United States
Geological Survey (USGS), which captures images of the Earth’s surface at a spa-
tial resolution of 30 m. Landsat 8 generates approximately 700 scenes per day,
with each scene covering an area of about 185 km by 185 km. Each scene con-
tains approximately 700 megabytes (MB) of data, which adds up to approximately
500 gigabytes (GB) of data per day.11
But the real deal is how to make sure this amount of data gets to the right
hands as fast as possible. On-board storage is not infinite, and more importantly,
customers are having problems which are waiting to be solved with the help of
geospatial data.
As we have described above, space systems have relied on radio links to transfer
payload and housekeeping—that is, health status—data to the ground stations, and
to end users. And we also saw the dark side of radio links: interference, noise,
and spectrum sharing. These are all somewhat intertwined—let’s unpack this a bit
more.
There is no such thing as a true point-to-point radio link. This means, one ideal
transmitting antenna sending electromagnetic photons to be exclusively captured
by a distant receiving entity. In reality, there are scattered photons all around the
place. Statistically, a substantial number of photons will arrive at the intended
receiving entity and the signal will be hopefully reconstructed—if the receiving
entity is active and capable of doing so, but what happens with those nomad radio
photons which went elsewhere? That is the key here. We as humans have managed
to create billions of radiating/transmitting artifacts around the world, all of them
scattering photons outside of their ideal boundaries, creating a soup of radio waves
wandering around, getting where nobody called them and knocking the door at the
wrong antennas. As we also described before, there are human-made transmitting
devices which are not purposely meant to be radio transmitters—a car engine
starting spreads radio energy in many different bands with various energies; such
radio energy is a by-product of the workings of internal electric devices of the
engine such as spark plugs, electric motors, etc.
Transmitting and receiving antennas have their own spatial preference when it
comes to collecting radio quanta—they collect more from some directions than
others—and such preference is never “laser focused” (pun intended I guess?). For
any receiving antenna collecting “right” and “wrong” electromagnetic energy, we
can call the former signal and the latter noise. The proportion between these two is
the already familiar signal-to-noise ratio. Due to the spectrum coordination needs
we talked about in the previous section, every satellite carrying a radio transmitter
requires approval from the International Telecommunications Union, or ITU. Such
need for coordination does not come without its dose of bureaucracy. As more
and more actors are coming into the space industry, with shorter timelines and the

10 Retrieved from https://sentinel.esa.int/web/sentinel/sentinel-data-access.


11 Retrieved from https://www.usgs.gov/media/files/landsat-8-data-users-handbook.
6.2 The Data Links: From Sparks to Mobile Networks, Lasers, and In-Orbit Networks 51

determination to rapidly get into the market, the radio licensing process can be
slow and cumbersome.
In summary, radio links are great and are still getting better as technol-
ogy matures, but the main downsides—notably interference and compliance/
regulation—are still there and will be there. The alternative to this? Laser
communications. But why lasers?
Laser communication is not regulated by the International Telecommunication
Union, which means it can be used without restrictions and does not require licens-
ing. The reason for this should be clear by now: laser-based links are comparatively
more focused than radio links—they have small beamwidths, and incredibly high
bandwidths—which means that between transmitter and receiver there are less
photons going astray. How? Lasers produce a narrow beam of light in which all
of its composing waves have very similar wavelengths, and they travel together
in phase—the emitted photons are “in step” and have a definite phase relation to
each other—concentrating a lot of energy on a very small area.
As there is no such a thing as a free lunch, adding laser communications to
satellites does not come without some challenges. Because the laser beam widths
are so narrow, this requires that the pointing capabilities of the satellites carry-
ing laser terminals be top-notch. What is more, satellites have to be structurally
optimized to carry the highly sensitive optical equipment, reducing, and filtering
mechanical jitter and unwanted vibrations, which calls for a solid structural anal-
ysis and design to precisely understand how the vibrations find their paths across
the satellite’s structure. Last but not least, the on-board data handling architecture
must be up to the task by allowing the high data rate frames to seamlessly flow
from the optical terminal to the on-board resources such as processing units and
data recorders, and vice versa.
Another important property of laser communications is that their beams cannot
pass through solid/opaque objects. Hence, manmade structures on earth as well as
cloud cover can cause some attenuation in earth to space communications. How-
ever, it is not a major deterrent, and can be overcome with careful planning and
considerations. More importantly, there are no such hurdles in space-to-space com-
munications and hence laser technology is ideal for inter-satellite links and other
space-based communication.
There is quite some activity in the laser comms sector happening at the
moment.12 NASA demonstrated an optical link from the Moon in 2018 with the
Lunar Laser Communications Demonstration (LLCD). The LLCD demonstration
consisted of a space terminal on the LADEE spacecraft13 and three ground ter-
minals on Earth. Together, they demonstrated that it was possible to transfer up
to 622 Mbps (megabits per second) of data from the Moon with a space termi-
nal that weighs less, uses less power, and occupies less space than a comparable

12 https://www.nasa.gov/feature/goddard/2021/laser-communications-empowering-more-data-
than-ever-before.
13 https://www.nasa.gov/mission_pages/ladee/main/index.html.
52 6 A Peek Under the Hood

RF system.14 Commercial actors in the market are increasingly incorporating laser


terminals in their constellations.15

6.2.3 Connectivity

We discussed spacecraft configuration in Chap. 5, where we said that configuring


a satellite is about defining the subsystems that will be sitting on-board. Ulti-
mately, combining subsystems from disparate origins inside a technical artifact
like a spacecraft means connecting them and making them exchange information.
The act of communicating is remarkable. Two people are capable of holding a
small talk about the weather in an elevator by making air vibrate in a layered
manner by means of the vocal cords of one whose airwaves are caught by the ears
of the other one, where different frequencies/pitches, pauses and timbers make up
phonemes which eventually become syllables, words, and ultimately intelligible
speech about how awfully humid the morning is going to be again. Why layered?
Layered because you can dissect the act of communication in different parts and
skins, like an onion. For starters, the air vibrations mean nothing by themselves—
a glass falling to the floor and crashing to pieces creates a lot of air vibrations
in the surroundings as well but that forms no speech—but the collection of those
vibrating patterns in “packets” in a coherent manner and eventually interpreted
by a human brain—in context—is what defines the messages being exchanged
and ultimately—by stitching them with some criterion—the act of communicat-
ing. Contracts and rules take place in the process of discerning what means what,
and there are contracts at each and every skin of the onion: phonemes, syllables,
words, phrases, and even semantic rules at the topmost layer, for example, by
being polite while the trivial conversation progresses and avoiding showing how
much you want to leave the elevator.
Imagine now that the two elevator passengers spoke different languages, and
that they used a smartphone for translation: in this case, the phone acts as a pas-
sive ‘bridge’ between them converting from one protocol—language—to another
without having the need to interpret or store what’s being said. The phone could
also store a recorded message in one language and send it to another actor later
on—something that is usually called store and forward. It could also happen that
the message is not well understood or gets corrupted or broken by a sudden noise
in the elevator, requiring one of the parties to ask for it to be re-sent. When we
communicate as humans, there is error-control at play, and we subconsciously try
to maintain a signal-to-noise ratio—think about how you automatically raise your
voice when you are talking to a friend on the metro as the train goes faster and
noise gets louder. And there is flow control: we don’t convey more information
than our receivers are able to take; ever asked a professor in university to go

14 https://www.nasa.gov/directorates/heo/scan/opticalcommunications/llcd/.
15 https://spacenews.com/all-future-starlink-satellites-will-have-laser-crosslinks/.
6.2 The Data Links: From Sparks to Mobile Networks, Lasers, and In-Orbit Networks 53

slower while writing something on the whiteboard? Too much data can easily clog
channels, and throttling down the rates is a way to go.
Elements inside a technical artifact like a spacecraft or a car imitate human
communication and have all the elements of the elevator small talk, only with
a more steampunk touch if you want. Elements on-board modern vehicles com-
municate by opportunistically manipulating signals through different transmission
media—air, vacuum, copper wires or light, among others—packing things in
meaningful “chunks” and defining a “dialect” or a contract on how the chunks
are supposed to be collected, packed, and unpacked by every actor who attends
the data party. What is more, computer communication defines what to do in pres-
ence of errors, how to manage things that are in principle too large to fit through
some particular channel in smaller chunks only to be then reassembled in the right
order at the receiving end.
Should be no surprise, then, that data communication is also layered. As it
happens, defining the layers is somewhat arbitrary, although there is one de-facto
definition which includes seven layers,16 but the concept can be largely explained
without falling into rigidly labeling things one way or another.
The lowest layer is called the physical layer, just like airwaves in the elevator
example. These are perturbations in some transmission media that can be even-
tually converted into ones or zeros a computer can manipulate better. It can be
electric pulses on a copper wire, electromagnetic waves in optical fiber or vac-
uum, or it could be smoke signals, for that matter. Then, these media perturbations
are converted to digital symbols and collected by the next layer—which probably
needs to include some sort of mechanism to discern where the message is starting
and ending, in case you were distracted looking at a bird when the smoke signals
kicked in—and some extra information to be added beforehand to ensure eventual
errors are corrected, if any. But this layer is strictly a local messenger and brings
no value if two networks are supposed to collaborate, or if data needs to be re-
transmitted, that’s someone else’s job.17 This layer strictly delivers a collection of
ones and zeros (“frames” in this layer’s jargon) from one point to another point
without really asking more questions. Examples in space are SpaceWire, SpaceFi-
bre, Ethernet, MIL-STD-1553, RS-422, AX.25, CCSDS AOS, etc. Many of these
may also specify the physical layer.
The next wrapper, the following layer of the onion—the network layer—is actu-
ally taking a peek into the bunch of ones and zeros because it knows that some
of those might be actually a destination address and it might also mean the thing
(“packet” in this layer jargon) could be targeted to another neighboring network.
The network layer may do clever things such as routing packets through differ-
ent paths according to some routing rules, and fancier operations such as ensuring

16https://en.wikipedia.org/wiki/OSI_model.
17For completeness, some data link layers may include features such as Automatic Retransmission
Request (ARQ) mechanisms and commutation, this is, how to use one single physical channel to
convey data from multiple “virtual” channels.
54 6 A Peek Under the Hood

packets are in the right order or fragmenting them if too large for what the net-
work can handle. An interesting fact starts to appear: the more ones and zeros
a layer needs to spy on—these are called headers in the jargon—the more such
data becomes unavailable as user payload data: there are metrics which tackle this
ratio of “useful data vs header data”, for instance goodput and protocol efficiency.
Examples of layer 3 in space are Space Packet and the good old Internet Protocol
(IP), with the particularity that if satellites want to talk IP, most likely the IP data-
grams need to be encapsulated inside more space-specific packets such as Space
Packets, but that’s a technicality we’ll leave out from the scope of this chapter.
Now, the fourth layer kicks in and says: “alright, it seems we have two entities
here trying to have a conversation, how can I help?”. Well, it does help by adding
everything a proper conversation would need: introduction, nods-with-the-head,
yeses, ahas, okays, i-don’t-understands, tell-me-mores, please-repeat. In short, all
things any normal chat has, but in a kind of more robotic way. Mind you, adding
all these facilities still creates more overhead—the need to transmit information
which is not strictly used for the conversation itself. Classical examples of layer 4
protocols are TCP and UDP.
From this point on, further layers are there to distill the information in a bit less
“machine-like” manner the higher they go, and making it increasingly a bit more
human, often paying the penalty of losing goodput and performance in exchange
of readability. For example, HTTP requests sending back and forth JSON objects
is nothing else than a cry for human readability kicking in higher layers, whereas
the underlying layers could not—and should not—care less about this; they are
comfortable working with obscure, meaningless ones and zeros.
By the way, the idea that boundaries between layers are strict and well-defined
is a naive utopia: there is a good deal of overlaps, which creates a massive amount
of confusion in the telecommunications engineering community. What is more,
lower, or higher layers may re-appear down the road to encapsulate back other
layers if need be, which is the governing principle of Virtual Private Networks, or
VPNs, where IP datagrams are encapsulated and encrypted in higher-layer packets.
There are also crucial matters such as authentication (ensuring you are talking
to the entity you think you are talking to) and encryption (making sure data is not
directly understood if it lands in the unintended hands) which must be considered
when establishing a communication channel which might be subject to spoofing
or sniffing (like radio links between ground stations and satellites are due to the
fact their physical layer is open for tampering18 ). It is a dangerous assumption to
believe only the expected, well-behaved parties are taking part in any exchange of
data. Every data communication channel must include the right amount of cyber
protective measures, with “right amount” being key here since overshooting will
create too much overhead and cost.

18See how amateur radio enthusiasts decode telemetry data from SpaceX rockets: https://www.
youtube.com/watch?v=74_N163HyhA&ab_channel=ScottManley.
6.3 The Software: Hello World in Space 55

All the satellites flying over your head as you read these lines are very talkative.
There are myriads of frames, packets, and data conversations happening inside and
outside spacecraft. From the inside, from computers to computers: health status
data, commands, files; a lot is being sent back and forth. As we speak, bit errors
are happening, hopefully getting corrected. But also, potentially uncorrectable data
corruptions that will make operators on the ground grab their heads. It’s been said:
space is hard and, up there, ones and zeros can randomly swap places without
notice due to the harsh environment we described in Sect. 3.3 and the delicate
nature of semiconductors we spoke at the beginning of Chap. 3.
There is a whole lot of talking from the satellites to the ground and vice
versa, by means of electromagnetic waves in different bands of the radio spec-
trum, with antennas obsessively stalking each other. And last but not least, there
are increasingly more conversations happening between fellow satellites while in
orbit, typically by means of lasers as we talked in Sect. 6.2.2.
Data communication is what makes satellites the indispensable tool they are. By
means of data channels of different speeds and through different media, satellites
can not only access their on-board resources to keep themselves safe and oper-
ative, but they also can convey information to ground and neighboring satellites,
leveraging the augmentation network effects bring, compared to what they could
achieve working in an isolated capacity. No one expects a computer not being
connected to some network these days. Satellites are following suit.

6.3 The Software: Hello World in Space

Software enjoys a strange reputation in the space industry. On one hand—let’s


call it the hero side—software tends to be the lifeline when it comes to solving
problems on a troubled flying satellite, given that there is basically nothing else
you can do after launch other than trying to fiddle with the on-board software and
see if you can bring things back to normal. On the other hand—the villain side—
software is considered the evil of all evils by non-software people. In any troubled
space mission, the fingers naturally tend to point to software as the probable cause
of failure (usually without tangible evidence). Any conversation next to a coffee
machine between two non-software engineers (say, a mechanical and a thermal
engineer) would either blame software, or radiation.
What does evidence say? In Fault-Tolerant Attitude Control of Spacecraft (Qin-
glei Hu, Bing Xiao, Bo Li, Youmin Zhang, Elsevier), the authors collected and
analyzed spacecraft data from databases such as the Satellite Encyclopedia (TSE),
Satellite News Digest, Mission and Spacecraft Library, Sera Data SpaceTrak,
Space Systems Engineering Database (SSED), and the Mission Failure Analysis
for NASA AMES Research Center.
As for the percentage of failures per subsystem type, the data collected shows
that ACS (Attitude Control Subsystem, the system in charge of controlling the
orientation of the spacecraft in space) accounts for 32% of failures in missions,
followed by Power subsystem with 27% (Fig. 6.2, left hand). This should not come
56 6 A Peek Under the Hood

Fig. 6.2 Subsystem faults and types of faults. Source Fault-Tolerant Attitude Control of Space-
craft (Qinglei Hu, Bing Xiao, Bo Li, Youmin Zhang)

Fig. 6.3 Types of faults in


the attitude control
subsystem. Source
Fault-Tolerant Attitude
Control of Spacecraft
(Qinglei Hu, Bing Xiao, Bo
Li, Youmin Zhang)

as a surprise,19 since the attitude control subsystem is the most complex subsystem
on a spacecraft (to find about why it’s so complex, see Sect. 6.4).
When it comes to the different types of faults, they can be further classified into
several types of failures: mechanical failure, electrical failure, software failure, and
other unknown failures. For mechanical failure, it mainly refers to the mechanical
structure deformation caused by temperature changes, external force, friction, and
pressure. An electrical failure accounts for power overload, overcurrents, short
circuits, and abnormal battery performance. A software failure mainly consists
of incorrect computer instructions and onboard software abnormalities. As you
can see in Fig. 6.2 (right hand), software faults account for only 6% of the total,
whereas electrical faults account for 45% of them. In Fig. 6.3, you can see the
types of faults for the attitude control subsystem only.
According to this data, there is no strong evidence that software faults are the
main causes of mission losses. Or, put more bluntly: if anything, the evidence
collected in that work indicates software is seldom the problem. There have been,
though, software failures which have gone down in history as very resonant, such
as the Ariane 5 failure,20 Mars Climate Orbiter,21 and others.22

19 A veteran space engineer once said: “there are two main reasons why space systems fail: one is
attitude control, the other one is funding”.
20 http://sunnyday.mit.edu/nasa-class/Ariane5-report.html.
21 https://en.wikipedia.org/wiki/Mars_Climate_Orbiter#Cause_of_failure.
22 https://en.wikipedia.org/wiki/List_of_software_bugs#Space.
6.3 The Software: Hello World in Space 57

Even if software were the problem, there is no chance a spacecraft can be


launched without any software on it. Satellites are becoming more and more
software-intensive and nothing indicates this trend will stop anytime soon. Then,
let’s discuss what’s so special about making and running software for space
missions. For example:

. What does it take to run software on a spacecraft? How does it differ from
running software on a laptop sitting in a coffee shop?
. How is software updated or changed in orbit?
. What type of languages are used for coding flight software?
. Can I run Linux on a satellite? What about Windows?
. Can I host a website on a satellite?
. What kind of skills are required for doing Flight Software?
. How is Flight Software designed?
. What does “software-defined satellites” mean?

Let’s go one by one.

6.3.1 What Does It Take to Run Software on a Spacecraft?

Well, for sure you need a microprocessor, and some memory. These can be dis-
crete—as in, individual devices—or they can be part of the same integrated circuit
(usually called microcontroller unit, or MCU). In any case, what we call software
is basically a set of instructions which are fetched from a memory and put in a
pipeline to eventually be interpreted by the microprocessor architecture executed.
Such instructions are typically mathematical operations of different kinds and data
movements from one place to another. This doesn’t really differ from the software
which runs on your laptop. The main difference is that the computing resources
spacecraft have tend to be more specialized compared to the resources consumer
electronics as laptops have. For example, spacecraft may use microprocessor archi-
tectures which consider reliability by design by adding voting capabilities in their
internal registers. Because energetic particles can interact with the semiconductors
that processors are made of (as we saw in Sect. 3.3), it means that an energetic
particle may alter the content of a memory cell or a status register at any given
time. One single bit being changed by radiation can cause serious consequences to
the execution of software; for instance, an if statement may suddenly become true
and execute a routine which could make the satellite change orientation. Hence,
adding voting capabilities ensures that, upon the upset of a memory location or
internal content of the processor, still the value finally used for evaluation will
require that 3 identical registers will show the same reading. This comes with
the penalty of a larger area of semiconductor used for implementing the voting
scheme. Because bits can be upset at any time while in orbit, several techniques
to mitigate this issue have been devised. For example, memory scrubbing, which
consists of reading from each computer memory location, correcting bit errors with
58 6 A Peek Under the Hood

an error-correcting code (ECC) and writing the corrected data back to the same
location. Parity bits, checksums, cyclic redundancy checks (CRC) and other error
detection and correction schemes are also used.
Computing resources on spacecraft have been historically more modest than
resources on ground-based computers, although this is a gap that has been shrink-
ing fast in recent years. On-Board computers are advancing rapidly, with more
resources than ever before: more storage, more floating-point operations per unit
of time (FLOPS), more millions of instructions per second (MIPS) executed. Still,
on-board computers remain closer to what are called embedded systems, which
are computers with specific roles and functionalities. Unlike a laptop—which is
sold without really knowing or caring for what the laptop will be finally used for,
whether it’s accounting, graphic design, or music production—spacecraft com-
puters tend to be optimized for the specific function they are ought to perform:
attitude control, thermal control, command and data handling, star trackers, pro-
tocols encoding and decoding, payload data acquisition, etc. Each one of these
requires specific interfaces, specific libraries, sensors, actuators, etc. For example,
attitude control needs to perform intricate calculations about rotations, matrix oper-
ations, sensor data fusion, and a lot more. And it has to talk to a variety of sensors
with different characteristics, such as sun sensors, star trackers, gyros, magnetome-
ters, and so forth. Microprocessors used in attitude control are typically optimized
with several ALUs and data paths to perform many of these computations in a few
CPU cycles.

6.3.2 How Is Software Updated or Changed in Orbit?

By means of…software. Yes, software can be designed to help modify software,


including modifying itself. Spacecraft computers are equipped with what is called
a bootloader. A bootloader is a program whose sole function is to fetch a piece of
executable code—typically application code—from some location or through some
data interface, for example a serial interface, and place it in a way the microproces-
sor can find it and run it. The bootloader is invoked during the power-on sequence
of the processor, and usually loads and executes the same application code, over
and over. Under special circumstances, the bootloader might be commanded to
load another image from some other location or take a new image through some
interface. Mind that bootloaders can also create some headaches. For example, if
during the process while a bootloader is loading an application memory something
odd happens and the image gets corrupted, then the application program will never
execute because it is only partially present. This is typically and somewhat col-
loquially called “bricking” the thing; because the device becomes as useful as a
brick. And it’s not really exclusive for space. You can brick your smart TV, your
Wi-Fi router or your phone while updating firmware. The key is to ensure the
bootloader can always recover from a failed flashing procedure. Goes without say-
ing, bootloaders should not be able to overwrite themselves, at least not their most
6.3 The Software: Hello World in Space 59

critical part (there can be more than one bootloader calling another bootloader and
another bootloader).

6.3.3 What Type of Languages Are Used for Coding Flight


Software?

In theory, any language can be used for coding flight software. Languages never
make it to orbit, only machine code does (unless the compilation happens in orbit,
which is somewhat rare, although possible). Then, if the processor being used
has a compiler which can translate whatever obscure language of your choice to
its machine code, you’re good to go. Do you want to code your flight software
in brainf*uck23 ? Go ahead. Anything is possible as long as you have a compiler.
This applies, well, for compiled languages. Because flight software has historically
stayed quite “embedded” and close to the hardware, compiled languages have been
more popular, notably assembly and C. These languages offer to the programmer
the tightest control of what’s happening as the software runs, while paying the
penalty of a lower code readability, reuse, and whatnot. Other compiled alterna-
tives to improve reusability is to employ Object Oriented Programming (OOP)
techniques and languages, although this claim tends to be disputed. For exam-
ple, in the book Coders at Work, Joe Armstrong—creator of Erlang programming
language—says on software reusability and OOP:

I think the lack of reusability comes in object-oriented languages, not functional languages.
Because the problem with object-oriented languages is they’ve got all this implicit environ-
ment that they carry around with them. You wanted a banana but what you got was a gorilla
holding the banana and the entire jungle.

On the other hand, interpreted languages such as Python are totally possible in
spacecraft. Provided the on-board computer is capable of running an interpreter, all
the rest comes reasonably easy. Interpreted languages tend to be relatively slower
and sit on higher abstraction layers; don’t ask an interpreted language to han-
dle hardware interrupts or hard-real time deadlines. But interpreted languages are
highly flexible for scripting and automating non-critical things on-board like daily
maintenance tasks.

23 Brainfuck is an esoteric programming language created in 1993 by Urban Müller. Notable for
its extreme minimalism, the language consists of only eight simple commands, a data pointer, and
an instruction pointer. While it is fully Turing complete, it is not intended for practical use, but to
challenge and amuse programmers. Brainfuck requires one to break commands into microscopic
steps.
60 6 A Peek Under the Hood

6.3.4 Can I Run Linux on a Satellite? What About Windows?

If you can run Linux, then you must. If the spacecraft computer is capable of
handling Linux—this means, it has a certain number of resources and a Memory
Management Unit, or MMU—then Linux is your friend. Even without an MMU,
Linux can be your friend. But, with great power comes great responsibility: a
full-blown Linux OS needs to be administered accordingly, and this may require
certain capabilities for accessing and monitoring the operating system’s resources.
Because satellites are remote systems—there is a lousy radio link between the
spacecraft and the ground—this means that such a noisy link will be the only way
to shell the remote operating system (more of this in Chap. 8). Therefore, remote
shells such as SSH (encrypted) or Telnet (unencrypted and very basic) have to be
used on top of radio, which needs IP stacks on top of space data link protocols
such as CCSDS. Not such a big deal if the link budget is generous and the latency
is small. This means, if the energy per bit sent back and forth is reasonably higher
than the overall noise, and the round trip for packets from satellite to ground is
reasonable for conversational protocols to work well. In high-latency links, for
example for distant orbits, protocols such as TCP can be problematic, therefore
shell access to a remote Linux-based spacecraft can be prohibitive. With proper
link established, handling of the remote Linux system running on board the satel-
lite becomes more of a sysadmin24 task. Using Linux and IP based links blurs the
line between operating a satellite and operating a server on a network.
As for Windows, technically speaking, you could run it on a satellite. But
because Windows is an operating system which heavily relies on graphical user
interfaces, the only way you could possible use it is to either use something like
VNC to remote desktop it (although considering the radio link speeds and the
amount of data VNC would generate would turn it very challenging to say the
least), or put a screen on board (which needs to withstand launch, vacuum and
radiation) and hook a camera looking at the screen while you send videos frames
at a lower rate to the ground. I am just being polite: just don’t’ do it. I recommend
you just discard the idea of running Windows on a satellite for now, but if you
insist, don’t go without tools like Cygwin25 if you want to live.

24 A sysadmin, or admin is a person who is responsible for the upkeep, configuration, and reliable
operation of computer systems, especially servers.
25 https://www.cygwin.com/.
6.3 The Software: Hello World in Space 61

6.3.5 Can I Host a Website on a Satellite?

Of course, you can. It is pretty straightforward provided you have some web
server like nginx or apache2 installed on-board, PHP installed (optional, but rec-
ommended) and some index.html file sitting somewhere with a basic code like
this:

<html
<head>
<title>ReOrbit: Software in Space</title>
</head>
<body>
<?php echo ’<p>Hello World from Space</p>’; ?>
</body>
</html>

Then, you can cURL the satellite, and the webpage should reply back. Again,
the question remains on how you will connect to the remote spacecraft’s web.
Since web pages rely on HTTP which in turn relies on TCP and IP, your radio
link will have to be comfortable with all these protocols for the website to work
and be responsive.

6.3.6 What Kind of Skills Are Required for Doing Flight


Software?

It does require a certain level of awareness of the typical satellite architectures,


interfaces, and whatnot. But all that can be learned. Flight software is nothing
fundamentally different than embedded software. Because a satellite is a physical
and dynamical system, the software developed for it must be designed to be aware
of such dynamics. In short, the (minimal) complexity of software which controls
a physical system is given by the complexity of the underlying physical system.
It says minimal because of course software engineers can always over engineer it.
And they will if given the chance.

6.3.7 How Is Flight Software Designed?

It’s an incremental work. But, if you take two satellites from two absolutely dif-
ferent operators and you somehow reverse engineer the machine code into source
code, more or less those two different flight softwares will show similarities. Every
flight software needs some of these modules or building blocks: telemetry han-
dling, telecommand handling, FDIR (failure detection, isolation, and correction),
thermal control, power control, scheduling, etc. Flight software can be designed
and developed from scratch—and software engineers will always try to convince
62 6 A Peek Under the Hood

you it’s the only way to go—or it can rely on existing software frameworks which
already contain some of the typical building blocks. Some of these frameworks
are even flight-proven and open source, which makes them very attractive for
low-budget NewSpace missions or University projects with very tight schedules.

6.3.8 What Does “Software-Defined Satellites” Mean?

The software-defined fad has gotten a bit out of hand lately. It is used too loosely
and gets stretched to levels where it stops having meaning. But, in the strict
sense of the term, when we say software-defined anything, we mean “something
which has been done traditionally by means of mechanical or electrical devices
but now is done by means of software”. Sounds obvious, but it’s not. For exam-
ple, take software-defined radio.26 Historically, modulating/demodulating, mixing,
and filtering radio signals has been performed using discrete—active and pas-
sive—devices. With the progress of semiconductor technology in areas such as
analog-to-digital converters and digital signal processing, now it is possible to
move such operations into software. How so? Well, manipulating electric signals
is ultimately a mathematical problem. Filtering a signal, mathematically speak-
ing, is done by computing the convolution of said signal (or, in the DSP27 jargon,
a sequence) with the step response of the filter we want to use. Or, equivalently,
multiplying their frequency responses (the signal’s and the filter’s) in the frequency
domain. Then, a superheterodyne receiver is nothing else but an arrangement of
mathematical manipulations of the input signal. So, if one manages to digitize the
signal as soon as possible in the signal chain (this is, as close to the antenna as
you can), then one can perform the math operations fully on the digital domain
and still obtain the same results.
On the same line, spacecraft have had certain things historically done by hard-
ware elements, for example communication protocols, like CCSDS.28 Traditionally
hardwired in highly specialized chips like ASICs and FPGAs, now the stacks can
perfectly sit on software, which adds a good deal of flexibility when it comes
to adding security layers and applying patches in case of cybersecurity threats.
Software-defined protocols allow to scale up or down resources depending on
the mission requirements. For data intensive space applications, software-defined
techniques allow to readjust mission configurations post-launch, providing a more
cost-effective way to optimize operations once the spacecraft faces the elements.
Although there are some particularities about running software inside a metal-
lic box that is orbiting at hundreds of kilometers above the ground while flying
at thousands of meters per second of velocity, software is still software. Software

26 https://en.wikipedia.org/wiki/Software-defined_radio.
27 DSP = Digital Signal Processing.
28 CCSDS is a collection of recommended standards issued by the Consultative Committee for

Space Data Systems which specify several layers of a protocol to command and process telemetry
data in space missions.
6.3 The Software: Hello World in Space 63

itself remains unaware about the altitude at which it’s executing. Variables are still
variables, and if statements work the same way as in office space. The software
engineers are the ones in charge of adding the correct safeguards to make the flight
software execute seamlessly while flying alone in space and while the underlying
hardware is being chastised by the space environment (see Sect. 3.3). Irrespective
of where the software runs, the work follows the same flow: defining data struc-
tures and their relationships in runtime. You can over-engineer it by adding fancy
methodologies and pile up dependencies, but flight software, at the end of the day,
still boils down to data and how it behaves over time. There’s a phrase attributed
to Linus Torvalds which goes:

Bad programmers worry about the code. Good programmers worry about data structures
and their relationships.

6.3.9 Bugs and Glitches in Orbit

Recently, I was reading an article from one major manufacturer of high-end cars
in which the company was humblebragging about their “zero-bug policy” in their
software development process. A claim that captured my attention. While the
article started bold and promising, it quickly mutated into some realistic optimism:

With ‘zero bug’, they actually meant zero bugs that *are known*. Well, fol-
lowing that reasoning, all software that we develop has zero bugs…if we ignore
them.
One thing was spot on in the article though and can be read from the screenshot
above; bug-free code and software production is impossible. Firstly, because we
are fallible. And second, because, as software developers and testers, we can never
foresee all the scenarios and situations our software will encounter once released.
And some of those unforeseen scenarios might29 (will) surely cause bugs to reveal
themselves. Therefore, as much as we spend on complex testing suites or the best
testing talent possible (which are good investments by the way) we can never fully
cover all potential situations.

29“Anything that can go wrong, will, and at the worst possible moment.”—Finagle’s Law of
Dynamic Negatives.
64 6 A Peek Under the Hood

Take commercial aircraft, with millions of lines of code running on-board. Air-
liners are tested very thoroughly. Still, nasty bugs happen. See the case of Flight
290430 in Warsaw, Poland where an A320 overran the runway at Ok˛ecie Interna-
tional Airport on 14 September 1993. Although the cause of the aircraft failing
to brake in time was due to a combination of factors—including human factors—
this accident involved certain design features of the aircraft. On-board software
prevented the activation of both ground spoilers and thrust reversers until a mini-
mum compression load of at least 6.3 tons was sensed on each main landing gear
strut, thus preventing the crew from achieving any braking action by the two sys-
tems before this condition was met. But because of the wind conditions present
that day—which impacted the way the A320 approached the runway—the aircraft
stayed for a few seconds standing only on one side of the main landing gear; pre-
cious seconds of braking time lost which directly contributed to the incident. A
scenario that clearly falls way outside of the typical testing strategy for this kind
of system. No one knew that scenario existed, until it did.
In fact, here we are not strictly talking about these kinds of very visible soft-
ware bugs. We spoke so far about a kind of software bug that always takes the
spotlight because it causes considerable losses and visible, sustained malfunction;
colloquially called in some bibliography hindenbugs.31 But there is a more special
and elusive kind of software bug: the glitch. A glitch is a slight and often tem-
porary fault. Hard to reproduce, at times almost imperceptible from the operator’s
point of view, glitches are the quintessential headache of any programmer. Now
you see it, now you don’t. Let alone if the glitch appears to be a heisenglitch32
which is a type of glitch where the observation changes the behavior: looking for
the glitch changes the results of the program, making the glitch not to reveal itself.
Software glitches are often opportunistically invoked by engineers as the poten-
tial cause of any mysterious occurrence when something happens—or someone
claims that something has happened—which cannot be easily explained.
—Weird. Maybe it was just a glitch.
“Just” a glitch?
Story time. Eons ago, while observing the telemetry dashboard of a satellite—a
screen full of time-series data represented in a myriad of colorful line plots depict-
ing the trends of all relevant variables monitored on-board—I noticed something
odd. Some plots were showing brief “jumps” in their values, only to come back

30 More details about this incident: https://en.wikipedia.org/wiki/Lufthansa_Flight_2904.


31 A Hindenbug is a catastrophic bug that destroys data and may also shut down systems or cause
other major problems with systems. It is a general IT slang word for a major bug that does more
than just create a nuisance or an annoyance for users.
32 In computer programming jargon, a heisenbug is a software bug that seems to disappear or alter

its behavior when one attempts to study it. The term is a pun on the name of Werner Heisenberg,
the physicist who first asserted the observer effect of quantum mechanics, which states that the act
of observing a system inevitably alters its state. In electronics, the traditional term is ‘probe effect’,
where attaching a test probe to a device changes its behavior.
6.3 The Software: Hello World in Space 65

to the trend it was following right before the jump. It was most visible on plots
which were smooth sine waves, like for example quaternion elements, where it
could be clearly observed the curve being very briefly interrupted but these abnor-
mal hiccups. The glitch was in some way “harmless” (only a visual artifact) which
contributed to the fact no one had complained about it, but it clearly indicated
there was something wrong. Which is the very nature of many software glitches.
They might not be causing a major problem, but their presence indicates there is
something that is not nominal. Glitches must be treated seriously, ruthlessly chased
down and sorted out.
Anyway, so I collected a good amount of telemetry data, and I went home
to analyze it in depth. The hiccups were seen on most values coming from one
particular subsystem connected to the on-board computer by means of a serial
port. So, the problem was more or less circumscribed to that particular subsys-
tem, which was good news. Said subsystem was packing all its telemetry frame
in a fixed-length binary chunk composed of something like 4 kilobytes, and the
unpacking of the frames was carefully done by the telemetry processing on the
ground, converting all values into “human readable” units.
Then, while mindlessly browsing through the binary data, I do not know really
how, I observed that one particular value of the 255 possible values of the bytes
sent by said serial channel that was oddly missing. There was no occurrence what-
soever of the byte valued 13 in decimal (0x0D in hex) at all in the several gigabytes
of telemetry I had taken home with me. Definitely strange. I had a lead. What is
more, the missing byte was an important value used in serial terminals as carriage
return (CR). Carriage Return is the ASCII character sent to the terminal when
you press the enter key on your keyboard while connected to a serial line, tracing
back to the times where typewriters roamed the earth. It looked as if something
was replacing the missing carriage return character for a newline character (0x0A).
Then, it all came together rather quickly by checking the configuration of the serial
port attributes (Fig. 6.4).
Bingo. It turned out the UART peripheral inside the on-board computer CPU
was doing some very inconvenient translations of characters by default. Disabling

Fig. 6.4 Funny little configuration to have by default


66 6 A Peek Under the Hood

this “feature” made the glitches disappear forever. It’s hard to explain how ecstatic
one feels when catching one of these things.
This story in particular had the advantage that the glitch was quite visual, and
it appeared repeatedly thanks to the “persistence” and trail the line plots provided.
Some other glitches are not as simple to spot or to remember, and they might also
be what some bibliography call a mandelbug33 which is a bug that is so complex
that it occurs in a chaotic or nondeterministic way. Consequently, a mandelglitch
is the hardest of them all: hard to detect, brief, chaotic, quasi-random. At times,
mandelbugs may show fractal behavior by revealing more bugs: analyzing the code
to fix a bug only uncovers more bugs.
All in all, software glitches are a serious problem, and they deserve the same
attention as their more famous cousins the bugs. In fact, there should not be a
separated terminology, a glitch is a bug, regardless of how brief or harmless it
might appear to be.
In general, software does exactly what it is told to do. The reason software
fails is that it is frequently told to do the wrong thing. That is not because of
programmers’ incompetence but because of one of the biggest, if not the biggest,
challenge in software development: the barrier between the way we create software
and its behavior. Creators like carpenters or sculptors benefit from an immediate
connection to what they’re creating. The challenge with programming is that it
violates this principle; the programmer, staring at a page of text, is abstracted
from whatever it is they are actually making. That’s why software systems are
hard and are becoming increasingly harder to understand.

6.4 The Orientation: Attitude Control

Imagine yourself in an empty room. Now picture you are somehow able to stay
suspended in the air, like magically floating. Imagine that someone exerts a gentle
torque on you, and this torque makes you start spinning. This mysterious person
leaves the room right after that, turning off the light. You can’t see or hear any-
thing, as you are spinning in this dark, weird place. How long would it take for you
to lose track of what your orientation is? What means up and what means down?
Right or left? How quickly would you get absolutely disoriented about where the
door, the roof or the floor are? Most probably, not too long.
We constantly rely on cues around us to situate ourselves in this world we
inhabit. Provided all our senses are available and healthy, we heavily use them
on a constant basis to navigate our environment. Visual cues, olfactory, acoustic,
and even thermal hints are used to understand where we are located and in what
direction our heads are oriented.
Another exercise. This time a bit less eccentric than the dark room. Pick a
chair in your office, cover your eyes and sit, not before asking a colleague to

33 https://ieeexplore.ieee.org/document/4085640.
6.4 The Orientation: Attitude Control 67

give you a nice spin. But before the spinning kicks in, carefully observe where
exactly you are looking at in the office: pick some feature from the landscape;
a column, a fern, a poster on the wall, wherever. As you spin—remember you
have your eyes covered—try to keep track of the chosen object or spot and how
it is supposed to move around as you go and try to estimate where your stopping
position will be with respect to that reference point. Chances are you will not be
able to precisely predict how the chosen spot has moved during your spinning trip,
and when your eyes will be uncovered, you’ll be more than surprised how different
the final orientation turned out to be.
Well, all these somewhat bizarre examples illustrate what spacecraft must deal
with in terms of orientation while in space. The orientation of spacecraft in three-
dimensional space is more specifically called attitude.
While in space, spacecraft are floating in a “dark room” (which is not totally
dark, as we shall see). Satellites are spinning one way or another while up there.
Spinning is a very natural motion in space. Here on the surface of the earth we deal
with friction forces which dramatically change the way we handle and experience
changes in orientation. Here, it somehow feels that motion is the exception, not the
rule. You have to do some work to move, as if somehow the “normal” state was
to be motionless. Well, that’s a bit of an illusion that friction forces create. The
philosopher Aristotle was confused by this as well, and it heavily influenced his
perspective of physics. Aristotle’s view of motion34 was that it required a force
to make an object move or, more simply, “motion requires force”. After all, if
you push a book, it moves. When you stop pushing, the book stops moving. But
Aristotle was wrong. Very wrong. Still, the reign of Aristotelian physics lasted
almost two millennia. Upon the work of many pioneers such as Copernicus, Tycho
Brahe, Galileo, Descartes and principally Newton, it became generally accepted
that Aristotelian physics was neither correct nor viable.
Perhaps somewhat counterintuitively, in the way physics actually works, if you
push a book, it shall keep on moving, and moving. And if you were to be set to
spin on an office chair without friction forces, you would spin there in eternum.
The guiding principle here is the conservation of momentum, linear and angular
respectively, as described beautifully by Newton’s First Law of Motion. Ubiquitous
external friction forces—which are the product of surfaces coming in contact with
each other and dissipating energy in the process—are the reason why everything
feels different from that and feels closer to what Aristotle believed. We cannot
blame the guy. You almost don’t think about these friction forces unless they are
absent. For instance, during some particular days in the beautiful Finnish winter
where it is so icy everywhere that you almost don’t get to choose where to go
anymore but the slopes along the way decide for you (Fig. 6.5).
In space, such friction forces almost do not exist or, more accurately, they are
considerably weaker than here on the surface of the earth. Therefore, spacecraft

34 More about Aristotle’s flawed take on physics: https://en.wikipedia.org/wiki/Aristotelian_phy


sics.
68 6 A Peek Under the Hood

Fig. 6.5 Good luck running for the tram in Helsinki without friction forces (Creative Commons)

must perform a set of, from a terrestrial perspective, weird operations to keep
control of their orientation in space.
We stated above that spinning happens very naturally in space. How so?
Well, what does it mean to spin in the first place? Spinning means rotating
around an axis passing through some internal point of interest of the body. For
generic bodies, let’s assume rigid bodies for simplicity—as opposed to flexible,
wobbly things—this point is called the center of mass. It shall be quickly clari-
fied that spinning is a subset of a more general motion called rotational motion.
Rotational motion is a circular motion around some arbitrary point which can be
perfectly located outside of the body, for example the earth rotating around the
Sun. In short, a body will spin around its center of mass either due to an external
force at a non-zero distance from that center or if its original state of rotation (or
the lack thereof) is changed internally in some way (Fig. 6.6).
Spacecraft are bodies with arbitrary form factors and uneven mass distribu-
tions; they are not perfect, ideal spheres like cows are.35 As they float (spacecraft,

35 The spherical cow is a humorous metaphor for highly simplified scientific models of com-
plex phenomena. Originating in theoretical physics, the metaphor refers to physicists’ tendency
to reduce a problem to the simplest form imaginable in order to make calculations more feasible,
even if the simplification hinders the model’s application to reality.
6.4 The Orientation: Attitude Control 69

Fig. 6.6 FA and FB are


forces applied at different
distances from the center O,
creating different torques.
Credit public domain

not cows) in the “dark room” in presence of very weak friction forces, there are
different torques exerted to them which make them change their orientation in
time. Such torques are from various sources, both internal and external. Let’s first
check the external ones. One external torque which comes up early in a space-
craft lifetime is the torque applied from the rockets which are delivering them to
space. In separation systems, the push typically comes from springs loaded specif-
ically for that purpose. If the push (force) happens at a nonzero distance from the
spacecraft center of mass—which means a torque if you remember the definition
above—there will be spinning. This spin is usually called tip off rate.
Other external sources of torques that can make a spacecraft spin are atmo-
spheric drag (there can be thin remnants of the atmosphere up to several hundreds
of kilometers in orbit which can create an aerodynamic torque which depends
on the object geometry and size/area), and gravity gradient. The gravity gradi-
ent disturbance can vary along its flight path, and it is caused by the variation in
the gravitational force experienced by the spacecraft due to the differences in the
gravitational field caused by the non-uniform distribution of mass in the nearby
environment. As the spacecraft moves through space, its distance and position rel-
ative to other massive objects, such as planets, moons, and another spacecraft, can
change. These changes can cause variations in the gravitational force experienced
by the spacecraft and therefore, the gravity gradient disturbance. Additionally, the
orientation of the spacecraft can also affect the gravity gradient disturbance. The
gravity gradient disturbance is strongest when the spacecraft is aligned with its
longest axis pointing towards or away from the massive object that is causing the
disturbance. As the spacecraft changes its orientation relative to the massive object,
the strength of the gravity gradient disturbance can change as well. Therefore, the
gravity gradient disturbance can vary along the spacecraft’s flight path depending
on its position and orientation relative to other massive objects in its environment.
There can also be radiation pressures associated with the interaction of elec-
tromagnetic radiation and extended body surfaces, which of course includes light
pressure, associated with the impact of light photons against surfaces. And for
planets where there is a magnetic field, such a field can also interact with electric
70 6 A Peek Under the Hood

currents circulating through the wires and devices of the spacecraft electronics,
creating a disturbance torque, in a similar manner to how a compass works.
As it can be seen, there are multiple external sources of torques trying to make
our poor spacecraft spin round like a record, and if we want them to stay looking
in a particular direction, for example to take a photo of a specific location or to
orient an antenna for communication purposes, something has to be done. And this
yields the floor to discuss internal torque sources to counteract the external ones.
The way a spacecraft can cope with the fact the universe seems to be machi-
nating to make them spin out of place, is by counteracting or balancing those
external effects with equal, but opposite effects. Just like a book stays still on
a table because the gravitational pull downwards—its weight—is perfectly bal-
anced by the reaction force provided by the table, a spacecraft can be made to
stay still pointing to a place of interest if all its torques are balanced, this means,
there is a net zero torque acting on the body, in some specific frame of reference.
But before talking about actions, more specifically about actuators (the devices on
board the spacecraft responsible for creating internal torques), how do spacecraft
know where they are looking to, so they can correct their orientation if they happen
to be looking in the wrong direction? By means of sensors, and some math.
We said before that space is similar to the “dark room” in the example which
opened this section. This is not entirely true. There is—luckily—light in space,
and such light comes from stars. Stars are, simplifying it a bit, nuclear reactors
burning hydrogen. In that process, a good deal of electromagnetic radiation is gen-
erated and radiated, and some of that radiation can be captured by some sensors,
for example our eyes, which are like passive antennas and makes it possible for
us to enjoy a starry night. Ancient navigators realized that, by eyeballing the stars
plus some measurements and calculations, they could understand where they were.
Space industry took note of this and developed a set of very smart “eyes”, that is,
cameras which can observe the stars, and by contrasting the current observation
with a stored internal catalog of stars and their possible rotations, these cameras
can infer where the lens axis is looking at with respect to a frame of reference. And
if you happen to know how the camera is oriented with respect to the spacecraft
body, you can then understand where the spacecraft is looking. These cameras are
called star sensors, or star trackers. Star sensors usually run fancy algorithms to
compute the multiple orientations of the star images, and to rule out unwanted
noise sources which could be misinterpreted as stars. In short, star trackers tend to
carry computers and run software algorithms to accomplish such tasks. A problem
quickly pops up: star sensors must be able to see a clear sky to be able to under-
stand the orientation. In short, if the spacecraft happens to be spinning somewhat
out of control and the star sensor cannot get to see a proper star field, it might not
be possible to use it to correct the situation. Additionally, start trackers tend to be
susceptible to SEEs, more specifically SEFIs (see Sect. 3.3.1.1 to know more).
So, there are other types of sensors as well which can come to help when
star sensors go blind or hangs. Spacecraft can measure the surrounding magnetic
field—if there is any field present in the first place—and use that information to
assess how their position is related to said magnetic field, which can be used to
6.4 The Orientation: Attitude Control 71

infer rates of rotation. The downside of magnetometers (the name of this sensor),
is that they are highly sensitive to magnetic fields created by the spacecraft itself:
how can magnetometers discern between “good” or “bad” magnetic fields? They
can’t.
There are some other ways of measuring spinning rates. For example, by using
gyroscopes. There are interestingly very different types of gyroscope technologies:
from some exploiting Coriolis force using very small micro-mechanical vibration
devices, up to more sophisticated ones which use interferometry and lasers, by
splitting a beam of light and making the two beams follow the same path but in
opposite directions, in a ring. On return to the point of entry the two light beams
are allowed to exit the ring and undergo interference. The relative phases of the
two exiting beams are shifted according to the angular velocity of the apparatus,
which can be sensed and reported. The cost between gyro technologies can be
abysmal and a key factor during the development stage.
Spacecraft are also very interested in knowing where the Sun is for both ori-
entation and ultimately survival purposes. The reader might quickly guess why:
power generation reasons. Solar power remains the most popular energy source
for spacecraft. Remember the exercise of the spinning office chair and how eas-
ily you can lose track where everything is located as you spin? Same happens
to spacecraft. Too much spinning and they may lose track where the Sun, or the
Earth is. And this is definitely problematic to keep batteries charged. So, how to
know where the Sun is? Readers may correctly say: ok, there are usually solar
panels on board of any spacecraft, then why not use them to guess where the Sun
is, since those are, well, sensitive to solar light by design? Fair. But wrong. That
would mean mixing cause and effect. Solar panels are the ones to be guided to
look at the Sun, not the ones guiding. Mainly because solar panels look in very
specific directions, hence from the whole sphere of possible rotations of a spinning
spacecraft, the direction where the panels are oriented is very unique, so to speak.
But, in fact, the reader was not entirely wrong. Solar cells are indeed involved in
searching for the Sun, just not the ones in solar panels. Typically, spacecraft are
equipped with tiny solar cells spread all across the body, and oriented in a way that
their sensitive axes, combined, can create a sort of “sensitive sphere” of sorts. This
way, the spacecraft are given “eyes all around their heads”, greatly simplifying the
Sun search challenge: those cells which are showing the highest current readings
are most likely the ones looking closer to the Sun, to better turn in that direction.
The collective of tiny solar cells is called Coarse Solar Sensor—coarse because
this method does not provide a terribly accurate position of the Sun, but more like
a “good enough” one, which is sufficient during contingencies.
Once the panels start pointing perpendicular to the Sun—as they should in
order to maximize power generation—survivability is ensured and operators on the
ground breath in relief. Of course, this is a reasonable simplification of the overall
process. There can be some problems in the process such as shadowing from
deployable appendages or booms, Albedo (reflection of sunlight on the earth’s
surface), etc. There are some other means of sensing the Sun’s presence, but those
are more for bigger, fancier missions.
72 6 A Peek Under the Hood

All in all, spacecraft have a rich set of sensors at hand to use for assessing where
the heck they are pointing to. But there’s some computing that must be done. All
sensors come with their own particularities, pros, and cons. Therefore, the software
running on the spacecraft shall ensure to use them as smartly as possible and that
depends on the situation on board. In other words, perform a sort of data fusion.
When everything is great and nominal, some sensors might make more sense (pun
not intended) than others, as opposed to when things go ugly where more basic
and less precise yet more reliable sensors are preferred. The software on board
must handle this trade-off continuously.
Mind that the position of stars and planets can also be obtained purely com-
putationally—the ephemeris of celestial objects can be calculated by means of
reasonably simple formulas. Why take the pain of adding expensive sensors if
all you need is a computer and some trigonometry? Basically, because all those
computations are theoretical, and require knowledge of time on board as well as
orbital geometry and that’s a luxury when things go south. Space engineers design
the control laws assuming that, under the worst conditions possible—for instance
tumbling randomly with no time keeping on board, no knowledge of position—the
spacecraft shall be able to charge batteries, detumble, and go back to business.
Now it’s time to act. Once the software on board performs the computations
needed to obtain a numeric representation which condenses where the spacecraft
is looking to—this numeric representation tends to take the form of an attitude
quaternion36 —the control software on board can compute the amount of torque
required to correct the current orientation versus what is desired if there is a sig-
nificant discrepancy between the two. The total torque computed is distributed
among the available actuators on board. Let’s dive into what these actuators are
and what they do.
Actuators are devices which can interact with the physical environment upon
demand. This is, we can command an actuator to create a force, torque or equiv-
alent to our plant (the plant being the physical system we want to control). In
spacecraft, actuators are typically reaction wheels, control moment gyroscopes,
magnetorquers and thrusters. And other types perhaps less frequently used, like
yo-yos.37
Reaction wheels are basically flywheels whose rotational speed is opportunisti-
cally changed for causing the spacecraft to counter-rotate proportionately through
conservation of angular momentum (the space equivalent of your home drill exert-
ing a torque in your wrist as it starts spinning). As wheels spin, they also store
angular momentum which can be used to stabilize rotations around the spin axis.
A somewhat related actuator is a control moment gyroscope (CMG). A CMG
consists of a spinning rotor and one or more motorized gimbals that tilt the rotor’s

36 Quaternions are a type of mathematical object that extends complex numbers to 4 dimensions.
Quaternions can be thought of as a way of representing rotations in 3D space, where the scalar part
represents the rotation angle, and the vector part represents the rotation axis.
37 https://ntrs.nasa.gov/api/citations/19620006811/downloads/19620006811.pdf.
6.4 The Orientation: Attitude Control 73

angular momentum. As the rotor tilts, the changing angular momentum causes a
gyroscopic torque that rotates the spacecraft.
A magnetorquer is a glorified solenoid, and their principle of work is extremely
simple. They create a magnetic dipole which is directly proportional to the elec-
tric current flowing through them and to the amount of turns of the coil, and
this magnetic dipole interacts with the surrounding magnetic field, creating a pure
torque.
Last but not least, thrusters. Thrusters are like small rocket engines. Rocket
engines produce thrust by the expulsion of an exhaust fluid that has been accel-
erated to high speed through a propelling nozzle. The exhaust fluid is usually a
gas created by high pressure combustion of solid or liquid propellants, consist-
ing of fuel and oxidizer components, within a combustion chamber. As the gasses
expand through the nozzle, they are accelerated to very high speed, and the reac-
tion to this pushes the engine in the opposite direction. Thrusters technically are
force-creating devices, but if conveniently placed around the center of mass they
can create torques as needed.
All in all, to control the orientation in space of a spacecraft, we require a
combination of devices: actuators, sensors, a computer running software that can
determine the current attitude and compute the corrections needed to reach the
targeted orientation (usually called controller), and, well, the plant itself, which is
the dynamical system under control.
Spacecraft are, fundamentally, cyber-physical systems, which is an glorified
term to describe a combination of computerized elements mixed with physical
elements, with some of those elements acting as bridges between both domains
(see the dotted line in Fig. 6.7 marking the boundary between the cyber and phys-
ical domains). This also means that, because controlling a spacecraft orientation
requires a fair deal of software and computer-machine interfaces, bugs and human
errors can make spacecraft go out of control: there is no way an actuator can dis-
cern if something or someone is asking it to act legitimately or not. An actuator
will act whenever requested. For instance, the ISS was recently unintentionally
steered due to a software bug in a docked module.38 Sorry, a glitch!
There are myriads of books written about attitude control. Hundreds of equa-
tions are related to the topic. But a good, birds eye’s level grasp of the underlying
principles is of paramount importance before diving into the details. Attitude con-
trol is one of most critical subsystems on board a spacecraft—it directly impacts
power, thermal, structures and mechanisms, and what have you. Almost goes with-
out saying, during the development phase of space systems, it is crucial to count
on a variety of simulation resources to create synthetic “dark room” environments
for testing and verification of the control software before launch. This is important

38https://www.theverge.com/2021/7/30/22601752/russia-nauka-module-software-glitch-iss-
space-station-nasa.
74 6 A Peek Under the Hood

Fig. 6.7 Building blocks of a computer-based control system

because we can’t easily reproduce on the ground the low-friction physics space-
craft experience in orbit, but one must ensure the control algorithms design is
sound before launch.

6.5 The Space Sauna: Thermal Control

We are very familiar with using electronics in our everyday lives. Laptops, smart-
phones, TVs are all around us. We pay little attention to the fact that electronics
are designed to work within narrow thermal ranges. For example, Helsinki winters
remind me of this every year when my phone (a phone that might not from the
highest end but also not the cheapest either) starts rebooting or turns off when I
am outside walking with the dog with – 18 °C outside. Electronics need a specific
thermal environment for it to work as intended.
Spacecraft electronics is no exception. It is the role of the thermal con-
trol subsystem to make sure all subsystems operate within their allowable flight
temperatures (AFT).
The thermal design engineer needs to consider all heat sources the spacecraft
will have, typically the Sun, the Earth, and the on-board electronics and subsys-
tems. All these sources or inputs are not steady but vary with the position in the
orbit or seasonally.
Thermal control manages the spacecraft temperatures by thermally isolating
the spacecraft from the environment, except in specific parts where radiators are
purposely placed. Thermal control engineering relies on specialized software tools
that provide simulations and numerical computations to have a good understanding
of the thermal scenarios the satellite will experience during its mission long time
before the spacecraft is brought to environmental testing to verify the thermal
design. In space projects, thermal control engineering runs for quite a long time
based solely on theoretical models and calculations until the chance to verify them
in a controlled environment comes.
As expected, the thermal subsystem interacts with many other subsystems, par-
ticularly with the power subsystem. This results from the need of accounting for
all dissipated electrical energy from the equipment around and transferring this
6.5 The Space Sauna: Thermal Control 75

energy to a radiator for rejection into space. Also, batteries generally present a
narrow temperature operating range and thus require special attention from the
thermal control engineers. The thermal subsystem interacts closely with on-board
software since the software to measure and control zones and rooms is typically
executed on one the main computers. Thermal engineers must work closely with
their mechanical and structure colleagues since fixations of heaters, thermostats,
and insulation blankets must be agreed, which also impacts the mass and power
budgets. Multidisciplinary engineering is at its best.
Heat in space does not transfer by convection but only by conduction and radi-
ation. This means that heat produced by on-board electronics needs to be guided
internally by means of conducting it through the proper physical channels toward
radiators so that it can be radiated away to space. In the same way, spacecraft can
(and will) absorb radiation from the external environment, namely from the Sun
and the Earth. This radiation can be either absorbed for practical purposes (for
heating things up) or reflected, to avoid overheating critical components.
Thermal control design internally discretizes the spacecraft volume in ‘zones’ or
‘rooms’ and makes sure all the equipment inside such areas will stay within their
allowable flight temperature (AFT) margins, which are specified by equipment
suppliers. Heaters are located in places where the thermal balance changes over
the mission lifetime. The causes of such changes are:

. Units’ dissipation changes. For example, an important heat load being turned
on or off varies greatly over time.
. External heat fluxes changes: spacecraft attitude is changed, eclipses.
. Radiator efficiency changes: changes of optical properties of radiators due to
radiation, etc.

Thermal control techniques can be:

. Passive: supported by fixed area radiators, thermal blankets, etc.


. Active: supported by electrical heaters controlled by thermostats and software.

Thermal control typically relies on the next set of devices to accomplish its task:

. Electrical heaters
. Thermistors
. Bimetallic thermostats
. Radiator surfaces
. Thermal blankets (MLI)
. Insulation materials
. Thermal fillers
. Paints
. Heat pipes.
76 6 A Peek Under the Hood

Fig. 6.8 A generic avionics block diagram

Desired orbits and payloads define which approach (passive, active) is more feasi-
ble or what type of hardware and devices to use to thermally control the spacecraft.
Active thermal control strategies increase the overall complexity of the architec-
ture, as they require more hardware to collect the necessary thermal telemetry
and run the software routines to execute the bang-bang control39 to keep the ther-
mal zones under safe margins. Said software executes in either a dedicated CPU
for thermal control—which is rarely the case—or in the main on-board computer.
An active and redundant thermal control system may severely impact the overall
power consumption of the bus and increase the mass.

6.6 The Avionics

In terms of spacecraft avionics architecture, there are many ways to skin the same
cat. Different mission requirements will drive the need for certain data interfaces,
computing power and data storage, and the avionics must be designed aligned
those needs.
When we talk about spacecraft avionics, engineers are very good at reinvent-
ing the carburetor. They—alas, we—tend to believe that everything is better if
designed from scratch. Which is a practice that is slowly phasing out, thanks to
the relatively recent appearance of standard form factors and a slowly growing
ecosystem of commercial-off-the-shelf bits and parts for satellites adopting these
form factors.
A very generic avionics architecture for a spacecraft looks as depicted in
Fig. 6.8.
The green boxes are different functional chains or blocks that provide some
essential capability for the spacecraft. Irrespective of what type of application or
mission the spacecraft is supposed to perform, those functional blocks are always
present; in other words, you cannot go to space without those functions on board.
The payload yellow box encloses the functionality that gives a purpose to the
mission.

39 In control theory, a bang–bang controller (also called on–off controller), is a feedback controller
that switches abruptly between two states.
6.6 The Avionics 77

Fig. 6.9 AOCS functional chain as a member of the avionics architecture

Some of the typical avionics functional chains or blocks do not need to have a
computer inside every time; they can be passive. For example, and as we just saw
in the previous section, thermal control can be passive, therefore computing will
not be present there, although it still exists as a function.
What stands out from the figure above, and we discussed in Sect. 6.2.3 in extent,
is that spacecraft avionics needs a great deal of interconnection. This means: the
different functional chains must exchange data with each other to couple their
functionalities for the global function of the spacecraft to emerge. That data is in
the form of commands, telemetry, or generic data streams such as files, firmware
binaries, payload data, and so forth. The architecture is recursive in the sense
that the functional chains have their own internal composition as well which will,
in most cases, also require interconnection. For example, for the attitude control
subsystem, the internal composition is exploded and depicted in Fig. 6.9.
As the picture shows, spacecraft avionics is an aggregation of data buses.40 A
spacecraft with poor interconnection between functional chains will see its perfor-
mance and operations greatly affected. This is a factor that is usually overlooked
by small, inexperienced space companies: low-speed, low-bandwidth data buses
are prematurely chosen at early stages of the project, only to find out later that the

40 A data bus is a communication pathway that allows data to be transmitted between different
components of a computer or electronic system. It consists of a set of parallel wires that connect
the various components, such as the central processing unit (CPU), memory, and input/output (I/
O) devices.
78 6 A Peek Under the Hood

throughputs are insufficient for what the mission needs. Changing interconnec-
tion buses at late stages can be costly, both in money and time. With high-speed
serial buses and peripheral-rich processors becoming more accessible and gaining
flight heritage, there is no real reason why not designing the avionics to be a big
interconnection matrix exploiting high-speed serial point-to-point connections.
Historically, spacecraft avionics has used hybrid interconnect approaches. Most
typically, ad-hoc, daisy-chained topologies, where cables come out from a box
and go inside the next one, only to go again out to the neighbor. Legacy spacecraft
avionics feature a fair deal of “private” buses; i.e., buses that are only accessible
by some subsystems and not from the rest of the architecture. When discussing
avionics interconnection, there are two different levels to discuss:

. Subsystem level: how a subsystem chooses to connect with its own internal
components. For example, how an attitude control subsystem talks to its sensors
and actuators.
. System level: how different subsystems connect with each other to materialize
the “global” function.

At either level, the approach chosen has been hybrid. Typically:

. Star topology: the subsystem main unit (which usually hosts its main CPU)
resides in a box of customized form factor and this box is the central “con-
centrator” of the subsystem. Everything flows towards it. The box exposes a
set of connectors. Then, different harnesses and cables come in and out from
those connectors, towards external peripherals. These peripherals can be either
point-to-point or connected through a bus, or both (Fig. 6.10). In this type of
design, the mechanical coupling between different parts of the subsystem is
likely different for the different peripherals; i.e., different types of connectors,
pinouts, harness, etc.
. Backplane: In this approach, the computing unit and the peripherals share a
mechanical interface which allows them to connect to a board (called back-
plane) acting as the mechanical foundation (see Fig. 6.11). The peripherals
connect by sliding in through slot connectors and mating orthogonally to the
backplane. The backplane not only provides the mechanical coupling but also
routes signal and power lines between all the different modules connected to it.
How to route the signals in the backplane is a design decision due to the fact
the backplane is basically a printed circuit board like any other.

When the backplane concept appeared, it quickly gained popularity among system
designers. The advantage of routing signals in a standardized way quickly became
attractive. This made it possible for multiple vendors to be able to interconnect
their products in backplanes and achieve interoperability. Several backplane stan-
dards proliferated, but one stood out as the most popular standard from those years:
VME (Versa Module Europe), and it is still in use today in some legacy appli-
cations. VME is one of the early open-standard backplane architectures. It was
6.6 The Avionics 79

Fig. 6.10 Subsystem federated architecture with a star topology

Fig. 6.11 A backplane


connecting 1 CPU unit and 2
peripheral boards

created to enable different companies to create interoperable computing systems,


following standard form factors and signal routing. Among the typical compo-
nents in the VME ecosystem, you can find processors, analog/digital boards, and
80 6 A Peek Under the Hood

the like, as well as chassis (housings), of course backplanes, power supplies, and
other subcomponents. System integrators benefited from VME because:

. It provided multiple vendors to choose from (supply chain de-risk)


. It provided a standard architecture versus costly proprietary solutions, lowering
switching barriers
. It gave a tech platform with a known evolution plan
. It provided shorter development times
. It lowered non-recurring costs.

Over the years, and as avionics architectures continued pushing the boundaries
in terms of bandwidth and performance, parallel bus based architectures started
to become obsolete. In fact, what became obsolete was the design philosophy:
the idea of a single bus with slave boards connected to the same lines wait-
ing for the master to dictate the overall operations does not hold to what the
modern applications need, which calls for more parallel processing with multi-
master architectures. Today, the approach has shifted into using point to point,
high-speed serial links and fiber optics to ensure the data rates needed for transfer-
ring and processing payload information. Examples of modern standard backplane
technologies are SpaceVPX (VITA 78) and CompactPCI Serial for Space.
SpaceVPX is a high-speed, ruggedized, and modular open standard for space
applications. It incorporates architectural measures to meet the requirements of the
harsh environment of space, including radiation, temperature extremes, and high-
vibration environments. SpaceVPX is based on the VPX (VITA 46) and OpenVPX
(VITA 65) standards, which are widely used in the defense and aerospace indus-
tries. SpaceVPX is intended to provide a common architecture for space-based
electronic systems, allowing for easier integration, maintenance, and upgrades. It
is also intended to promote interoperability and reusability of components and
subsystems, reducing development costs, and improving time-to-market for space-
based systems. The SpaceVPX standard is maintained by the VITA Standards
Organization, a non-profit trade association that develops open technology stan-
dards for critical embedded systems. In the SpaceVPX approach, the signal routing
in the backplane is defined by the system designer, therefore mission dependent.
CompactPCI Serial Space (CPCI-S.0) is another open standard for high-
performance, ruggedized computing systems designed for space applications. A
tad more popular in Europe, it is based on the CompactPCI Serial standard
(PICMG CPCI-S.0) and incorporates additional requirements to meet the demands
of space environments, including radiation tolerance, thermal management, and
specification for low outgassing. The CompactPCI Serial Space standard defines
a modular architecture that allows for easy integration of custom and off-the-
shelf components, including CPU boards, memory modules, storage devices, and
input/output (I/O) modules. The standard supports high-speed serial intercon-
nects, including PCI Express, Serial RapidIO, and Gigabit Ethernet, providing
high-bandwidth connectivity between system components. CompactPCI Serial
6.7 The Payload 81

Space systems are designed to operate in harsh environments, including high-


vibration, shock, and temperature extremes. The standard specifies requirements
for radiation-hardened components and materials, as well as for thermal manage-
ment and EMI/EMC shielding. In CPCI Serial Space, the backplanes have fixed
topologies and signal routings, therefore stay the same from mission to mission.
The CompactPCI Serial Space standard is maintained by the PCI Industrial
Computer Manufacturers Group (PICMG), a non-profit consortium that devel-
ops open standards for industrial computing and communication systems. It is
intended to provide a common architecture for space-based electronic systems,
allowing for easier integration, maintenance, and upgrades, as well as promoting
interoperability and reusability of components and subsystems.

6.7 The Payload

All the complexity we explored in the previous sections meets its main purpose
here. The payload is the component on board the satellite which gives meaning to
it all; it is the VIP passenger on board. But it all starts with the mission.
The mission means: what is the spacecraft intended for. The reason why you
happen to need a metallic box flying at unholy velocities, far from the ground.
Because there are simpler ways of doing things than having to launch something
into space, so there must be a good reason one must choose to go for that, and
that’s what’s called the mission. The mission defines the application: be it com-
munications, earth observation, surveillance, science, astronomy, or deep space
exploration. There are multiple different ways to use satellites, and the mission
must clearly state what it aims for. With this information in the bag—which tends
to take time to figure out and refine—then comes the part of understanding which
sensor or equipment can materialize the freshly captured mission requirements.
Zooming out, the mapping appears as rather simple: the mission drives the
payload and the orbit, which in turn drives the overall spacecraft design, or the
tailoring of the design in case of a multi-mission bus. Here is a slightly expanded
list of spacecraft payload types (the list is only notional and not exhaustive):

. Earth Observation: These payloads are used to monitor, track, and study the
Earth’s surface, atmosphere, and oceans. They typically include sensors, cam-
eras, radars, and other instruments that can capture data in different wavelengths
of the electromagnetic spectrum. Earth observation payloads can be used
for a wide range of applications, including surveillance, weather forecasting,
environmental monitoring, and disaster response.
. Communication Payloads: These payloads are used to ensure the flow of data
between the spacecraft and ground stations (including end user terminals) or
other spacecraft. They typically include antennas, transmitters, processors, and
receivers that can manipulate and process signals in different frequency bands.
See Sect. 6.2 for more details about how telecommunications work).
82 6 A Peek Under the Hood

. Scientific Payloads: These payloads are used to conduct scientific experiments


and research in space. They can include telescopes, spectrometers, particle
detectors, and other instruments that can measure and analyze various phe-
nomena in space. Scientific payloads can be used for a wide range of research
areas, including astronomy, astrophysics, and planetary science.
. Navigation Payloads: These payloads are used for precise determination of
position of assets. They typically include hardware that can determine the
spacecraft’s position and velocity relative to other objects in space, which in
turn may be transmitted to the ground for user terminals to determine their own
positioning.
. Technology Demonstration Payloads: These payloads are used to test new tech-
nologies in space. They can include new materials, subsystems, biological
experiments, and other technologies that can be tested in the space envi-
ronment. Technology demonstration payloads are essential for advancing the
state-of-the-art in space technology and enabling future space missions.

6.8 Putting It Together: Assembly, Integration and Test

As space projects mature, a situation starts to build up as the different elements,


equipment and subassemblies begin to materialize: it must be all put together. This
activity is called assembly, integration, and test (AIT), or assembly, integration,
and verification (AIV).41 Acronyms vary in bibliography or from organization to
organization, but they tend to refer to the same: the act of putting the system
together carefully, step by step, verifying that everything works according to the
plan while stressing the system under controlled test conditions.
To verify that the system is able to withstand the space environment, a set of
system-level tests shall be performed which will replicate, to certain extent, the
conditions the spacecraft will face during launch and operations. The facilities
needed for these tests are somewhat special, such as test rigs, shakers and vacuum
chambers, and seldom owned by the system designers.

6.8.1 Mechanical Tests

For mechanical tests, bus designers must specify a set of verifications through
test to demonstrate and assure the compliance of strength and static stiffness of
the primary and secondary structures throughout the design lifetime. For elements
supplied by third-party providers, the subsystem-level compliance with the envi-
ronmental requirements must be ensured, otherwise the system designer shall test

41The word ‘verification’ is more general than ‘test’. According to bibliography, there are several
way of verifying things, and testing is just one of them.
6.8 Putting It Together: Assembly, Integration and Test 83

and qualify these items separately in case insufficient evidence from suppliers is
provided. The mechanical verifications performed at the system level are briefly
described in this section.

6.8.1.1 Sine Sweep Test


A sine sweep test is a type of dynamic testing that is commonly used to evaluate
the structural integrity and dynamic response of spacecraft components, such as
antennas and solar arrays. The test involves subjecting the system to a series of
vibrations that increase in frequency over a specified range, typically from a few
hertz to several hundred hertz.
The typical steps in sine sweep tests are:

. Set up the Test Equipment and Environment: This involves using a shaker table
or other vibration equipment to apply controlled vibrations to the spacecraft.
. Define the Test Parameters: The test parameters are defined based on the specifi-
cations and requirements for the system. These parameters typically include the
frequency range, amplitude, and duration of the vibrations that will be applied.
. Apply the Sine Sweep: Once the test parameters are defined, the sine sweep test
is performed by applying a series of vibrations to the system. The amplitude of
the vibrations is typically held constant throughout the test.
. Monitor the System’s Response: During the test, the system response to the
stimulus is monitored using accelerometers or other sensors. This data is used to
evaluate the system’s dynamic response and to ensure that it meets the required
specifications.
. Analyze the Test Results: After the test is complete, the data collected from the
sensors is analyzed to evaluate the component’s structural integrity and dynamic
response.

A sine sweep test is an important step in the development and testing of space-
craft. By subjecting the system to controlled vibrations over a controlled range of
frequencies, the system’s structural integrity and dynamic response can be better
understood.

6.8.1.2 Random Vibration Test


This test shall verify the strength and structural response under loading by intro-
ducing random vibration through the mechanical interface. For the best outcome
of this tests, the random spectrum shall be notched accordingly to limit vibra-
tion levels at certain frequencies during testing which may impose stress levels on
selected units.
As we saw in Chap. 4, the launch vehicle generates significant levels of ran-
dom vibration during the launch phase. These vibrations can potentially damage
sensitive components and subsystems on the spacecraft. Therefore, it is relevant to
perform a random vibration test on the spacecraft to ensure that it can withstand
such vibrations.
84 6 A Peek Under the Hood

The launch vehicle defines the random vibration test by specifying the vibra-
tion levels and frequencies that the spacecraft will experience during the launch.
These levels are typically determined through extensive testing of the launch vehi-
cle and analysis of the data collected during previous launches. The launch vehicle
manufacturer provides the spacecraft manufacturer with a vibration specification
document that includes the vibration levels and frequencies that the spacecraft
will experience during launch. The spacecraft manufacturer then uses this specifi-
cation to design and test the spacecraft to ensure that it can withstand the vibration
profiles; this is a critical part of structural design process.
During the random vibration test, the spacecraft is mounted on a shaker table
that generates the vibrations simulating the launch environment. The vibration lev-
els and frequencies are gradually increased until the spacecraft has been exposed
to the maximum levels specified in the vibration specification document. The
spacecraft is then inspected for any damage or anomalies.

6.8.1.3 Shock Test


A shock test is an important part of the environmental verification for a space-
craft that will ride on rocket vehicles in atmospheric flight. During launch, the
spacecraft experiences sudden, high-intensity shocks due to events such as stage
separation, ignition of rocket engines, and other abrupt changes in the launch envi-
ronment such as the ‘max q’ or maximum dynamic pressure condition which is
the point when an aerospace vehicle’s atmospheric flight reaches the maximum
difference between the fluid dynamics total pressure and the ambient static pres-
sure. These shocks can potentially damage the spacecraft and its components, so
it’s important to ensure that the spacecraft can withstand them. The launch vehicle
defines the shock test on a spacecraft by specifying the magnitude, duration, and
frequency content of the shock loads that the spacecraft will experience during
launch. The launch vehicle manufacturer also provides the spacecraft manufac-
turer with a shock specification document that includes the shock loads that the
spacecraft will experience during launch.
During the shock test, the spacecraft is subjected to mechanical shocks simu-
lating the launch environment. The shocks are typically delivered using a shock
table, which is designed to reproduce the shock loads specified in the shock speci-
fication document. In more NewSpace approaches, the structure might be literally
hit with a hammer. The spacecraft is then inspected and tested for any damage or
anomalies.

6.8.1.4 Extras
There might be some other tests as well, depending on the schedule and the budget.
“Classic” space companies may execute acoustic tests which involves putting the
satellite in front of a wall of speakers just like the ones used in concerts and
blasting a few “songs” which tend to sweep different frequencies to simulate the
acoustic environment during launch. You can hear and feel these tests even if you
are miles away from the facilities performing them. Deployment tests simulating
6.8 Putting It Together: Assembly, Integration and Test 85

zero gravity conditions are also done in classic space using highly sophisticated
rigs or even helium-filled balloons.

6.8.2 Thermal Vacuum Test (TVAC)

Spacecraft designers must perform a Thermal Vacuum (TVAC) test on the


spacecraft. This type of testing simulates the space environment by subject-
ing the satellite to extreme temperatures and vacuum conditions. Here’s a brief,
step-by-step guide on how the TVAC test on a satellite is typically executed:

. Vacuum Chamber Set Up: The first step in performing a TVAC test shall be
to set up the chamber. The vacuum chamber is a large, sealed container that
creates a vacuum by removing all the air and other gases from the chamber.
The satellite is placed inside the vacuum chamber, and the chamber is then
sealed.
. Install and Verify Thermal Control Systems: The thermal control systems are
responsible for maintaining the satellite’s temperature during the TVAC test.
These systems include heaters, radiators, and thermal shrouds that help regu-
late the satellite’s temperature. The thermal control systems are installed before
the TVAC test to ensure that the satellite’s temperature is properly maintained
during the test.
. Begin Thermal Cycling: The thermal cycling process involves subjecting the
satellite to a range of extreme temperatures, from extreme positive (hot case)
to extreme negative (cold case). The temperature shall be varied gradually, and
the satellite shall be held at each temperature for a set period according to a
predefined profile to be defined during critical design stage.
. Begin Vacuum Exposure: Once the thermal cycling process is complete, the
vacuum chamber is evacuated to create a vacuum similar to that found in space.
The vacuum exposure phase is designed to simulate the vacuum conditions that
the satellite will experience in space. During the vacuum exposure phase, the
satellite’s thermal control systems are used to maintain its temperature.
. Conduct Performance Testing: Once the satellite is in a vacuum, the per-
formance testing phase shall begin. This phase is designed to evaluate the
satellite’s performance under simulated space conditions. The performance test-
ing includes a range of tests, including electrical, mechanical, and thermal
tests.
. Monitor the Satellite: Throughout the TVAC test, the satellite shall be closely
monitored to ensure that it is functioning properly. The thermal control systems
are monitored to ensure that the satellite’s temperature is maintained within
the specified range, and the performance tests are monitored to ensure that the
satellite is performing as expected.
86 6 A Peek Under the Hood

. End the Test: Once the TVAC test is complete, the vacuum chamber is re-
pressurized, and the satellite is removed. The satellite is inspected for any
damage or other issues that may have arisen during the test.

TVAC test is an essential step in the development and testing of the spacecraft. As
the project progresses into critical stage, and as more information about equipment
allowable flight temperatures (AFT) and the thermal control subsystem design
matures, a more detailed TVAC profile will be devised.

6.8.3 Software Verification

Besides shaking and cooking the satellite like there’s no tomorrow, the on-board
software must be thoroughly verified before launch. For verifying the software,
the satellite is connected to a set of simulators which feed the software with the
right stimulus to make it believe it’s flying. A variety of procedures will be run
to ensure the attitude control, and command and data handling and other parts
of the software are working as designed. More about simulating spacecraft can
be found in bibliography.42 ,43 The test environments to verify software might be
purely digital, or a combination of real on-board equipment connected to simula-
tors and stimulators. These setups are usually called “Flatsat”, engineering models
or Hardware-in-the-loop, depending on who you ask.

6.8.4 Concluding and Shipping

The AIT phase ends with a pre-ship review (PSR) which assesses the acceptability
of the spacecraft in the light of test data.
A largely overlooked area in space projects is related to shipping the satellites
to the launch site. Shipping complex goods, and goods which may contain haz-
ardous materials (such as large Lithium-Ion batteries) tends to be a bit more than
a headache. There is a fair amount of paperwork and compliance involved, and the
regulations for moving such things around the world are strict and complex. It is
very easy to overlook shipment matters when you are minding some other “more
technical” business, but it is equally, or even more important than anything else.
There is little point working hard for years on software, structures, and attitude
control systems if at the end of the day the satellite cannot go through the door
and get its passport stamped due to shipment or export regulations.

42 https://link.springer.com/chapter/10.1007/978-3-030-66898-3_6.
43 https://link.springer.com/book/10.1007/978-3-642-01276-1.
Satellites and Machine Learning
7

In God we trust. All others must bring data.


—W. Edwards Deming

Abstract
Spacecraft are not only eye-catching because they go up riding loud rockets
spitting fire. Spacecraft are also remarkable data sources. The sea of data they
generate is what essentially allows them to be reliably teleoperated. Space-
craft are time-series data goldmines thanks to a network of on-board computers
and sensors that create a tide of multivariate information to be consumed. The
question is: to be consumed by whom? Only by humans with a functioning
brain capable of “connecting the dots”? This has been the approach for the last
64 years since Sputnik 1. Can’t algorithms connect those dots? This chapter
dives into how algorithms must be equipped with the nuances needed to under-
stand the dynamic of the processes hidden behind the numbers, and how those
numbers and figures relate with each other.

We wouldn’t be very bold if we said satellites are data sources. For sure they
are, and that is perhaps one—if not the—reason why we use them in the first
place. But the term data can have different meanings. Or, in some other words,
spacecraft can generate data of different flavors and for different uses. On the one
hand, spacecraft such as observation satellites generate data which is marketable to
customers. Depending on the type of mission, this can be a depiction of an area of
the earth in the format of an image, or a scientific monitoring of the atmosphere,
or the sea.
But there is some other type of data generated on board spacecraft which is not
directly consumed by customers, although this does not make it any less relevant:

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 87


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_7
88 7 Satellites and Machine Learning

Fig. 7.1 An unannotated graph

Fig. 7.2 Google searches over time about something unspecified

spacecraft health data, also known as housekeeping data. This data may be as
simple as one bit indicating the open/close status of a valve or as complex as
biometric data of an astronaut.
Now, imagine that someone comes and puts this plot right in front of you
(Fig. 7.1).
And imagine the person showing you this does not say a single word about
what the plot represents. Well, that curve could be anything: a measurement of
temperature from an industrial process, the inflation rate of a country, the price of
some stock, the blood pressure of a panda bear, or the deformation of a beam in
certain axis. As you look at it, you can start to see some features. You can see
a change of trend at some point—it was more or less flat in the first third, then
started to ramp up. You can observe some spikes here and there. Still, this plot
says very little: we don’t even know what the axes represent.
Let’s add more information to the mysterious plot (Fig. 7.2).
Now, we see the x axis appears to be time. By now, you also most likely
recognize it comes from Google Trends, which means this plot represents searches
over time—in Google terms, “interest over time”—about something. This plot now
starts to say a bit more: it represents how much Google users have been googling
about something, in a time range. But we remain largely ignorant, as we do not
know how many people (is it one person or millions?), and we do not know the
topic searched. But we do know more about the axes now, the X axis is clearly
time, and the range of this plot goes from 1 Jan 2004 until somewhere in 2018.
This is, then, time-series data. Time-series data is a collection of discrete obser-
vations, each accompanied by a timestamp which clearly indicates when that
sample was obtained. Is it very natural to plot this kind of data versus time, this
means, y axis being the value of the variable of interest, and x axis being the time
7 Satellites and Machine Learning 89

Fig. 7.3 Plot is about people searching about dogs

at which the sample of the variable was obtained. Let’s complete the plot now for
clarity (Fig. 7.3).
The plot is about worldwide Google searches of the animal “Dog”, from 2004
to 2018. Now you could start to believe that you know everything you needed to
know about the data which originated the plot. But do you? Well, no. The data is
still leaving many questions unanswered. All you know is that people have been
more or less searching for the animal “dog” with the same “intensity” from 2004
to somewhere in late 2009, until the trend mysteriously changes. Why did the trend
change? Here lies the core of the challenge with time-series data analysis, whether
the plot is about google searches on pets, or astronaut biometric telemetry. One
thing is to see that a trend has changed, a different story is to know why it changed;
i.e., the dynamics behind it. Imagine you feel ill, you grab a thermometer, and you
can see that you have a fever. But that’s not enough to solve the root issue; the
key is to know why your body temperature has changed. Is it a flu?
As we discussed before, Spacecraft, internally, are networks of computers and
electronic devices hooked together. And because satellites are teleoperated arti-
facts, it means that once launched, you can only get to know their status by means
of measurement data sent to ground. Operators rely entirely on time-series data
collected as the object flies: currents, voltages, temperatures, pressures, different
types of counters, Boolean status bits. And here, it is important to observe the
following: measurements represent discrete “snapshots” of physical variables that
evolve in a continuous capacity. But the actual physical variables themselves, we
shall never be able to directly observe. It is only through transducers (aka sen-
sors), which are devices sensitive one way or another to the physical variables of
interest, that we can obtain indirect observations of such variables. For example: a
temperature sensor does not output a temperature but a voltage, a current or a dig-
ital word which represents the temperature. When you check your liquid-in-glass
thermometer out of your window to see how you need to dress in a cold morning,
what you are reading is the expansion/contraction of a liquid in a capillary all put
90 7 Satellites and Machine Learning

Fig. 7.4 Google searches about Christmas are obviously seasonal

in a convenient scale which makes you believe you are reading a temperature, but
you are not, nor you ever will be able to read a temperature directly. You can only
“feel” temperatures directly through your body, but not in a quantifiable way. And
machines, at least for now, are unable to “feel” like we do.
A problem starts to reveal itself: because between our control needs and the real
physical variables we need to gauge for control there will always be something,
measurement devices such as sensors can—and will—introduce noise and artifacts
which are not present in the physical variable being monitored, but how to know?
A faulty sensor may interfere with the time-series data by introducing a trend
which is not present in the variable it is sensing or add out-of-family values which
might look like serious failures.
Is time-series data on-board a spacecraft generated from stochastic processes?
Well, most physical processes in the real world involve a random element in their
structure. A stochastic process can be described as ‘a statistical phenomenon that
evolves in time according to probabilistic laws’. Time-series data generated on
board is a combination of deterministic processes (a power system, batteries, ther-
mal control, star tracker sensors observing the sky) combined with a random part
which stems from the contribution of non-ideal nature of sensing devices, thermal
noise, etc.
Time-series data can show different types of variations: seasonality, trends, and
other cyclic variations. Let’s use Google trends again to exemplify seasonality with
an obvious one: Christmas (Fig. 7.4).
With little surprise, we see that the term “Christmas” is very popular once per
year, unsurprisingly peaking at the end of December. On board satellites, there
are plenty of variables which are ‘Christmas-like’ variables, that is, variables with
obvious seasonality. What marks a season on a satellite? It depends on the orbit,
7 Satellites and Machine Learning 91

Fig. 7.5 A bit of less obvious seasonal spikes in data

but typically the sunlight/eclipse cycle.1 As the satellite goes from being under full
sunshine to the darkest shadow, many variables on board will reflect this, notably
temperatures, as the system may experience extreme thermal swings due to this
cycle, given that the sun’s radiated heat is absorbed by the structure and then
spread across the body through its different conductive surfaces. Areas looking
straight to the Sun can experience wild swings of several tens of degrees Celsius
in a very short time.
Let’s see now a bit of a less obvious example of seasonality, using the term
“Football” (Fig. 7.5).
There are spikes all over the place, but those in June 2006, June 2010, June
2014, and June 2018 are sensitively higher. What is the reason behind those?
You guessed it right: World Cups. Now, we should note that people around the
world may like different types of football: American football, Australian football,
Gaelic football, etc. All those searches are most likely also part of this plot. If we
were strictly searching for what is also called miscalled “soccer”, all those other
“footballs” can be considered noise, but they can’t be ignored.2 A man’s signal
is another man’s noise. Satellite time-series data can show plenty of ‘world-cup-
like’ spikes. For instance, when loads are turned on or off, there can be peaks
associated with capacitors suddenly charging (inrush currents), inductive loads,
etc. If needed, spikes or other artifacts can be smoothen-out from the data with
different signal processing techniques such as filtering, moving averages, etc. But

1 Certain orbits may not experience eclipses at all or may face eclipses during certain times of the
year.
2 The plot shows an interesting “negative” spike around April 2020. Most likely related to the

COVID-19 pandemic most likely related to the fact most football leagues around the world went
to a full stop during lockdown.
92 7 Satellites and Machine Learning

Fig. 7.6 Pythagorean theorem searches versus time

watch out: processing time-series data may also create mirages.3 A proper data
scientist must be able to look at the numbers with a good dose of skepticism while
supporting her work with the necessary domain knowledge.
One thing might be very clear by now: the multivariate nature of satellite on-
board telemetry. Very rarely a variable on a satellite is absolutely “self-contained”,
this means, its behavior entirely depending on itself, with all the information about
it strictly packed in its value and its trends, and nothing else. Self-contained vari-
ables are—by far—the exception. Then, understanding data correlation on-board
spacecraft is paramount. Let’s see an illustrative example, using the data about
google searches on “Pythagorean Theorem” versus time, worldwide, 2004–2021
(Fig. 7.6).
It shows a clear yearly seasonality, where every July of every year since 2004
the “interest” about Pythagorean Theorem drops substantially. If this plot were to
hold absolutely all the information needed, we could conclude that when it gets
hot, everyone stops caring about the Pythagorean Theorem. Clearly this is absurd,
and the seasonality obviously responds to the school season being over at that time
of the year (Fig. 7.7).
On board a spacecraft, time-series data variables depend on a variety of other
variables which in turn depend on other variables, and the analysis shall be done
taking this dependence into account, otherwise the conclusions drawn might be
dangerously wrong. Drawing wrong conclusions on top of bad measurements or

3Eugen Slutsky showed that by operating on a completely random series with both averaging and
differencing procedures one could induce sinusoidal variation in the data. Slutsky went on to sug-
gest that apparently periodic behavior in some economic time series might be accounted for by the
smoothing procedures used to form the data. More info: https://mathworld.wolfram.com/Slutzky-
YuleEffect.html.
7 Satellites and Machine Learning 93

Fig. 7.7 A probable correlation: Pythagorean theorem searches and school season (note the inter-
esting noise during the COVID-19 pandemic)

misidentifying correlations and covariances with causation can make the situation
spiral down into a disaster, at times with terrible consequences.4 Often, humans in
the loop can make the situation worse by adding risky time delays before acting;
we are very smart, but we take our time to think in the face of ambiguity.
Although spacecraft operation remains a very human-centered activity, equip-
ping machines with decision-making power during emergency situations does
not come without its challenges as well. If anything, machines can be equipped
with failure isolation capabilities which, at least, will minimize the probability of
situation worsening or snowballing.
Time-series data may show slow, long-term trends which may be difficult to
identify. These kinds of trends are particularly tricky on spacecraft telemetry. Such
trends can indicate a slow “build up” of a situation leading to a failure. For exam-
ple, a steadily increasing current consumption of some equipment on board, for
instance a radio, could indicate the device is requiring more and more power to
function due to degradation produced by TID, which may eventually put too much
stress on the power regulators feeding the device if the power drain crosses certain
boundaries. Long term trend analysis may be hard to perform as it may require
processing very large amounts of data and be deprioritized in favor of more imme-
diate, shorter-term, and easier to identify data features. Slow trends are perhaps
the most unsettling type of anomaly to track: hopelessly watching a telemetry
variable consistently approaching a critical threshold is not precisely an amusing
experience.
Now, the big question: can we predict time-series data? The quick answer is:
it depends. The somewhat longer answer is that, by knowing the features of a
variable (seasonal, cyclic and other trends) and by having good domain knowledge
(see next section for more on this) about the process which generates such variable,
we can assess with some level of confidence that the variable might continue

4See a tragic case of actions taken on top of bad measurements: https://pubsonline.informs.org/


doi/10.1287/orsc.2017.1138.
94 7 Satellites and Machine Learning

Fig. 7.8 Life of a Turkey: all


is great, and nothing indicates
the trend will change; until it
does. Source The Black
Swan, by Nicholas Nassim
Taleb; Wikimedia Commons

evolving one way or another, but we shall never be able to fully predict—this
means, have 100% confidence—its behavior in the longer shot because there can
always be “black swans”. Let’s use a famous example.
Consider a turkey that is fed every day. Every single feeding will firm up the
bird’s belief that it is the general rule of life to be fed every day by friendly
members of the human race. This feeding process goes on for days, weeks, months,
and years. If a remote operator were able to monitor the turkey’s weight or quantify
its happiness as a telemetry variable, it would be a very boring variable, showing
a very “predictable” increasing trend. The operator could even be able to do some
hand calculations and assess her predictive skills by estimating how much weight
the bird will gain next week. Absolutely nothing indicates the bird’s situation will
change anytime soon, and the chubby feathered fella believes life couldn’t get any
better. Until some Wednesday before Thanksgiving.
The history of a variable or a process tells only a partial story about what is
going to happen next. Naive projections of the future from the past can be very
misleading (Fig. 7.8).
Perhaps somewhat less dramatically than what happened to the poor turkey,
spacecraft can grow statistical models where initial “black swans” can become
“whiter swans” as more data is gathered. For example, any satellite in a Low-Earth,
Sun Synchronous orbit will periodically visit a region called the South Atlantic
Anomaly,5 which is a region where the magnetosphere allows for an increased
flux of high-energy particles. As it is known (and we spoke in extent in Sect. 3.3),
such particles affect the performance of on-board electronics, creating unwanted
resets, bit-flips, and other nasty effects (see Sect. 3.3.1 in particular). Any untrained
algorithm, as well as an inexperienced human operator, might be puzzled the first

5 The South Atlantic Anomaly (SAA) is an area where Earth’s inner Van Allen radiation belt
comes closest to Earth’s surface, dipping down to an altitude of 200 km (120 mi). This leads to
an increased flux of energetic particles in this region and exposes orbiting satellites to higher-
than-usual levels of ionizing radiation. The exact extension of the SAA varies with magnetosphere
conditions. An algorithm preventing from failures during SAA flyovers shall be able to ingest this
variable geometry as a configuration input.
7.1 Can There Be Too Much Data? 95

time the spacecraft is affected with SEFIs while flying over there. It could be
still somewhat perplexing by the second and the third time. By the fourth time,
and because orbits can be predicted with reasonable accuracy, no one should be
surprised, and both an algorithm or an operation shall be able to assess if the
current geolocation of the satellite will approach this critical area and take proper
precautions, for example avoiding critical operations or going into safe mode. Fool
me once…
In conclusion, spacecraft are not only eye-catching because they go up riding
loud rockets spitting fire. Spacecraft are also remarkable sources of data about
their own status in the form of a legion of sampled variables sent to ground.
This sea of data is what essentially allows spacecraft to be reliably teleoperated.
Spacecraft are time-series data goldmines, continuously monitored by a myriad of
computers and sensors creating a tide of multivariate information to be consumed.
The question is: to be consumed by whom? Only by humans with a functioning
brain capable of “connecting the dots”? This has been the approach for the last
64 years since Sputnik-1. Can’t algorithms connect those dots? They can, but such
algorithms shall be equipped with the nuances needed to understand the dynamic
of the processes hidden behind the numbers, and how those numbers and figures
relate to each other.

7.1 Can There Be Too Much Data?

As we speak, petabytes of data are stored only to be ignored forever. It may be


telemetry from a gas turbine, sales data from an e-commerce platform, or video
footage from a camera pointing to a bucolic street. Neglecting data is a growing
problem: we appear to be generating more and more data, but we humans are still
instrumental in performing all sorts of data cleansing, wrangling, and tidying for
algorithms meant to work with such data to perform better. No matter how sophis-
ticated or complicated our machine learning algorithms might be, they still require
a human brain equipped with good domain knowledge to help the algorithm be
aware of the nuances they cannot parse by themselves. In the words of Robert
Monarch6 : no algorithm survives bad data.
But data wranglers and feature engineering experts are not growing at the same
rate as data is growing. Therefore, just like an overflowing sink, raw data is filling
hard disk drives everywhere.
This data surplus appears as a ‘good problem’ to have: it is better to have more
data than less data. But is it? Data seems to be showing what economics calls
marginal utility. This factor has been illustrated by the famous diamond-water
paradox, where an essential element for life like water is comparatively cheaper
than an object with less practical use like diamonds. The paradox is explained

6Robert Monarch is an expert in combining Human and Machine Intelligence, working with
Machine Learning approaches to Text, Speech, Image, and Video Processing.
96 7 Satellites and Machine Learning

around the concepts of marginal utility and total utility. Marginal utility refers to
the additional satisfaction or usefulness a person derives from consuming one more
unit of a good, while total utility refers to the overall satisfaction or usefulness
derived from consuming all units of a good.
Water has a high total utility because it is essential to human life. However, its
marginal utility is low because people have access to enough water to meet their
basic needs. Therefore, the additional satisfaction from consuming an additional
unit of water is low, and its price is correspondingly low.
On the other hand, diamonds have a low total utility because they are not essen-
tial to human life. However, their marginal utility is high because they are scarce,
and people value them highly for their beauty and rarity. Therefore, the addi-
tional satisfaction from consuming an additional diamond is high, and its price is
correspondingly high.
In other words, the value of a good is determined not only by its usefulness, but
also by its scarcity and the additional satisfaction it provides. The diamond-water
paradox is a reminder that prices are determined not only by the intrinsic value
of a good, but also by the subjective values and preferences of the people who
consume it.
Data follows a similar pattern. The more data there is, the less the satisfaction
of a single ‘unit’ of extra data generated. A century ago, the satisfaction of taking
a photo must have been something. Nowadays, we take 5 photos of the same thing,
just to be sure. You feel no extra satisfaction as you check those 5 same pictures,
besides the preference of choosing the one you hair looks best. During a data
drought—in satellites, for instance during failures—every single bit of information
becomes a matter of life or death.
You would think data is always valuable, as everyone appears to make it look.
But data can also show effects intangible assets show, such as sunkenness. That
means that certain types of data is difficult to sell, because such data is only worth,
if worth at all, to certain audiences. Think of an IoT equipped fridge temperature
data stored in the cloud. Would anyone be willing to pay for it?
Also, research7 has provided some proof that as data sets grow larger they tend
to contain arbitrary correlations. These correlations appear simply due to the size
of the data, which indicates that many of the correlations will be spurious. Too
much information tends to behave like very little information.
And, last but not least, data gets old. For instance, earth observation data rapidly
loses value as time passes and as the scenery under observation changes. Try using
Google Street when the city imagery available is 10 years old. Only good for
nostalgia, but not for navigation as the scenery has most likely changed a lot.
In times where Machine Learning and Artificial Intelligence are the buzzwords
everyone is wanting to spout in their marketing materials, it is good to think

7“The deluge of spurious correlations in Big Data”: https://www.di.ens.fr/users/longo/files/Big


Data-Calude-LongoAug21.pdf.
7.1 Can There Be Too Much Data? 97

about the practical limits behind data accumulation and the effort—human effort—
required to improve data quality for algorithms to perform better. Satellites create
invaluable data about their own performance which cannot go ignored, and with
more processing power on-board such data can be used to empower the spacecraft
to take more autonomous actions and let the human operators on the ground mind
more intellectually challenging tasks than monitoring a battery charge on every
pass.
Operating Distant Machines Floating
in Space 8

The difference between a developer and a sysadmin is


that the developer wants the system to work, whereas the
sysadmin wants the system to keep working.
—Unknown

Abstract
Although most computer-controlled cyber-physical systems share a lot in com-
mon—after all, they all carry CPUs, memories, sensors, actuators, operating
systems, and application software—the fact their underlying behavior is largely
architecture and mission dependent requires creating unique sets of commands
and responses to manage this unique behavior. This chapter delves into the
challenges present while operating bad-tempered systems over unreliable, noisy
channels, and how designers’ choices on commands and response’s become
critical to ensure a sound operation.

Picture yourself sitting in front of a computer system. And imagine you want to
interact with it. You may want to know its status, or command it to do something
you need. What’s next? Well, we’ll have to assume the system offers a User Inter-
face (UI) which includes a mechanism for inputting and expressing what you want
to do (a keyboard, a touchscreen, etc.), and a way of displaying outputs (a screen,
some LED lights, some audible buzzer). Nothing too complicated or unfamiliar
thus far.
Now imagine that, in turn, the system is remotely located. You are in a specific
longitude and latitude, and the system is diametrically on the opposite side of the
globe. Then, you have no keyboard, no screen, no possibility of eyeballing any
LED lights whatsoever, let alone hearing a beep.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 99


I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_8
100 8 Operating Distant Machines Floating in Space

What now? We may need to rule out telepathy or the possibility of hiring a
psychic.
This is, in a nutshell, the challenge of administering remote systems that contain
computers. It could be a server, a satellite orbiting Venus, a power plant in a
distant location, an ultra-low-power IoT sensor gauging the humidity of a crop in
a distant field, or a drone patrolling a border. How do we interact and manage
distant, bad-tempered systems when all we have is a noisy, unreliable channel that
sends information as a stream of faint electromagnetic energy? The type of tasks
we would like to perform on this remote system are:

. Analyzing system logs and telemetry identifying potential issues.


. Applying updates, patches, and configuration changes.
. Adding, removing, or updating passwords, encryption keys, etc.
. Tuning system performance.
. Configuring, adding, and managing memory contents and file systems.
. Performing tests, commissioning procedures, etc.

And here lies an interesting issue. We’ve talked about “computer systems” so
far. But there are different kinds of them. On one hand, a remote server is basi-
cally pure software running either a web server, a database, a backend, or similar,
without much interaction with physical reality other than needing a power input
and perhaps sensing the surrounding temperature for informing air conditioning
requirements. On the other hand, we may have what’s we have called cyber-
physical systems, that we said is a combination of software plus an underlying
collection of special equipment to gauge the environment—sensors—and act upon
said environment by means of actuators. Then, when it comes to administering
these special kinds of computer systems, we need to interact with their spe-
cial commands, telemetry, and internal state machines. What is a state machine?
Shortly, a state machine is a computational abstraction which defines different
modes and transitions between those modes that define the system’s behavior in
time. For example, a fuel pump has a reasonably simple state machine (Fig. 8.1).
Now, as a cyber-physical system complexity grows, it becomes a composition of
a myriad of state machines running in parallel, although there is always a “system
level” state machine which aggregates those smaller, local ones and defines the
behavior of the system from an external observer’s perspective.
When cyber-physical systems adopt operating systems, part of the problem is
(kind of) solved. Operating systems come with a rich set of pre-loaded commands
and responses (the usual ls, nc, pwd, dd, chmod, or similar), so system devel-
opers and operators do not have to reinvent the wheel and can focus on application
behavior. Operating systems have undergone a good deal of standardization with
notable efforts such as the Portable Operating System Interface (POSIX). But—
there’s always a but—most operating systems expect commands and responses to
be sent and received by means of the user interfaces of the style mentioned at the
beginning of the chapter. Then, when tele operating an system that contains an
OS, we must replicate the “keyboard + screen” scenario somehow, even if we are
8 Operating Distant Machines Floating in Space 101

Fig. 8.1 Fuel pump state machine (you probably don’t think of this while you’re topping your car,
but it’s what the pump needs to deal with)

remotely located. If you happen to be on a good network, this is just a remote


shell. Nothing too fancy, and it works well. Fun kicks in when you are NOT on
a good network, and you need to type human-readable text and read text back
from the OS. When the channel is bad, shells are a suboptimal way of adminis-
tering remote operating systems. One solution for this is to define very efficiently
packed commands with carefully defined responses. In short, human-readability is
abandoned in favor of reducing overhead and optimizing the use of channel band-
width. But, because the standard commands in operating systems are designed for
humans, you still need to parse their standard outputs to understand the status of
their execution. There is a reality: making things more “human-eye friendly” tends
to bring a good dose of informational redundancy. For example, informing about
the quantity zero over a serial line for a human to interpret it takes 8-bits (ASCII
code 0x30), whereas a computer can manage with just 1-bit.1 Operating systems
and their interfaces to the outside world still remain highly human centered.

1 Actually, with 1 bit, more than a “quantity” you can represent a status (true/false, hi/low, on/off).
102 8 Operating Distant Machines Floating in Space

Needless to say, an operating system cannot predict the behavior that under-
lies the cyber-physical system it’s installed on. A fuel pump, an electric scooter, a
drone, a rover exploring Mars; they all have very different architectures and pur-
poses and therefore they interact very differently with the environment. Even if all
those systems run the exact same operating system distribution and version, the
system designers must create their own application software and system-specific
commands to alter and gauge the execution of the underlying state machine. How
to define all this application-specific things? There is no recipe for this; it’s up
to the designers and largely system dependent. This also means there is very lit-
tle standardization when it comes to defining commands and responses to handle
remote systems’ time evolution.
To summarize, although most cyber-physical systems share a lot in common—
after all, they all carry CPUs, memories, sensors, actuators, operating systems,
and application software—the fact their underlying behavior is largely architecture
and mission dependent requires creating unique sets of commands and responses
to manage unique behavior. What’s more, when tele operating these somewhat
bad-tempered systems over unreliable, noisy channels, designers’ choices on com-
mands and response’s structure become critical to ensure a reliable operation.
Choices that tend to be underestimated until too late in the system life cycle.
Making Reliable and Dependable
Spacecraft 9

Reliability is not something you can add later.


—SpaceX Founder and CEO

Abstract
System dependability is a design decision, and it requires architectural flexibil-
ity while equipping the flight software with the proper decision making power
to act and avoid depending on a human operator to decide what to do or make
the situation worse. This chapter describes the difference between reliability,
availability, and dependability, and how all that applies to satellites.

In high school, during the first classes of programming, it is quite typical to discuss
“infinite” loops. These are the classic sequence of instructions which are repeatedly
executed as long as a condition remains true. Regardless of if you finished school
25 years ago or last week, if those loops were truly infinite, they should still be
running by now. Are they still looping by any chance? No, hence the “infinite”
label was a tad too much. A case of fake marketing?
In computing—and in general—nothing “loops” forever. Things eventually stop
running. First and foremost, because machines require energy from finite sources
which will eventually be exhausted. There are pretty good reasons why machines
don’t work perpetually.
But it can also be that someone just pulls the plug or just presses ctrl+C1 —
which is probably what happened to all our “infinite” loops back in school. Or

1 In many command-line interface environments, control+C is used to abort the current task and
regain user control. It is a special sequence that causes the operating system to send a signal to the

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 103
I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_9
104 9 Making Reliable and Dependable Spacecraft

it can be an internal failure like a power supply short circuiting, an external fac-
tor like an electricity blackout. Computers and systems eventually give up. Bad
use, defective materials, flawed designs, or simply because we flip the switch off.
Chances are that next time we need to use a system, the system will disown us
and refuse to do anything useful. We saw in Chap. 3 how susceptible to upsets
and disruptions are the chips we put in orbit, so we cannot hope a chip in space
to work forever.
This opens up the discussion for a set of concepts which are related to the
probability of artifacts operating as intended under certain circumstances. Concepts
that tend to be used interchangeably: dependability, reliability, and availability.
Do they all mean the same? Let’s see.
In literature—for example in Dependable computing and fault-tolerance2 —
dependability is defined as an umbrella concept that encompasses reliability,
availability, safety, integrity, and maintainability. The IEC international vocabu-
lary defines the dependability (of an item) as the “ability to perform as and when
required”.
Something dependable means that it is reliable AND of high availability. But
also, maintainable, and safe. What is reliability, more specifically? Reliability is
the ability to perform a required function under given conditions for a given time
interval; the continuity of correct service. There are some relevant reliability met-
rics such as MTBF (Mean Time Between Failures) and MTTR (Mean Time to
Repair) which describe reliability of a system in figures, as well as FIT (Failures
in Time) number which is the number of failures that can be expected in one
billion (10E9) device-hours of operation.
On the other hand, availability is the ability to be in a state to perform a required
function, under given conditions, at a given instant of time, or after enough time
has elapsed, assuming that the required external resources are provided. In short,
availability is readiness for correct service. Another take states that availability
is the degree to which a system is in an operable and committable state when it
is called for at an unknown—i.e., a random—time. Availability tends to be also
called uptime, although it is very easy to see uptime mixed up with reliability.
High availability tends to be specified and quantified in an unusual unit called
nines. Percentages of a particular order of magnitude are sometimes referred to
by the number of nines or “class of nines” in the digits. For example, electricity
that is delivered without interruptions 99.999% of the time would have 5 nines
availability, or class 5. This unit finds application in enterprise computing and
telecommunication equipment, often as part of a service-level agreement. Mind
that availability and the number of nines a service complies to is a somewhat
slippery metric which can be hard to gauge, depending on the scope of the service
and how the downtimes are—sometimes opportunistically—counted.

active process. Usually the signal causes it to end, but the program may “catch” it and do something
else, typically returning control to the user.
2 https://ieeexplore.ieee.org/document/532603.
9 Making Reliable and Dependable Spacecraft 105

Table 9.1 Availability and downtime as function of “class of nines”


Availability (%) Downtime per Downtime per Downtime per Downtime per
year month week day (24 h)
90% (“one nine”) 36.53 days 73.05 h 16.80 h 2.40 h
99% (“two 3.65 days 7.31 h 1.68 h 14.40 min
nines”)
99.9% (“three 8.77 h 43.83 min 10.08 min 1.44 min
nines”)
99.99% (“four 52.60 min 4.38 min 1.01 min 8.64 s
nines”)
99.999% (“five 5.26 min 26.30 s 6.05 s 864.00 ms
nines”)
99.9999% (“six 31.56 s 2.63 s 604.80 ms 86.40 ms
nines”)
99.99999% 3.16 s 262.98 ms 60.48 ms 8.64 ms
(“seven nines”)
99.999999% 315.58 ms 26.30 ms 6.05 ms 864.00 ms
(“eight nines”)
99.9999999% 31.56 ms 2.63 ms 604.80 ms 86.40 ms
(“nine nines”)

Table 9.1 shows some examples of high availability for different number of
nines, with an indication of the “downtime” per relevant unit of time (year, month,
week, day) which the number of nines represents.
A fundamental difference between reliability and availability is that reliabil-
ity refers to failure-free operation during an interval, while availability refers to
failure-free operation at a given instant of time, usually the time when a device or
system is requested to provide a required function or service. Reliability is a mea-
sure that characterizes the failure process of an item and the probability of failure,
while availability combines the failure process with the restoration or recovery
process and looks at the probability that at a given time instant the item is opera-
tional independently of the number of failure/repair cycles already undergone by
the item. This is interesting because it means that something can be considered
reliable and yet of low availability. Let’s see an example using one of the most
dreadful machines ever conceived: printers.
Are printers available or reliable? You would probably say, neither. But, more
accurately, provided we patiently provide them with all the elements they need
(toner, paper, electricity, love, and comprehension) and we ensure the right con-
figuration (IP address, drivers, etc.) they may work for a given amount of time, at
least as long as all those conditions are met. This is somehow enough to call them
“reliable”; remember that reliability is the ability to perform a function for a given
period of time. But are printers available? Definitely not. Most of the times when
you randomly need them, and you happen to be in a terrible urgency to get some-
thing printed—which the machine happens to be able to sense somehow—they’ll
106 9 Making Reliable and Dependable Spacecraft

turn their back on you. Therefore, printers cannot be called dependable devices,
for they are not reliable *and* available.
Designing for dependability requires a set of engineering measures which not
only include using the right materials to ensure constructive reliability, but also
ensuring that failures, should they happen, will be sorted out in good time in
order to always keep the system in its right state to respond if needed. What
are typical examples of dependable systems? A few very quickly pop up: life
support equipment, surgery rooms, power grids, aircraft. The implications of these
systems being undependable are intolerable. In order to make them dependable,
a set of technologies and architectural decisions must be combined: alternative
power sources, hot/cold redundancies, cross-strapping circuitry, self-diagnostics,
failure detection and handling, etc.
But what about spacecraft? Is dependability a typical design driver? The answer
is (no pun intended): it depends.
Small satellites, for example for LEO missions, tend to be designed for
cost-effectiveness, and it is because of this—among other factors—that their
dependability performance can be, to say the least, mediocre. Small satellites may
be equipped with redundancies on board, but this still does not imply high avail-
ability. Why? Principally because they typically require intervention of a human
operator to assess on-board anomalies and command the satellite back to nominal
state, which can only be done during the discrete windows of opportunity when
the satellite is passing over its ground station. In short, LEO satellites’ uptime—
this means, being in a state of readiness to generate revenue when requested—is
not precisely impressive and rather a function of the independent probabilities of
failure and contact with ground occurring around the same time: the only way
to minimize downtime would be if the failure happens right at the time when a
human eye is looking. Unlikely. Customers of LEO constellations must usually
deal with low availability figures of constellation services: it can take several tries
for a customer to obtain an acquisition of a location of interest because of satel-
lites not being ready when acquisition request arrives, missing good opportunities
of capturing relevant, time-sensitive information. No one would accept having to
try several times to take a picture with their phones: a good reason why we take
pictures is to capture something which is unlikely to happen again, at least to hap-
pen in the same way. Of course, except for people who like to take pictures of
their breakfast.
On the other hand, telecommunications satellites tend to be designed for high-
availability and high reliability, typically in the range of two-nine and a half figure
(~ 99.5%3 ), which means having only around 2 days per year maximum of down-
time allowed. SATCOM spacecraft are mandated to be dependable due to the fact
they typically serve many customers at once due to the multiple-access nature of
their payloads. For SATCOM missions, failure can still be an option—space is hard

3 https://artes.esa.int/european-data-relay-satellite-system-edrs-overview.
9 Making Reliable and Dependable Spacecraft 107

after all—but their architectures allocate resources to ensure the system will maxi-
mize availability and be able to quickly go back to provide service, should internal
subsystems act up. This is done by having automatic FDIR (failure detection, iso-
lation, and recovery) software on board, combined with extensive redundancy in
the avionics architecture.
In conclusion, system dependability is largely a design decision, and it requires
the right combination of architectural flexibility and equipping the flight software
with the proper decision making power to avoid depending on a human operator
to decide what to do.
TL:DR; Frequently Asked Questions
About Space 10

10.1 Q0: Why Launch a Metal Box into Space?

Answer: Satellites have the high ground and the advantage of being able to sweep
and periodically visit large areas of the globe compared to, say, aerial platforms.
This finds multiple uses, notably Earth Observation (i.e., for taking images of the
planet using cameras or radars), or communications where on-board antennas can
have a footprint with a vast area underneath. For taking fully advantage of their
flight paths, satellites must be able to control their orientation in space to point
these payloads—cameras and antennas—to areas of interest. The type of payload
used is defined by the mission, with a mission being the reason why you are taking
the burden of launching a satellite.

10.2 Q1: What Are the Rules for Launching Something


into Space?

Answer: There is a legal framework for doing so. In summary, states are respon-
sible for their space activities and for the international registration of their space
objects and must accept liability for any damages they may cause. International
space law also requires that each state ensure that their national space activi-
ties be conducted in accordance with international law, even when these national
activities are conducted by non-state actors such as corporations, institutions, uni-
versities, and amateurs. Also, the use of the radio spectrum has to be authorized
and coordinated in order to avoid interference.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 109
I. Chechile, Space Technology, https://doi.org/10.1007/978-3-031-34818-1_10
110 10 TL:DR; Frequently Asked Questions About Space

10.3 Q2: How Are Satellites Designed and Developed? Also,


Is It Done Differently in NewSpace Versus Classic Space?

Answer: Satellites are developed more or less like any other vehicle (like a
car, a drone, an aircraft, or a truck). The design process is separated in stages.
During early stages, requirements are collected, candidate architectures are eval-
uated, make versus buy evaluations are performed where designers choose what
to procure and what to make in-house. As design matures, some prototypes are
put together and tested under simulated conditions, which can lead to a general
rethinking of the architecture if results are non-satisfactory, resetting the pro-
cess back to a previous stage. The design process is highly iterative, and once
“production-grade” maturity is reached, the more fluid stages are fully left behind.
The cost of developing a satellite is usually split between “recurrent” (things
that have to be done every time we want to manufacture a satellite) and “non-
recurrent” (things done only during conceptual design and development to never
be done again). The design process is not dramatically different between NewS-
pace and “classic” Space. Difference lies in the overall budgets for the materials,
the intensity and time spent in the development stages, and the resources avail-
able for testing, simulating, and verifying the designs. NewSpace makes extensive
use of non-radiation hardened, Commercial-off-the-Shelf (COTS) devices and
components. Fundamentally, NewSpace and classic space differ on the risks taken.

10.4 Q3: What’s Typically Under the Hood of a Satellite?

Answer: Mostly computers and cables connecting those computers together. But
also, solar panels, batteries, radios, antennas, fuel tanks, pipes, valves, cameras,
sensors, and actuators and at times some other more specific or eccentric equip-
ment like pressure vessels or radioisotope thermoelectric generators (RTGs). And
of course, a structure which provides the housing that keeps everything together
during launch and operations and at the right temperatures throughout the lifetime
of the satellite. And lots of software to make the thing perform as a functional,
integral unit. See Chap. 6 for details.

10.5 Q4: How Are Satellites Launched?

Answer: By means of rocket launchers which provide the velocity required to reach
a stable orbit. It is quite a rough ride. Rockets vibrate profusely and expose the
satellites they carry to high shocks and accelerations; satellites must be designed
to survive these harsh conditions during launch. Rockets are split in stages, and
the satellites sit on the last stage. This stage eventually gently pushes the satellites
away so they can become independent and fly on their own. That gentle push can
make the satellite spin, which is an interesting way of starting your first day at
work in space: tumbling randomly, all alone in space. It is important to clarify that
10.7 Q6: How Do Satellites Orient Themselves in Space? 111

rockets do not always take off from the ground (even though it’s the most com-
mon approach). There are air-based launchers, sea-based launchers, and submarine
based launchers. Even balloon-based launchers. See Chap. 4 for more details.

10.6 Q5: Ok the Thing Is up in Space, Now What?

Answer: Now comes the part where the satellite needs to be powered on, stabilized
(detumbled), and brought up to its fully functional state where it can deliver a
service or a product as soon as possible, because time is money. This includes
booting computers up, deploying solar panels (if needed), pointing to the Sun to
charge the battery, setting up the propulsion system, checking that thermal ranges
are nominal, calibrating instruments and more. All in all, making sure everything
is functional after the launch. This may also include injecting the satellite into
a different orbit compared to the orbit the launcher delivered it in, which is a
complex thing because high amounts of propellant is used if the target orbit is too
different from the current one, which makes the satellite experience big changes
in its dynamic properties (center of gravity, tanks fill factor, etc.).
All these activities are called commissioning or LEOP (Launch and Early Orbit
Phase). After LEOP is done, the satellite can start providing its nominal service.
There are always surprises during LEOPs that engineers have to sort out.

10.7 Q6: How Do Satellites Orient Themselves in Space?

Answer: They determine their orientation (called attitude) by observing a certain


set of variables of interest such as position of stars, magnetic field, angular speed,
position of the Sun, all measured by a set of sensors. Then, after determining their
pointing situation, satellites compute the amount of torque needed to minimize
the difference between the current orientation and the desired one and apply the
torques accordingly using actuators. To define the amount and type of correcting
action, a control “law” is coded somewhere in the on-board software.
Typical sensors used are star trackers (which are cameras facing the sky
programmed to identify star patterns, i.e., celestial navigation), gyroscopes (to
gauge angular velocity), sun sensors (which help to assess where the Sun is) and
magnetometers which measure Earth’s magnetic field, among others.
The typical actuators used in Low Earth Orbits are reaction wheels and mag-
netorquers. The former applies a reaction torque to the satellite body by means
of accelerating a rotating mass (a flywheel). The latter exploits the interaction (a
torque) between a magnetic field created by an electric current applied through a
coil and the Earth magnetic field. Thrusters are also used as actuators sometimes.
The control law is in charge of computing the amount of torque needed to achieve
a certain desired orientation, and it applies such torque by distributing it to the
available actuators. More in Sect. 6.4.
112 10 TL:DR; Frequently Asked Questions About Space

10.8 Q7: How Are Satellites Operated?

Answer: There’s a data link between the satellite and the ground, where a radio
signal is manipulated (modulated) by a data signal in order to transfer bits of said
data across the channel. Links can be bidirectional, with an uplink (from ground to
the satellite) and a downlink (from satellite to the ground). The uplink is typically
used to send commands (orders or actions) to the satellite, whereas the downlink is
typically used to receive telemetry (health data) and payload data (the data which
pays the bills). Links can also be unidirectional, as it is the typical case for payload
data links, where the satellite broadcasts data to a ground receiver, without any
feedback. Space radio links can work at different “speeds” (as in, amount of data
transferred per unit of time, or bit rate), but higher data rates do not come without
penalties. Higher bit rates require higher bandwidths and higher bandwidths can
push the envelope of the overall radio system design.
Commercial satellites establish their radio links towards third-party ground sta-
tions, which are operated by companies offering the infrastructure (big antennas
with tracking capabilities and modems), in a way satellite operators do not have to
grow this heavy infrastructure themselves. Schematically, satellite operators con-
nect to the ground stations’ modems (usually by means of a private network
over Internet) where they receive the bytes coming from the satellite and send
their bytes up to the satellite. The modulation and demodulation is handled by
the ground station supplier; this includes tracking the satellite as it flies over the
antenna, sweeping the frequency to lock onto the satellite carrier, synchronizing
and reconstructing the data from the radio channel, and storing the payload data
in local data recorders or, more recently, in cloud-based servers.
The way bits are formatted in commands and telemetry tend to be standardized
in a way satellites and ground stations can exchange data reliably considering
the inherent noisy nature of radio links. The true semantic “meaning” of the data
remains proprietary and every operator parses/decodes the data according to their
format and using their private tools.
A considerable deal of routine operations of satellites are still manual, that is,
they involve large teams of human operators working 24 × 7 (Fig. 10.1).

10.9 Q8: What Does the Software on Board of a Satellite Do


Exactly?

Answer: Multiple things. Mainly, software acts as an on-board “manager”. It col-


lects, stores, and retrieves health data from all the equipment, it processes the
commands sent from the ground, it flags, logs, and handles faults while preventing
them from spreading. And, perhaps more importantly, the on-board software han-
dles on-board behavior or “business logic” by means of running a set of state
10.10 Q9: How Do Satellites Generate Power? 113

Fig. 10.1 A ground antenna (photo by Donald Giannatti on Unsplash)

machines.1 Behavior heavily depends on the architecture, the type of mission


and the equipment present, although a great deal of software can be reused from
missions to missions, differing only in the uppermost “application” layer.

10.10 Q9: How Do Satellites Generate Power?

Answer: There are several ways, but by far the most popular is using solar power,
which requires solar panels to capture photons from the Sun, converting this
electromagnetic energy into electrical power, and storing some of that power in
batteries. Solar panels are tricky devices: largely inefficient, which means that large
areas are needed to generate power for even small satellites. Since the area needed
usually exceeds the area available from the satellite body area, and since rockets
constrain the volume a satellite can take during launch, engineers need to “pack”
(fold, stow) panels and make them “unpack” (also unfold or deploy) in orbit; which
is a risky operation. If panels fail to deploy, the mission can be greatly affected
since not enough power will be generated. Solar panels degrade substantially with
throughout the mission lifetime, so their ability to generate power at beginning of

1 A state machine is a computational abstract machine that can be in exactly one of a finite number
of states at any given time. The FSM can change from one state to another in response to some
inputs; the change from one state to another is called a transition. An FSM is defined by a list of
its states, its initial state, and the inputs that trigger each transition.
114 10 TL:DR; Frequently Asked Questions About Space

Fig. 10.2 A sketch illustrating deployable solar panels

life is not the same as their ability to generate power at end of life, and the design
engineers must take this aging effect into account when dimensioning the power
subsystem (Fig. 10.2).

10.11 Q10: How Big and Heavy Are Current Commercial


Satellites?

Answer: It is largely mission dependent. Can be as small as a few kilograms cube,


up to 6000 or 8000 kg and several meters wide (Note I am not counting here the
ISS, Hubble nor big scientific spacecraft). Currently, the payload heavily influences
the on-board resources needed, impacting SWaP (Size, Weight, and Power), but
also dictating the overall physical configuration: for example, where solar panels,
sensors and thermal radiators need to be situated. This is the more “classic” ratio-
nale of space systems configurations, where payload and bus are tightly coupled,
and the bus is designed as a one-off. More modern approaches include designing
multi-mission buses which can accommodate for different payloads.

10.12 Q11: What Is the Lifetime of a Satellite?

Answer: Lifetime is a design subject, a requirement. Therefore, mission dependent.


Satellite lifetime is more about probabilities than certainties. Satellite designers can
make some decisions during the design process to make sure the probability of the
satellite surviving the required amount of time in the specified orbital environment
is of certain confidence. Things can still fail due to early failures, which are sta-
tistically possible.2 Lifetime design decisions may involve duplicating elements

2The bathtub curve is a failure rate graph which is used in reliability engineering and deterioration
modeling and quantitatively depicts the initial non-negligible probability of “infant mortality” of
components and parts of a system.
10.14 Q13: What Is an Orbit Exactly? 115

on-board (adding redundancies), adding shielding, thermally isolating and cou-


pling, and making the software ready to keep the satellite always flying in a safe
envelope and to be ready to handle faults in a way they do not propagate. Exten-
sive testing and verification while the satellite is still sitting on the ground are key.
Single string (this means, no redundancy) architectures can be used In space with
effective risk management, architecture flexibility, extensive testing, use of proven
designs, and a rigorous approach to fault protection.

10.13 Q12: What Can Affect the Lifetime of a Satellite?

Answer: Notably the space environment, which is wild (see Chap. 3). As we said
there, there are high concentration of high-energy particles going around orbits. We
are fairly (and luckily) protected at the surface of the Earth due to the atmosphere
and the magnetosphere, but satellites do not have this luxury. The semiconductors
we fly on-board satellites are delicate microscopic structures prone of disruptions
from the environment. Radiation can affect them in different ways, potentially
inflicting permanent damage but also causing momentary upsets which, if not prop-
erly handled, may evolve into dangerous situations. Sudden software reboots and
memory bits flipped are frequent, and more frequent around some specific orbital
areas. But radiation is not the only factor that can affect lifetime. Thermal condi-
tions in orbit rapidly change due to the fact most satellites experience an aggressive
repetitive cycle of sunlight and shadow (eclipse). Such a cycle can create ample
swings in temperatures on board that can damage equipment if unhandled, and
since materials expand and contract due to temperature variations, the structure
and equipment on board can experience thermoelastic stresses which can greatly
affect, degrade, impair, or even kill a mission. Last but not least, software bugs
may affect mission success.

10.14 Q13: What Is an Orbit Exactly?

Answer: It is the trajectory the satellite follows as it flies around the globe, and
it is predominantly defined by the launch, although Earth orbits change and drift
substantially as time goes by post-launch. They drift due to perturbations from
the environment (Earth oblateness, atmospheric drag, other celestial bodies’ grav-
itational pull, sun radiation, etc.), but also satellites may purposely modify their
orbits on-demand. They do so by applying forces in specific directions and around
points of interest of the orbit using thrusters they carry. Note that not all satel-
lites are equipped with the capability of changing orbits. For those which are not
equipped with propulsion, their orbits are just what they are, and they go with the
flow as perturbations run the show and their orbits happily drift. Note that the orbit
chosen may greatly affect an earth observation mission, since it defines the way a
satellite will visit certain areas of the globe.
116 10 TL:DR; Frequently Asked Questions About Space

10.15 Q14: Why Are Orbits Crowded and How Is This an Issue?

Answer: Not all orbits are crowded, and luckily the Universe is vast. There are,
though, some particular Earth orbits which are quite popular and more interesting,
commercially speaking. For example, polar low Earth orbits are chosen by the
majority of Earth Observation organizations because a satellite orbiting in such
an orbit can sweep the globe in a short time by exploiting Earth’s rotation in
their advantage. As satellite operators fill the most interesting orbits with their
assets, probabilities of conjunction (i.e., crashing or coming uncomfortably close
to another object) increases. There is no worldwide space traffic control authority
as of today, so each satellite operator is on its own when it comes to securing their
property in orbit. If the conjunction risk is against another functional satellite,
companies can coordinate a joint course of action privately.
Satellites are launched under the promise they will deorbit (as in, reenter the
atmosphere) before a maximum specified timeframe. The constantly increasing
number of satellites being launched is raising concerns about the risk of a cas-
cade effect called Kessler Syndrome which could render space activities in some
specific orbits very difficult, including human spaceflight.

10.16 Q15: Why Are Satellites Assembled in Clean Rooms?

Answer: First, to keep particles in the air under control. Dust is everywhere. You
just cannot get a dust-free room on Earth; but you can keep the particle count
per cubic centimeter comparatively low, and that’s what clean rooms are about.
Why do dust and satellites not get along well? First, optics are very sensitive to
dust. Payload cameras, star trackers; you don’t want to have their lenses affected
by unwanted micrometer-sized particles and put the whole mission at risk; a star
tracker with a contaminated lens could be unable to track real stars in orbit. But
also, mechanisms and electronics can be affected by air impurities. Second, ESD
(electrostatic discharge) may damage satellite electronics beyond repair. What is
more, ESD may not necessarily break electronics right away but affect or degrade
its performance only to finally give up in orbit. Therefore, clean rooms are strictly
ESD-controlled environments. Clean rooms maintain a constant temperature and
humidity, which helps to minimize ESD risks. The floor and equipment around
clean rooms is specially constructed and chose so that static electricity has a hard
time building up. Clean rooms do not only keep air clean with sophisticated air-
renovation systems, but it also needs collaboration of the workers. Everyone has to
go around clean rooms with special clothes, masks, hairnets, and shoe protection
(Fig. 10.3).
10.17 Q16: How Are Satellites Currently Distributed Across Different Orbits? 117

Fig. 10.3 A team in action in a cleanroom (photo by Laurel and Michael Evans on Unsplash)

10.17 Q16: How Are Satellites Currently Distributed Across


Different Orbits?

Answer: Let’s first categorize the orbits. In general, orbits are categorized in Low
Earth Orbits, Medium Earth Orbits and Geostationary Orbits. There are other less
common orbits as well, which are out of the scope of this text.

. LEO (Low Earth Orbit): The term “LEO region” is used for the area of space
below an altitude of 2000 km (1200 mi). Approximately one-third of the radius
of Earth. Objects in orbits which pass through this area, even if they have an
apogee further out, or are sub-orbital, are carefully tracked because they present
a collision risk to the many satellites in LEO. All crewed space stations to date
have been in LEO. From 1968 to 1972 the Apollo program’s lunar missions
sent humans beyond LEO. Since the end of the Apollo program there have
been no human spaceflights beyond LEO.
. MEO (Medium Earth Orbit): MEO is an orbit with an altitude between 2000 km
(1243 mi) and 35,786 km (22,236 mi) above sea level. The boundary between
MEO and LEO is an arbitrary altitude chosen by accepted convention. All satel-
lites in MEO have an orbital period of less than 24 h, with the minimum period
(for a circular orbit at the lowest MEO altitude) about 2 h.
118 10 TL:DR; Frequently Asked Questions About Space

. GEO (Geostationary Orbit): A geostationary orbit, also referred to as a geosyn-


chronous equatorial orbit (GEO), is a circular orbit 35,786 km (22,236 miles)
in altitude above Earth’s equator (42,164 km from Earth’s center) and follow-
ing the direction of Earth’s rotation. An object in such an orbit has an orbital
period equal to the Earth’s rotational period, and so to ground observers it
appears motionless in a fixed position in the sky. The first satellite to be placed
in this kind of orbit was launched in 1963. Communications satellites are often
placed in a geostationary orbit so that ground-based satellite antennas do not
have to track them but can be pointed fixed at the position in the sky where the
satellites are located. Note that geostationary satellites’ orbits drift and must be
periodically adjusted, which consumes considerable amounts of propellant and
affects their lifetime. Weather satellites are also placed in this orbit for real-time
monitoring and data collection.

Orbits distribution.3 ,4
In the next graph, it can be seen the greatly uneven distribution of satellites
across the orbits previously described.
In Fig. 10.4, one can observe the distribution of satellites in a wide area between
very low earth orbits up to 50,000 km apogee. One can see that there is a big
concentration of satellites until 1300 km apogees, then very little amount, until
a spike appears somewhere above 35,000 km apogee. These are the geostation-
ary satellites use for telecommunications. In Fig. 10.5 we see the distribution
of satellites between very low earth orbits up to 2000 km. We observe a crowd
around 550 km (basically Earth observation satellites and Starlink), and then a
spike around 1200 km, belonging to the OneWeb constellation.
Finally, in Fig. 10.6 we see a zoomed in perspective in LEO, between very low
earth orbits and 700 km, showing again the spike out of Sun-Synchronous Earth
Observation (EO) satellites and the Starlink constellation.

3 Source: https://www.ucsusa.org/resources/satellite-database.
4 The number of satellites in LEO is an approximation and is continuously increasing.
10.17 Q16: How Are Satellites Currently Distributed Across Different Orbits? 119

Fig. 10.4 Distribution of satellites for altitudes between 0 and 50,000 km altitude

Fig. 10.5 Distribution of satellites for altitudes between 0 and 2000 km (LEO)
120 10 TL:DR; Frequently Asked Questions About Space

Fig. 10.6 Distribution of satellites for altitudes between 400 and 700 km (LEO)

You might also like