Writing Papers 5
Writing Papers 5
Articles
Students perspective
http://mmlab.ie.cuhk.edu.hk/ 2
The good news
Reading
Reviewing
Writing (and trying to improve)
3
The good news continues ...
Planning
Starting – Dirty your hands
Create a detailed outline of
Implement state-of-the-art
the entire paper
Colleagues
Writing
Checking
Use tools, e.g. dictionaries,
Rewriting
cross-referencing…
4
Today’s take home message
= Jobs =
5
Theme of the talk
6
Abstract
Typical abstract
Subject
Contributions (if any) and methods
Results
7
Abstract example (Viola & Jones 2001)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers [6]. The third contribution is a method for
combining classifiers in a “cascade” which allows background regions of the
image to be quickly discarded while spending more computation on promising
object-like regions. A set of experiments in the domain of face detection are
presented. The system yields face detection performace comparable to the best
previous systems [18, 13, 16, 12, 1]. Implemented on a conventional desktop,
face detection proceeds at 15 frames per second.
8
Abstract example (Viola & Jones 2001)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers [6]. The third contribution is a method for
combining classifiers in a “cascade” which allows background regions of the
image to be quickly discarded while spending more computation on promising
object-like regions. A set of experiments in the domain of face detection are
presented. The system yields face detection performace comparable to the best
previous systems [18, 13, 16, 12, 1]. Implemented on a conventional desktop,
face detection proceeds at 15 frames per second.
9
Abstract example (Viola & Jones 2001)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers [6]. The third contribution is a method for
combining classifiers in a “cascade” which allows background regions of the
image to be quickly discarded while spending more computation on promising
object-like regions. A set of experiments in the domain of face detection are
presented. The system yields face detection performace comparable to the best
previous systems [18, 13, 16, 12, 1]. Implemented on a conventional desktop,
face detection proceeds at 15 frames per second.
10
Abstract example (Viola & Jones 2001)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers [6]. The third contribution is a method for
combining classifiers in a “cascade” which allows background regions of the
image to be quickly discarded while spending more computation on promising
object-like regions. A set of experiments in the domain of face detection are
presented. The system yields face detection performace comparable to the best
previous systems [18, 13, 16, 12, 1]. Implemented on a conventional desktop,
face detection proceeds at 15 frames per second.
11
Abstract example (Viola & Jones 2001)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers [6]. The third contribution is a method for
combining classifiers in a “cascade” which allows background regions of the
image to be quickly discarded while spending more computation on promising
object-like regions. A set of experiments in the domain of face detection are
presented. The system yields face detection performace comparable to the best
previous systems [18, 13, 16, 12, 1]. Implemented on a conventional desktop,
face detection proceeds at 15 frames per second.
12
Abstract example (Better one?)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers. The third contribution is a method for combining
classifiers in a “cascade” which allows background regions of the image to be
quickly discarded while spending more computation on promising object-like
regions. A set of experiments in the domain of face detection are presented. On
MIT+CMU database, the system achieves a correct detection rate of 0.91 and
false detection rate of 5e-6, which is comparable to the best previous systems.
Implemented on a conventional desktop, face detection proceeds at 15 frames
per second.
13
Abstract example (Better one?)
This paper describes a visual object detection framework that is capable of
processing images extremely rapidly while achieving high detection rates. There
are three key contributions. The first is the introduction of a new image
representation called the “Integral Image” which allows the features used by our
detector to be computed very quickly. The second is a learning algorithm, based
on AdaBoost, which selects a small number of critical visual features and yields
extremely efficient classifiers. The third contribution is a method for combining
classifiers in a “cascade” which allows background regions of the image to be
quickly discarded while spending more computation on promising object-like
regions. A set of experiments in the domain of face detection are presented. On
MIT+CMU database, The system achieves a correct detection rate of 0.91 and
false detection rate of 5e-6, which is comparable to the best previous systems.
Implemented on a computer with the Intel PIII 455 Mhz processor, our face
detector runs at 15 frames per second, approximately 10 times faster than the
alternative approaches.
14
Introduction
Your contributions
key components
15
Introduction – a more direct approach
Your contributions
key components
16
Introduction – a more direct approach
17
Introduction – a more direct approach
(Sochman & Matas 05)
In many computer vision problems such as detection, both error
rates and computational complexity reflected by time to
decision, characterize the quality of a given algorithm. We
show that such problems can be formalized in the framework of
sequential decision-making. The optimal strategy in terms of
the shortest average decision time subject to a constraint on
error rates (false positive and false negative rates) is the Wald’s
sequential probability ratio test (SPRT). In the paper, we build
on Wald’s theory and propose an algorithm for two-class
classification problems with near optimal trade-off between
time and error rate.
18
Title
19
Interesting titles
20
Interesting titles
21
A few words about math…
22
Scalars
Vectors
Elements of a vector
Iterators
Generic indexing
Referring to equations …
28
Presenting results
Figures
Tables
A few words about figures…
30
Comparison of methods (bad)
Original Printed
How to waste space 101 (Gauvrit 05)
How to waste space 101 (Satoh et al. 03)
Incorrect model – we know aneurysms are smooth
(Hassan et al. 03)
Captions
39
Captions (bad)
Messy
Not so much information for such large space
Table Design (better)
Simpler
Cleaner
More information, less space
How to waste space 101 (McKinney 07)
Language
Interesting read:
“The science of scientific writing” (Gopen & Swan 90)
47
Language
Be polite
48
Verb Tenses
Past Tense
Example
Present Tense
Example
Passive:
Active:
The “energy crisis” and the demand for the efficient utilization of
abundant natural gas reserves renewed the interest…
Active vs. Passive Voice
55
“And” and “But”
Example:
57
Other mistakes (cont.)
It was found that only manganese and tin oxides were stable…
58
Other mistakes
An example of an bad paper:
Estimates are usually made with some sampling error, and limits of
agreement are no exception. We showed how to estimate CIs for the
limits of agreement. Another regret is that these Cis are seldom
quoted. For the data of Cicero et al.12 the mean difference was 0.2
mm with SD 3.0 mm, giving 95% limits of agreements −5.8 to +6.1
mm. There were 231 cases. The standard error of the limits is
approximately3s2/n. This gives3 × 3.02/231 = 0.34. The 95% CI for
the limits of agreement is given by ±1.96 standard errors = 0.67, so for
the lower limit the CI is −6.5 to −5.1 and for the upper limit the 95%
CI is +5.4 to +6.8. Not so hard, really!
60
Language (Bland & Altman 03)
Estimates are usually made with some sampling error, and limits of
agreement are no exception. We showed how to estimate CIs for the
limits of agreement. Another regret is that these Cis are seldom
quoted. For the data of Cicero et al.12 the mean difference was 0.2
mm with SD 3.0 mm, giving 95% limits of agreements −5.8 to +6.1
mm. There were 231 cases. The standard error of the limits is
approximately3s2/n. This gives3 × 3.02/231 = 0.34. The 95% CI for
the limits of agreement is given by ±1.96 standard errors = 0.67, so for
the lower limit the CI is −6.5 to −5.1 and for the upper limit the 95%
CI is +5.4 to +6.8. Not so hard, really!
61
Language (Bland & Altman 03)
62
Conclusions
Never cite a paper for which you haven’t read at least the
relevant part.
67
References
We all know how to use variables… BibTeX provides the @string entry
Like @inproceedings, @article,
@techreport, @thesis
68
Using table of contents
Show it to colleagues
and get feedback
71
Today’s take home message
72
Today’s take home message
Thank You!
74
More Materials
How To Write A Scientific Paper
http://www.improb.com/airchives/paperair/volume2/v2i5/ho
wto.htm