On Asymptotic Distribution Theory in Segmented Regression Problems
On Asymptotic Distribution Theory in Segmented Regression Problems
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at .
http://www.jstor.org/page/info/about/policies/terms.jsp
.
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of
content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms
of scholarship. For more information about JSTOR, please contact [email protected].
Institute of Mathematical Statistics is collaborating with JSTOR to digitize, preserve and extend access to The
Annals of Statistics.
http://www.jstor.org
The Annalsof Statistics
1975,Vol. 3, No. 1, 49-83
GeneralElectricCompany
Thispaperdealswiththeasymptotic theoryofleastsquares
distribution
estimatorsin regressionmodelshavingdifferent analyticalformsin dif-
ferentregionsof the domainof the independent variable. An important
specialcase is thatof brokenline regression,
in whicheachsegmentofthe
regression functionis a different line. The residualsumof squares
straight
functionhas manycorners,and so classicalleastsquarestechniquescannot
be directlyapplied. It is shown,however,thatthe problemcan be trans-
formedintoa newproblemin whichthesumofsquaresfunction is locally
smoothenoughto applytheclassicaltechniques.Asymptotic distribution
theoryis discussedforthenewproblemand it is shownthattheresultsare
also valid forthe originalproblem. Resultsrelatedto the usual normal
theoryare derived.
thetruestateof nature.
Asymptotic propertiesof theleastsquaresestimators of so will be examined.
Since the change-over points r1(0, .. **,1 are unknown, the derivation of
asymptotic properties moreinvolvedthanone mightat firstex-
is considerably
of r) is chosento minimize
pect. For each fixedr, theestimate6 (a function
tll(C;t) s(c)
t C
r tNn
tnltn2 tnitn)+l
FIG. 2.1. FIG. 2.2.
SEGMENTED REGRESSION PROBLEMS 53
0 = (01 . 9r); 03 = .
( 0jll I jK(j))
E) = {admissible
O's}; E = {(0, r(6); 6 e Ej; U = {He(;t); e }
T' = (T'i, * 7 r-i) ; = (@, ); 5 = (e, y2)
600?',r'?' = "true" statesof nature; P0 = P at 6 = 60() T ()
;
e(0) (O (0), r(0)) f(0) 2) (0)Io
= (6, r) = leastsquaresestimator (l.s.e.) of e(O)
= (0,,,r,,) = l.s.e. restricted
to c E
(0*, r*) leastsquaresestimator in thepseudoproblem(see Sec-
tion4)
* = (0,,',,r,w*)= restrictedl.s.e. in the pseudo problem
fj(6j ; t) = E k=1 _jkfjk( t); fjk(t) fjk f (0). t) = f.j0(t) =
fjo
X = [(; tni)+ eni
Eeni = 0; Var (enj)= 2; Elej,%12(la+)
< 00 forsome 6 > 0
H,,(s) = distributionof {tni}; Hn(s) ---d H(s) as n -+ oo
H,(A) = 5AdHn (s)
D+(h, j, k)-D+ = the kthleftand rightt-derivatives of
respectively
fh(Oh(0);t) at t= . If D+ = D- their common value is
denotedas D.
n* sample size in the pseudo problem; n** = n - n*
* = summationover the n* termsof thepseudo problem; E **
Pn *
( ) H,.t:
~~inf(,0:110-V0()JJ>d) Ijy(e;t) - p(e'?; t)J> K} > .
Thenthereexistsa d* such that
limn 0 P0{01l - 6(0)11
< d*} - 1
PROOF. Take K = 3q0o/i. For n > n(K), d > d(K), and ff6 - 6(0)1> d,
inf6,
E I (Xni - (e; tni))2_ inf 1 (et - ni)
The triangleinequalityimplies
(e,i - pi)2 > [(E OJni)- ( E e2i) ]2
- ( E 2) [(E )i 2(E - e2 )']
+E eX
> (E 02)I[30ni - +
2a,0ni o,(nl)] + E e .
Sincetheleastsquaresestimator
0 minimizes
theresidualsumofsquaresfunction
it followsthat with probabilityapproaching 1 as n -> oo 110- 60()11 ? d(K). L
LEMMA 3.5. If tois a centerof observations,
( > 0, i7 > 0, and condition(*) of
Lemma 3.4 holds,then
>
P0{1,(4;t) - p(e(0); t)l r for all t suchthat It- tol O
< .
PROOF. Let S denote{0: 110- 0')11 < d*}. Let e denotethe least squares
estimatorof e with0 restricted to S. Lemma3.4 impliesthatwithlargeprob-
abilityas n -oo,o, the unrestrictedleast squares estimator,is equal to e, the
leastsquaresestimator.We utilizeTheorems1 and 4 of Jennrich
restricted [15]
to discuss the behavior of e.
Since tois a centerofobservations,
thenumberofobservations intheinterval
[to-3, to+3]is 1n+o(n) forsome12>O. LetF= {v: Ij(s; t)? >r for all t
withIt- tol< ( and 118- 0(0)11< d*J. Now
3(Xni- p(e;t,i))= E (e,j- j)n
= ^
eli + i - 2 *e,j
if v e F, then 2u.i > h2n[I + o(l)] where o(l) does not depend on 0.
Furthermore, the argumentsof Theorems1 and 4 ofJennrich [15] implythat
n-1z eniUniconvergesto 0 in probability
uniformly for e S, sinceS is a com-
0
pact set. Thus infGF E (ei - _ni)2> Z el, + 2i2n + op(n). Since the least
squaresestimateminimizes theresidualsumof squares,it followsthat
z (ni -
it(e;tn))2 <
pM(e(?); tnt,))' e2i.
(X,,i
\/(t)
t
0 1/201/10 1/5 2/5 3/5 4/5 t I
FIG. 3.1.
= Po(t*) t,ni = t*
l.s.e. This is
of O.U,the restricted
Theorem3.6 directlyimplytheconsistency
statedas
formally
COROLLARY 3.7. If el" e co c i;, then underthe hypothesesof Theorem3.6
,- 0'? op(1), ""WI2
- - a = l.s.e. of 0, and aR2
op(l), where0,,,is therestricted
l.s.e. of a2.
is therestricted
Thus far,no mentionhas beenmadeof thebehaviorof r. It turnsout that
undersuitableconditions
r' converges and therateof con-
to r(?) in probability
vergencedependson the numberof t-derivatives thatadjacentsegmentshave
in common at the change-overpoints r1()*,
r(),
(0) T) and which
Let A denotethesetof r's such thatfj(01j(); r) = fj+1(O(0'41;
lie to therightof all centersof observation involvedin theidentification
of 0
and totheleftofall centersofobservation involvedin theidentification
of 0,+.
(For brevity, one can describeAJ(0) as thesetof T's whichare compatible,with
respectto 00(,) 0(,), withthecentersof observation.)It will be shownthatT
lies near an elementof A,') with large probabilityas n -> oo.
It is shownin Lemma3.8 thatif03 and 0,+8 are closeto 0 (O) and 0(.4 respec-
tively,theneach elementof A, (theset of rj's compatible,withrespectto 0,,
89+J,withthecentersof observation) lies close to an elementof A,(').
LEMMA 3.8. Let j-1?be anycollectionof neighborhoods whichcoversthesetA '(0).
Thereexistneighborhoods N1 and N,+1about O,(0)and 0(?+ respectively,
such thatif
0 e) 0and 0j e Nj,0j+1 e N1+1respectively,thenif vj is compatiblewithrespectto
0j, 0j+1 withthecentersof observation, v1 mustbe containedin an elementof jIft.
PROOF. Supposeto the contrary thatthereexistsa collectionof neighbor-
hoods IA, such thatforall Nj, Ni+1as describedin thestatement,
thereexists
0 e e with 0j e Nj, 0j+1 E Ni+1 and such that vj's compatiblewith respectto 0op
0j+1 existoutsideof theelementsof -xVj;. Thenthereexistsa sequence0(n) e 93
such thatOf,,,and 01j+,,, convergeto Oi(o)and 0O("1respectively and a sequence
vrj.J}such that for each n, 1j,,,belongsto no element of A,j is compatible,
withrespectto 0jn,,,0j-1,n, withthe centersof observation, and is such that
fj(60j,,; tj,n) = f+1(6j+?1,n; v,j, ). Thus thereexistsa subsequenceof Tj,?,'s which
convergesto T* 2 Aj1) and whichis compatiblewithrespectto 0j(0), 094, with
the centers. By continuity,fi(Oj(O);T*) = fj+1(0.41; T*), which contradictsthe
definitionof A (0). E
An important
specialcase of Lemma3.8 occurswhenAi(') is the one point
set, tvj(0)}. This suggests
DEFINITION 3.9. The parameter at prl by t if
0 is well-identified
(i) 0 is identifiedat p(?) by t
(ii) foreachj, 1 < j < r - 1, Aj(') is theone pointset,{rj'ol}.
of I is thenan immediateconsequenceof Theorem3.6,
The consistency
Lemma3.8, and Definition3.9. Thisis statedin Theorem3.10.
60 PAUL I. FEDER
4~~~~~~~~~~~~~~~~
2~~~~~~~~~ 3
I
EXAMPLE EXAMPLE
2
= _
2 I 2 3
EXAMPLE
3a EXAMPLE
3b
FIG. 3.2.
< m i(EJiYle
o=(l)]
e2i), < mn e2j)'[ + = Op(1) -
Now consider VN=- E1 Cie,.J(E=l 1 j)i. For every s > 0 there exists a
K = K(e) such thatN < Kp(n) < n withprobabilitygreaterthan 1 - e as n -+ oo.
Thus, withprobabilitygreaterthan 1 - e as n -* oo, IVNI < maxl?k?Kp(-) IVkl. If
Z2 is bounded as n -> oo, then Kolmogorov's inequalityimpliesthat that
zKp(n)
op(l) as n -*oo. Ei
Let noj,Nj denotetheindicesof the observations whichoccur immediately
precedingrj, j j = 1, ..., r 1. Assumethatthespacingof theobserva-
-
REMARK. If e(t)is a brokenline and the tji are equally spaced, then equation
(3.7) holdstruewithh(i) = i-a/2, p(n) = n,and N, Cnidefinedas in thepreceding
paragraph.To see this,expressfjf?)(t)= ai(?' + bj'?'(t-
-1 f,(t) = a +
-
b'.?+1(t (j0). Then Cni is (bj'() - b(,94I)i/n(1
+ o (1)) and so n" -> (bj 0) - b(.?)i
Ci. The assertionin (3.7) is now easily verified.
Next,a keytheorem in therateofconvergence argument is statedandproved.
The theoremguaranteesthatwithinany subsetof t values whichcontainsa
"substantial"portionoftheobservations, theestimatedregression function
must
be "quiteclose" to thetrueregressionfunction,at least at one point. An as-
sumptionis requiredto theeffect that"enough"observations are takenwithin
each (true)segmentof theregression function.
Suppose Sj is a subset of (zr1)j,Zrj'0), j = 1, ***, r and that with probability
approachingone as n -+ oo, S; e (7_l rj), j] 1, *, r. Let M, denotethe
K(j) x K(j) matrix(1/nES fjh(tn,)fjk(tni)), h, k 1, 2, . . ., K(j). Assumethat
theminimum eigenvalueof M. is 2. + o(l ) as n oo where2i > 0,1 1, =* ..,
(tnlg..., t,") and let p, denote the n-vector(p(e(01;t1) ..*, jj(e?); t,,)) within
(and onlywithin)thisproof.
Thereare n + 1 waysin whichthet71imaybe dividedamongtwo segments.
Considerthekthof thesepartitions:
(ti,, * *, tn,k_1), (tnk, . *, t..). Let _k be
thelinearspacespannedbytheq (_ K(l) + K(2)) n-vectors wheretheithcom-
ponentof thejth vectoris
fij(tn) if i ? k-1 and j < K(l)
O if i > k and j < K(l)
O if i ? k-1 and jIK(l) + h > K(l)
f2h(t,i) if i > k and j _ K(l) + h > K(l)
and let -k+ = ?k [p#(t)]denotethe directsumof thetwovectorspaces.
Let Qk+ denotetheorthogonalprojectiononto ,k+ Qk theorthogonal
projec-
tiononto Wk
Let X (Xnl, * , X,)' and let Pk* be the orthogonal projection of X onto
f1k the closestpointto X in ,k+ subjectto theunderlying
+ continuity
restrictions.This is displayedschematically
in Figure3.3.
XE-Lote
I-lo Ilk Pk k
FIG. 3.3.
Then
IIX - ?l
|jIk* - PkiI = |IX "kli ? lIX- Poll
whichimplies
IX - Poll2- IIPk - Yoll2+ lPSk -
PollX
- 2(Pk* Ok-09 ) +-
OIO IXI-
Il2
|k I
Thus
- ol12 < 2(pk* - PO k - Po) < 21|P* - PolI I I -k PolI
and so
poll < 211Pk* - foll = 21IQk+eII
-
(3.10) llp-k
stepinshowingthat i/
isan important
Thiscomputation ,lI= Op((loglogn)i).
SEGMENTED REGRESSION PROBLEMS 65
1
That Qell= O(l) is a consequenceof Lemma 3.12 and the assumptions
regardingS3,Mj, j = 1, 2. Fromstandardleastsquaresresults
IQell2< S2 v2'M,-1[1? o(1)]v3
=A v3'M3-'v,}(l+ op(l))
where niv1'= (T 10f1(tni)eni, Z
=JflK(l1(tn)en) and where niv2'-
( ==Nl+lf2l(tni)e-nil . .. I 1=N1+lf2K(t2ni)eni) Lemma 3.12 implies that
P{v,'Mylv3? 22} < i-2 tr{M*My}j- where M3* is the K(j) x K(j) matrix
(n 1 z1 fjh(t.i)fjk(tni)), h,k = 1, 2, = 0(1), it
, K(j). Since tr{1M*MNJ1}
follows that v,'Mj'iv, = 0p(l) and so IIQell= 01(1).
It now remainsto show that(C, e)/l
hCh =-O((log logn)l). But thisfollows
as a direct consequence of Lemma 3.13. Thus the proof of the theoremis
complete.
66 PAUL I. FEDER
is suchthattheassumptions
aroundtheeachofthebreakpoints ofLemma3.13 are
then
satisfied,
(3.11) 0 - 6'0? = O(n-'(log log n)').
PROOF.Theorem3.14 impliesthatwithinanysmallneighborhood
ofa center
of observations
thereexistsa tni suchthat
(tni) - po(tnj) = Op(n i(log log n)')
=
a8i
+ 0(1)] (6,-O '?))- Fa8~~~~~~~j+1
+ o(1)1 (6j+1- ?+4).
Equation(3.13) and Theorem3.16 imply
(3.14) (r -rJO)).=Op (n-i(log log n)i).
This is statedformally
in thetheorembelow.
68 PAUL I. FEDER
dj(n) about each of the -rj'?. The intervalsLj(n) are chosen so that dj(n) -> 0
but (n/loglog n)(l/2m)jdj(n)
-+ oo. The term pseudo problem is used because in
practicethe statisticiandoes not know (rz-j') and thus cannot delete such
observations.
Assumethatthet,i aredistributed ina mannerthatimpliesonlyo (n/loglogn)
observations are deletedby thisprocess. (If H(t) is continuousand has finite
slopeat each zj(", thenthiswillbe thecase.) Intuitively,deletingo(n/loglogn)
observations eliminates a percentage of theinformation whichapproacheszero
as n -- oo. Thus,froman asymptotic pointof view,thedeletionof theseobser-
vationsshouldnot affect anyof thedistributiontheory.It will be shownthat
thisis in factthecase.
Moreprecisely, it will be shownthat
-* _ 8'?) = OP(n-e), (* - ,j0')mi = O(n-i)
t9*- 0 = op(n-i) , (T * - TJ0?)mi (7 _3(o))ms
-- - op(n-i)
t) haveat mostm,-1
are satisfied(in thepseudoproblem),andfj(6j'?'; t),fj+1(6;041.;
t-derivatives in commonat t - (0'),then
pointoffj(Oj; t) and
PROOF.Recall equation(3.13): If r, is theintersection
fj+l(60j+l;t)
in 6. For the
is a paraboloidin 6. In particular,it is twice differentiable
remainderof the proofdenotes*(e) by s*(6). Thus,withlargeprobability as
n -0oo
O > S*(O*) - 5*(0(0))
(4.6) {1 z
n ~
p
(a1(')(
~ ~
;)
,I~ a6
t)}i (0* - (0))
_ apc(V);
t1i)e,
n a6
The proofof Lemma3.15 impliesthatthesmallesteigenvalueof thematrixon
the left-handside of equation(4.6) is boundedaway fromzero. Since this
matrixconvergesto G, by continuity G mustbe positivedefinite.Thus
r 0.
Let e - 6(0 denote the translateof e by 6(0); i.e., E -( '0) = to - 0)?; 6E
O}. This amountsto translating
theoriginof theparameter
space to 6(0).
LEMMA 4.8. If (i) 6(0) is a boundarypointof 9, (ii) 601) is approximated9 -
by the convex,positivelyhomogeneousset C, thenthe asymptoticdistribution of
ni(9* - 6(0)) is thatof theclosest(in themetric
determined
bytheasymptotic infor-
mationmatrix,G) pointin C to Z, whereS?'(Z) = _-(0, G-1).
The p.l.s.e. 0* is thatelementof 9 whichminimizess*(f). Let 6*
PROOF.
denotethe parameterwhichminimizes s*(e) withoutregardforthecontinuity
restraintsat the change-over points. Then _2n[ni(P*- 6()] -+ _xV(0,G-'),
whereG is specifiedin thestatement of Lemma4.4.
Recall thatwithintheregionof interestof the parameterspace, s*(e) does
not dependon r and is a paraboloidin 6. It will thusbe convenient forthe
remainder oftheprooftodenotes*(e) bys*(6). The change-over pointsbetween
segmentswill be understoodto lie in L,(n), j = 1, , r - 1 regardlessof whe-
theror not 6 e O.
For all 6 in theneighborhood of 6(0)
-
(4.7) S*(6) s*(6*) + (6* -
8)'Gn(d* - 6)
whereG. is thematrixon theleft-hand sideof equation(4.6). Gnconverges to
G, whereG is givenin equation(4.5). Withlargeprobability 0* is thatelement
of 9) whichminimizes (6* - 8)fG,(8*- 6). It maybe shownby an argument
thatutilizestheconvexity of C that6* r*+ o(n-i) wherer*- (61 mini-
-
forf e
uniformly .8 and Se S ..
LEMMA 4.12. If (i) 6 is well-identifiedat p(O) by t and the componentsof t are
centersof observations,(ii) theconditionsof Lemma 3.13 are satisfied(in boththe
originaland pseudoproblems),(iii) 0 is locallyconvexat 6(0), then
(4.11) ,0(n
PROOF.It followsfromthedefinition
of identification
that0 is also identified
at p,u?byt in thepseudoproblemand thecomponents of t are centersof obser-
vationswithrespectto thepseudoproblem.It followsfromTheorems3.16 and
4.1 that 0 - 00 -_ O(n-i(log log n)f) and * -6(0) = O(n-i(log log n)i).
Selecta. > 0 suchthata,/(log log n)* oo and a, = o((n/n**)i). Let X, -
Equation(4.12) thusimplies
THEOREM 4.15. (i) If m, is odd and D,+ = Dj-, = 1,2, *.., r - 1, then
(-O 1+
(10r O0(n-1)
-t(?) (022- 012 __ 021) _ (022- 012)
00 0()
22 - 12
(O/L(0 11 -621)
(0) (fl(O)
("22 O (0))
-"12)-
2 21
-~~~~~
~(001
FIG. 4.1.
Si dt== 2
-210
and I(t - )4 dt= T16 and offdiagonal
iaon elementsequal to 0.
Lemma 4.8 assertsthat ni(O* - 0(0)) has thesame asymptoticdistributionas the
pointin C closestto Z, wherei29(Z) = _41f(0,
G-1) and closenessis measured
in themetricdetermined by G. Thus
1)} - (Z1+, Z2)
2
whereZ + =_ max(0, Z1) and ;9{Z1, Z2)} = V(O,diag(2, 160)).
The change-over
point,r*, is estimated in theobviousmanner.Thus (?* -
1)2 0* and so
02 - l/2-@ \i
02- -20\
I
1/2 0
.1/2
-4.1
FIG. 4.2.
80 PAUL I. FEDER
Therefore
IT-21 - (T - 4)02(0) = [(6, + 262) - (0,(0) + 02)02())] + Op(n1).
This equationhas two solutionsT+~,
r_, wherer+ >, r_? <4.
Thus
This eqai- 2 -so )] r[(+,(Z) +
hre2(Z))/(1 0 )] '
(A ni_- )] -+ 196[-( + 02)]
1(Z) + 10f2(Z))/(1
l1 72?0 if + _
Thisimplies
T re 4)] 9(t) - where Z
-[=-(Z61
+ Z2)I(1(+ a) if Z+ 2Z2?0)
4 ?
= (Z, + 4Z2)I(1-a) if Z, + 2Z,?0 ?
wherea = 102(1 Similarly,if02(0) <- 1,n(-2
2 1
)]
+ (-t). In both 0
instances69,suitablynormalized, however
normnal;
is asymptotically TiSnot.
SEGMENTED REGRESSION PROBLEMS 81
REFERENCES
R. and ROTH,R. (1969). Curvefitting
[1i BELLMAN, by segmentedstraightlines. J. Amer.
Statist.Assoc.64 1079-1084.
[2] CHERNOFF, H. (1954). On the distributionof the likelihoodratio. Ann.Math. Statist.25
573-578.
[3] CRAMfER, H. (1966). Mathematical Methodsof Statistics.PrincetonUniv. Press.
Processes.Wiley,New York.
[4] DOOB, J. L. (1953). Stochastic
[5] DUNICZ,B. L. (1969). Discontinuities in the surfacestructure of alcohol-watermixtures.
Sond.aus derKolloid-ZeitschriftundZeitschrift furPolymere. 230 346-357.
[6] FARLEY,J. U. and HINICH,M. J. (1970). A testfora shifting in a linear
slopecoefficient
model. J. Amer.Statist.Assoc.65 1320-1329.
[7] FARLEY,J. U., HINICH,M. and McGUIRE,T. W. (1971). Testingfora shiftin the slopes
of a multivariate lineartimeseriesmodel. TechnicalReportWP 77-70-1,Graduate
Schoolof IndustrialAdministration, Carnegie-Mellon Univ.
[8] FEDER,P. I. (1967). On the likelihoodratiostatisticwithapplicationsto brokenline re-
gression.TechnicalReportNo. 23, Department of Statistics, StanfordUniv.
[9] FEDER,P. I. (1975). The log likelihoodratioin segmentedregression.Ann.Statist.3 84-
97.
[10] GREVILLE, T. N. C., ed. (1969). Theoryand Applications of SplineFunctions.Academic
Press,New York.
[11] HINKLEY, D. V. (1969). Inferenceabout the intersection in two-phaseregression.Bio-
metrika 56 495-504.
[12] HINKLEY, D. V. (1969). On the ratioof two correlatednormalrandomvariables. Bio-
metrika 56 635-639.
[13] HINKLEY, D. V. (1970). Inferenceaboutthechange-point in a sequenceofrandomvariables.
Biometrika 57 1-17.
[14] HUDSON, D. J. (1966). Fittingsegmentedcurveswhosejoin pointshave to be estimated.
J. Amer.Statist.Assoc.61 1097-1129.
[15] JENNRICH, R. I. (1969). Asymptotic propertiesofnon-linear leastsquaresestimators.Ann.
Math. Statist.40 633-643.
[16] LOtVE, M. (1960). ProbabilityTheory,(2nded.). Van Nostrand,New York.
[17] POIRER, D. J. (1973). Piecewiseregressionusingcubic splines. J. Amer.Statist.Assoc.68
515-524.
[18] PRATT, J. W. (1959). On a generalconceptof "In probability."Ann.Math. Statist.30
549-558.
[19] QUANDT, R. E. (1958). The estimationof the parametersof a linear regressionsystem
obeyingtwo separateregimes.J. Amer.Statist.Assoc.53 873-880.
[20] QUANDT, R. E. (1960). Testsof the hypothesis thata linearregression systemobeystwo
separateregimes.J. Amer.Statist.Assoc.55 324-330.
[21] ROBISON, D. E. (1964). Estimatesforthe pointsof intersection of two polynomialregres-
sions. J. Amer.Statist.Assoc.59 214-224.
[22] SCHOENBERG, I. J., ed. (1969). Approximations withSpecialEmphasison SplineFunctions.
AcademicPress,New York.
SEGMENTEDREGRESSIONPROBLEMS 83