Reliability of the Scores for the Finger-to-Nose Test in

Research Report
Reliability of the Scores for the Finger-to-Nose Test
in ~ d u l t swith Traumatic Brain Injury
Background and Purpose. The purpose of this study was to determine the intrarater and interrater reliability of measurements of three clinicalfeatures of coordination based o n the perfomnce of the '~nger-tonose"test. Subjects. Thirtyseven persons with traumatic brain injury (26 male, 11 female), aged 17 to G4
years @=29.1, SD=9.9),participated in the study. Mdbocis. Each subject's pHorrnance was videotaped and evaluated for the right and left upper extremities (CIEs)
(two trials each) with respect to thefollowing vanables: time of execution, degree of
dysmetria, and degree of tremor @our-pointordinal ratings). One year later,jive
mpmener1ced
physcal therapists (including the orig2'nal investigator) independently
rated each patient's videotaped pe?ji~nnancein the same manner as described
above. Remlis. Intraclm correlation coeficients (ICC[3,1])for intrarater reliability
were .971 and .986 and ICCs for interrater reliability were ,920and ,913for nght
and lefr UEs, respectively,for the time of execution. A generalized ffippa statistic of
.54 was calcuhted for the scoring of dymtria (both UEs), and Kappa stat&&
calculatedfor the scoring of tremor were .18 and .31for right and left UEs, respectively. Interrater reliability was lowerfor the scoring of these vanables and varied
from .36 to .40for dysmetria andfrom .27 to .26for tremor (nght and left UEs,
respectively). C d u s l o n and Dascusston. These results indicate that physical
tberapkts demonstrate low reliability in assesment of the presence of dysmetria and
~revwrusing videotaped pe&nmnces of thejinger-to-nose test. The results suggest,
b o w e ~that
~ , therapkts reliably measure the time of execution of this test. gthe
limitations associated with therapists' capacityfor objective measurement of subjecby establishment of more deJnitive scartive phenomena cannot be overcome (e,
ing cn'tma for the measures of dysmeria and tremor), tben therapisa should seek
alternative methods of evaluation of UE coordination. [SwaineBR, Sulliuan SJ
Reliability of the scoresfor the finger-tonose test in adults with traumatic brain
injury. Phys Tber. 1333;73:71-78.]
Bonnle R Swalne
S John Sullhran
Key Words: Coordination, Reliability, Tests and measurements, Traumatic brain
injury.
BR Swaine, FT,is a student in the Doctor of Biomedical Sciences Program, L'Ecole d e Readaptstion, Facultk d e Medeane, UniversitC de Montreal, Montreal, Quebec, Canada H3C 3J7. Address all
correspondence to Ms Swaine at Centre de Recherche, Institut d e Readaptation de Montreal, 6300
Darlington Ave, Montreal, Quebec, Canada H3S ZJ4.
SJ Sullivan, PhD, is Associate Professor, Department of Exercise Science, Concordia University, 7141
Sherbrooke St W,Montreal, Quebec, Canada H3G 1MS; Researcher, Centre de Recherche, Institut
d e Readaptation d e Montreal; and Adjunct Professor, L'Ecolc de Readaptation, Facult6 de Medecine,
Universite de Montreal.
This study was approved by the institutional review boards of the Institut de Readaptation de Montreal, the Centre de Readaptation Lucie Bruneau, and the Centre de Rkadaptation Estrie Inc.
This project was Funded in part by the Societe de L'Assurance Automobile du Quebec and the
Fonds de la Recherche en Sante du Quebec. Ms Swaine was supported by the Fonds Pour la Formation de Chercheurs et L'Aide a la Recherche.
This article was submitted January 30, 1 9 2 , and was accepted September 28, 1992.
Physical Therapy /Volume 73, Number 2Pebruar-y 1993
Coordination disturbances are only
one of the clinical manifestations of
the complex motor disorders affecting
persons with traumatic brain injury
PI).Disturbances in coordination
among this population are frequent14
and can severely compromise motor
function. The documentation of coordination disturbances (eg, the inability
to execute smooth, accurate, and
controlled movementss) forms an
integral part of the assessment of
sensorimotor function in the population of patients with TBI.
71 / 17
Clinicians usually assess coordination
disturbances by observation, noting
the quality of movements performed
by the patient. Instrumented, quantitative methods have recently been used
to measure coordination in patient
populations with varied neurologic
disorders. These methods include
tracking tasks,6,7 reciprocal tapping
tasks? and kinematic analysis.9 A traditional and long-standing method of
the measurement of upper-extremity
WE) coordination, however, consists
of the clinician observing the patient's
performance during the "finger-tonose" testlo and noting certain characteristics of the movement such as
time of execution and presence or
absence of dysmetria and tremor. The
time of execution of this test can be
recorded using a stopwatch. Dysmetria is characterized as an impaired
ability to judge the force and range of
a movement; tremor refers to an
involuntary oscillatory movement.5
Information regarding the degree of
dysmetria and tremor is usually
scored using ordinal rating scales and
is heavily dependent on the clinician's
judgment.
Although the protocol for recording
results from the finger-to-nose test
may vary among institutions and individual clinicians, the literature indicates that it is used routinely in the
clinical setting and can perhaps be
identified as the "gold standard" measure of UE coordination.5~10J~
The use
of ordinal rating scales such as those
used in the finger-to-nose test has
recently been questioned, however,
because of their reliability and sensitivity limitations.12
Recently, Mayo et all3 examined the
interrater reliability of several clinical
measures (including the finger-tonose test) from a standard neurophysical evaluation. Four experienced
physical therapists rated the finger-tonose test performances of 18 adults
with neurological impairment (including 16 patients with TBI). The degree
of agreement among raters was found
to be 49.7% for dysmetria and 76%
on the scoring of tremor; each variable was evaluated using a three-point
ordinal rating scale. Kappa statistics
calculated for dysmetria and tremor
were .17 and .08, respectively, indicating "poor" interobserver agreement.
The authors did not, however, examine the reliability of the timed measure of this test, nor did they examine
intrarater reliability. We believe,
therefore, that further research must
be conducted to fully examine both
the intrarater and interrater reliability
of this clinical measure of coordination in persons with TBI.
The purpose of this study was to
determine the intrarater and interrater reliability of the scoring system
used by experienced physical therapists who independently rated the
videotaped performances of the
finger-to-nose test among persons
with TBI.
subjects who consented to be filmed.
Subjects ranged in age from 17 to
64 years @=29.1, SD=9.9). The duration of coma varied considerably
among subjects, ranging from 1 to
135 days @=38.2, SD=32.7), as did
the time elapsed since the subjects'
injury, ranging from 3 to 117 months
@=34.0, SD=30.9).
Raters
Five physical therapists (including the
original investigator [BRS]) participated as raters in this study. All raters,
except the original investigator, were
recruited from a neurology team of a
physical therapy department within a
Montreal rehabilitation center. The
raters, each with between 4 and
15 years @=10.8,SD=4.5) of clinical
experience in the treatmeni of patients who have neurological impairment, were all very familiar with the
application of the finger-to-nose test
in persons with TBI.
Method
Equipment and Procedure
Subjects
Thirty-seven persons with TBI
(26 male, 11 female) participated in
the study. The sample was one of
convenience, with subjects recruited
from three different rehabilitation
centers* on the basis of the diversity
of the levels of severity of sensorimotor deficits. Inclusion criteria were
(1) an adequate UE motor function
(suficient active range of motion and
voluntary movement) to perform the
finger-to-nose test with at least one
UE and (2) sufficient comprehension
of simple verbal commands. The
protocol for this study was approved
by each institution from which the
subjects were recruited (and the data
were collected). Informed consent to
be filmed was obtained from all
subjects.
The data analyzed in this study represent a subset of data (n=37) from a
previously published study14 (n =40).
The present study included only those
'The Institut de Rkadaptation de Montreal, the Ceritre d e Readaptation Lucie Bruneau, and the
Centre d e Readaptation Estrie Inc.
The finger-to-nose test was administered to all subjects during a previous
study conducted approximately 1 year
before the current study.'* At that
time, subjects were instructed to perform the finger-to-nose test in the
typical clinical manner. With his or
her eyes open, each subject began the
test with an arm extended in front of
the body at shoulder level. The subject then flexed the elbow, touched
his or her nose as accurately as possible, and returned the arm to the fully
extended position. The need for both
speed and accuracy of performance
was emphasized to the subjects. Subjects performed the task with each UE
whenever possible (two trials with
each arm). The total time for the
completion of five complete cycles of
movement was recorded using a
stopwatch. In addition, the degree of
both dysmetria and tremor observed
during the movement was recorded
for each LlE. These variables were
rated on separate four-point ordinal
scales using the following criteria:
3=normal performance or the absence of a deficit, 2=slight deficit,
Physical Therapy /Volume 73, Number 2/February 1993
1=moderate deficit, and 0 =severe
deficit.
Each subject's performance (two trials
for each UE) was videotaped while
the investigator administered and
scored the test. The subject was videotaped from a lateral view so that
the rater could clearly observe
whether subjects accurately touched
their finger to their nose. The videotaped performances were then analyzed 1 year later by the original investigator and by four other physical
therapists.
Rater trainlng. No specific training
of the therapists was conducted. Prior
to the viewing of the videotape, however, the raters participated in a brief
orientation session consisting of the
presentation of the evaluation form,
the scoring protocol, and the procedure for the rating session. The scaring system (ordinal rating scales) was
provided for all raters on a printed
sheet. The rating scales were undefined for dysmetria and tremor because they are typically not defined
when using this clinical test. The
variable time of execution (for each
trial), however, was defined for the
raters as the time for the completion
of five complete cycles of movement.
A cycle began with the arm extended
in front of the body at shoulder level,
included the flexion of the elbow so
as to bring the index finger to touch
the nose, and finished with the return
of the arm to the fully extended position. During this session, the therapists were also given time to practice
and become comfortable with the
manipulation of their stopwatches.
lntrarater scoring procedure. The
original investigator scored the subjects' performances (two trials) on the
finger-to-nose test using the previously described procedure on two
different occasions: (1) during the
original data collection and (2) during
the viewing of the videotaped performances 1 year later. This rater did not
review any of the previous scores
collected 1 year earlier, prior to the
viewing of the videotape a year later.
The videotaped performances were
presented in such a way that the order of subjects tested differed from
that in the original data collection.
Therefore, for the analysis of intrarater reliability with 1 year between
tests, there was one examiner.
lnterrater scoring procedure. The
videotape was viewed at the Institut
de Readaptation de Montreal during
two 1-hour sessions. Each subject's
videotaped performance was presented to the group of five raters
(including the original investigator).
The raters scored each subject's performance (two trials for each UE)
simultaneously and independently
and did not consult with each other
during or after the evaluations. The
raters were given the opportunity to
review a particular segment of the
videotaped performance only if necessary. The total time for the completion of five complete cycles of movement and the degree of both
dysmetria and tremor observed during the movement performed by each
UE (two trials for each UE) were
recorded by each therapist using the
scoring system described earlier.
Data Analysis
This study sought to determine the
reliability of measurements obtained
by the "typical" user of the test. It was
assumed that the typical therapist (in
terms of years of clinical experience)
knows how to skillfully score this test;
therefore, this session did not include
a training demonstration with examples of a performance characterizing
each of the four ratings. Likewise,
therapists were not given the opportunity to view any videotaped performances prior to beginning the rating
procedure.
The generalized Kappa statistic was
chosen to determine the intrarater
and interrater reliability of the measurements of dysmetria and tremor.
This statistic, developed by Cohen,l5 is
a coefficient of agreement (among
two or more raters) for categorical
data that corrects for chance agreement. Kappa values less than .OO have
been characterized arbitrarily as indicating "poor" agreement, values between .OO and .20 have been characterized as indicating "slight"
Physical Therapy/Volume 73, Number 2/February 1993
agreement, and those between .21
and .40 are said to represent "fair"
agreement. Kappa values between .41
and .60 have been characterized as
"moderate," those between .61 and
.80 have been characterized as "substantial," and values greater than .81
are said to indicate "almost perfect"
agreement.16
Although Kappa was the preferred
statistic, agreement was also described
as the total percentage of subjects in
which both of a pair of raters agreed
on the score. The agreement among
rater pairs, however, does not indicate the degree of reliability that can
be attained.
In order to establish the intrarater
and interrater agreement for the measurement of "time of execution" (interval data), intraclass correlation
coefficients (ICC[3,1.])were calculated.
This particular model was used with
the assumption that judges are considered fixed effects1' as the selection
procedure of raters did not meet the
requirements of randomness. Tests of
significance using an F test were performed at the .O1 level.
lntrarater agreement. Intrarater
agreement for this study was determined by having the same person
(BRS) measure the same three variables on two different occasions. The
patients' scores obtained by the original investigator were compared with
those recorded 1 year later from the
videotaped performances of the
finger-to-nose test.
lnterrater agreement. In order to
assess the interrater agreement
among the pairs of raters for the
scoring of dysmetria and tremor,
generalized Kappa coefficients were
calculated for each of the 10 possible
pairs of raters (for trials 1 and 2) and
are reported as average values across
all possible pairs.
Results
The distributions of the data from the
finger-to-nose test for the 37 subjects
are shown in Tables 1 and 2. Only
data for right-sided performances
-
Table I. Descriptive Characteristics and Distributions of Data Obtained with the
Finger-to-Nose Test When the Same Rater Evaluated Each Subject (N=37) on Two
Dzyerent Occasionsa
Number of Obsewatlonsb
Varlable
Trial 1
Trlal 2
Dysmetria
O=severe
3=normal
Tremor
(trials 1 and 2) are presented because
the distributions for right- and leftsided performances were very similar.
Reported in Table 1 are the distributions of the combined data obtained
when the same rater (BRS) measured
the same three variables on two different occasions for each of the
37 subjects (maximum of 74 obsemtions). For example, the scores for all
subjects from trial 1 on the first testing occasion (test) have been combined with the scores from trial 1 on
the second testing occasion (retest)
1 year later, and likewise for trial
2. Not all subjects could perform the
finger-to-nose test with their right UE;
therefore, only data from 67 performances are presented.
Time of execution (s)
X
SD
Range
-
"Right-sided performances only.
'percentage of evaluations shown in parentheses.
Table 2.
Descriptive Characteristics and Distributions of Data Obtained with the
Finger-to-Nose Test When Five Raters Evaluated Each Subject (N=37)"
Number of Obsewatlonsb
Varlable
Dysmetria
O=severe
1=moderate
2=slight
3 =normal
Trlal 1
Trlal 2
In general, scores for dysmetria and
tremor did not vary greatly among the
subjects. The majority of subjects
tested were scored as having normal
performances when dysmetria
(>55% of the subjects) and tremor
(92.5% of the subjects) were evaluated. This scoring, however, creates a
distribution problem that makes our
use of the Kappa statistic problematic, because the Kappa corrects for
chance agreement. Mean times of
execution of the finger-to-nose test
were 4.98 and 4.58 seconds for trials
1 and 2, respectively, with the range
of times being similar for both trials.
Furthermore, the distribution of the
scores did not vary over the two
trials. This finding may indicate that
this particular rater consistently
judged subjects as having the same
degree of dysmetria or tremor while
performing each trial, o r that the
subjects' performance was indeed
consistent over the two trials.
Tremor
O=severe
1 =moderate
2=slight
3=normal
Time of execution (s)
X
SD
Range
"Right-sided performances only.
'percentage of evaluations shown in parentheses.
20 / 74
The distribution of scores obtained
when the five raters evaluated each of
the 37 subjects, for a maximum of
185 possible observations, is shown in
Table 2. Different numbers of observations were recorded for individual
variables and trials. These differences
were due to either a subject's inability
to perform both trials of the finger-tonose test with the right UE or a therapist's inability to rate a particular
subject. Although the variability of
Physical Therapy /Volume 73, Number 2Pebruary 1993
1
-
Table 3. Intrarater Reliability Coeficients
Mean
Kappa"
Percentage of
Agreement
Varlable
ICC(3,1)b
Dysmetria
RC
72
Ld
78
Tremor
R
90.5
L
89.5
Tables 3 and 4 present the results of
three indexes of agreement-percentage of agreement, Kappa statistic, and
ICC-for the three variables (ie, degree of dysmetria, degree of tremor,
and time of execution). The values for
these indexes of agreement were
originally calculated with the data
from each trial; however, the values
reported here represent the mean of
the values from both trials for each
UE. For simplicity, data are presented
separately for the intrarater and interrater reliabilities.
Time of execution (s)
lntrarater Agreement
R
L
aMean Kappa represents the mean of the Kappa coefficients calculared for trials 1 and 2.
b~<.005.
'R=right upper extremity.
d ~ = l e f tupper extremity.
scores exceeds that found for the data
obtained by the same rater (Tab. I),
scores for dysmetria and tremor still
did not vary greatly among the subjects. This finding again makes interpretation of values obtained with the
Kappa, a statistic that corrects for
chance agreement, problematic, with
an uneven distribution likely as chance
agreement increases. A greater degree
of variability was found for the scoring
-
Table 4.
of the presence of dysmetria, with only
45.9% (trial 1) and 48.8% (trial 2) of
the subjects being scored as having
normal performances. The majority of
subjects (74.6% and 81.3% for trials
1 and 2, respectively) were scored as
having normal performances when
tremor was evaluated. Consistency
between tnals was again very high.
lnterrater Agreement
Interrater Reliability Coeficients
Varlable
Percentage of
Agreement
Percentage of agreement was 72% or
higher for the scoring of dysmetria
and tremor (Tab. 3). In general,
Kappa values for the scoring of dysmetria were higher than those for
tremor, indicating moderate agreement on the scoring of dysmetria but
only slight to fair agreement on the
scoring of tremor. This result, however, may also be a factor of the differences in distribution for the scores
on the two variables. Both ICCs were
above .95 for time of execution and
were determined to be significant
(PC ,005).
Mean
Kappa"
ICC(3,l)b
Percentage of agreement among the
five raters exceeded 62% (Tab. 4).
Kappa statistics for the scoring of
dysmetria were again slightly higher
than those for tremor but indicated
only fair agreement for both variables.
All ICCs were above .89 for time of
execution and were significant
(PC .005).
Dysmetria
RC
62.9
Ld
68.2
Tremor
R
78.6
L
80.8
Time of execution (s)
R
L
aMean Kappa represents the mean of the Kappa coefficients calculated for trials 1 and 2
b~<.005.
'R=right upper extremity.
d ~ = l e f tupper extremity.
Physical Therapy /Volume 73, Number 2/February 1993
Discussion
Generally, the Kappa coefficients
(agreement statistic) indicated that the
five therapists did not agree well on
the scoring of dysmetria and tremor.
In contrast, agreement was somewhat
higher when the same therapist (intrarater) evaluated these variables on
two different occasions. The greatest
agreement among raters was observed in the assessment of the time
of execution of the finger-to-nose test,
with both intrarater and interrater
reliability being very high. As expected, the scoring of the qualitative
aspects of the performance was more
difficult than the recording of the
quantitative measurement of time. It
should be noted, however, that there
was not an equal representation of all
observations and, because Kappa corrects for chance agreement, our reliability coefficients may be low, reflecting not poor reliabhty but rather the
poor distribution of observations.
These results are similar to those
reported elsewhere.l3 Interrater
agreement (Kappa) for the assessment
of dysmetria and tremor remained
low. Because most observations of
tremor were normal, however, the
Kappa values really d o not accurately
reflect reliability. The agreement of
these measures improved slightly (by
.20 for tremor) when the team of
raters consisted of therapists practicing within the same rehabilitation
setting. In the previous study, both
the raters and the subjects were
drawn from two different rehabilitation centers and one acute care setting. Another difference, perhaps also
contributing to the improved reliability, is the nature of the sample. In the
present study, the sample was restricted to only subjects with TBI and
in particular to those who were in the
later stages of the physical rehabilitation process. It is perhaps noteworthy
that in this study, in which a slightly
higher agreement was obtained, a
four-point rating scale was used (as
opposed to a three-point scale). The
choice of the four-point scale was
predicated by the clinical protocol
used by the therapists participating in
the study.
There may be several reasons for the
slight to moderate reliability (Kappa)
reported for the variables of dysmetria and tremor. A number of causes
of low reliability have been discussed
by Rothstein18 and include those attributable to (1) the person administering the test, (2) the instrument
itself, or (3) the characteristics of the
subject(s) being tested. We will address each of these causes as they
pertain to this study.
First, in this study, the same person
administered the finger-to-nose test to
all 37 subjects, thereby introducing a
constant source of error. Five different raters, however, scored (one aspect of the administration of the test)
each performance. The purpose of
this study was not to determine the
reliability of scores on the finger-tonose test among physical therapists
who had specialized training in its
application and to conclude that expert users of the test can obtain reliable measurements. Instead, its purpose was to determine the reliability
of scores on the finger-to-nose test for
assessing the ability of the "typical"
user of the test. The typical therapist
(in terms of years of experience)
knows how to skillfully perform this
test and frequently uses this measure
of coordination in his or her clinical
practice.
We assumed that the raters from this
study, by virtue of their training and
clinical experience, had the necessary
skills to score the finger-to-nose test.
Perhaps the raters had difficulty noting the deficits from the videotaped
performances (a method that they are
not familiar with), or they misinterpreted the scoring criteria because
the criteria were not sufficiently defined. Furthermore, the interrater
agreement may perhaps be different
among therapists with less clinical
experience in the treatment and evaluation of patients who have TBI. Likewise, the reliability reported in this
article may have been overestimated,
because the five therapists were recruited from the same physical therapy department (sample of convenience). The results might perhaps
have been quite d8erent if the raters
had been selected from a number of
rehabilitation centers.
A second possible cause of the low
reliability relates to the instrument
itself. The finger-to-nose test uses
ordinal rating scales to evaluate the
presence of dysmetria and tremor,
thereby providing an appreciation of
the qualitative nature of the movement performed by the patient. Descriptors such as "moderate deficit,"
however, lack precision and conse-
quently may not be sensitive to small
and perhaps meaningful changes in
patient ~tatus.5~~9
These descriptors are
vague and could, in theory, be based
on more definitive criteria that might
improve the reliability of the scores
on the finger-to-nose test (eg, number
of times the subject misses touching
his or her nose).
A probable cause of the low reliability
in this study (and a limitation) was
restrictions in observed scale variability. The values of the Kappa statistic
depend on the prevalence of the
scores over all subjects.20When the
proportion of observed values is
heavily skewed toward one of the
scale values (Tabs. 1 and 2), Kappa
values can be quite low, even with
little rater error. If the percentage of
agreement is high and Kappa is relatively low, then a restriction in the
distribution of judgments is a likely
explanation for the low reliability. For
example, for the rating of tremor
(Tabs. 3 and 4), low Kappa values
were associated with a high percentage of agreement, indicating that the
raters were observing subjects with
little or no presence of tremor consistently and were in fact agreeing on
that score.
Attempts were made to obtain a sample of subjects with a diversity of
levels of severity of sensorimotor
deficits. We expected our sample to
be similar to those reported in the
literature,21122 in which at least 33% of
the subjects presented "cerebellar
syndromes" and coordination deficits.
As measured by the finger-to-nose
test, however, only a small percentage
(maximum of 6%) of our sample was
judged by the five therapists as having
moderate and severe coordination
deficits. In actuality, the sample may
have been composed of a greater
percentage of subjects with coordination deficits, but the deficits were not
identified using the finger-to-nose
test.
The use of video recordings has provided an alternative method for the
examination of interrater reliability
i~sues.~3.~*
Furthermore, by eliminating the intertest interval (when estab-
Physical Therapy /Volume 73, Number 2Pebruary 1993
lishing intrarater reliability), no
change in the variables to be measured can occur, and the measurement obtained on the second occasion is not influenced by the first
measurement. Such an approach has
allowed the examination of the reliability of the scoring protocol of the
finger-to-nose test-the focus of this
report. This approach, however, does
not necessarily allow generalizations
to the traditional test-retest situation
in which the test is administered and
scored on two occasions. In addition,
the use of videotaping permits a number of therapists to observe and score
a performance without disturbing the
patient. Patients with TBI may not
perform optimally while being observed by a team of raters.
Cllnlcal lmpllcatlons
The resi~ltsof this study provide some
important information regarding the
amount of error associated with the
measurement of coordination using
the finger-to-nose test. Therapists can
now better interpret the measurements associated with the finger-to
nose test in view of the reliability
associated with it. In turn, therapists
may then be able to make more appropriate inferences or judgments
regarding patient progress in coordination performances or treatment
intervention. Therapists should realize, however, that the Kappas reported in this study may be underestimating reliability because of the
distribution problems we discussed.
Therapists should be encouraged in
learning that they can reliably measure the time of execution of the
finger-t:o-nosetest using a stopwatch.
The reliable data obtained with this
method offer therapists, and possibly
third-party payers, important information regarding a particular dimension
of performance (eg, speed of movement) in coordination deficits and its
resolution in patients with TBI. Improvement in this aspect of performance, however, may not necessarily
reflect an improvement in the overall
coordination abilities of a patient,
because the accuracy of performance
may be sacrificed for the increased
speed of movement.
Knowing the limitations of the fingerto-nose test should prompt therapists
to seek more objective and quantitative methods for evaluating coordination performances. Computerized
methods for the measurement of
coordination are currently available to
therapist~,~5326
and the relationship
between this method and the standard clinical finger-to-nose test has
recently been established.14
Suggestions for Further
Research
We suggest that if therapists continue
to use the finger-to-nose test to measure UE coordination among patients
with TBI, efforts should be made to
establish clearer scoring criteria for
the rating of dysmetria and tremor.
The need for accuracy criteria for the
timed measure, for example, is clearly
evident.
The results of this study d o not allow
generalizations to other patient populations, and therapists cannot assume a
similar level of reliability in other
patient groups.18If therapists wish to
know the reliability of data obtained
with the fmger-tonose test among
other patient populations, further
research must be conducted. Additional research with patients who have
TBI is also needed. Before the reliability of scores in patients with TBI can
be understood, a study must be conducted in which the Kappa values are
not influenced by the types of distribution problems we encountered.
Physical therapists appeared to demonstrate low reliability (intrarater and
interrater) for scoring the presence of
dysmetria and tremor during the
videotaped performances of the
finger-to-nose test. The therapists,
however, reliably measured the time
of execution of this test (using a stopwatch). These results suggest a need
to establish more definitive scoring
criteria in order to improve the reliability of these measures. Therapists
Physical Therapy /Volume 73, Number 2February 1993
should be aware of the amount of
error associated with the measurements from the finger-to-nose test and
should be very cautious with interpretations of this test in the population of
patients with TBI.
Acknowledgments
We thank Linda Pezzi for her valuable
assistance during the interrater reliability rating session. We also thank
Helen Corriveau, PT, Maria Carangelo,
PT, Christine Chessex, PT, and Felicia
Guama, PT, for their participation in
the scoring part of this study and
Ginette Gravel for her assistance with
the statistical analysis.
References
1 Jennett B. Some aspects of prognosis after
severe head injury. S c a d J Rehabil Med. 1972;
4:16-20.
2 Rinehart MA. Considerations for functional
training in adults after head injury. Pbys Ther
1983;63:1975-1977.
3 Caronna JJ. The neurologic evaluation. In:
Rosenthal M, Bond MR, Griffith ER, Miller JD,
eds. Rehabilitation ofthe Adult and Child with
Traumatic Brain Injury. Philadelphia, Pa: FA
Davis Co; 1984:5%73.
4 Whyte J, Rosenthal M. Rehabilitation of the
patient with head injury. In: DeLisa JA, ed. Rehabilitation Medicine: Principles and Practice.
Philadelphia, Pa: JB Lippincott Co; 1988:585
611.
5 Schmitz TJ. Coordination assessment. In:
O'Sullivan SB, Schmitz TJ, eds. Physical Rehabilitation: Assessment and Treatment. 2nd ed.
Philadelphia, Pa: FA Davis Co; 1988:121-133.
6 Behbehani K, Kondraske GV, Tintner R, et
al. Evaluation of quantitative measures of upper extremity speed and coordination in
healthy persons and in three patient populations. Arch Phys Med Rehabil. 1990;71:10&111.
7 Jones RD, Donaldson IM. Measurement of
integrated sensory-motor function following
brain damage b y a computerized preview
tracking task. International Journal ofRehabililation Medicine. 1981;3:7143.
8 Tunon A, Fraser C. The use of a simple aiming task to measure recovery following stroke.
Phsiotherapy Practice. 1987;3:117-125.
9 Rosecrance JC, Giuliani CA. Kinematic analysis of lower-limb movement during ergometer
pedaling in hemiplegic and nonhemiplegic
subjects. Phys Ther. 1991;71:334-343,
10 Grinker RR, Sahs AL. Neurology. 6th ed.
Springfield, Ill: Charles C Thomas, Publisher;
1966:4041.
11 Adams RD, Victor M. Princgles of Neurology. 4th ed. New York, NY: McGraw-Hill Inc;
1989.
12 Merbitz C, Morris J, Grip JC. Ordinal scales
and foundations of misinference. Arch Phys
Med Rehabil. 1989;70:308-312.
1 3 Mayo NE, Sullivan SJ, Swaine BR. Observer
variation in assessing neurophysical signs
among patients with head injuries. Am J Phys
Med Rehabil 1991;70:118-123.
14 Swaine BR, Sullivan SJ. The relation between clinical and instrumented measures of
motor coordination in traumatically brain injured persons. Arch Phys Med Rehabil. 1992;73:
55-59.
15 Cohen JA. A coefficient of agreement for
nominal scales. Educational and PJychological
Measurement. 1960;20:3746.
16 Iandis JR,Koch GG. The measurement of
observer agreement for categorical data. Biomerrics. 1977;33:159-174.
17 Shrout PE, Fleiss JL. Intraclass correlations:
uses in assessing rater reliability. PJychol Bull.
1979;86:42@-427.
18 Rothstein JM. ,Measuremat in Physical
Therapy. New York, NY: Churchill Livingstone
Inc; 1985:%14.
19 Sullivan SJ, Swaine BR Objective measures
in physiotherapy: utility and advantages. In:
Proceedings: L'Evaluation Mesurable en Mbdecine de Rbadaptation. Montreal, Quebec, Canada; 1989.
20 Feinstein AR, Cicchetti DV. High agreement
but low kappa, I: the problems of two paradoxes.J Cljn Epidemiol 1990;43:543-549.
21 Held JP, Mazeau M, Rodineau J. Les troubles d e la motilite et du language chez les
traumatises craniens sevkres. Annales de la
Mkdecine Physique. 1975;18:337-354.
22 Cohadon F, Richer E. Evolution et devenir
des comas traumatiques graves. Neurochimrgie. 1983;29:303-325.
can enjoy-ln one volume-the wisdom of those who have
steered the profession and established a pattern of service and
purpose. Includes the lectures from 1964 to 1991 as originally
published in Physical Tlteuppy.
(250@Zg~s,
24af&k.?, 1992)
Ordrr bAPTA-I2
24 I78
M m k . ..................$19.95
Nonmembers: ............ .$27.95
23 Eastlack ME, h i d s o n J, Snyder-Mackler L,
et a]. Interrater reliability of videotaped observational gait-analysis assessments. Phvs Ther.
1991;71:465472.
24 Henderson L, Kennard C, Crawford TJ, rt
al. Scales for rating motor impairment in Parkinson's disease: studies of reliability and convergent validity.J Nertrol Neurosurg Psychiat y . 1991;54:1a24.
25 Kondraske GV, Potvin AR, Tounellotte WW,
Syndulko K. A computer-based system for automated quantification of neurologic Function.
IEEE Trans Biomed Eng 1984;31:401414.
26 Smith SS, Kondraske GV. Computerized
system for quantitative measurement of sensorimotor aspects of human performance. Phys
Thm 1987;67:1860-1866
q
ZIP
State
Membership # and (otegoy
-
@ Check enclosed popble to APTA @ MosterCord @ VIS4
Credit Card#
Physical Therapy/Volume 73, Number 2February 1993