A Sample of Classroom Assessment Instruments Attitude Inventories

A Sample of Classroom Assessment Instruments
Attitude Inventories
Biology
Biology Self-Efficacy Instrument for Nonmajors (BSEIN). The Biology Self-Efficacy Instrument for
Nonmajors is a 23-item scale that asks students to rate on a five-point scale their level of confidence in three
key areas: 1) in writing and critiquing biological ideas through the use of laboratory reports; 2) in general
skills used in a typical biology course (e.g., analyzing data, asking meaningful questions); and 3) in ability to
apply biological concepts to every day life.
Baldwin, J.A., Ebert-May, D., & Burns, D.J. (1999). The development of a college biology self-efficacy
instrument for nonmajors. Science Education, 83, 397-408.
Biology Attitude Scale. The biology attitude scale investigates students attitudes toward biology with 14
Likert-like questions (five points, strongly agree to strongly disagree) and eight "semantic differential" scale
questions. An example of a Likert-like question used is: "Biology is fascinating and fun." An example of a
"semantic differential question used is: "Biology is Cruel A B C D E Kind" (students are asked to circle the
letter that reflects their view of where biology fits relative to the end points).
Russell, J. & Hollander, S. (1975). A biology attitude scale. The American Biology Teacher, 37 (5),
270-273.
Chemistry
Chemistry Attitudes and Experiences Questionnaire (CAEQ). The CAEQ is a closed-ended survey that asks
students to rate their attitudes about chemistry (22 items), their self-efficacy with respect to chemistry (20
items), and their learning experiences in their chemistry courses (35 items). The article below includes the
CAEQ and describes the instrument development and testing.
Coll, R.K., Dalgety, J., & Salter, D. (2002). The development of the Chemistry Attitudes and
Experiences Questionnaire (CAEQ). Chemistry Education: Research and Practice in Europe,
3(1),19-32.
Engineering
Pittsburgh Freshman Engineering Attitudes Survey (PFEAS). The Pittsburgh Freshman Engineering
Attitudes Survey (PFEAS) is a 50 item multiple-choice survey that gathers information about incoming
engineering students' attitudes in the 13 key areas.
URL: http://www.engr.pitt.edu/%7Eoutcomes/#
Besterfield-Sacre, M.E., Atman, C.J., & Shuman, L.J. (1998). Student Attitudes Assessment, Journal of
Engineering Education, 87(2), 133-141.
17
Mathematics
Questionnaire on attitudes about mathematics. This 81-item questionnaire (with 70 closed and 11 open
items) asks students to answer questions regarding elements which contribute to success/failure in
mathematics, perceptions of mathematics, english, and social science (as comparison), how mathematics is
taught, the nature of proofs, reasoning and constructions; motivation; and personal performance. There is no
discussion in the article of how the questionnaire was developed, and how it was tested for
validity/reliability.
Schoenfeld, A.H. (1989). Exploration of students' mathematical beliefs and behavior. Journal for
Research in Mathematics Education, 20 (4), 338-355.
Modified Fennema-Sherman Scale. Elizabeth Fennema and Julia A. Sherman constructed an attitude scale in
the early 1970's that capture students views on mathematics and views on gender and mathematics. . The
scale consists of four subscales (i.e., set of questions that is intended to reflect a concept, like confidence): a
confidence scale, a usefulness scale, a scale that measures mathematics as a male domain and a teacher
perception scale. Each of these scales consists of 12 items. Six of them measure a positive attitude and six
measure a negative attitude. The URL below provides the Fennema-Sherman Scale and an adaptation of the
scale to science.
URL: http://www.woodrow.org/teachers/math/gender/08scale.html
Fennema, E., & Sherman, J. (1978). Sex-related differences in mathematics achievement and related
factors: A further study. Journal for Research in Mathematics Education, 9(3), 189-203.
Physics
Maryland Physics Expectation (MPEX) Survey. The Maryland Physics Expectation (MPEX) survey has been
developed by the Maryland Physics Education Research Group (PERG) as part of a project to study the
attitudes, beliefs, and expectations of students that have an effect on what they learn in an introductory
calculus-based physics course. Students are asked to agree or disagree on a five-point scale with 34
statements about how they see physics and how they think about their work in their physics course.
URL: http://www.physics.umd.edu/rgroups/ripe/perg/expects/mpex.htm
Redish, E.F., Steinberg, R.N., & Saul, J.M. (1998). Student Expectations in Introductory Physics. Am. J.
Phys., 66, 212-224. <http://www.physics.umd.edu/rgroups/ripe/papers/expects/expects1.htm>
Science
Epistemological Beliefs Assessment for Physical Science (EBAPS) Instrument. The EBAPS is a forcedchoice instrument designed to assess students' epistemologies (i.e., their views about the nature of knowledge
and learning) in the physical sciences. EBAPS has 30 items and most students need 15-22 minutes to
complete the EBAPS.
URL: http://www2.physics.umd.edu/~elby/EBAPS/home.htm .
"The Idea behind EBAPS." 12 December 2001.
<http://www2.physics.umd.edu/~elby/EBAPS/idea.htm>.
Views About Sciences Survey (VASS). The Views About Science Survey (VASS) is designed to survey
student views about knowing and learning science, and to assess the relation of these views to student
understanding of science and course achievement (Grades 8-16).
URL: http://modeling.la.asu.edu/R&E/Research.html.
Halloun, Ibrahim. Views About Sciences Survey – VASS P20 Synopsis. 12 December 2001.
http://www.inco.com.lb/halloun/VASS20Snps.PDF
18
Views on Science-Technology-Society (VOSTS). The Views on Science-Technology-Society (VOSTS)
survey that investigates students' perceptions of the social nature of science and how science is conducted.
The VOSTS has 114 multiple-choice questions and the multiple-choice responses are grounded in the
literature and on an extended process of analyzing student comments, creating potential responses, and
verifying those comments with students. (The question in the figure below is illustrative of the kinds of
questions VOSTS asks.) The validity of the survey arises from its development, which is grounded in
student views (Aikenhead & Ryan, 1992).
URL: http://www.usask.ca/education/people/aikenhead/.
Aikenhead, G., & Ryan, A. (1992). The development of a new instrument: 'Views on ScienceTechnology-Society' (VOSTS). Science Education, 76, 477-492.
Scientific Attitude Inventory II. The SAI II asks students to rate their level of agreement on a five-point scale
to 40 statements about the nature of science and potential attitudes about science, both positive and negative.
An example statement is: "The laws and/or theories of science are approximations of truth and are subject to
change."
Moore, R.W. and Hill Foy, R.L. (1997). The scientific attitude inventory: A revision (SAI II).
Journal of Research in Science Teaching, 34(4), 327-336.
Modified Nature of Scientific Knowledge Scale (MNSKS). The Nature of Scientific Knowledge Scale
(NSKS) was revised to create the MNSKS scale. The MNSKS asked students to rate their level of
agreement on a five-point scale to 32 statements about four dimensions of nature of science (creativity of
science, development of science, testability of science, and unity of science). As an example, one
statement from the MNSKS is "A scientific theory and a work of art are similar in that they both
represent creative work of people." The MNSKS does not appear in the article.
Meichtry, Y. (1992). Influencing student understanding of the nature of science: Data from a case of
curriculum development. Journal of Research in Science Teaching, 29(4), 389-407
Views on the Nature of Science Questionnaire (Forms A-C). The VNOS-A, B, or C ask students write their
responses to a series of between 6 and 10 questions (depending upon the form). An example question is, "Is
there a difference between a scientific theory and a scientific law? Illustrate your answer with an example."
Abd-El-Khalick, F., Lederman, N.G., Bell, R.L., & Schwartz, R. (2001). Views of nature of science
questionnaire (VNOS): Toward valid and meaningful assessment of learners’ conceptions of nature
of science. Paper presented at the annual meeting of the Association for the Education of Teachers in
Science, Costa Mesa, CA
URL: http://www.flaguide.org/tools/diagnostic/views_of_nature_questionnaire.php
Statistics
Survey of Attitudes Toward Statistics (SATS). The Survey of Attitudes Toward Statistics (SATS) contains 28
seven-point Likert items designed to assess four components of students' attitudes toward statistics.
URL: http://www.unm.edu/~cschau/infopage.htm.
Schau, C., Stevens, J., Dauphinee, T., & Del Vecchio, A. (1995). The development and validation of the
Survey of Attitudes Toward Statistics. Educational & Psychological Measurement, 55 (5), 868-876.
Writing
Views about Writing Survey (VAWS). The VAWS is 26-item survey that asks students to rate on an eightpoint scale their attitudes and beliefs about writing and composition. The structure is based on the
"contrasting alternative design" developed for to measure students' attitudes towards physics (Halloun &
Hestenes, 1996). The paper that presents VAWS data does not describe testing or validation of the VAWS
instrument (Rhoads et al., 1998). The authors do indicate that they will conduct validation via focus group
interviews with students. The instrument was used to measure change in attitudes toward writing of students
19
enrolled in English class that was geared for students in an engineering learning community (50 students)
and of students in a traditional English class (155 students).
Rhoads, T.R., Duerden, S.J., & Garland, J. (1998) Views about Writing Survey - A New Writing
Attitudinal Survey Applied to Engineering Students. Paper presented at the Frontiers In Education
Conference, Tempe, AZ.
Halloun, I., & Hestenes, D. (1996). Views about Sciences Survey: VASS. Paper presented at the Annual
Meeting of the National Association for Research in Science Teaching, St. Louis, MO.
http://fie.engrng.pitt.edu/fie98/papers/1220.pdf
Classroom Environment
Learning Context Questionnaire. This 50-item questionnaire investigates students' level of agreement with
statements describing student's view of himself/herself and his/her education (e.g., "I enjoy courses most
when they are more theoretical.").
Griffith, J. V. & Chapman, D. W. (1982). LCQ: Learning context questionnaire. Davidson, North
Carolina: Davidson College.
Constructivist Learning Environment Questionnaire (CLES). The CLES is a 42-item questionnaire that asks
students to rate on a five-point scale where and how they learn science; the extent to which their classroom
environment enables them to present their own views, manage their own learning, and talk with other
students about science; and their interest and motivation in the classroom.
URL: http://www.letus.northwestern.edu/msta/surveys/source-documents/cles.html
Tayler, P.C., Fraser, B.J., and White, L.R. (1994). CLES: An instrument for monitoring the development
of constructivest learning environments. Paper presented at the annual meeting of the American
Educational Research Association, New Orleans. Available at:
http://www.letus.northwestern.edu/msta/surveys/source-documents/index.html
Learning Approach Questionnaire (LAQ). This 49-item questionnaire was developed to measure students’
“preferences of using learning strategies that are consistent with either rote or meaningful learning
orientations” (Bretz, 1994). Students are asked to rate their level of agreement/disagreement to 44
statements, and the remaining 5 items ask demographic information. An example item is: “I try to relate new
material, as I am reading it, to what I already know on that topic.”
Bretz, S.L. (1994). Learning strategies and their influence upon students’ conceptions of science
literacy and meaningful learning: the case of a college chemistry course for nonscience majors.
Dissertation: Cornell University.
Metacognition
Metacognition Awareness Inventory (MAI). The Metacognition Awareness Inventory (MAI) is a 52-item test
intended to measure the metacognitive awareness. Students are asked to rate statements on a 7 point Likertlike scale (1=not at all true of me, to 7=Very true of me). The MAI includes two factors: knowledge about
cognition and regulation of cognition.
Shraw, G., & Dennison, R. S. (1994). Assessing metacognitive awareness. Contemporary Educational
Psychology, 19, 460-475.
Motivated Strategies for Learning Questionnaire (MSLQ). The Motivated Strategies for Learning
Questionnaire (MSLQ) is an 81-item self-report instrument that has two sections: a motivation section (31
items) and a learning strategies section (50 items). Students are asked to rate statements on a 7 point Likertlike scale (1=not at all true of me, to 7=Very true of me). The learning strategies scales incorporate the
20
metacognition construct and seek to measure rehearsal, elaboration, organization, critical thinking,
metacognitive self-regulation, effort regulation, peer learining, and help seeking.
Pintrich P., Smith D., Garcia T., & McKeachie W. (1991) A Manual for the Use of the Motivated
Strategies for Learning Questionnaire (Technical Report 91-B-004). The Regents of The University
of Michigan.
Duncan, T.G., & McKeachie, W.J. (2005). The making of the Motivated Strategies for Learning
Questionnaire. Educational Psychologist, 40(2), 117-128.
Motivation
Achievement Goal Questionnaire (ACG). The Achievement Goal Questionnaire (ACG) has 12 items that
look at motivation. Three items assess mastery-approach, three items assess mastery avoidance, three items
measure performance approach, and three items measure performance avoidance. Students are asked to rate
statements on a 7 point Likert-like scale (1=not at all true of me, to 7=Very true of me).
Elliot, A., & McGregor, H.A. (2001). A 2x2 achievement goal framework. Journal of Personality and
Social Psychology, 80, 501-519.
Concept Maps
Biology
Using concept maps to assess declarative knowledge. The authors of the paper below used concept maps to
assess lower-order knowledge and compared the concept maps against multiple-choice tests, finding high
reliability and inter-rater agreement. Students were provided with a small number of terms (5-7) and had to
link them in appropriate ways. For example, to "replace" a multiple-choice item which asked, "Which of the
following is not a type of protozoan? A) paramecium; b) amoeba; c) Euglena; d) bacteria" (d is correct), a
student would be provided with the terms (kingdom, Protozoa, Monera, Euglena, paramecium, amoeba,
bacteria) and asked to draw a concept map using these terms. Three scoring rubrics were used to measure
understanding and had high inter-rater agreement. Scoring rubric A assigned a "0" if students used all
pertinent answers and the stem (question) on their map (and assigned a "-1" otherwise). Scoring rubric B
assessed whether students linked relationships to the appropriate kingdoms in their map: a "+1" was given if
they correctly linked bacteria to kingdom Monera (and not Protozoa); a "–1" was given if they incorrectly
did so; and a "0" was given if the terms were not on the map. Finally, scoring rubric C was used to assess
misinformation: a "–1" was given if the student linked anything incorrectly to the kingdoms, and a "0" was
given otherwise.
Rice, D.C., Ryan, J.M., & Samson, S.M. (1998). Using concept maps to assess student learning in the
science classroom: must different methods compete? Journal of Research in Science Teaching,
35(10), 1103-1127.
Using a concept mapping to identify students' misconceptions in biology. This article presents numerous
student-generated concept maps and the misconceptions that they reveal.
Lavoie, D.E. (1997). Using a modified concept mapping strategy to identify students' alternative scientific
understandings of biology. Paper presentation at the 1997 Annual Meeting of the National Association
for Research in Science Teaching, Chicago, Illinois. Accessible at:
http://www.educ.sfu.ca/narstsite/conference/97conference/lavoie.pdf
Concept maps as a diagnostic tool. The researchers in the study below investigated the sensitivity of
student-generated concept maps as methods for documenting changes in student understanding. The
researchers used an experimental pre/post design. Students in an elementary science methods course were
randomly divided into two groups. Both sets took a "Life Zones in the Ocean" multiple-choice exam and
21
constructed concept maps on the topic at the start of the experiment. The experimental group received 45
minutes of computer instruction on the marine life zones; the control group received similar instruction on an
unrelated topic. Following this, both sets took a "Life Zones in the Ocean" multiple-choice exam and
constructed concept maps on the topic at the start of the experiment. The researchers found that the
experimental students' concept maps were much more detailed than the control students. Example studentgenerated concept maps are provided, as is an explanation of the scoring rubric used.
Wallace, J.D., & Mintzes, J.J. (1990). The concept map as a research tool: exploring conceptual change
in biology. Journal of Research in Science Teaching, 27 (10), 1033-1052.
Concept Tests
Astronomy
ConcepTests for Astronomy. A library of ConcepTests for astronomy is available at the URL below.
Users/contributers need a login id and password and can access this library by contacting Paul Green
([email protected]).
URL: http://hea-www.harvard.edu/~pgreen/educ/ConcepTests.html
Green, P. (2002, to be published). Peer Instruction for Astronomy. Prentice Hall: Upper Saddle River,
NJ
Chemistry
Journal of Chemical Education On-Line Library of Conceptual Questions (CQs). The American Chemical
Society's Division of Chemical Education offers a library of Conceptual Questions (or CQs) in their Journal
of Chemical Education OnLine that would be useful as ConcepTest questions. The site offers questions in a
variety of topics, techniques for creating CQs, and accepts submissions of CQs.
URL: http://jchemed.chem.wisc.edu/JCEWWW/Features/CQandChP/CQs/CQIntro.html
Chemistry ConcepTests. The ConcepTests website offers a range of conceptest questions on a variety of
topics. Readers can download sets of questions on various chemistry topics in Adobe Acrobat (*.pdf) and
Microsoft Word (*.doc) format.
URL: http://www.chem.wisc.edu/~concept/
Landis, C.R., Ellis, A.B., Lisensky, G.C., Lorenz, J.K., Meeker, K., Wamser, C.C. Chemistry
ConcepTests: A Pathway to Interactive Classrooms, Prentice Hall, Inc., in preparation. Preprints of
this booklet are available; email [email protected] for more information.
ConcepTests for General Chemistry. This site offers additional conceptests (without answers) for classroom
use. URL: http://people.brandeis.edu/~herzfeld/conceptests.html
Mathematics
Better file cabinet. This site has a searchable database of questions covered in calculus classes.
URL: http://betterfilecabinet.com/
22
Physics
Project Galileo: Your gateway to innovations in science education. The Project Galileo site has a variety of
teaching materials that are based on research in science education. The site, maintained by Harvard
Professor Eric Mazur and colleagues, includes an extensive set of resources for teaching with ConcepTests in
astronomy, biology, chemistry, and physics. Physics ConcepTests are stored in a database, from which users
can select sets of ConcepTests for use in their classes.
URL: http://galileo.harvard.edu
Mazur, E. (1997) Peer Instruction: A User's Manual; Prentice Hall: Upper Saddle River, NJ.
Conceptual Diagnostic Tests
Astronomy
Astronomy Diagnostic Test (ADT). The ADT consists for 33 questions, with the first 21 questions collecting
the content portion, and the final 12 questions collecting demographic information. The test can be accessed
at the URL below. This site also offers comparative data by institution to the ADT question and links to
research articles on the development of the ADT.
URL: http://solar.physics.montana.edu/aae/adt/
Lunar Phases Concept Inventory. The LPCI was developed based upon interviews with 14 undergraduate
students around their conceptual understanding of lunar phases. The LPCI consist of 14 multiple-choice
items.
Lindell, R.S. (2001). Enhancing College Students' Understanding of Lunar Phases. Unpublished doctoral
dissertation, University of Nebraska – Lincoln.
Biology
Conceptual Inventory of Natural Selection (CINS). The Conceptual Inventory of Natural Selection (CINS)
was developed to identify student’s knowledge about natural selection, as well as their misconceptions. The
20 item multiple choice asks students to reflect on scenarios regarding Galapagos finches, Venezualan
guppies, and Canary Island lizards, and select options that best reflect what an evolutionary biologist would
select.
Anderson, D.L., Fisher, K.M., & Norman, G.J. (2002). Development and evaluation of the conceptual
inventory of natural selection. Journal of Research in Science Teaching, 39, 952-978. Available on-line at:
http://scires.educ.msu.edu/EnvironmentalLiteracy/sgnatsel/conc_inv_natsel.pdf .
URL: http://www.biologylessons.sdsu.edu/CINS6_03.pdf
Chemistry
Chemical Concepts Inventory. The Chemical Concepts Inventory (CCI) can be used to identify chemistry
misconceptions held by students. The inventory is a multiple-choice instrument composed of nonmathematical conceptual questions (22 questions total). The questions are based on commonly-observed
student misconceptions about topics generally covered in the first semester of a college chemistry course. A
question from the inventory appears in the figure below.
URL:
http://jchemed.chem.wisc.edu/JCEWWW/Features/CQandChP/CQs/ConceptsInventory/CCIIntro.ht
ml
Mulford, D.R. (1996). An Inventory for Measuring College Students' Level Of Misconceptions in First
Semester Chemistry. Unpublished doctoral dissertation, Purdue University. Available at:
http://faculty.pepperdine.edu/dmulford/thesis/title.html
23
Earth Science
Test on students' alternative conceptions of earth and space. This article lists the common misconceptions as
identified by a survey on students' alternative conceptions of earth and space. The survey was described as
having 18 multiple-choice questions, but was not listed in the article.
Schoon, K.J. (1992). Students' alternative conceptions of earth and space. Journal of Geological
Education, 40, 209-214.
Engineering
The Foundation Coalition has a variety of concept inventories spanning a wide variety of areas in
engineering: chemistry, circuits, computer engineering, dynamics, electromagnetics, electronics, fluid
mechanics, heat transfer, materials, material strength, signals and systems, and thermodynamics.
Information about these instruments can be accessed at:
http://www.foundationcoalition.org/home/keycomponents/assessment_evaluation.html
Environmental Science
Test on the Greenhouse Effect, Ozone Depletion, and Acid Rain. This diagnostic test asks students to report
"yes," "no," or "don't know" to a set of 29 research-based misconception questions about the greenhouse
effect, ozone depletion, and acid rain. The survey was adopted from a questionnaire developed by Jane
Dove (1996).
Khalid, T. (1999). The Study of Pre-Service Teachers' Alternative Conceptions Regarding Three
Ecological Issues. Paper presented at the Annual Meeting of the National Association for Research
in Science Teaching, Boston, Massachusetts.
<http://www.educ.sfu.ca/narstsite/conference/khalid/khalid.html>
Dove, J. (1996). Student teacher understanding of the greenhouse effect, ozone layer depletion and acid
rain. Environmental Education Research. 2,89-100.
Mathematics
Basic Skills Diagnostic Test (BSDT). The Basic Skills Diagnostic Test (BSDT) is a 24-item free-response
test on basic mathematics understanding. Eleven questions (two with two parts) assess pre-algebra level
(division, multiplication and ordering of numbers represented in fractions and in decimals), concept of area
and volume, proportional reasoning. Eight questions assess basic algebra (concept of a variable, order of
operations, simplifying expressions, particularly inappropriate cancellation, linear equations), and 3
questions assess intermediate level, concept of a function, graph of a function, concept of a logarithm.
The tool can be sent to colleges and universities by contacting the author, Jerome Epstein ((718) 2603572, [email protected]).
Epstein, J. "What is the real level of our students? or What do diagnostic tests really measure?" to be
published.
Physics
Force Concept Inventory (FCI) Instrument. The Force Concept Inventory (FCI) instrument is designed to
assess student understanding of the most basic concepts in Newtonian physics. This forced-choice
instrument has 30 questions and looks at six areas of understanding: kinematics, Newton's First, Second, and
Third Laws, the superposition principle, and types of forces (such as gravitation, friction). Each question
offers only one correct Newtonian solution, with common-sense distractors (incorrect possible answers) that
are based upon student's misconceptions about that topic, gained from interviews.
URL: http://modeling.asu.edu/R&E/Research.html.
Hestenes, D., Wells, M., & Swackhamer, G. (1992). Force Concept Inventory. The Physics Teacher, 30,
141-158.
24
The Mechanics Baseline Test (MBT). The Mechanics Baseline Test (MBT) instrument is an advanced
companion to the Force Concept Inventory (FCI). The forced-choice MBT has 26 questions that were based
upon interviews with students about their misconceptions on basic topics in Newtonian mechanics. The test
covers concepts in kinematics (linear and curvilinear motion), basic principles (Newtons' First, Second, and
Third Laws, superposition principle, energy conservation, impulse-momentum, and work) and special forces
(gravity and friction).
URL: http://modeling.asu.edu/R&E/Research.html.
Hestenes, D. & Wells, M. (1992). A Mechanics Baseline Test. The Physics Teacher, 30, 159-165.
Conceptual Survey of Electricity and Magnetism (CSEM). The 32-item multiple-choice CSEM was
developed to assess students' knowledge about electricity and magnetism.
Maloney, D.P., O'Kuma, T.L., Hieggelke, C.J., & Heuvelen, A.V. (2001). Surveying students'
conceptual knowledge of electricity and magnetism. American Journal of Physics, 69 (S1), S12-S23.
Other diagnostic tests. Physicists are quite active in the development of diagnostic tests. A large number
can be found at the North Carolina State University's Physics Education Research group at
http://www.ncsu.edu/per/ and specifically at their test instruments page at www.ncsu.edu/per/TestInfo.html.
There are many other Physics Education Research groups around the country.
Known misconceptions in physics. Additionally, the "Student Difficulties in Physics Information Center"
lists common misconceptions students have in areas of physics, as well as the research articles that have
investigated/uncovered those misconceptions.
URL: http://www.physics.montana.edu/physed/misconceptions/index.html
Reasoning
Classroom Test of Scientific Reasoning (revised). 12-item test asks about conservation of weight/volume,
proportional thinking, probabilistic/proportional thinking, combinatorial thinking, and correlational thinking.
Scores on test classify students into three groups: empirical/inductive reasoners, transitional reasoners, and
hypothetical/deductive reasoners.
Contact Anthony Lawson directly for tool at: [email protected]
Lawson, A.E. (1987a). Classroom test of scientific reasoning: revised pencil-paper edition. Arizona State
University.
Interviews
Biology
Interviews of students' conceptions of genetics. The article provides an interview protocol, as well as
excerpts from interviews, of the concepts of inheritance, genetic structure, genetic processes, and extension.
Venville, G.J., & Treagust, D.F. (1998). Exploring conceptual change in genetics using a
multidimensional interpretive framework. Journal of Research in Science Teaching, 35 (9), 10311055.
25
Chemistry
Interviews of students' conceptions of equilibrium and fundamental thermodynamics. Article provides a set
of interview questions surrounding a chemical reaction and the codes used to score those interviews.
Interviews provided information on the type and extent of student misconceptions regarding the first and
second law of thermodynamics, among others.
Thomas, P.L. & Schwenz, R.W. (1998). College physical chemistry students' conceptions of
equilibrium and fundamental thermodynamics. Journal of Research in Science Teaching, 35 (10),
1151-1160.
Performance Assessment
Critical Thinking
Performance Assessments of Critical Thinking. The article below describes the development and testing of
four critical thinking performance tasks, and provides a rubric for scoring the tasks. The rubric used had high
inter-rater reliability, with 95% of papers having exact or close agreement on scores between raters.
Correlations of the tasks against various closed-ended critical thinking tests were low (0.25-0.30); however,
high school teachers who used the tasks found them to provide good information on students' skills.
Christen, R., Angermeyer, J., Davison, M., & Anderson, K. (1994). Performance assessment of critical
thinking: the reflective judgment approach. Research/Practice, 2(1)
Scoring Rubrics
Rubrics for Oral Presentations
Rubric for Oral Presentation. The scoring rubric is an adaptation of several evaluation instruments that have
been demonstrated strong inter-rater reliability and validity (Carlson and Smith-Howell, 1995).
Carlson, R.E., & Smith-Howell, D. (1995). Classroom Public Speaking Assessment: Reliability and
Validity of Selected Evaluation Instruments. Communication Education, 44, 87-97.
Quigley, B.L. (1998) Designing and Grading Oral Communication Assignments. New Directions for
Teaching and Learning, 74, 41-49.
Rubrics for Written Presentations
"Presentation Sandwich" Checklist. The "Presentation Sandwich" checklist is the mnemonic used in an
introductory engineering course to describe the presentation requirements for all technical work. The
checklist has two major sections: Section I concerns the Presentation Sandwich itself and Section II concerns
Graphical Work. Section I has three sub-sections (A, B, and C), one concerning problem context, one
concerning the technical work (i.e., the "sandwich filling"), and one concerning the discussion. Students who
do not meet all of the yes/no are required to re-submit their work. Space is left for students demonstrating
exemplary work. The checklist is provided to students prior to their assignment.
McNeill, B., Bellamy, L., & Burrows, V. (1999). A Quality Based Assessment Process for Student
Work Products. ASEE Journal of Engineering Education, 88(4).
http://ceaspub.eas.asu.edu/mcneill/word_documents/papers/using_checklists_v10.doc
Scoring Rubric for Evaluating Writing.
Smith, K.A. (1998) Grading Cooperative Projects. New Directions for Teaching and Learning, 74, 41-49.
Scoring Rubric for an Assignment which Examines a Consumer Product. The rubric shown here involves an
assessment of student reports on a product (like a consumer report). These tools were originally developed
by pharmacy faculty at Eastern Illinois University to develop and grade students' reports about a health-care
product.
26
Hobson, E.H. (1998) Designing and Grading Written Assignments. New Directions for Teaching and
Learning, 74, 51-57.
Scoring Rubric for Evaluating a Persuasive Argument Paper. This rubric evaluates a paper that is intended
to present a persuasive argument on some topic. The rubric is divided into four categories (organization,
content, usage, and mechanics), each of which is weighted to reflect the relative importance of that category.
Johnson, D.W., & Johnson, R.T. (1996). Meaningful and Manageable Assessment Through Cooperative
Learning. Edina, MN: Interaction Book Company.
Computer Modeling Checklist. The "Computer Modeling" checklist is used for computer modeling
assignments. This checklist has five major sections. Section I concerns the correctness of the model. Section
II defines what is expected when charts are present. Section III states how Spreadsheet models must be
presented. Section IV is concerned with the written responses given to the questions posed in the assignment.
McNeill, B., Bellamy, L. & Burrows, V. (1999). A Quality Based Assessment Process for Student Work
Products. ASEE Journal of Engineering Education, 88(4).
http://ceaspub.eas.asu.edu/mcneill/word_documents/papers/using_checklists_v10.doc
Scoring rubrics for projects that model a real life solution. Charles Emenaker (University of Cincinnati,
Raymond Walters College) outlines two scoring rubrics for projects which require students to model a reallife situation mathematically, to interpret and to present their results. One rubric sums the score from four
categories: problem understanding, problem plan, problem solution and problem presentation. The other
rubric uses a "holistic" four-point scoring system and can be used for problems that require a less-detailed
assessment.
Emenaker, Charles E. Assessing modeling projects in calculus and pre-calculus: Two approaches.
Supporting Assessment in Undergraduate Mathematics website. Accessed on 4/3/02 at
http://www.maa.org/SAUM/maanotes49/116.html
Rubrics for Problem Solving
Rubrics for Assessing Skill in Addressing Open-Ended Problems. This rubric for the assessment of students'
skills in addressing open-ended problems is founded on empirical data on the developmental process in
problem solving. This four-step process is outlined in Figure 1 below. Essentially, these steps are: Step 1 -Identify the problem, relevant information, and uncertainties; Step 2 – Explore interpretations and
connections; Step 3 – Prioritize alternatives and communicate conclusions; and, Step 4 – Integrate, monitor,
and refine strategies for re-addressing the problem (Lynch and Walcott, 2001). The rubric measures
students' skill level against each of the four steps outlined above.
URL: http://www.wolcottlynch.com/index_files/page0002.html
Lynch, C.L. & Wolcott, S.K. (2001). Helping your students Develop Critical Thinking skills. IDEA
Paper, 37. [On-line]. Available: http://www.idea.ksu.edu/papers/Idea_Paper_37.pdf
Scoring Rubric for Evaluating an Inquiry Project. This rubric can be used to evaluate an inquiry project:
namely, a project in which students formulate and pursue questions using multiple sources of information
and multiple investigative methods. This rubric is based on the research on the nature of inquiry in the
learning of professional practitioners and is intended to give weight to the critical processes of investigation
while preventing the uncontrolled factors of working in a real-world environment from counting against the
student.
Busching, B. (1998). Grading Inquiry Projects. New Directions for Teaching and Learning, 74, 89-96.
27
Scoring Rubric for Writing in a Mathematics Class. Annalisa Crannell (at Franklin & Marshall College) has
developed writing assignments and a rubric for grading them in her mathematics classes.
Rubric URL: http://server1.fandm.edu/departments/Mathematics/writing_in_math/checklist.html
Assignment URL:
http://www.fandm.edu/DEPARTMENTS/MATHEMATICS/writing_in_math/writing_index.html
Crannell, Annalisa. (1994). How to Grade 300 Mathematical Essays and Survive to tell the Tale.
PRIMUS, 4(3), 193-201.
Rubrics for evaluating questions students ask. This rubric evaluates the number and type of questions
students pose in thinking about a problem-based case study situation. Students were asked to read a
scenario and then to write down every question they had about the scenario. The set of questions students
posed were evaluated along three dimensions: Orientation (i.e., whether question is descriptive, or whether
asks about potential solutions/treatments) Relation to Case Study (i.e., whether question could be answered
by the case study); and Complexity (i.e., whether question involved a response that involved higher-order
thinking). A metric was assigned to the questions that enabled a numeric grading of the set of questions.
Dori, Y.J., & Herscovitz, O. (1999). Question-posing capability as an alternative evaluation method:
analysis of an environmental case study. Journal of Research in Science Teaching, 36(4), 411-430.
Rubrics for Discussions
Scoring Rubric for Evaluating a Discussion Conducted via a Listserv. This rubric evaluates a list-serv
discussion in four areas: mechanics (grammar), participation (five postings required per week), content, and
critical thinking. One can evaluate all the posts or a random sample of posts. The selected postings are rated
on the four constructs and then summed for a total score. An instructor can also create two scores from the
rubric: the sum of mechanics and participation describes the students' use of technology and the sum of the
content and critical-thinking ratings describes the students' content understanding.
Morrison, G.R., & Ross, S.M. (1998) Evaluating Technology-Based Processes and Products. New
Directions for Teaching and Learning, 74, 69-77.
Scoring Rubric for Whole-Class Participation. This rubric is a holistic scoring rubric to be used in classes
that use a combination of whole-class discussion with occasional small group work. A holistic scoring rubric
rates student skills, attitudes, knowledge, and behavior in one score, rather than as a sum of different scores
measuring specific competencies.
Bean, J.C. & Peterson, D. (1998). Grading Classroom Participation. New Directions for Teaching and
Learning, 74, 33-40.
Rubrics for Evaluating a Portfolio
Scoring Rubric for Evaluating a Portfolio. These two rubrics were developed in a course as a negotiation
between the students (pre-service elementary education students) and the faculty, and were based upon the
recent literature on how to grade proposals. The rubrics emphasize in particular students' growth over time,
and attempt to take into consideration contextual factors that students may have faced in the class (which
involved work in their field). In addition, having the students grade themselves promoted their ability to
reflect on their own learning, a life long skill.
Scanlan, P.A. & Ford, M.P. (1998). Grading Student Performance in Real-World Settings. New
Directions for Teaching and Learning, 74, 97-105.
28
29
30
Recommended Reading List
(Submitted by Conference Speakers)
Alters & Nelson. 2002. Teaching Evolution in Higher Education. Evolution 56:1891-1901
Black, P. & William, D. (1998). Assessment and classroom learning. Assessment in Education: Principles,
policy, and practice, 5(1), 7-74.
Bransford, J.D., Brown, A.L., Cocking, R.R. Eds. (1999). How People Learn: Brain, Mind, Experience, and
School. Washington, D.C.: National Academy Press. Available on line at: http://www.nap.edu
Bruer, J. T. (1993). Schools for Thought. Cambridge, MA: MIT Press.
Cameron, W. & Chudler, E. (2003) A role for neuroscientists in engaging young minds. Nature Reviews
Neuroscience, 4: 1-6.
DeHaan, R.L. (2005) The Impending Revolution in Undergraduate Science Education. Journal of Science
Education and Technology, 14 (2): 253-269.
Donovan, M. S., Bransford, Pellegrino, Eds. (2001). How People Learn: Bridging Research and Practice,
Washington, D.C.: National Academy Press.
Available on line at: http://www.nap.edu
Gabel, D.L., Ed. (1994). Handbook of Research on Science Teaching and Learning. New York: MacMillan.
Klymkowsky, Garvin-Doxas & Zeilik. 2003. Bioliteracy and Teaching Efficacy: What biologists can learn from
physicists. Cell Biology Education, 2:151-163
Lottridge, S.M. & Seymour, E. (1999). The Student Assessment of Learning Gains Website.
[http://www.wcer.wisc.edu/salgains/instructor/default.asp] [An on-line course evaluation website]
McClymer & Knoles. 1992. Ersatz Learning, Inauthentic Testing. Journal of Excellence in College Teaching
3:33-50.
National Institute for Science Education, College Level 1 Team. (2002). The Field-tested Learning
Assessment Guide Website. [http://www.flaguide.net] [This site provides introduction to assessment,
self-instructional modules on different assessment techniques, and links to assessment instrument].
Seymour, E. (2001). Tracking the processes of change in US undergraduate education in science,
mathematics, engineering, and technology. Science Education, 86, 79-105. [This article outlines the
changing views on assessment and teaching and learning in science, math and technology.]
Skrabanek. 1986. Demarcation of the Absurd. The Lancet No. 8487, pp. 960-61.
Springer, L, Stone, M.E., & Donovan, S. (1997). Effects of Small-Group Learning on Undergraduates in
Science, Mathematics, Engineering, and Technology: A Meta-Analysis (NISE Research Monograph
No. 11). University of Wisconsin, Madison: National Institute for Science Education [This is a
meta-analysis on the impact of small groups on learning]
31
White, R. and Gunstone, R. (1992). Probing Understanding. Falmer Press. Contains a chapter on the POE
(Predict, Observe, Explain) Framework.
Approaches to Biology Teaching (with key words and web sites)
Allen, D. & Tanner, K. (2004). Approaches to cell biology teaching: Mapping the journey—Concept maps
as signposts of developing knowledge structures. Cell Biology Education, 2, 133-136.
Available at: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=192451
Baines, A. T., McVey, M., Rybarczyk, B,, Thompson, J.R., & Wilkins, H. R. Mystery of the toxic flea dip:
An interactive approach to teaching aerobic cellular respiration. Cell Biology Education, 3, 62-65.
... Diagnosing secondary students’ misconceptions of photosynthesis and respiration ... Teaching
Available at:
http://scholar.google.com/scholar?hl=en&lr=&q=cache:Y1u5ROf4xDwJ:cellbioed.org/articles/vol3no1/pdfs/
03-06-0022.pdf+biology+%26+misconceptions+%26+teaching+OR+learning+%22research%22
Dancy, M. H. & Beichner, R. J. (2002). But Are They Learning? Getting Started in Classroom Evaluation.
Cell Biology Education, 1, 87-94.
Available at: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=128540&tools=bot
Engle, D. (2004). The biological justification for transforming our teaching. Cell Biology Education,
3(4),233-234. Review of Zull, James E. The art of changing the brain: Enriching the practice of
teaching by exploring the biology of learning. (Stylus Publishing. ISBN: 1-5792-2054-1.
http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=533124&tools=bot
Feig, A. (2004). Challenge your teaching. Nature Structural & Molecular Biology. 11(1), 16-19.
... ingrained learning patterns or misconceptions distilled from ... the allied departments biology, chemistry,
mathematics ... learning resulting from teaching innovation ...
Available at: www.anahuac.mx/medicina/archivos/dra_anafuentes.pdf
Hall, A. C. & Harrington, M.E. (2003). Experimental Methods in Neuroscience: An Undergraduate
Neuroscience Laboratory Course for Teaching Ethical Issues, Laboratory Techniques, Experimental
Design, and Analysis. Journal of Undergraduate Neuroscience Education, 2(1), A1-A7.
Available at: http://www.funjournal.org/results.asp?juneid=73
Kitchen, E., Bell, J.D., Reeve, S. & Sudweeks, R.R. & Bradshaw, W.S.(2003). Teaching cell biology in the
large-enrollment classroom: Methods to promote analytical thinking and assessment of their
effectiveness. Cell Biology Education2, 180-194.
Available at:
http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&list_uids=14506506&dopt=Ab
stract
Khodor, J. Halme, D. G., & Walker G. C. (2004). A Hierarchical Biology Concept Framework: A Tool for
Course Design, Cell Biology Education, 3, 111-121.
... erroneous answers diagnose the misconceptions held by ... effectiveness of different
teaching methods ( Hake ... cellular, and developmental biology” ( Klymkowsky et ...
Klymkowsky, M.W., Garvin-Doxas, K., & Zeilik, M. (2003). Bioliteracy and Teaching Efficacy: What
Biologists Can Learn from Physicists, Cell Biology Education, 2, 155-161.
32
Available at:
http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&list_uids=14506504&dopt=Ab
stract
Markwell, J. (2004). The human side of science education. Biochemistry and Molecular Biology Education,
32, 323-325.
... has permitted an evolution of teaching style ... material, attempt to uncover misconceptions, and
orchestrate ...
Morse, M. P. (2003). Preparing biologists for the 21st century. BioScience, 53 (1), 10.
... students enter our classrooms with misconceptions about the ... is deeper when students learn biology by
acting ... agrees that the most effective teaching strategy is …
Staub, N.L. (2002). Teaching evolutionary mechanisms. Genetic drift and M&Ms, Bioscience 52(4), 373377.
... Teaching Evolutionary Mechanisms: Genetic Drift and M&M's. ... to correct such
misconceptions about evolution ...
Sundberg, M.D. (2002). Assessing student learning. Cell Biology Education, 1, 11-15.
... once had two large sections of nonmajor biology students (more ... categories ranging
from common misconceptions about the ... learning to our classroom teaching as we ...
Available at: http://www.pubmedcentral.gov/articlerender.fcgi?tool=pubmed&pubmedid=12587030
Tanner, K., Chatman, L.S. & Allen, D. (2003). Approaches to cell biology teaching: Cooperative learning in
the science classroom – Beyond students working in groups. Cell Biology Education, 2, 1-5.
Available at: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=152788&tools=bot
Tanner, K. & Allen, D. (2004). Approaches to biology teaching and learning: From assays to assessments—On
collecting evidence in science teaching. Cell Biology Education, 3, 69-74.
Tanner, K. & Allen, D. (2005). Moving from knowing facts toward deep understanding through conceptual
change. Cell Biology Education, 4(2), 112-117.
... a longitudinal study of conceptual change in biology. ... tests to evaluate students'
misconceptions in science ... Handbook of Research on Science Teaching and Learning ...
Available at: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1103711#N0x88e6dc0.0x840c45c
Tidon, R. & Lewontin, R. C. (2004) Teaching evolutionary biology. Geneics and Molecular Biology, 27 (1),
124-131.
… Issues related to biological misconceptions, curriculum and didactic material are
discussed, and some proposals are presented with the objective of aiding .... (2004).
33
Cooperative/Collaborative/Group Learning:
L. D. Fink. (2002) Beyond small Groups: Harnessing the extraordinary power of learning teams. In L. K.
Michaelsen, A. B. Knight, & L.D. Find (Eds). Team-Based Learning: A Transformative Use of Small
Groups, Westport, CT: Praeger Publishers.
Johnson, D. W., Johnson, R. T., and Smith, K. A. (1998). Active learning: Cooperation in the college classroom,
2nd ed. Edina, MN: Interaction Book Company.
Millis, Barbara J. and Cottell, Philip G., Jr. (1998). Cooperative Learning for Higher Education Faculty.
American Council on Education. Oryx Press; Phoenix, AZ.
(For higher education faculty in all areas. The authors have used cooperative learning in diverse settings.
Good balance of research results and practical application.)
Springer, L., Stanne, M.E., Donovan, S. S.(1999). Effects of Small-Group Learning on Undergraduates in
Science, Mathematics, Engineering, and Technology: A Meta-Analysis. Review of Educational Research, 69
(1), 21–51.
(Examines 39 high-quality studies from 1980 and later in STEM areas. Finds overall, robust gains for
achievement, persistence, and attitudes).
Misconceptions:
Bodner, G.M. (1991). I have found you an argument, Journal of Chemical Education, 68 (5), 385-388.
Gabel, D. (1998). The complexity of chemistry and implications for teaching. In B.J. Fraser and K.G. Tobin
(eds). International Handbook of Science Education, Kluwer, Academic Publishers, Great Britian, 233-248.
Hestenes, D. (1995) What do graduate oral exams tell us?, American.Journal of Physics. 63, 1069.
Lewis, E. L., & Linn, M. C. Heat energy and temperature concepts of adolescents, naïve adults, and experts:
Implications for curricular improvements. Journal of Research in Science Teaching, 31(6), 657-677 (1994).
Mulford, D.R.; Robinson, W.R., (in press) An inventory for misconceptions in first-semester general chemistry.
Journal of Chemical Education. Submitted (2001). Materials on:
http://faculty.pepperdine.edu/dmulford/thesis/title.html
Osborne, R. & Cosgrove, M. (1983). Common misconceptions among college science students. Journal of
Research in College Science Teaching, 20, 825-830.
Smith, K.J. and Metz, P.A. (1996). Evaluating student understanding in solution chemistry through microscopic
representations, Journal of Chemical Education, 73(3), 233-235.
Students’ Alternative Frameworks and Science Education, 4th Edition, H. Pfundt and R. Duit, (IPN Reports-in-Brief,
Kiel Germany, 1994). Available on-line at: http://www.ipn.uni-kiel.de/aktuell/stcse/stcse.html
Misconceptions Web Sites:
http;//www.oise.utoronto.ca/~science/miscon.html
http://www.physics.montana.edu/phyed/misconceptions
http://www.ipn.uni-kiel.de/aktuell/stcse/stcse.html
http://faculty.pepperdine.edu/dmulford/thesis/title.html
34
Other Useful Web Sites:
FLAG (Field-tested Learning Assessment Guilde: http://flagguilde.org Web site contains assessment
primer, CATs (Classroom Assessment Techniques, Matching Goals to Assessments, Resources, etc.
SALG (Student Assessment of Learning Gains): http://www.flaguide.org/cat/salg/salg1.php or
http://www.wcer.wisc.edu/salgains/instructor/ Web site that contains a powerful tool that allows assessment
of any/all components of classroom/laboratory practices’ impact on learning. Instructors create their own
versions of the SALG from template and can save over them for multiple courses over multiple semesters.
Students take SALG on-line and data is returned to faculty as raw or statistically analyzed data.
Bibliography STCSE (Students' and Teachers' Conceptions and Science Education). Documents research on
teaching and learning science. Role students' and teachers' conceptions in the teaching and learning process
given particular attention. http://www.ipn.uni-kiel.de/aktuell/stcse/stcse.html Searchable and downloadable
in html or Endnote.
http://www.first2.org/resources/wrkshopmaterials/workshop_table.html
Richard Hake’s web site: http://www.physics.indiana.edu/~sdi/
Rich Felder’s web site: http://www2.ncsu.edu/unity/lockers/users/f/felder/public/
1999-2000 SUCCEED Faculty Survey of Teaching Practices and Perceptions of Institutional Attitudes
toward Teaching. SUCCEED Coalition Report, December 2001. View full report (95 pages) or executive
summary (9 pages). Reports frequencies of use of active and cooperative learning, learning objectives,
writing assignments, and faculty development services, and perceptions of the value ascribed to teaching
quality by faculty, administrators, and the faculty reward system. Survey respondents were 586 engineering
faculty members at 8 institutions.
http://www2.ncsu.edu:8080/unity/lockers/users/f/felder/public/#WhatsNew
Project Galileo has ConcepTests that have been field-tested. http://galileo.harvard.edu/
Joe Redish’s web site: http://www.pysics.umd.edu/perg/papers/redish/cogsci.html
ABC's of teaching with excellence: A compendium of suggestions for teaching with excellence," B.G. Davis,
L. Wood, R.C. Wilson; online at: http://teaching.berkeley.edu/compendium including the Minute Paper
description.
PKAL (Project Kaleidoscope: http://www.pkal.org - [email protected]) has an on-line publication on "What
Works, What Matters and What Lasts." The goal of this collection of stories and essays is to capture, analyze
and disseminate lessons learned from the experience of leading agents of change: institutional, organizational
and individual so to inform the work of the larger community.
Teaching Large Lecture Classes, TA Professional Development, Cognitive Apprenticeship Model, etc. with
downloadable files: http://groups.physics.umn.edu/physed/
More on Bloom’s Taxonomy: http://www.kcmetro.cc.mo.us/longview/ctac/blooms.html
35
Web Sites With Scientific Teaching Examples:
Group problem-solving in lecture
www.ibscore.org/courses.htm
http://yucca.uoregon.edu/wb/index.html
http://mazur-www.harvard.edu/education/educationmenu.php
Problem-based learning
www.udel.edu/pbl/
www.microbelibrary.org
www.ncsu.edu/per/scaleup.html
http://webphysics.iupui.edu/jitt/jitt.html
Case studies
www.bioquest.org/lifelines/
http://ublib.buffalo.edu/libraries/projects/cases.case.html
http://brighamrad.harvard.edu/education/online/tcd/tcd.html
Inquiry-based labs
http://chemistry.Beloit.edu
http://mc2.cchem.berkeley.edu
www.plantpath.wisc.edu/fac/joh/bbtl.htm
www.bioquest.org/
http://biology.dbs.umt.edu/biol101/default.htm
http://campus.murraystate.edu/academic/faculty/terry.derting/ccli/cclihomepage.html
Interactive computer learning
www.bioquest.org/
www.dnai.org
http://evangelion.mit.edu/802TEAL3D/
http://ctools.msu.edu/
General Papers:
D. Ebert-May et al., Bioscience 47, 601 (1997).
P. Laws, Phys. Today 44, 24 (1991).
D. Udovic et al., Bioscience 52, 272 (2002).
J. C.Wright et al., J. Chem. Educ.75,986 (1998).
J. Trempy et al. Microbiol. Educ. 3, 26 (2002).
L. Springer et al., Rev. Educ. Res. 69, 21 (1999).
Bonwell, Charles C. & Eison, James A. (1991). Active learning: Creating excitement in the classroom. ASHEERIC Higher Education Reports; Report No. 1. Washington, D.C.: The George Washington University,
School of Education and Human Development.
Hake, R.R. (1998a). "Interactive-engagement vs traditional methods: A six-thousand-student survey of
mechanics test data for introductory physics courses," American Journal of Physics 66, 64-74 (1998); online
at <http://www.physics.indiana.edu/~sdi/>.
Handelsman, J. et al., Biology Brought to Life: A Guide to Teaching Students How to Think Like Scientists.
McGraw-Hill, New York, 1997.
36
Hestenes, D. (1979). Wherefore a science of teaching?, The Physics Teacher, 17, 235-242.
Karplus, R. (1977). Science Teaching and the Development of Reasoning, Journal of Research in Science
Teaching 14(2): 169-175. (Currently online at <http://www.kluweronline.com/issn/1059-0145/contents> as
an Appendix to Fuller (2003). Also in Fuller (2002), pages 70-76.
Lawson, A. E., Abraham, M. R,. & Renner, J.W. (1989). A theory of instruction: Using the learning cycle to
teach science concepts and thinking skills. National Association for Research in Science Teaching (NARST)
Monograph, Number one. ED324204.
Lemke, J. (1990). Talking Science - Language, Learning, and Values. Norwood NJ: Ablex publishers.
Learning Cycle: Schneider, L.S. & J.W. Renner (1980). Journal of Research in Science Teaching, 17 (6), 503517.
McDermott, L.C. (1991). Millikan Lecture 1990: What we teach and what is learned–Closing the gap, American
Journal of Physics, 59, 301-315.
Musheno,B., Lawson,A. B. (1999). Effects of learning cycle and traditional text on comprehension of science
concepts by students at differing reasoning levels. Journal of Research in Science Teaching, 36(1), 23-27.
Robinson, W. R. (1999). A view of the science education research literature: Student understanding of chemical
change. Journal of Chemical Education, 76(3), 297-298.
Smith, E., Blakeslee, T, & Anderson, C. (1993). Teaching strategies associated with conceptual change learning
in science," Journal of Research in Science Teaching 30, 111-126.
Strike, K. A. & Posner, G. J. (1992). A revisionist theory of conceptual change. In R. A. Duschl & R. J.
Hamilton (Eds.), Philosophy of science, cognitive psychology, and educational theory and practice (pp. 147176). Albany, NY: State University of New York Press.
Wyckoff, S. (2001). Changing the culture of undergraduate science teaching, Journal of College Science
Teaching, February: 306-312.
Scholarship of Teaching and Learning
Boyer, E.L. (1990). Scholarship Reconsidered: Priorities of the Professoriate. Princeton, NJ: Carnegie
Foundation for the Advancement of Teaching.
Cambridge, B. (December 1999). The Scholarship of teaching and Learning: Questions and Answers From the
Field. AAHE Bulletin 52(4): 7-10.
Carnegie Academy for the Scholarship of Teaching and Learning. (1999). Informational Program. Booklet.
Menlo Park, CA: Carnegie Foundation for the Advancement of Teaching.
Glassick, C.E., M.T. Huber, and G.I. Maeroff. (1997). Scholarship Assessed: Evaluation of the Professoriate.
Special Report of The Carnegie Foundation for the Advancement of Teaching. San Francisco: Jossey-Bass.
Huber, M.T. (July/August 2001). Balancing Acts: Designing Careers Around the Scholarship of Teaching.
Change 33(4): 21-29.
37
Hutchings, P. & Shulman, L.S. (September/October 1999). The Scholarship of Teaching: New Elaborations,
New Developments. Change, 31(5): 10-15.
Shulman, L.S. 2000. From Minsk to Pinsk: Why a Scholarship of Teaching and Learning. The Journal of
Teaching and Learning (JoSoTL) 1(1); online at: http://www.iusb.edu/~josotl/search_archives.ht
Quotable Quotes
“Science not as a noun…but as a process, a set of activities, a way of proceeding and thinking.”
(Tinker & Thornton, 1992, p. 155).
"There is no neutral pedagogical practice. Every single one is based on a given conception of the learning
process and of the object of such a process. Most probably, those practices much more than the methods
themselves are exerting the greatest lasting effects in the domain of literacy, as in any field of knowledge."
(Ferreiro, Emilia, 1991).
Elementary and high school students perceive science largely as a passive process of observing and recording
events. They view good scientists as ones who make careful observations and keep complete and accurate
records of all they observe.”
(Carey, Evans, Honda, Jay & Unger, 1989; Songer & Linn, 1991).
‘The structure of our institutions conveys a more powerful message than their contents."
(Jay Lemke)
38