1.1 Course structure p. 1
1.2 Course objectives
p. 1
1.3 Learning
objectives p. 2
1.4 Assessment tools p.
2
2.1 Methods p. 3
2.2
Results p. 3
Table 1. Results of the entrance and
exit tests, broken down by topic p.
4
Figure 1. Entrance
and exit test results, by topic p.
5
Figure 2. Frequency histogram of the changes from entrance to exit test
p. 6
2.3 Discussion p. 7
3.1 Methods . p. 7
3.2
Results . p. 8
Table 2. Relative rankings of course
components.. p. 9
3.3
Discussion p. 9
Table 3. Average student course
evaluation scores, Clare and Bio. 101 p.
10
This report presents an assessment of the classroom component of Clare 102, Inquiry in the Natural World, as taught in Spring semester of 2000. The laboratory component will be assessed separately. The course was team-taught by the following individuals:
Class (lecture) component: |
|
Laboratory component: |
Dr. Walter Budzinski |
|
Dr. Paul Joireman |
Dr. David DiMattio |
|
Prof. Jim Miller |
Dr. Ted Georgian |
|
Prof. Kevin Vogel** |
Dr. George Lapennas |
|
|
Dr. Larry Wier* |
|
|
* Lecture coordinator ** Lab coordinator
1.1 Course
structure:
1.11
Class (lecture) component:
3 credit hours. The class
met three times each week, 50 minutes per meeting.
There were 13 topics, each covered in one week.
The first class meeting for each topic occurred as a single, large
lecture for all 5 sections. Various
instructors presented these lectures (see list of topics and presenters,
Appendix 1). The second class
meeting used a variety of formats, including cooperative group exercises, to
review the material covered in the large lecture and the assigned textbook
reading. The third class was
dedicated primarily to student discussion of assigned readings.
There were 4 tests given during the semester; all sections of the course
used identical tests.
1.12 Laboratory component: 1 credit hour. Laboratory groups meet once a week for 110 minutes. Laboratory exercises were different than other laboratory courses because the laboratory procedures were rarely be specified. Instead, groups of either two or four students were given a problem, and expected to figure out how to solve it. The laboratory experience was designed to be one of discovery rather than recapitulating worn out "follow the instructions" labs. Students were asked to proceed in the same way that scientists make new discoveries, by making observations, forming, testing, and evaluating hypotheses. Grades were based on in-lab evaluations (65%), a written paper based on a student-initiated investigation (25%) and class participation (10%).
Syllabi for the class and laboratory parts of the course are in Appendix 2.
The objectives approved by the Faculty Senate for Clare 102 (numbered for clarity) are:
The laboratory component of the course has its own objectives:
1. Students will understand that science is an objective process that allows them to distinguish between adequate and inadequate explanations of natural phenomena and that scientific inquiry can be applied to everyday phenomena that are often taken for granted.
2. Students will be able to perform a complete scientific investigation of an appropriate, testable, and measurable phenomenon of interest.
3.
Students will be able to apply the appropriate statistical test to a data
set and correctly interpret the result.
Specific learning objectives for each week’s topics were given to the students. Class exercises and tests were based directly on these objectives. The learning objectives for Spring 2000 are listed in Appendix 3. When analyzed from the perspective of the three course objectives, 53% of the learning objectives related most closely to the first objective (to introduce the mode of inquiry of the natural sciences), 1% to the second (to enable students to understand and apply basic investigatory skills in a problem solving context), and 46% to the third (to examine a sample of fundamental discoveries of the natural sciences).
How well did the tests reflect the course objectives? Fifty-two% of the points on the semester’s tests related most closely to the first objective, 4% to the second, and 44 to the third. The distribution of test questions closely matched the learning objectives, and supported objectives 1 and 3 to approximately equal degrees. Objective 2 was not well represented in learning objectives or test questions, but this is because it was intended to be satisfied primarily by means of the laboratory component of the course.
Two formal assessment tools were used:
(1)
Matched entrance/exit tests were administered to assess both the level of
understanding of students beginning the course and the extent to which student
understanding was improved by the course. The
questions were designed by the instructors to reflect general knowledge of
science as a process and as a body of historical results.
A definite attempt was made not to
tailor test questions to the specific topics and ideas which are covered in the
course, but rather to test fundamental scientific knowledge that we might hope
any educated person would possess. Copies
of the entrance and exit tests are in Appendix 4.
(2)
A detailed student evaluation instrument was used at the end of the
semester to assess levels of student satisfaction and to what extend students
believed that the course objectives had been met.
The instrument is shown in Appendix 4.
In addition to
these tools, both the class and laboratory teaching teams met weekly during the
semester to evaluate the success of the previous week’s exercises and to plan
for the following week. These team
meetings gave instructors a chance to discuss student reactions to the course
and helped give us a sense of the effectiveness of each course component.
2.1 Methods:
The entrance test was given
during the first week of classes and students were told that their scores would
not affect their class grades. Perhaps
as a result, quite a few students left answers blank.
The exit test used identical questions and was administered for credit as
part of the final examination. Students
were not told that that the questions were from the entrance test or that they
would be used for course assessment as well as for grading students.
Two of the 25 questions from the entrance test (Multiple choice, # 8 and
12) were not used on the exit test because they were judged by the teaching team
to include material that hadn’t been covered in the course.
The tests did not cover the 13 topics in the course at all evenly:
three topics (#6, 11 and 12) were not tested at all and three others (#4,
8 and 9) were tested by only one question.
A few students missed either the entrance or exit test (<10 in each
case); their tests were not included in this analysis. Results of all the other tests (n=128) were entered into a
spreadsheet, with correct answers coded as “1”, wrong answers as “0” and
answers left blank by a blank. It’s
not clear how to handle answers left blank.
It seems reasonable to assume that students left blank the answers they
were least sure of, so that blanks should be treated as wrong answers.
This approach would probably underestimate student understanding at the
beginning of the course and exaggerate the degree of improvement caused by the
course. Blanks were treated both
ways in this analysis and the results (presented below) indicate that the
difference was small.
2.2
Results:
Average scores for each question and for each topic are shown in Table 1.
The same results are shown graphically in Figure 1.
Student performance on the entry test was quite poor.
When blanks were counted as wrong, the overall score was 35.3%, ranging
from a low of 7% on the questions about biology and evolution (topics #8 and 9)
to a high of 56.0% for the scientific method (topic 1).
Not counting blanks increased the entry scores to 40.7%, still an
amazingly low figure for such general science knowledge.
Scores on the exit test improved dramatically, to 66.6% when blanks were
counted and the same when they weren’t. The
degree of improvement was somewhat higher (31.3%) when blanks were counted than
when they weren’t (26.3%).
A frequency histogram of improvement by individual students is shown in
Figure 2. As can be seen, most
students improved by a substantial amount; only 8 of 128 showed an improvement
of less than 10%. It is equally
true that few students (15) showed very high improvements (> 50%), presumably
because the better students scored so well on the entrance test that high levels
of improvement were impossible. The
overall improvement averaged 31.3%, with a standard deviation of 15.54.
Improvement was highly significant statistically:
the chance that the level of improvement observed could happen by chance,
without a significant effect from taking the course, is 1.0 x 10-46 (one-tailed
t-test, df=127).
Table
1. Results of the entrance and exit
tests, broken down by topic. TF =
true/false, MC = multiple choice questions.
Refer to Appendix 4 for the actual questions.
|
|
|
Entrance |
Test |
|
Exit
|
Test |
|
|
# |
Topic |
Syllabus
title |
Quest.
# |
Average |
|
Quest.
# |
Average |
|
Change |
1 |
Nature
of sci. inv. |
How
do we find out about the |
TF
#1 |
79 |
|
TF
#1 |
80 |
|
1 |
|
|
world? |
MC
#1 |
33 |
|
MC
#1 |
52 |
|
19 |
|
|
|
average: |
56.0 |
|
|
66.0 |
|
10.0 |
|
|
|
|
|
|
|
|
|
|
2 |
Astronomy |
Where
are we in the Universe? |
MC
#2 |
44 |
|
MC
#2 |
49 |
|
5 |
|
|
|
MC
#4 |
28 |
|
MC
#4 |
77 |
|
49 |
|
|
|
average: |
36.0 |
|
|
63.0 |
|
27.0 |
|
|
|
|
|
|
|
|
|
|
3 |
Mechanics |
Why
do heavy things fall and |
TF
#2 |
62 |
|
TF
#2 |
72 |
|
10 |
|
|
planets revolve? |
MC
#3 |
13 |
|
MC
#3 |
40 |
|
27 |
|
|
|
average: |
37.5 |
|
|
56.0 |
|
18.5 |
|
|
|
|
|
|
|
|
|
|
4 |
Thermodynamics |
What
is energy? |
MC
#5 |
17 |
|
MC
#5 |
65 |
|
48 |
|
|
|
|
|
|
|
|
|
|
5 |
Chemistry
I |
What
is matter? |
TF
#3 |
64 |
|
TF
#3 |
79 |
|
15 |
|
|
|
MC
#6 |
43 |
|
MC
#6 |
86 |
|
43 |
|
|
|
MC
#7 |
42 |
|
MC
#7 |
73 |
|
31 |
|
|
|
MC
#8 |
34 |
|
--- |
|
|
|
|
|
|
average: |
45.8 |
|
|
79.3 |
|
29.7 |
|
|
|
|
|
|
|
|
|
|
6 |
Chemistry
II |
What
are atoms? |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
Quantum
mech. |
What
is light and what are |
MC
#9 |
37 |
|
MC
#8 |
96 |
|
59 |
|
|
electrons? |
MC
#10 |
30 |
|
MC
#9 |
85 |
|
55 |
|
|
|
average: |
33.5 |
|
|
90.5 |
|
57.0 |
|
|
|
|
|
|
|
|
|
|
8 |
Intro.
To Biology |
What
is life? |
MC
#12 |
7 |
|
--- |
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
Evolution |
How
do living things evolve? |
MC
#13 |
7 |
|
MC
#11 |
16 |
|
9 |
|
|
|
|
|
|
|
|
|
|
10 |
Genetics |
Why
do we (usually) resemble |
MC
#14 |
32 |
|
MC
#12 |
65 |
|
33 |
|
|
our parents? |
MC
#15 |
36 |
|
MC
#13 |
77 |
|
41 |
|
|
|
MC
#16 |
45 |
|
MC
#14 |
66 |
|
21 |
|
|
|
MC
#17 |
15 |
|
MC
#15 |
52 |
|
37 |
|
|
|
MC
#18 |
39 |
|
MC
#16 |
62 |
|
23 |
|
|
|
average: |
33.4 |
|
|
64.4 |
|
31.0 |
|
|
|
|
|
|
|
|
|
|
11 |
Respir./photosyn. |
How
does life use energy? |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
Global
warming |
How
is Earth's climate regulated? |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
Cosmology |
How
do we view the universe |
TF
#5 |
55 |
|
TF
#4 |
82 |
|
27 |
|
|
now? |
MC
#19 |
29 |
|
MC
#10 |
93 |
|
64 |
|
|
|
MC
#20 |
26 |
|
--- |
|
|
|
|
|
|
average: |
36.7 |
|
|
87.5 |
|
45.5 |
Figure 1. Entrance and exit test results, by topic. Topics 6, 11 and 12 were not tested. Topics 4, 8 and 9 were tested with only one question, Topics 1-3, 7 with 2 questions, Topic 13 with 3 questions, Topic 5 with 4 questions, and Topic 10 with 5 questions.
Figure 2.
Frequency histogram of the changes from entrance to exit test for
individual students. The columns are coded as follows:
Column 1 = Change < -10%
Column 6
= Change 30 to 40%
Column 2 = Change -10 to 0%
Column 7
= Change 40 to 50%
Column 3 = Change 0 to 10%
Column 8
= Change 50 to 60%
Column 4 = Change 10 to 20%
Column 9
= Change 60 to 70%
Column 5 = Change
20 to 30% Column
10 = Change
70 to 80%
2.3
Discussion:
Although the test results show a high level of improvement, there are
some limitations that should be noted:
(1)
The assessment tests are
not independent of the course. To
some extent they test the topics and even the specific facts covered in Clare
102. This possible bias was
worsened by the decision of the teaching team to eliminate two questions from
the exit test that addressed material that hadn’t been covered in class.
A more valid assessment could be attained if Clare 102 students were
given standardized entrance/exit tests from some external source such as a
national testing agency. This approach would also allow us to compare our students’
exit performance with external benchmarks.
(2)
The test questions do not
adequately test the first two course objectives.
Of the 25 questions used in the entrance test, only two focus primarily
on the process of inquiry in the sciences; two others test the history of
science, and none test students’ ability to apply investigatory skills. It is simply easier to write content-oriented test questions.
Also, the results for topics #1, 8 and 9 are based on only one question
and cannot be considered adequate assessments of how much students learned about
these topics. In general, more
questions are needed for each topic. Also,
the use of essays should be considered. Most
of the methodological or historical concepts in the course were tested during
the semester with essays rather than objective questions. However, grading essays would add substantially to the
effort required for course assessment.
(3)
The exit results, while
showing definite improvement, could be higher.
It would be useful to compare them to results from other general science
courses or even other Clare courses.
The teaching team is continuing to discuss which topics should be
included in the course and the degree of detail in each topic.
3.1 Methods:
A course evaluation was given to
students during the final exam period. The
evaluation uses the standard system of ratings on a scale of 1-5, with 1 =
strongly disagree, 3 = neutral and 5 = strongly agree.
The evaluation instrument (copy in Appendix 5) is modeled closely after
the one used in the Biology Department. We
did add three items related to Clare 102’s objectives:
# 3.
I have a better understanding of the scientific process after taking this
course.
#12. This course helped me
understand how scientists answer questions about the natural
world.
#13. This course introduced
me to many of the fundamental discoveries of the natural
sciences.
We were concerned
that students may not respond carefully to each item in course evaluations,
tending instead to give high scores to all items if pleased with the course and
low scores if not. To offset this
possibility, we added a section (“Aids to Learning”) that asked students to rank
the various components of the course rather than rate
each component separately (as for items 6-11 under the Course Evaluation
section). This way, we reasoned, we
would receive a relative evaluation of the parts of the course, whether a given
student liked or disliked the overall course.
The scheme worked reasonably well but despite careful directions, about
one-third of the students rated the items in items #23-25 rather than ranking
them. These responses were not
included in the statistical analysis.
Each course instructor tallied their own evaluations and turned in means
and sample numbers for the Course Evaluation and Aids to Learning items (#1-14
and #23-25, respectively). Responses
to the Instructor Evaluation items (#15-22) and written comments (#26-30) were
kept by individual instructors and not tabulated.
Grand averages for each item were calculated by multiplying the average
for each section by the number of evaluations, totaling these products, and
dividing the total by the number of evaluations for all sections combined.
3.2
Results:
The average evaluation scores for each section, and the grand averages
for all sections combined, are given in Appendix 7.
Overall ratings of the course fell in the mid-twos:
students rated their personal interest in the course material (item 2) at
2.58, the fairness of the grading (item 5) at 2.83, and whether they would
recommend the course to other students (item 14) at a low 2.12.
In terms of how they believed Clare 102 met its course objectives,
students were somewhat in agreement (ave. = 3.72) with the statement (#1) that
the topics covered in class fit the course description and about neutral as to
whether they have a better understanding of the scientific process after taking
the course (item #3, ave. = 3.19) and whether the course helped them understand
how scientists answer questions about the natural world (item #12, ave. = 3.16).
Students rated their coverage of the content of science somewhat higher:
they gave a rating of 3.41 to the statement (#14) that the course
introduced them to many of the fundamental discoveries of the natural sciences.
Some clear differences emerged in terms of which components of the course
students believed to have helped them learn.
Items #6-11 in the Course Evaluation section asked students to rate the
large lectures, class discussions, active learning exercises, textbook, assigned
readings, and laboratories as aids to understanding the course material.
Items #23-25 asked them to rank the large lectures, textbook, assigned
readings, classroom presentations/discussions, active learning exercises, and
review sessions in terms of their ability to grasp basic scientific facts (item
#23), understand science as a human activity (#24), and prepare for tests (#25).
The two sets of questions overlap except for the items on laboratories
(#12) and review sessions (#23-25). Accordingly,
I have ranked student evaluations of the remaining 5 course components in Table
2. Students were remarkably
consistent in ranking class discussions as the most helpful course component
(after review sessions) in both sections of the evaluation form. Active learning exercises were consistently ranked next
highest. This result is in
considerable contrast to the opinions of the course instructors, who generally
found these parts of the course the least effective.
|
Rank based on: |
|
|
|
Average of |
Component |
Items
#6-10 |
Item #23 |
Item #24 |
Item #25 |
Items
#23-25 |
Large lecture |
1 |
3 |
2.5 |
1 |
3 |
Textbook |
3 |
2 |
1 |
3 |
2 |
Assigned readings |
2 |
1 |
2.5 |
2 |
1 |
Class discussions |
5 |
5 |
5 |
5 |
5 |
Active learning ex. |
4 |
4 |
4 |
4 |
4 |
Review sessions |
--- |
6 |
6 |
6 |
6 |
The results for large lecture, textbook, and the assigned readings showed
more variability (Table 2). Overall
rankings, based on items #6, 9 and 10, placed the textbook next highest, the
readings next, and the large lectures the least useful.
Students showed some discrimination, however, when responding to items
#23-25. Item # 23 asked students
which components were most helpful for learning the basic scientific facts:
here they ranked the large lectures above the textbook, and the readings
last. When asked about which helped
them understand science as a human activity, students ranked the large lecture
and readings the same, and place the textbook last.
Finally, for help in preparing for tests, students ranked the textbook
the highest of the three, the readings lower, and the large lecture last.
By far the strongest fact which emerged from the student evaluations was
the high value awarded to review sessions as aids to learning. This is interesting, as the review sessions were voluntary
activities carried out by the instructors but not part of the planned structure
of the course. It may reflect a
regrettable focus of students on last-minute “cramming” rather than diligent
completion of weekly assignments.
3.3
Discussion:
Students show moderate levels of agreement (3.16 to 3.41) with statements
that involve the fulfillment of course objectives.
They agree overall (3.72) that the topics covered in the course meet the
course description. Active modes of
learning (class discussions and active learning exercises) are ranked about
pedagogies such as large lectures and assigned readings that are more
traditional in science courses. These
findings validate, to some extent, the decision to emphasize active student
learning in the course. The modes
of learning that the students found most helpful are, however, viewed by the
instructors as the least effective means of communicating the course material.
The evaluations reflect student attitudes but do not necessarily tell us
which components of the course actually produced positive learning.
The discrepancy between student and faculty assessment of pedagogy needs
to be explored and understood if learning outcomes in the course are to be
improved.
Overall, students did not like the course. This observation is based not only on formal evaluations but on many conversations with students during the semester. The evaluation instrument used doesn’t shed much light on why the course was disliked: none of the 13 items under Course Evaluation had as low a rating as #14 (“I would recommend this course to other students”). In the future Clare 102 instructors should pursue more “nuanced” ways of assessing student reactions to the course.
Two considerations may balance the low overall student evaluation of the
course. First, the course was
dramatically successful at improving students’ grasp of science.
It worked! Second,
non-science majors generally do not enjoy science courses.
To put the Clare 102 evaluations in perspective, they are compared in
Table 3 with comparable questions from student evaluations of selected sections
of Biology 101 here at St. Bonaventure over the past 10 years.
Responses to the first two items, which involve suitability of topics and
tests, are not statistically different for the two courses (2-tailed t-test, P =
0.20 and 0.32). Students
rated Clare 102 significantly lower on the next two items, however:
fairness of grading (2-tailed t-test, P = 0.010) and usefulness of
readings (2-tailed t-test, P = 0.011). Finally,
the overall rating of Clare 102 (“I would recommend this course to other
students”) was very much lower than for Biology 101 (2-tailed t-test, P =
0.001).
Table
3. Average student course
evaluation scores for the 5 sections of Clare 102, Spring 2000 compared with 5
selected sections of Biology 101. 1
= strongly disagree, 3 = neutral, 5 = strongly agree.
------ Clare 102 ------
difference -------- Bio.
101 -------
Question |
mean |
std.
dev. |
n |
(Clar
– Bio) |
mean |
std.
dev. |
n |
Did topics fit course description? |
3.74 |
0.321 |
5 |
-0.24 n.s. |
3.98 |
0.205 |
5 |
Did
tests match course material? |
3.64 |
0.336 |
“ |
+0.24 n.s. |
3.40 |
0.374 |
“ |
Was
the grading procedure fair? |
2.82 |
0.432 |
“ |
-0.88 **
|
3.70 |
0.387 |
“ |
Were
the readings helpful? |
2.82 |
0.415 |
“ |
-0.76 *
|
3.58 |
0.277 |
“ |
Would
you recommend the course? |
2.12 |
0.303 |
“ |
-1.00 ** |
3.12 |
0.342 |
“ |
Significance of the observed difference, as the probability of
their occurring when the null hypothesis of no difference between the two
courses is true: n.s. not
significant (P > 0.05), * P < 0.05, ** P < 0.01.
Based on t-tests, unequal variances.
The following conclusions and recommendations are made by me, without
necessarily being supported by the other Clare 102 instructors:
I am very grateful to Patsy O’Brien, the building secretary for DeLaRoche Hall, for doing the tedious work of coding all the test results. Thanks to the other Clare 102 class instructors for providing tabulated results of student evaluations for their sections. This report benefited from the comments and suggestions of others, but the analyses and opinions are ultimately my responsibility.