An introductory cell biology laboratory course was redesigned using two inquiry-based modules to align with the goals of scientific inquiry as described in Vision and Change. To evaluate the lab's efficacy, we used a broad range of materials, including pretests and posttests, online surveys, focus group interviews, and course evaluations. Although our students produced significant learning gains on technical laboratory skills, methods, and data analyses, during the first two years their affective dispositions toward the experience were more negative. By evaluating our class data in light of insights from the scholarship of teaching and learning, we were able to provide better guidance to students and subsequently to persist past this implementation dip, showing continued positive gains in student learning outcomes, and eliminating the negative impact on student affective outcomes. Our experience underscores the value of scientific teaching, using class data and evidence-based practices to persist beyond the implementation dips that come with adopting new curricula.
Introduction
Improving scientific literacy is a national priority for undergraduate education. Foundational to informed citizenship in a democratic society, scientific literacy includes core competencies highlighted by Vision and Change (AAAS, 2011) such as proficiency in using scientific processes, quantitative reasoning, and the ability to tap into the interdisciplinary nature of science. These competencies are particularly crucial with the current generation of STEM students pursuing vocations that address climate change, genome editing, chronic diseases, pollution, and other complex issues in contemporary society.
An introductory laboratory class, in particular, has advantages for instilling these core competencies in students. It is a natural fit to teach scientific processes during a laboratory. The experiments in the laboratory can have an inquiry-based design to mimic the authentic experience of scientists (Freeman et al., 2014). With the right choice of focus, quantitative reasoning can be an important part of the experience (Baumgartner et al., 2015). Connections to the interdisciplinary nature of science can be made by incorporating concepts and practices from related fields such as other natural and social sciences, a need that is receiving renewed attention and emphasis (Garlick & Levine, 2016).
Many of us, however, are also familiar with the implementation challenges associated with instilling core competencies, even in laboratory settings. If the design is inquiry-based and requires student self-direction, students can quickly become overwhelmed and frustrated by the lack of direction (Meyer et al., 2008). Affective outcomes can suffer with an inquiry-based design (Gormally et al., 2009), and if the students do not “like” the laboratory experience, the negative affective outcome can have a negative impact on the overall effectiveness of the learning (Trujillo & Tanner, 2014). Similarly, attempts to make connections to quantitative reasoning and other sciences can fall flat, as students often fail to transfer learning from one setting to another (Mestre, 2005). These challenges and others can result in an implementation dip (Fullan, 2007). Research shows that instructors who are using evidence-based pedagogical changes should expect that the first years of a new curriculum will be unpopular with students and may not show the learning effectiveness gains promised by the research literature (Fullan, 2007). This can undermine the success of such revisions to the point where they are deemed counterproductive and are abandoned (Felder & Brent, 1996), rather than revised to be more effective.
Here, we will describe new inquiry-based modules for an introductory biology course for majors. We will demonstrate how the laboratory activities are aligned with the goals of scientific literacy as described in Vision and Change. We will also show, through our assessment data, how this course was subject to an implementation dip. We will further describe how, by learning from our assessment data and revising the course, we were able to persist past the implementation dip, show net positive gains in student learning outcomes, and eliminate the negative impact on student affective outcomes. We hope that this article will provide the details of an implementation-ready laboratory investigation as well as an encouraging narrative of how to persist beyond the implementation dips that come with adopting new curricula.
Taken in the second semester of the first year, the Cellular and Genetics Systems lab uses investigatory modules to teach essential laboratory skills, prevailing methods, and core competencies. Recognizing that the typical flow of a research project – reading literature, designing and conducting experiments, analyzing data – parallels the learning cycle (Lawson, 1995), we surmised that modules aligned with this flow would promote effective learning. We recognized a growing body of literature – accessible to first-year students – showing that cooking affects the nutritional properties of foods (e.g., López-Berenguer et al., 2007). To take advantage of innate student interest in food, nutrition, and health, we selected the preparation of vegetables, specifically broccoli, for consumption as the research context for these modules. We believed this would provide a platform for achieving significant engagement with our target core competencies: scientific process, quantitative reasoning, and interdisciplinary science skills. The literature forms the basis for designing interesting experiments. Affordable, durable instrumentation and easy-to-use methodologies, even for students with little or no prior experience, enable a focus on quantitative skills, data standardization, and interpretation, based on descriptive statistics and graphs. Inherent interdisciplinary connections to chemistry, statistics, and social issues (including food systems and heath policy) facilitate integrative skills.
Materials
For a list of instruments and materials needed each week, please consult the Instructor's Manual provided in the online Supplemental Material.
Methods
This laboratory course comprises two multiweek modules, each centered around a research question pertaining to the production of isothiocyanates, nutraceutical compounds produced from glucosinolates in a reaction catalyzed by the enzyme myrosinase (Vermeulen et al., 2008). Initial weeks introduce relevant literature and essential laboratory skills. Next, students learn key experimental methods by performing them. Finally, student teams formulate a focused hypothesis (based on the literature), design and conduct an experiment to test this hypothesis, and then standardize their data and perform descriptive statistical analyses. (Note: a course in statistical analysis is not a prerequisite to our course.)
The following is an outline of our two modules. Please see the Instructor's Manual in the Supplemental Material for a more descriptive summary of each week's activities.
Module 1: How do food preparation parameters affect nutraceutical properties of broccoli?
Weeks 1–3: Explore literature and master essential skills
Conducting literature searches, reading scientific papers, microliter pipetting, preparing dilutions, using spectrophotometers, graphing
Weeks 4 and 6: Learn methods
Protein extraction and Bradford assay, myrosinase (glucose production) assay
Weeks 5 and 7: Design and conduct experiments to test hypotheses; analyze data
Measure protein content and myrosinase activity of broccoli cooked using various methods, standardize data, conduct descriptive statistical analyses, formulate conclusions based on data interpretation
Module 2: How do isothiocyanates (ITCs) affect the proliferation of human (Jurkat) cancer cells?
Weeks 1–2: Explore literature and master essential skills
Conducting literature searches, reading scientific papers, using microscopes, using sterile technique
Weeks 2–4: Learn methods
Human cell culture, cell counting using a hemocytometer, microplate-based cell viability assay, DNA extraction, gel electrophoresis
Weeks 2 and 4: Design and conduct experiments to test hypotheses; analyze data
Measure cell death and DNA fragmentation to assess the effect of ITCs on apoptosis, standardize data, conduct descriptive statistical analyses, formulate conclusions based on data interpretation
Methods: Assessment Strategy
Our assessment of the revised curriculum had several components. In the first week of the course, students completed a pre-assessment consisting of open-ended questions that involved quantitative reasoning skills (such as calculating dilutions) that are taught in a prerequisite chemistry course, and graphical interpretation. At midterm, we administered a lab practical exam based on Module 1 and, at the end of the course, a comprehensive final exam that included items that matched the pretest as well as other items assessing mastery of core competencies and essential skills, familiarity with literature, understanding of methodologies, and mastery of data standardization, graphing, and interpretation skills. We directly measured student learning gains by comparing the mean scores on questions presented to students on the pre-assessment to the mean scores on matched questions (see Appendix B) presented on a midterm and a final exam (Table 1).
Learning Outcome . | Percent Correct . | Pa . | ḡ . | |
---|---|---|---|---|
Pretest . | Posttest . | |||
Scientific literature search | 59.0% | 87.0% | 0.000b | 77.0% |
Simple dilutions | 14.7% | 68.3% | 0.000b | 64.3% |
Serial dilutions | 3.3% | 64.7% | 0.000b | 70.1% |
Using a standard curve | 26.3% | 77.8% | 0.000b | 36.7% |
Quantitative transfer with micropipette | 2.0% | 85.0% | 0.000c | 85.5% |
Test total | 19.9% | 72.7% | 0.000d | 70.6% |
Learning Outcome . | Percent Correct . | Pa . | ḡ . | |
---|---|---|---|---|
Pretest . | Posttest . | |||
Scientific literature search | 59.0% | 87.0% | 0.000b | 77.0% |
Simple dilutions | 14.7% | 68.3% | 0.000b | 64.3% |
Serial dilutions | 3.3% | 64.7% | 0.000b | 70.1% |
Using a standard curve | 26.3% | 77.8% | 0.000b | 36.7% |
Quantitative transfer with micropipette | 2.0% | 85.0% | 0.000c | 85.5% |
Test total | 19.9% | 72.7% | 0.000d | 70.6% |
aα = 0.05 adjusted with a Bonferroni correction.
bWilcoxon signed-rank test.
cSign test.
dPaired-sample t-test.
We also assessed affective outcomes, another key indicator of the efficacy of our reforms. Students' overall attitude toward the lab was measured with campus-wide course evaluations ( Appendix C) administered at the end of the semester. To measure students' self-perception of learning, we administered customized Student Assessment of Learning Gains (SALG) surveys ( Appendix A) to the control (2012) and treatment (2013–16) groups (Figure 1). Finally, we used focus group interviews at the end of the first two treatment years (2013 and 2014) to solicit specific suggestions for improvement from students (Table 2).
Students' perceptions of their core competency gains. Four SALG survey questions were used to assess perceived gains, by way of a Likert-like scale ranging from “no gains” (1) to “great gains” (5), with respect to scientific process, basic laboratory skills, analytical and quantitative skills, and interdisciplinary integration (see Appendix A for question details). Sample sizes varied by year: 86 (2012), 93 (2013), 63 (2014), 71 (2015), and 58 (2016). Bars represent means ± SE. Data from 2012 serve as the pre-implementation controls; cohorts from 2013–16 are independent treatments reflecting ongoing adjustments to the course.
Students' perceptions of their core competency gains. Four SALG survey questions were used to assess perceived gains, by way of a Likert-like scale ranging from “no gains” (1) to “great gains” (5), with respect to scientific process, basic laboratory skills, analytical and quantitative skills, and interdisciplinary integration (see Appendix A for question details). Sample sizes varied by year: 86 (2012), 93 (2013), 63 (2014), 71 (2015), and 58 (2016). Bars represent means ± SE. Data from 2012 serve as the pre-implementation controls; cohorts from 2013–16 are independent treatments reflecting ongoing adjustments to the course.
Focus Group Questions . | Top Three Responses . | |
---|---|---|
2013 . | 2014 . | |
What aspects of the lab helped you learn? |
|
|
What aspects of the lab hindered your learning? |
|
|
What specific suggestions do you have for modifying the lab? |
|
|
Focus Group Questions . | Top Three Responses . | |
---|---|---|
2013 . | 2014 . | |
What aspects of the lab helped you learn? |
|
|
What aspects of the lab hindered your learning? |
|
|
What specific suggestions do you have for modifying the lab? |
|
|
Results
Significant Learning Gains Realized
By comparing responses to identical questions on pre- and post-assessments (midterm and final exam), we directly measured learning gains in core competencies (Table 1). In all cases, normalized learning gains (ḡ) were significant. In particular, we examined science process skills such as ability to use the literature and perform lab techniques, quantitative reasoning as used in dilution calculations and with a standard curve, and interdisciplinary connections with chemistry and statistics. Most students entering this course have some knowledge of Internet search tools that are pertinent to scientific literature. Nevertheless, it was clear from the comparison of students' responses to the relevant pretest and posttest question (see Appendix B) that they made learning gains that reflect a greater ability to distinguish primary from review articles and more attentiveness to the types of key words to use when performing literature searches. In the prerequisite General Chemistry course, students were taught to perform simple dilutions and use standard curves; yet only a minority of our students were able to explain these tasks on the pretest. Even so, through repeated performance of these tasks throughout the biology laboratory course, the majority of students demonstrated mastery of these crucial skills. Our assessments also revealed that few students came into the course with any experience in quantitative transfer with micropipettes or with serial dilution; indeed, neither of these was taught in our General Chemistry course. Despite this fact, the majority of our students mastered these skills. This is especially true of micropipetting, a technique introduced early and employed nearly every week; the normalized learning gain for this skill was the highest observed, at 85.5%.
Student Perceptions Reveal an Implementation Dip
Student perceptions of their mastery of core competencies, as determined by analyzing their responses on end-of-course SALG surveys (Figure 1), tended to decline in the first two years of implementation (2013 and 2014) compared to our control (2012). Students were asked to rate their improvement in four bioscience competencies on a scale from 1 to 5 (1 = no gains, 3 = moderate gains, 5 = great gains): mastering the scientific process of investigation; developing basic lab skills; developing analytical and quantitative skills; and understanding the connection between the concepts in biology, chemistry, and mathematics. Figure 1 reveals that for all competencies, except developing basic lab skills, mean perception scores declined in the first two years (2013 and 2014) following the curriculum revision in comparison to our pre-implementation control (2012). However, as improvements were made in subsequent years (2015 and 2016), there was a trend toward higher ratings in all of these areas. These ratings tended to mirror the kinds of comments voiced by students during the course, with frustrations most commonly expressed in 2013 and 2014.
To further illuminate the nature of our implementation dip (Fullan, 2007), we also scrutinized student ratings from end-of-semester course evaluations. Figure 2 reveals some very informative results. First, students perceived that the intellectual challenge and effort required in our project-based modules had increased compared to the level of challenge and effort in more traditional labs (which these students experienced the prior semester in General Chemistry). In fact, since project-based experiences provide a richer learning environment, we expected students to find this more challenging. Secondly, these data revealed a mismatch between the significant learning gains made by students in 2013–14 (Table 1) and these same students' low perceptions of their learning gains (Figure 2). Finally, our results clearly show a correlation between higher perceived levels of challenge and effort, lower levels of confidence in the amount learned, and lower perceptions of the course and instructor. Student opinion of the course as a whole declined significantly from a mean of 3.8, which is on par with department averages for all courses, to a mean of 2.6 in 2013 and 2.5 in 2014 (Figure 2). Similarly, perceptions of the course's organization declined from a mean of 4.0 to 2.7–2.8. These spilled over into lower opinions of the instructor's teaching effectiveness, clarity of instruction, fairness, and (to some extent) even the instructor's enthusiasm. Together, these data led us to initiate a series of revisions in attempts to recover from this demoralizing implementation dip.
Students' perceptions and affective outcomes. End-of-semester course evaluation questions were used to assess perceptions of this course in relation to other courses by way of a Likert scale ranging from “much lower than average/poor” (1) to “average” (3) to “much higher than average/excellent” (5) (see Appendix C for question details). Data are from one instructor who taught one or two sections of the course each year. Sample sizes varied by year: 16 (2012), 38 (2013), 34 (2014), 35 (2015), and 16 (2016). Bars represent means ± SE. Data from 2012 serve as the pre-implementation controls; cohorts from 2013–16 are independent treatments reflecting ongoing adjustments to the course.
Students' perceptions and affective outcomes. End-of-semester course evaluation questions were used to assess perceptions of this course in relation to other courses by way of a Likert scale ranging from “much lower than average/poor” (1) to “average” (3) to “much higher than average/excellent” (5) (see Appendix C for question details). Data are from one instructor who taught one or two sections of the course each year. Sample sizes varied by year: 16 (2012), 38 (2013), 34 (2014), 35 (2015), and 16 (2016). Bars represent means ± SE. Data from 2012 serve as the pre-implementation controls; cohorts from 2013–16 are independent treatments reflecting ongoing adjustments to the course.
Scientific Teaching Aids Recovery
Scientific teaching entails the combined use of evidence-based pedagogies and course data to promote better learning (Handelsman et al., 2004). Our data reveal that even though our students were making significant learning gains (Table 1), their perceptions were declining (Figures 1 and 2). This is consistent with commonly held faculty concerns about student resistance as a barrier to reform efforts (Seidel & Tanner, 2013). Nevertheless, we took these as evidence of the need for corrective action. Qualitative data from open-ended SALG survey questions and focus group interviews revealed the top areas of student dissatisfaction (Table 2):
Lack of a clear connection between lecture and lab material
Lack of understanding of significance of learning goals
Expectation of more direction from faculty, less student self-direction (i.e., less inquiry)
Too much effort required to meet lab expectations
In response to these student concerns, we made several revisions to the curriculum. We modified lab manual introductions to more clearly explain how weekly activities pertain to the scientific research process and the driving research question. We added scaffolding, such as calculation prompts, to guide quantitative analysis (Hmelo-Silver et al., 2007). We added text boxes to highlight learning outcomes, critical-thinking tips, and vocational connections. Questions that facilitate reflection at the end of each lab period were also added to the lab manual. Collectively, these steps were correlated with marked improvements in student perceptions and affective outcomes in 2015 and 2016 (Figures 1 and 2). It is important to note that desired learning outcomes and perceived intellectual challenge were sustained during the 2015 and 2016 iterations of the course. Thus, it is unlikely that the affective outcomes increase can be attributed to any perception of the course being made “easier.”
Implementation Notes
After our first year with the new course, the implementation dip guided revisions that we now see as key for other potential adopters to consider. The signature component of our implementation dip was the mismatch between student perceptions of their learning (Figures 1 and 2) and their actual learning gains (Table 1). This contributed to lower student self-efficacy, science identity, motivation, and negative attitudes (as is evident in written comments on SALG surveys, course evaluations, and focus group interviews; Table 2), all of which pertain to the affective domain (Trujillo & Tanner, 2014). This type of mismatch between perception and achievement is not uncommon when instructors adopt learner-centered pedagogies that alter the classroom dynamics (Van Sickle, 2016). Inquiry-based learning often requires the students to invest more time and effort, which the students may think could have been avoided if they were “told what to learn” (Loughran & Derry, 1997). We surmise that this contributed to our markedly lower scores for course organization and clarity of instruction in 2013 and 2014 (Figure 2). Indeed, providing prompts for performing a complex set of calculations for data standardization (involving unit conversions and adjusting for dilutions) – rather than guiding students to discover the required steps themselves by constructing a flowchart of the experimental manipulations (Figure 3) – contributed significantly to our implementation dip recovery in 2015 and 2016 (Figures 1 and 2).
Type of instructional scaffolding influences learning outcomes. In 2014 three different instructors tested different types of instructional scaffolding in their sections of the course. One section (18 students) constructed experimental flowcharts to reveal quantitative manipulations. Another section (24 students) was given instructional prompts outlining the necessary steps in quantitative reasoning. Other sections (38 students) used both types of instructional scaffolding. One week later, a midterm exam assessed performance on an analogous quantitative problem (see Appendix B for details). Scores for that problem are shown; bars represent means ± SE. One-way ANOVA revealed significant differences in performance as denoted by the asterisk (p = 0.006). Scheffe post hoc comparisons indicated that the mean score for the “flowchart only” section was significantly lower than the mean scores for the “prompts only” section (p = 0.011) and the “flowchart + prompts” sections (p = 0.036).
Type of instructional scaffolding influences learning outcomes. In 2014 three different instructors tested different types of instructional scaffolding in their sections of the course. One section (18 students) constructed experimental flowcharts to reveal quantitative manipulations. Another section (24 students) was given instructional prompts outlining the necessary steps in quantitative reasoning. Other sections (38 students) used both types of instructional scaffolding. One week later, a midterm exam assessed performance on an analogous quantitative problem (see Appendix B for details). Scores for that problem are shown; bars represent means ± SE. One-way ANOVA revealed significant differences in performance as denoted by the asterisk (p = 0.006). Scheffe post hoc comparisons indicated that the mean score for the “flowchart only” section was significantly lower than the mean scores for the “prompts only” section (p = 0.011) and the “flowchart + prompts” sections (p = 0.036).
Between the first and second iterations of the course, we took steps to address the concern that the lab contained too much material and was overwhelming. We attended to this by reducing the number of modules from three to two (Table 3), thereby providing more time for in-class guidance when performing more complex calculations (quantifying protein levels and enzyme activities; see scaffolding in Appendix D) and making appropriate inferences that take into account the variance in the data (see Instructor's Manual, p. 9, in Supplemental Material). After this modification, we noted that student affective dispositions stopped declining in 2014 (Figures 1 and 2). However, they did not improve significantly until we had taken additional pedagogical steps to provide sufficient guidance in the research process and data analysis stages. Our data demonstrate the necessity of providing direct instruction in teaching unfamiliar methodologies. For examples of this direct instruction, see teaching tips in the Instructor's Manual for procedures such as micropipetting (p. 8), gel electrophoresis (p. 29), microscopy (p. 24), the use and calibration of a spectrophotometer (p. 8), the creation and implementation of Beer's Law standard curves (p. 9), and performance of the Bradford assay (p. 13). As a result of making these pedagogical improvements, students indicated in SALG surveys that these lab skills were taught well and would be helpful in the future.
Week . | 2012: Last Iteration of Old Lab Activities . | 2013: First Implementation of Project-Based Lab Modules . | 2014–16: Adjustments to Facilitate Desired Outcomes . | ||
---|---|---|---|---|---|
1 |
| Effects of cooking on broccoli proteins |
| Effects of cooking on broccoli proteins |
|
2 |
|
|
| ||
3 |
|
|
| ||
4 |
|
|
| ||
5 |
|
|
| ||
6 |
| ITC effects on bacteria |
|
| |
7 |
|
|
| ||
8 |
|
|
| ||
9 |
| Effects of ITCs on cancer cell death |
| Effects of ITCs on cancer cell death |
|
10 |
|
|
| ||
11 |
|
|
| ||
12 |
|
|
| ||
13 | – |
|
|
Week . | 2012: Last Iteration of Old Lab Activities . | 2013: First Implementation of Project-Based Lab Modules . | 2014–16: Adjustments to Facilitate Desired Outcomes . | ||
---|---|---|---|---|---|
1 |
| Effects of cooking on broccoli proteins |
| Effects of cooking on broccoli proteins |
|
2 |
|
|
| ||
3 |
|
|
| ||
4 |
|
|
| ||
5 |
|
|
| ||
6 |
| ITC effects on bacteria |
|
| |
7 |
|
|
| ||
8 |
|
|
| ||
9 |
| Effects of ITCs on cancer cell death |
| Effects of ITCs on cancer cell death |
|
10 |
|
|
| ||
11 |
|
|
| ||
12 |
|
|
| ||
13 | – |
|
|
The way we developed scaffolding to assist with quantitative tasks was fortuitous in that it enabled us to measure its impact directly. Prior to its development, even our best students were often confused and frustrated during the data standardization and analysis stage of the process. This led one instructor, who taught the first section of the course, to abandon the existing strategy – guiding students to discover the required steps themselves by constructing a flowchart of the experimental manipulations – and instead prompt them through a series of steps in the calculations. The prompts (PowerPoint slides given in Appendix D) then became a scaffolding, which the instructor shared with colleagues. One colleague implemented the switch; the other did not. Subsequently, in the analysis of a complex midterm question, we observed a marked difference between sections that was directly correlated with exposure to the scaffolding (Figure 3). To eliminate the possibility that this difference was due to a baseline difference in students' inherent abilities, pre-assessment mean scores were also compared with a one-way ANOVA and revealed no significant difference between the three sections (p = 0.798). These data are consistent with the idea that in inquiry-based learning, students require more guidance to successfully navigate complex data analysis problems, a process that requires higher-order cognitive skills encompassing the application of knowledge and critical thinking and involving deep conceptual understanding (Zoller, 1993).
Finally, in 2015 and 2016 the instructor who taught at least one section of this course each year also made a conscious effort to engage in more non-content dialogue with students (Seidel et al., 2015). This included building rapport with students and within student teams, sharing personal experiences in research, explaining pedagogical choices, and clarifying expectations. As a result, this instructor realized a full recovery from our implementation dip. Other instructors, who have not been consistently assigned to this course, have seen improvements since the implementation dip, but not a full recovery.
Conclusion
Our experiences underscore the value of evidence-based curriculum development and revision (i.e., scientific teaching; Handelsman et al., 2004). In addition to using evidence-based pedagogies, our use of assessment data to drive interventions in the learning process were crucial in our recovery from an implementation dip (Fulan, 2007). Our experiences also attest to an essential feature of guided inquiry: since different levels of guidance are needed at different stages, its success is dependent on high-quality student–faculty and student–student partnerships. Our experiences demonstrate the importance of persistence in scientific teaching practices in order to find the appropriate level of guidance. In terms of educational theory, our scaffolding reduces the extraneous cognitive load (challenges posed by instructional choices), making it easier for students to manage the intrinsic cognitive load (imposed by the complexity of the calculations), and resulting in a higher germane cognitive load that facilitates effective learning (Sweller, 1988).
Reading the literature about inquiry-based learning, it is easy to get the impression that “if you build it, they will come.” But perhaps the comparison between traditional didactic teaching and inquiry-based learning is more like dancing. Instead of insisting that the role of the instructor is to lead and the role of the students is to follow, the reality is that each has to constantly pay close attention to the other. For the dance to be beautiful, both have to practice working together. If one dancer tries to do it all, then there is no dance. “It takes two to tango.”
This project was funded by the National Science Foundation (TUES grant no. 1140767). We are indebted to our biology department colleagues for their contributions in the planning and implementation of the revised lab curricula. We also thank our Calvin College collaborators – Paul Moes and Randall Pruim – for assistance with statistical analyses.
References
Appendix A
Assessment of Students' Perceptions of Their Core Competency Gains: SALG Survey Questions (see Figure 1)
Students were asked: Please rate your GAINS in each of the following BIOSCIENCE COMPETENCIES.
[Scale = 1 (no gains), 2 (a little gain), 3 (moderate gain), 4 (good gain), 5 (great gain)]
Scientific Process:
Mastering the scientific process of investigation (designing and conducting experiments)
Laboratory Skills:
Developing basic lab skills (pipetting, making dilutions, following procedures)
Analytical/Quantitative Skills:
Developing analytical and quantitative skills (statistical analysis, graphing)
Interdisciplinary Integration:
Understanding the connection between the concepts in biology, chemistry, and mathematics
Appendix B
Assessment of Student Learning Gains: Matched Pretest & Posttest Questions (see Table 2)
Scientific Literature Search: If you had to find a scientific review article about the health effects of different “nutraceutical” (having nutritional and pharmaceutical properties) compounds in broccoli, how would you go about doing this? Explain in as much detail as you can.
Simple Dilutions: You are given a stock solution of 1 M glucose and told that you need to prepare 10 mL that has a final concentration of 10 mM glucose. How do you go about doing this?
Serial Dilutions: By using a standard curve, we will be able to quantify the amount of ATP in our cultures. The manufacturer of our assay kit suggests preparing ATP standards with a final volume of 100 μL and concentrations of 0.01–1000 nM. Starting with a 1 μM stock solution, what would be the most accurate way to prepare these standards? Show all your work.
Using a Standard Curve: Biologists often use a spectrophotometer to measure the absorbance of a solution and thereby determine the concentration of a particular solute in that solution. To do this requires the use of a standard curve in which different concentrations of known quantities of solute are measured.
What is a standard curve and how do you prepare one?
Absorbance is a ratio of the level of light passing through a sample over the level of incident light coming from a light source. As the concentration of solute increases, the absorbance _______________ (choose one: increases/decreases).
How can one determine the concentration of the solute in an experimental tissue sample in comparison with the standard curve?
Quantitative Transfer with a Micropipette: Suppose you have to measure out 250 μL of a 10 μM protein solution and transfer that into a microcentrifuge tube containing 1 mL of water. What is the best way to accurately measure out the 10 μM protein solution?
Appendix C
Assessment of Students' Perceptions & Affective Outcomes: Course Evaluation Questions (see Figure 2)
For the following questions, response options were “Much Higher,” “Higher,” “Average,” “Lower,” and “Much Lower”:
The intellectual challenge was …
The amount that I learned was …
The amount of effort expected of students in the course was …
For the following questions, response options were “Excellent,” “Very Good,” “Good,” “Fair,” and “Poor”:
The course as a whole was …
The effectiveness of the instructor's teaching methods in actively engaging me with the ideas of this course was…
The enthusiasm the instructor showed for the content of this course was …
The course organization and planning were …
The clarity of the instructor's interactions with students was …
The fairness of the instructor's evaluation of student performance was . …
Appendix D
Assessment of Type of Instructional Scaffolding That Influences Learning Outcomes: Scaffolding Prompts (see Figure 3)
The following three PowerPoint slides were used as scaffolding.
Procedure . | Starting material or aliquot (incl. amount) . | Sample dilution during process . | Ending material (incl. amounts) . |
---|---|---|---|
Cooking | __ g of broccoli | none | __ g of broccoli |
Protein extraction | __ g of broccoli in __ μL of extraction buffer | __ μL of broccoli extract | |
Bradford assay | __ μL of broccoli extract | __ μL in each well | |
Myrosinase assay | __ μL of broccoli extract | __ μL in each well |
Procedure . | Starting material or aliquot (incl. amount) . | Sample dilution during process . | Ending material (incl. amounts) . |
---|---|---|---|
Cooking | __ g of broccoli | none | __ g of broccoli |
Protein extraction | __ g of broccoli in __ μL of extraction buffer | __ μL of broccoli extract | |
Bradford assay | __ μL of broccoli extract | __ μL in each well | |
Myrosinase assay | __ μL of broccoli extract | __ μL in each well |
STANDARDIZING BRADFORD DATA
What do we have? | Protein concentration (μg/mL) | |
What do we need? | Protein amount (mg) / gram of broccoli |
Steps needed to get us there:
Determine what is known from the standard curve & sample data.
Take into account the sample dilution (V1/Vf). For example, if you diluted 1:10, then you need to multiply by 10 to figure out the protein concentration of your broccoli extract.
Convert your protein concentration to a weight (mg) of protein. Multiply the concentration by the volume and make sure your units “factor out.”
Divide by the amount of broccoli tissue in your extract.
STANDARDIZING AMPLEX RED DATA
What do we have? | Glucose concentration (μM) | |
What do we need? | Units of myrosinase / gram of broccoli | |
Units of myrosinase / mg of protein |
Steps needed to get us there:
Determine what is known from the standard curve & sample data.
Take into account the sample dilution (V1/Vf).
Convert your glucose concentration to an amount (nmol) of glucose produced by the myrosinase. Multiply the concentration by the volume and make sure your units “factor out.”
Divide by the total reaction time (min). This yields a rate of glucose production per minute. Note: 1 unit of myrosinase = 1 nmol of glucose/min
Divide by the amount of broccoli tissue (or protein) in your extract.