Friday, June 13, 2008

Applying Dan's assessment system, Part I

Dan Meyer breaks his courses into some 35 discrete skills and concepts, keeps separate records on students' performance on each skill, and keeps retesting students and counting their highest scores. The following two entries are some notes on things I learned while applying an adapted version of his system to my Algebra and Intermediate Algebra this year. The second entry is a dryly technical discussion of scoring.

In accordance with my Department Head's recommendation, I did not entirely replace traditional comprehensive tests with this more piecemeal system. For Algebra 1, these concept quizzes were weighted at 40% of students' grades while comprehensive tests made up the remaining 30% of the assessment grade. For Intermediate Algebra I weighted the two types of assessments at 35% each. My experiences were that...

...this system worked significantly better for Algebra 1 than for Intermediate Algebra.

In Algebra 1, I felt that pretty much everything the students really needed to know was covered by the concept quizzes – I might as well not have done chapter tests at all. For Intermediate Algebra, however the skills tended to get cumbersomely complex or impossibly many, and the supplemental chapter tests were necessary and useful.

One reason is that Intermediate Algebra, which is essentially the first 70-80% of Algebra 2, covers much more1. Another reason is that synthesis and solution of multi-step problems are inherent, irreducible goals of Algebra 2, and these skills need to be assessed, too.2

... for diagnosing and remedying deficiencies in basic skills, this system was beautiful.

At some point early in the semester I realized that a number of incoming Algebra 1 students did not know the concept of place value and could place neither decimal numbers nor fractions on a number line. Writing an assessment on placing decimals on the number line made it possible to separate out who was having trouble with this, and to know when a critical mass of students had caught up in this area. As a tool for probing missing background skills and for placing these skills clearly and definitely on the agenda this was powerful.

... writing effective assessment items was harder than I thought.

When an assessment may potentially be repeated two, three, even five or six times, what it measures had better be really important, and the assessment had better actually capture the intended skill. It is not as easy as it may sound to decide which elements of the course really are that important; which are the parts on which other understanding hinges. My list of concepts to be assessed always tended to get too long, and trimming down to the real essentials was a constant challenge. As for designing valid measurements of students' skills, I guess only experience makes it possible to figure out what kinds of problems will really show that they know what they need to know, what kinds of problems plough just deep enough without getting too involved, what kinds of misunderstandings are typical and must be caught in order to make necessary remediation possible.3

... assessments are not enough. Improvement is not automatic.

That's obvious, of course. How silly to think otherwise. Frankly, part of what I found attractive about this assessment system was the idea that with goals broken down into such small, discrete pieces, students would become empowered and motivated and take the initiative to learn what they needed to make the grade. That was actually to a significant extent the case. Tutoring hours were far more efficient due to the existence of these data, and students knew what to do to "raise their grade." However, a lot of students continued to score poorly, repeating the same mistakes, after three, four, five rounds of assessment on the same topic. Some would come during tutoring hours to retake a quiz and still make exactly the same mistakes... For weaker students especially, then, it is important to remember that the assessment data are tools for me to actually use. There is no automaticity in the translation of this very specific feedback into actual understanding.

... the transparency of the system means bad things are out there for everyone to see.

That's what we want, don't we? The direct and honest reporting involved was a major appeal of this system. However, it takes some foresight for this not to lead to discouragement. While it is pretty common practice among math teachers, any teachers, to rescale test scores so that the class average turns out okay, this could not be done in any simple way with these conceptwise assessments. The only way to improve class grades was by reteaching the material and testing again. This involved a time delay during which the grades, which were published in an online gradebook, could be quite low. This was especially true during the first month or two of school, when the grades were constituted by relatively few entries, and - well - the first months of school may not be the time you want parents to worry about what you're doing when you're a new employee. In the early stages I ended up scaling chapter tests a good deal in order to compensate for some low concept quiz scores and make the overall grades acceptable. With time, a combination of rewriting certain concept quizzes that were needlessly tricky and teaching some topics better made this less necessary. 4

In conclusion, I am definitely keeping this system for Algebra 1, probably increasing the weighting of these assessments and reducing the number and importance of comprehensive tests. For Intermediate Algebra I am keeping chapter tests, and writing a new set of piecemeal assessments to cover just the basics, so that I can have the hard data on who is really lost, but without even trying to force these assessments to cover the entire curriculum. I'll need to make sure that the first skills are very well taught and mastered before the first round of assessments: thinking a little strategically to make sure the early results are good increases buy-in, and student ownership is after all much of the point here.


Notes

1 By way of example, a comparison of the content of the chapters on exponents in the two courses: To assess mastery of this chapter for Algebra 1, I needed to check that students knew the definition of a natural power as repeated multiplication, that they could apply the power rules to simplify expressions, that they could deal with negative and zero powers, that they could complete a table of values of a simple exponential function such as 2x and plot the points to sketch a simple exponential graph. For the chapter on exponential and logarithmic functions for Intermediate Algebra, however, I needed to check whether students could do all of the above, plus convert between exponential and logarithmic form, apply the properties of logarithms, solve exponential and logarithmic equations by applying one-to-one properties, solve such equations by applying inverse properties, apply the change-of-base formula, apply the compound interest formula, identify transformations of the exponential function, understand that exponential and logarithmic functions are inverses of each other, plus a few other things that I just skipped. The number of chapters to be covered is pretty much the same for both courses, but the number of concepts and skills? Different stories. Writing broader concept tests for more advanced courses is a possibility, but the advantages of this piecewise assessment system over the usual comprehensive test system is quickly lost this way.

2 For an example of how some core skills of Intermediate Algebra are by nature multi-step and integrative, consider the case of solving a third degree polynomial equation by first finding a root by graphing, then dividing by the corresponding linear factor, then applying the quadratic formula to find the remaining roots. This task is too complex for a concept wise assessment to be very useful. I had separate assessments on 1) identifying factors given the graph of a polynomial, on 2) polynomial division and rewriting a polynomial using the results of the division process, on 3) stating and applying the factor theorem, and 4) applying the quadratic formula. I still wanted to check whether the students could put it all together.

3 As for the assessment being valid, actually capturing the important skill, here's an example of a failed attempt: I wrote one concept quiz about identifying the difference between an equation and an expression, about distinguishing the cases where you solve for a variable from the case where you can only simplify – but success on this assessment did not mean an end to confusing these two cases. Does that mean that the assessment was poorly written, or rather that this distinction just doesn't lend itself to being assessed once and for all in a little concept quiz? Is understanding equivalence, and distinguishing equations as statements that are true or false from expressions that just are, too abstract ideas to be covered this way? I don't know, but my impression is that the quiz did little to eradicate the common mistake of treating expressions as if they were equations, for example by adding or subtracting new terms in order to simplify.

4 This is at a private school, where determining the required level of mastery of each standard is to a larger extent up to the teacher, since no state testing is involved in defining the bar.

12 comments:

Anonymous said...

Going to try this in an Algebra 1 and a Geometry class this year.

For my Algebra-2 and Pre-Calc, a combo of skill assessment and chapter assessment (putting it all together) I think are the way I will go.

I teach at a small school (average class size-4), so this might help me make huge strides this year.

Dan Greene said...

H, I'm interested in trying this for my Algebra 2 classes next year (not that I am sure yet what I'll be teaching). Sounds like the same material as your intermediate algebra. Do you want to collaborate on developing skills and items? Let me know.

Dan

H. said...

Dan, I would like that very much, thanks! My concept tests for linear functions and polynomials worked well, I think, the ones on quadratics need tweaking, and the ones on exponentials and logarithms need to be reworked completely from scratch. I also want a couple of items on using graphing calculators - do you emphasize that? I'll by back in the Bay Area in the middle of July.

bcarrera said...

I'd like a look at what you come up with as well (ie Alg-2 Skills List). I'm switching schools and won't get at my books until the beginning of July.

H. said...

bcarrera, just remind me about making the files available if I forget to get back to you in the latter half of July.

sam shah said...

Hi,

It's coincidental and fortuitous that you posted about this now, because when I was grading my Alg II final exams, I decided that I might want to try doing something like you: some comprehensive tests and some dy/dan basic skills tests that can be retaken (http://samjshah.com/2008/06/10/final-exam-grading-marathon-is-over/).

I'm so glad you posted your evaluation of this system and your problems with implementing it -- it will definitely help me out next year.

I would love to be in on any collaboration if you're looking for yet another teacher to work with.

I made a list of the Alg II skills/topics from the first semester for another project I worked on this year (http://samjshah.files.wordpress.com/2008/06/video-topics-and-rubric.pdf) so if the course seems at all similar I'd love to join in the fun! (Second semester are more exponents/logs and the dreaded trigonometry.)

Sam.

H. said...

Sam, the link to the video topics doesn't seem to work, but I found a topic list (the second document - is that the right one?) via this page. There's a very large overlap of topics, so do join the fun.

Started looking at some of those videos a few days ago - very neat project.

sam shah said...

Yes, that's the one! Sorry about the non-working link.

Dan Greene said...

Cool. Send me an email when you get back.

Cipolla Dolce said...

I am teaching Algebra 2 for the first time next year and started making a "concept list" a la Dan Meyer...I quit when I already had over 100 for the first marking period alone. So thanks for posting about how the system worked better for Algebra 1 than Algebra 2...I feel more confident about implementing it in Algebra 1 next year, and more validated about doing some sort of hybrid system for Algebra 2.

I will have some questions for you when I digest your post mortem about the grading nitty gritty, but I need to read it a few more times. :-)

Tracy said...

What a good self-assessment - implement a new system and then do a detailed debrief.

It strikes me, reading through your post and Dan's posts, that writing good questions for test purposes is a very particular skill, and a vital one for effective teaching. I wonder if there is some guidance somewhere on how to do it.

Jackie said...

Great point Tracy. I struggle over this with every question I write. What am I assessing? Is it problem solving, a skill? Why am I assessing it?

I wish there were more options for PD on this topic.