Regents Recap — June 2013: Another Embarrassment

Here is another installment in my series reviewing the NY State Regents exams in mathematics.

Sometimes a poorly-written question together with an ill-conceived scoring rubric creates a truly embarrassing situation for those involved with the New York math Regents exams.

Consider the following question from the June 2013 Geometry exam.

2013 June Geometry 33

This two-point problem seems straightforward enough, but things take a turn in the scoring rubric.

2013 June Geometry 33 -- Partial Rubric

According to the rubric, the only way a student can earn full credit here is to graph both loci.  Unfortunately, the student was only asked to graph one locus.

The problem directs students to graph the set of points that are both four units from the x-axis and equidistant from the two given points.  This is a single, compound locus.  Only two points satisfy both conditions simultaneously, so the graph of the locus looks like this:

regents problem -- one locus

Unfortunately, if a student graphs only the locus they were asked to graph, the rubric awards them a maximum of one out of two points.  Why?  Because the rubric mistakenly interprets this compound locus as two loci:  the set of points that are four units from the x-axis and the set of points that are equidistant from (-2,0) and (8,0).  The graph of these two loci look like this:

regents problem -- two loci

The correct compound locus is the intersection of these two loci.  But the students were only asked to graph the points that satisfy both conditions; not the points that satisfy either condition.  It’s absurd to penalize students for not providing information that wasn’t asked for.

I brought this to the attention of our grading site supervisor, but nothing was done.  The site supervisor ultimately defended the rubric by claiming the graphs of the two loci constituted appropriate work, and so were necessary for full credit.  But this ad hoc argument barely warrants a response.  Someone here wrote a bad question, a bad rubric, or both.  Mistakes happen, but in the world of high-stakes testing, students and teachers end up paying the price, while the test-makers avoid accountability.

Other math teachers I spoke to had similar experiences at their grading sites.  And unfortunately, this isn’t the first time state administrators were reluctant to address an absurdly erroneous question.

A Conversation About Rigor, with Grant Wiggins — Part 1

This is the first post in a dialogue between me and Grant Wiggins about rigor, testing, and the new Common Core standards.  Each installment in this series will be cross-posted both here at MrHonner.com and at Grant’s website.  We invite readers to join the conversation.

Patrick Honner Begins

After a spirited exchange on Twitter regarding New York State’s new Common Core-aligned tests, Grant Wiggins cordially invited me to continue our conversation about rigor in a collaborative blog post.  It wasn’t until I saw his first suggested writing prompt—What is rigor?—that I suspected that perhaps I was being rope-a-doped by a master.  But the opportunity was too intriguing to pass up.

While I don’t expect to flesh out a fully-formed, operational definition of rigor, our conversation brought a couple of important ideas to mind regarding what it means for a mathematical test question or task to be rigorous.

Higher grade-level questions are not necessarily more rigorous

Our exchange began after I wrote a piece for Gotham Schools about the 8th grade Common Core-aligned test questions that were released by New York state.  In reviewing the items, I noticed some striking similarities between these 8th grade “Common Core” questions and questions from recent high school math Regents exams.

function quesions side-by-side

Above on the left we see a problem from the set of released 8th grade questions, and on the right, we see a virtually identical question that appeared on this January’s Integrated Algebra exam, an exam taken by high school students at all grade levels.  (This was one of many examples of this duplication).

The annotations that accompany the released 8th grade questions suggest that this question requires that a student understand that “a function is a rule that assigns to each input (x) exactly one output (y)”.  In reality, this question simply tests whether or not the student knows to apply the vertical line test to determine whether or not a given graph represents a function.  If the student recalls this piece of content, the question is simple; if not, there is little they can be expected to do.  It’s hard to imagine this particular question meeting anyone’s standard for rigor:  it is a simple content-recall question.  It is not designed to elicit any deep thinking or creative problem solving.

I was highly critical of the exam-makers for simply putting high school-level problem on the 8th grade exam and calling it more rigorous.  In our Twitter exchange, Grant pointed out that a 10th grade question could be considered more rigorous, however, if the 8th grader had to reason the solution out rather than simply recall a piece of content.  This excellent point led me to think about the subjective nature of rigor.

Rigor depends on the solution, and thus, the student

The following trapezoid problem from the 6th-grade exam is a good example of what Grant was talking about.  It also illustrates the subjective nature of rigor.

6th grade trapezoid question

Here, the student is asked to find the area of an isosceles trapezoid.  According to the annotations, the student is expected to decompose the trapezoid into a rectangle and two triangles, use mathematical properties of these figures to determine their dimensions, find the areas, and then combine all of the information into a final answer.  This definitely sounds like a rigorous task:  the student is expected to think of a mathematical object in several ways, connect multiple ideas through several steps, and be precise in putting everything together.

The problem would be perceived much differently, however, if the student knew the standard formula for the area of a trapezoid (that is, area is equal to the product of the average of the bases and the height).  Knowing that formula would make this a recall-and-apply-the-formula question, and not especially rigorous.  Thus, this question might be considered rigorous in the context of 6th grade mathematics, but not in the context of 8th grade mathematics.

But if a 6th grade student happens to know the formula for the area of a trapezoid, they’ll get the answer faster while sidestepping all the messy details; that is, the rigor.  And it won’t take long for students and teachers to realize that memorizing higher-level formulas might help them navigate these rigorous exams more efficiently and effectively.  These rigorous questions, whose rigor depends in part on a lack of advanced content knowledge, might actually encourage some decidedly un-rigorous behavior in students and teachers.

Grant Wiggins Responds

I agree with your first two points. Just because a question comes from a higher grade level doesn’t make it rigorous. And rigor is surely not an absolute but relative criterion, referring to the intersection of the learner’s prior learning and the demands of the question. (This will make mass testing very difficult, of course).

To me, rigor has (at least) 3 other aspects when testing: learners must face a novel(-seeming) question, do something with an atypically high degree of precision and skill, and both invent and double-check the approach and result, be it in math or writing a paper. The novel (or novel-seeming) aspect to the challenge typically means that there is some new context, look and feel, changed constraint, or other superficial oddness than what happened in prior instruction and testing. (i.e. what Bloom said had to be true of any “application” task in the Taxonomy).

I would go further: depending upon context, a problem can go from hard to easy and easy to hard. Case in point – a great example from Michalewicz and Fogel’s book on heuristics:

“There is a triangle ABC, and D is an arbitrary interior point of this triangle. Prove that AD + DB < AC + CB. The problem is so easy it seems as if there is nothing to prove. It’s so obvious that the sum of the two segments inside the triangle must be shorter than the sum of its two sides. But this problem is now removed from the context of its chapter, and outside of this context the student has no idea of whether to apply the Pythagorean Theorem, build a quadratic equation, or do something else!

triangle and interior point -- GW

“The issue is more serious than it first appears. We have given this very problem to many people, including undergraduate and graduate students, and even full professors of mathematics, engineering and computer science. Fewer than 5% of them solved this problem within an hour, many of them required several hours, and we witnessed some failure as well.” (pp. 4-5)

It is helpful here to bring in Paul Zeitz and his clear account of the difference between an exercise and a (real) problem to flesh out my claim that rigor requires the latter, not the former (which is your point, too): “An exercise is a question that you know how to resolve immediately. Whether you get it right or not depends upon how expertly you apply specific techniques, but you don’t need to puzzle out what techniques to use. In contrast, a problem demands much thought and resourcefulness before the right approach is found.”

The first two authors go on to say that real problems are very difficult to solve, for several reasons:

  • The number of possible solutions in the search space is so large as to forbid an exhaustive search for the best answer.
  • The problem is so complicated…we need to use simplified models
  • The evaluation function is ‘noisy’ or varies in time.
  • The possible solutions are so heavily constrained that constructing even 1 feasible answer is difficult.
  • The person solving the problem is inadequately prepared or imagines some psychological barrier that prevents them from discovering a solution.

To me the 2nd and 4th elements are key.

Here is one of my favorite such problems, from the TIMSS over a decade ago:

TIMSS problem -- GWHere’s what a NY Times Reporter wrote about this problem:

 “The problem is simply stated and simply illustrated. It also cannot be dismissed as being so theoretical or abstract as to be irrelevant for the technocrats of tomorrow. It might be asked about the lengths of tungsten coiled into filaments; it might come in handy in designing computer chips where distances are crucial. It seems to involve some intuition about the physical world and some challenge about how to determine something about that world.

It also turned out to be one of the hardest questions on the test. The international average of advanced mathematics students [in 12th grade] who got at least part of the question correct was only 12 percent (10 percent solved it completely). But the average for the United States was even worse: just 4 percent for a complete solution (there were no significant partial solutions).”

By EDWARD ROTHSTEIN (NYT); Business/Financial Desk, March 9, 1998, Monday

So, the challenge in math teaching – always! – is to come up with real problems, puzzling challenges that demand thought, not just recall of algorithms (mindful of the fact that in mass testing some kids will get lucky and have instant recall of such a problem and its solution).

My favorite sources? The Car Talk Puzzlers, and Math Competition books. But there is more to be said on the subject of ‘novel’ problems via mass testing.

Read Part 2 of the conversation here.

Are The New Tests More Rigorous?

The release of student test data from 2013 has educators, administrators, politicians, and parents abuzz in New York.  These are the first state exams aligned to the Common Core standards, and as widely predicted, proficiency rates have plummeted, leaving everyone scrambling to explain what has happened.

The most common explanation offered is that these new tests are substantially more rigorous than the old ones, so lower student performance is to be expected.  I was curious about the claim that the new tests are more rigorous, and while the state does not release the exams to the public, they do publish a small number of questions from each grade level.

The new tests were administered in grades 3-8.  As a high school teacher, I am not well-versed in elementary school tests, but I have spent a substantial amount of time scrutinizing New York state math Regents exams, so I thought I’d look at the 8th grade math questions that were released to the public.  I was quite surprised by what I saw.

The “representative sample” of 8th grade math questions does not seem more rigorous to me.  They do not seem to emphasize “deep analysis” or “creative problem solving over short answers and memorization”, which is often how the new standards are characterized.  I can’t say I was surprised to discover this.

What did surprise me, however, was how many of these 8th grade math questions were virtually identical to questions that have recently appeared on high school math Regents exams.

Here is the first example from the set of 8th grade math questions released to the public:

New Test Sample Q 1

This problem is essentially the same as #4 from the January, 2013 Integrated Algebra exam

January 2013 IA 4

The second example from the set of 8th grade math questions released to the public

New Test Sample Q 2

is quite similar to #4 from the January, 2013 Geometry exam

January 2013 G 4

And the fourth example from the set of 8th grade math questions released to the public

New Test Sample Q 4

is essentially the same as #9 from the January, 2013 Integrated Algebra exam

January 2013 IA 9

This surprising discovery left me with a lot of questions.

First, why are 8th graders facing the same kinds of questions on this state exam that 9th, 10th, 11th, and even 12th graders faced this year?  Were teachers and students prepared to see this kind of content on the 8th grade exam?

Second, how can it be argued that this new test is more rigorous if it is comprised of the same kinds of questions that appear on the old tests?  Simply moving a question from a 10th-grade test to an 8th-grade test doesn’t transform the question into one that requires deep analysis or creative problem solving.  More rigorous questions would ask students to construct mathematical objects, explore concepts from different perspectives, and demonstrate mathematical reasoning.  None of the above questions do this:  they are not especially challenging, deep, or novel.  In short, they are typical standardized test fare.

And perhaps the most important question is this:  if these are the hand-picked exemplar questions released to the public, what must the rest of the test look like?  Only by releasing the entire test to the public can we truly assess what we are assessing.

A version of this post appears at GothamSchools.

Presenting at MOVES Conference

moves logoI am very excited to be a part of the inaugural MOVES conference at the Museum of Mathematics in New York City!

The focus of the conference is the Mathematics of Various Entertaining Subjects, and it features an amazing lineup.  Erik Demaine, Dave Richeson, and Henry Segerman are among invited speakers, and Tim Chartier and Colm Mulcahy will be part of special evening of mathematical entertainment!

I will be running a Family Track activity at the Museum on Monday afternoon.  This workshop, Sphere Dressing, is inspired by the activity I submitted for the 2012 Rosenthal Prize.

The conference runs August 4-6.  You can find out more information here, and see the entire conference program here.

Regents Recap — June 2013: Encouraging Bad Habits

Here is another installment in my series reviewing the NY State Regents exams in mathematics.

Some of the worst exam questions aren’t merely erroneous, but actually encourage students to exercise bad mathematical habits.  Consider these questions from the June 2013 Algebra 2 / Trig exam.

2013 June A2T 4

Notice that the problem doesn’t specify what kind of series this is:  the student is expected to assume that the series is geometric.  This is a terrible habit to encourage, and I wrote about this the last time it happened on a Regents exam.  I guess no one is listening.

2013 June A2T 3

In order to determine whether a relation is one-to-one or onto, it is necessary to know the relation’s domain and range.  Here, the student is expected to assume that the oval on the left represents the domain of the relation and the oval on the right represents the range.  Perhaps these assumptions are reasonable given the nature of the diagrams, but this just seems sloppy to me.  I wouldn’t accept imprecise formulations of functions and relations like this from my students; I would demand they be more explicit.  (It’s also worth noting that the relation in (2) is one-to-one and onto its image in the right-hand oval.)

Here’s another example of sloppiness in question construction.

2013 June A2T 19

Is it supposed to be obvious that the i here is the imaginary unit?  The letter i could be just a variable, like, say, the m that also appears in the question.  The available answers support the assumption that i^2 = \sqrt{-1}, but why are we forcing students to play test-detective?

The Regents exams also continue with their long-standing tradition of presenting unscaled graphs, another bad mathematical habit to encourage.

I believe these tests should stand as exemplars of proper mathematics.  Maybe I’m alone in thinking this, but it seems to me that repeated exposure to these sloppy exam questions might actually interfere with a student’s ability to truly understand the underlying mathematics.

Follow

Get every new post delivered to your Inbox

Join other followers: