Danielson Framework criticized by Charlotte Danielson
I’ve been writing about the Danielson Framework for Teacher Evaluation for a couple of years, and in fact my “Fatal Flaws of the Danielson Framework” has been my most read and most commented on post, with over 5,000 hits to date. I’ve also been outspoken about how administrators have been misusing the Framework, resulting in demoralized teachers and unimproved (if not diminished) performance in the classroom. (See in particular “Principals unwitting soldiers in Campbell Brown’s army” and “Lowered teacher evaluations require special training.”) At present, teachers are preparing — at great time and expense — to embark on the final leg of the revamped teacher evaluation method with the addition of student performance into the mix (see ISBE’s “Implementing the Student Growth Component in Teacher and Principal Evaluation”). I’ve also written about this wrongheaded development: “The fallacy of testing in education.”
Imagine my surprise when I discovered an unlikely ally in my criticism of Charlotte Danielson’s much lauded approach: Charlotte Danielson herself. The founder of the Danielson Framework published an article in Education Week (April 18 online) that called for the “Rethinking of Teacher Evaluation,” and I found myself agreeing with almost all of it — or, more accurately and more egocentrically, I found Charlotte Danielson agreeing with me, for she is the one who has changed her tune.
My sense is that Ms. Danielson is reacting to widespread dissatisfaction among teachers and principals with the evaluation process that has been put in place which is based on her Danielson Framework. Her article appeared concurrently with a report from The Network for Public Education based on a survey of nearly 3,000 educators in 48 states which is highly critical of changes in teacher evaluation and cites said changes as a primary reason for teachers exiting the profession in droves and for young people choosing not to go into education in the first place. For example, the report states, “Evaluations based on frameworks and rubrics, such as those created by Danielson and Marzano, have resulted in wasting far too much time. This is damaging the very work that evaluation is supposed to improve . . .” (p. 2).
Ms. Danielson does not, however, place blame in her Framework, at least not directly. She does state what practically all experienced teachers have known all along when she writes, “I’m deeply troubled by the transformation of teaching from a complex profession requiring nuanced judgment to the performance of certain behaviors that can be ticked off a checklist.” Her opinion is a change from earlier comments when she said that good teaching could be easily defined and identified. In a 2012 interview, Ms. Danielson said that her assessment techniques are “not like rocket science,” whereas “[t]eaching is rocket science. Teaching is really hard work. But doing that [describing what teaching “looks like in words”] isn’t that big a deal. Honestly, it’s not. But nobody had done it.”
Instead of her Framework, then, Ms. Danielson places the lion’s share of the blame with state legislators who oversimplified her techniques via their adoptions, and — especially — with administrators who are not capable of using the Framework as it was intended. She writes, “[F]ew jurisdictions require their evaluators to actually demonstrate skill in making accurate judgments. But since evaluators must assign a score, teaching is distilled to numbers, ratings, and rankings, conveying a reductive nature to educators’ worth and undermining their overall confidence in the system.”
Amen, Sister Charlotte! Testify, girlfriend!
Ms. Danielson’s critique of administrators is a valid one, especially considering that evaluators were programmed, during their Danielson training, to view virtually every teacher as less than excellent, which put even the best-intentioned evaluators in a nitpicking mode, looking for any reason, no matter how immaterial to effective teaching, to find a teacher lacking and score them “proficient” instead of “excellent.” In her criticism of administrators Ms. Danielson has touched upon what is, in fact, a major shortcoming of our education system: The road to becoming an administrator is not an especially rigorous one — especially when it comes to academic rigor — and once someone has achieved administrative status, there tends to be no apparatus in place to evaluate their performance, including (as Ms. Danielson points out) their performance in evaluating their teachers.
Provided that administrators can keep their immediate superior (if any) content, as well as the seven members of the school board (who are almost never educators themselves), they can appear to be effective. That is, as long as administrators do not violate the terms of the contract, and as long as they are not engaging in some form of obvious harassment, teachers have no way of lodging a complaint or even offering constructive criticism. Therefore, if administrators are using the Danielson Framework as a way of punishing teachers — giving them undeservedly reduced evaluations and thus exposing them to the harms that can befall them, including losing their job regardless of seniority — there is no way for teachers to protect themselves. They cannot appeal an evaluation. They can write a letter to be placed alongside the evaluation explaining why the evaluation is unfair or invalid, but their complaint does not trigger a review of the evaluation. The evaluator’s word is final.
According to the law of averages, not all administrators are excellent; and not all administrators use the evaluation instrument (Danielson or otherwise) excellently. Some administrators are average; some are poor. Some use the evaluation instrument in a mediocre way; some use it poorly. Hence you can quite easily have an entire staff of teachers whose value to the profession is completely distorted by a principal who is, to put it bluntly, bad at evaluating. And there’s not a thing anyone can do about it.
Another crucial point that Charlotte Danielson makes in her Education Week article is that experienced teachers should not be evaluated via the same method as teachers new to the field: “An evaluation policy must be differentiated according to whether teachers are new to the profession or the district, or teach under a continuing contract. . . . Once teachers acquire this status [i.e. tenure], they are full members of the professional community, and their principal professional work consists of ongoing professional learning.” In other words, experienced teachers, with advanced degrees in their content area and a long list of professional accomplishments, shouldn’t be subjected to the same evaluation procedure as someone who is only beginning their career and has much to learn.
In fact, using the same evaluation procedure creates a very odd dynamic: You oftentimes have an administrator who has had only a limited amount of classroom experience (frequently fewer than ten years, and perhaps only two or three) and whose only advanced degree is the one that allows them to be an administrator (whereby they mainly study things like school law and school finance), sitting in judgment of a teacher who has spent twenty or thirty years honing their teaching skills and who has an advanced degree in their subject area. What can the evaluator possibly say in their critique that is meaningful and appropriate? It is commonplace to find this sort of situation: A principal who was a physical education or drivers education teacher, for perhaps five years, is now sitting in an Advanced Placement Chemistry classroom evaluating a twenty-year veteran with a masters degree or perhaps even a Ph.D. in chemistry. The principal feels compelled to find something critical to say, so all they can do is nitpick. They can’t speak to anything of substance.
What merit can there be in a system that makes evaluators omnipotent judges of teachers in subject areas that the evaluators themselves literally are not qualified to teach? It isn’t that veteran teachers don’t have anything to learn. Far from it. Teaching is a highly dynamic, highly challenging occupation; and the successful teacher is constantly learning, growing, self-reflecting, and networking with professional peers. The successful principal makes space for the teacher to teach and for the student to learn, and they protect that space from encroachment by anyone whose design is to impede that critical exchange.
Ms. Danielson offers this alternative to the current approach to evaluation: “An essential step in the system should be the movement from probationary to continuing status. This is the most important contribution of evaluation to the quality of teaching. Beyond that, the emphasis should be on professional learning, within a culture of trust and inquiry. . . . Experienced teachers in good standing should be eligible to apply for teacher-leadership positions, such as mentor, instructional coach, or team leader.”
Ironically, what Ms. Danielson is advocating is a return to evaluation as most teachers knew it prior to adoption of the Danielson Framework.
(Grammar alert: I have opted to use the gender-neutral pronouns they and their etc. even when they don’t agree in number with their antecedents.)
Principals unwitting soldiers in Campbell Brown’s army
(This is a long post — and for that, my apologies. But it’s important, and I encourage you to take your time and read it thoroughly.)
Because of my interest in the subject (as demonstrated in my blog posts over the past few months), I was invited to participate in a video roundtable via Skype with administrators from several schools about implementing the Danielson Framework for Teacher Evaluation, and I found many of the comments, well, bewildering. Even though it was a select group, I strongly suspect that their attitudes and approaches are representative of not only administrators in Illinois, but across the country — as the Danielson Framework has been adopted by numerous states. Before I go any further I must stress that these are all good people who are trying to do their job as they understand it from the State Board of Education, their own local school boards and the public at large. Around the video table were a superintendent of a k-12 district, building principals of elementary, middle, junior high and high schools, and even a k-12 curriculum director, along with three teachers — elementary, junior high and high school (yours truly). I’m going to try to represent their words accurately, but without attribution since their comments were not on the record. In fact, as the two-hour video chat became more heated, several people were speaking with a good deal of candor, and clearly their remarks were not intended for all ears. (By the way, kudos to the tech folks who brought us all together — it worked far better than I would have suspected.)
I considered not writing about the video conference at all, but ultimately felt that I owe it to the profession that I’ve devoted my adult life to (as I enter my 31st year in the classroom), a profession that has been beleaguered in recent years by powerful forces on every side: attacking teachers’ integrity, our skills, our associations, our job security, our pensions. We feel we have so many enemies, we don’t even know where to focus our attention.
What is more, most teachers are afraid to speak candidly with their own administrators and they’re especially afraid to speak out about what’s going on in their buildings. In spite of education reformers blanketing the media with the myth of “powerful teachers unions,” the truth is that associations like the National Education Association and American Federation of Teachers aren’t all that powerful — if they were, would teachers be in the plight we are now? — and individual teachers are very vulnerable. Nontenured teachers can be terminated without cause, and tenured teachers can be legally harassed right out of the profession. In fact, it happens all the time. Moreover, teachers tend to be naturally non-confrontational, which is why they chose to go into teaching in the first place. People with more aggressive personalities will seek other kinds of professions. As a result, we’ve been lambs to slaughter at the hands of reformers, legislators, school board members, administrators … at the hands of anyone who wants to take a whack at us. Rather than fight back, it’s easier to keep quiet and bear it, or to move on.
I’ve been writing about educational issues for the past several months — the unfair termination of young teachers, the inherent flaws of the Danielson Framework, the way the Framework affects teachers, and my issues with PARCC and the Common Core. My posts have been garnering hundreds of hits, and a few online likes, but many, many private, under-the-radar thumbs-ups and thank-yous. Teachers appreciate that someone is speaking out, but they’re not only afraid to speak out themselves, they’re even afraid to be seen agreeing with my point of view. If this isn’t evidence of the precariousness of being a teacher and the overall weakness of “teachers unions,” I don’t know what is.
Public Opinion and the Rarefied Air of Excellence
Much of the round-table discussion had to do with the Framework’s insistence that very, very few teachers rank in the top category (identified as “Excellent” in many districts’ plans). Before Danielson, districts tended to have three-tier evaluation instruments, which were often labeled as “Excellent,” “Satisfactory” and “Unsatisfactory.” Danielson adds a tier between “Excellent” and “Satisfactory”: “Proficient.” Many veteran teachers who had consistently received an excellent rating under the previous model were downgraded to merely proficient under Danielson. This downgrading was predicted as early as two years ago when the new instrument emerged on the educational horizon. I didn’t want to believe it would be that severe, but it has been this past year, the year of implementation, with very few teachers being rated as excellent. For the record, I was rated as proficient — not as excellent for the first time since I was a nontenured teacher, more than 25 years ago.
In fact, as I wrote in a previous post, the Illinois Administrators Academy offered a special workshop this past summer to train administrators how to deliver the unpleasant news that a veteran teacher has been downgraded to proficient — the downgrading was so pervasive across the state. The Framework was originally developed by Charlotte Danielson in 1996 as a way to evaluate first-year teachers, so it made perfect sense that a single-digit percentage would be deemed as excellent. The Framework has undergone three revisions since then and now purports to be an instrument that can assess every teacher, K-12, every subject, and even nonclassroom professionals like librarians and school nurses. Nevertheless, the notion that very, very few teachers will rate as excellent has clung tenaciously to the Framework throughout each revision.
I asked the administrators why that aspect of the Framework remains even though the Framework’s purpose has been expanded dramatically since it was conceived in the mid-1990s. I was told by the k-12 superintendent that the Framework has gained such wide acceptance in large part due to that very aspect. Under previous evaluation instruments, 90% of teachers were judged to be excellent, and the public doesn’t accept that as true. In fact, the public believes (and therefore school boards, too, since they, like the public at large, are almost always noneducators with no classroom experience) that the traditional bell curve should apply to teachers. The bell curve, or Gaussian function, is of course the statistical representation that says the fewest examples of anything, qualitatively speaking, are at either extreme of the gathered samples, and the vast majority (let’s say 80%) fall somewhere in the middle, from below average to above average.
According to the superintendent, then, the public believes that the bell curve should apply to experienced, career teachers as well — that only a small percentage are truly excellent, and the vast majority fall somewhere in the middle (to use Danielson terms, in the satisfactory to proficient range). First of all, who cares what the uninformed public thinks? In our country we have a fascination with asking pedestrians on the street what they think of global warming, heightened military involvement in the Middle East, and allowing Ebola victims to enter the country. John Oliver of “Last Week Tonight” did a segment on this phenomenon that went viral on social media:
Assuming this is true — that the public believes only a small percentage of teachers are excellent based, unconsciously, on the principle of the bell curve (and I’m willing to believe that it is true) — the belief yet again speaks to the ignorance of the “man on the street.” In this instance, the bell curve is being fallaciously applied. If you take a random sampling of people (let’s say, you go to the mall at Christmas time and throw a net around a random group of shoppers) and task them with teaching some random topic to a random group of students, then, yes, the bell curve is likely to be on target. In that group of shoppers, lo and behold, you netted a couple of professional teachers, so they’re able to teach the material pretty effectively; another much larger group of shoppers who are decently educated and reasonably articulate could do a passable job imparting the information; and a smaller group on the other extreme would really make a botch of it.
But career teachers are not a random sampling of shoppers at the mall. They’re highly educated professionals who have devoted their lives to teaching, who have constantly worked to improve their craft, and who have honed their skills via thousands of contact hours with students. It stands to reason, in fact, that career teachers should be excellent at what they do after all that training and experience. No one, I suspect, would have an issue with the statement that all Major League baseball players are excellent at baseball — some may be bound for Cooperstown and some may go back to the minors or to some other career altogether after a season or two, but they’re all really, really good at playing baseball compared to the average person. Why is it so hard to believe that 90% of career teachers are excellent at what they do?
The Fallacy of the Bell Curve and Nontenured Teachers
Unfortunately, the acceptance of the bell-curve fallacy has an even more devastating impact when applied to teachers in the beginning of their careers. One administrator shared that her board expects a few nontenured teachers to be terminated every spring, that the board implies the administrators aren’t doing their jobs if every nontenured teacher is retained. I was dumbfounded by this statement. It’s barely a figurative comparison to say that it’s like having to sacrifice a virgin or two to appease the gods at the vernal equinox. It’s no wonder that many young teachers feel as if they’re performing their highly complex duties with a Damoclesian Sword poised above their tender necks. I know firsthand one young teacher who resigned last spring after two years in the classroom to pursue another career option because she’d seen the way other young teachers were treated and had already experienced some administrative harassment. And this was a teacher who by all accounts was doing well in the classroom (in a specialized area in which there aren’t a lot of qualified candidates). She didn’t even know what she wanted to do for a living, but it will have to be better (and professionally safer) than teaching, she believed. I have to believe she’s right.
But, again, in the case of young teachers, the bell curve is being applied erroneously. Generally speaking, when teachers are hired, administrators are drawing from an applicant pool in the hundreds. They’re college educated, trained in their field, and they’ve passed their professional exams. They often have to go through multiple rounds of interviewing before being offered a position. Of course, even after all of this, there can be young teachers who have chosen their profession poorly and in fact they’re not cut out for teaching — but school board members shouldn’t just assume a certain number should be cut from the herd to make room for potentially more effective young professionals — and if that sort of pressure is being applied to administrators, to be the bearers of the bloody hatchet every spring, that is grossly unfair, too.
The Danielson Group’s Indoctrination
The evaluation training that administrators have to undergo, all forty hours of it, indoctrinates them to the Danielson Framework’s ethos that excellent is all but attainable, and it has led to all kinds of leaping logic and gymnastic semantics. An idea that was expressed multiple times in various ways during the roundtable was that proficient really means excellent, and a rating of excellent really means something beyond excellent — what precisely is unclear, but it has to do with teachers going above and beyond (above what? beyond where? … no one seems to know or be able to articulate). The Framework was often referred to as “fact-based” and “objective,” yet administrator after administrator couldn’t put into words what distinguishes a “proficient” teacher from an “excellent” one. It’s just a certain feeling — which is the very definition of subjectivity. The Framework for Teacher Evaluation approach is fact-laden, but it is far from fact-based.
The Danielson model is supposed to be an improvement over previous ones in part because it requires evaluators to observe teachers more than in the past. In the old system, typically, tenured teachers were observed one class period every other year. Now they’re observed one class period plus several pop-in visits, which may last only a few minutes, every other year. The Framework recommends numerous visits, even for veteran teachers, but in practicality evaluators are doing well to pop in a half dozen times or so because they have so many teachers to evaluate. Nevertheless, the increased frequency seems to give administrators the sense that they have a secure hold on the behaviors of their teachers and know with confidence what they’re doing in their classes. This confidence, frankly, is troubling. Let’s be generous and say that a principal can observe a teacher for a total of three class periods (one full period, plus bits of four or five other ones). Meanwhile, the typical teacher teaches, say, six periods per day for 180 days, which equals 1,080 periods. Three class periods represent less than one percent (0.3 percent, rounding up) of that teacher’s time with students during the year. How in the world can an evaluator say with confidence Teacher A is excellent and Teacher B is really close, but definitely only proficient based on seeing them teach less than one percent of the time?
Yet one principal said with confidence, bravado even, that he could observe two high-performing teachers who had always been rated as excellent in the past, and based on his Danielson-style observations he could differentiate between the excellent high-performing teacher and the proficient high-performing teacher, because, he said, the excellent teacher was doing something consistently, whereas the proficient teacher was doing that something only some of the time — what that something is was left undefined. If a writer submitted an academic article to a peer-reviewed journal and was drawing rock-solid conclusions based on observing anything .03% of the time … well, let us say that acceptance for publication would be unlikely.
The same standards of logic should be applied to judging teachers’ careers and assessing their worth to the profession. Period.
The Portfolio Conundrum
The confident administrator may point to another component of the Danielson model that is supposed to be an improvement over the previous approach: a portfolio prepared by the teacher. Teachers are supposed to provide their evaluator with evidence regarding their training and professionalism (especially for Danielson domains 1 and 4, “Planning and Preparation” and “Professional Responsibilities”), but there are some inherent problems with this approach and a lot of confusion. As far as confusion, principals seem to be in disagreement about how much material teachers should provide them. Some suggest only a few representative items, but the whole idea is for the portfolio to fill in the blanks for the evaluator, to make the evaluator aware of professional behaviors and activities that he or she can’t observe in the classroom (especially when they’re observing a teacher less than one percent of their time with students!). However, if teachers hand in thick portfolios, filled with evidence, the overburdened principal (and I’m not being sarcastic here), the overburdened principal hardly has time to pore through dozens of portfolios that look like they were prepared by James Michener (I debated between Michener and Tolstoy) — which leaves teachers in a conundrum: Do they turn in a modest amount of evidence, thereby selling themselves short, or do they submit copious amounts of evidence that won’t be read and considered by their evaluator anyway?
And it’s a moot question, of course, since nearly all teachers are going to be lumped into the proficient category to satisfy the public’s erroneous bell-curve expectations.
The Undervaluing of Content
I’ll add one bit more from the conversation because it leads to another important point — perhaps the most important — and that is one principal’s statement that he mainly focuses on a teacher’s delivery of the material and not the validity of the content because he usually doesn’t have the background in the subject area. In larger school districts, there may be department chairs who are at least in part responsible for evaluating teachers in their department (so an English teacher evaluates an English teacher, or a math teacher, a math teacher, etc.), but the vast majority of evaluations, for tenured and nontenured teachers alike, are performed by administrators outside of the content area. This, frankly, has always been a problem and largely invalidates the entire teacher evaluation system, but when the system was mainly benign, no one fussed too much about it (not even me). Now, however, when tenure and seniority laws have been weakened, and principals are programmed to be niggardly with excellent ratings, the fact that evaluators oftentimes have no idea if the teacher is dispensing valid knowledge or not undermines the whole approach.
Not to mention, the Danielson Framework claims to place about fifty percent of a teacher’s effectiveness on his or her knowledge of the subject. The portfolios are supposed to help with this dilemma (the portfolios that aren’t being read with any sort of care because of time issues). I’m dubious, though, that this is a legitimate concern of the framers of the Danielson Framework because they definitely privilege an approach to teaching that places the burden of knowledge production with the students. That is, ideally teachers are facilitating their students’ acquisition of knowledge through self-discovery, but they’re not imparting that knowledge to them directly. Indeed, excellent teachers do very little direct teaching at all.
This devaluation of content-area knowledge has been a growing trend for several years, and it’s not surprising that administrators are easily swayed toward this mindset. After all, teachers who go into administration have made the choice to pursue knowledge not in their subject-area field. Very, very few administrators have a masters degree in their original content area in addition to their administrative degrees and certificates. In theory, they may accept the idea that broader and deeper knowledge in your subject area is important, but they can’t truly understand just how valuable (even invaluable) it is since they didn’t teach as someone with an advanced degree in their field. They’re only human after all, and none of us can truly relate to an experience we haven’t had ourselves.
Campbell Brown and Her Unwitting Campbell Brown-shirts
We didn’t talk about this during the video round-table, but it seems clear to me that none of the administrators had any sense of the role they’re playing in the larger scheme of things. The players are too numerous and the campaign too complex to get into here in any depth, but there’s unquestionably a movement afoot to privatize education — that is, to take education out of the hands of trained professionals and put it in the hands of underpaid managers so that corporations can reap obscene profits, and turn traditional public schools into barely funded welfare institutions. The well-to-do will be able to send their sons and daughters to these corporate-backed charter schools, and middle-class parents can dig their infinite hole of financial debt even deeper in an effort to keep up and send their children to the private, corporate schools as well.
Campbell Brown and the Partnership for Educational Justice were behind the lawsuit that made teacher tenure unconstitutional in California (the Vergara decision), and they’re at it again in New York (Wright v. New York). The Danielson Framework, wielded by brainwashed administrators, is laying the groundwork for Vergara-like lawsuits across the land. Imagine how much easier it will be for Brown and partners in “reform” like David Boles to make the case that public schools are failing because, see, only a handful of teachers are performing at the top of their field. The rest, 90-something percent, are varying shades of mediocre, with powerful teachers unions shielding their mediocrity from public view.
Superintendents and principals have drunk the Campbell Brown-colored Kool-Aid. In this instance the metaphor is especially apropos because there are already movements underway to dismiss traditionally trained administrators as underqualified. In Illinois, the State Board of Education is changing from certificates to licenses and in the process requiring additional training to become an administrator. It is a recent change, but already there are insinuations that administrators who received the traditional training are going to be underqualified compared to their newly licensed colleagues.
Moreover, what does it say about a principal as recruiter of young talent when a significant number of his new hires have to be terminated year after year? What a waste of money and resources, and what a disservice to children! And what does it say about a principal as educational leader of his building when he can’t even shape the majority of his veteran teachers into excellent practitioners? Clearly, he’s not especially excellent either. And all those well-paid superintendents who hired all those lackluster principals, well … And all those publicly elected boards of education who hired all those lackluster superintendents, well … the gross mismanagement of taxpayer dollars is bordering on criminal fraud.
As I see it, the Partnership for Educational Justice’s grand scheme is to have principals help them dismantle professional associations like the NEA and AFT via their use of the Danielson Framework, state by state. Then they’ll systematically replace public schools with corporate-backed charter schools which will be staffed by undertrained, low-paid “teachers,” and instead of principals, each school/franchise will be overseen by a manager — just as it works in the corporate world now. Instead of boards of education who answer to taxpayers there will be boards of directors who answer to shareholders. Brilliant.
So every time principals sign an evaluation that undervalues their teachers, they’re also signing their own resignation letter. It’s all right: they’ll look quite fetching in their Brown-shirts as they wait in the unemployment line.
Here’s my beef with PARCC and the Common Core
Beginning this school year students in Illinois will be taking the new assessment known as PARCC (Partnership for Assessment of Readiness for College and Careers), which is also an accountability measure — meaning that it will be used to identify the schools (and therefore teachers) who are doing well and the ones who are not, based on their students’ scores. In this post I will be drawing from a document released this month by the Illinois State Board of Education, “The top 10 things teachers need to know about the new Illinois assessments.” PARCC is intended to align with the Common Core, which around here has been rebranded as the New Illinois Learning Standards Incorporating the Common Core (clearly a Madison Avenue PR firm wasn’t involved in selecting that name — though I’m surprised funds weren’t allocated for it).
This could be a very long post, but I’ll limit myself to my main issues with PARCC and the Common Core. The introduction to “The top 10 things” document raises some of the most fundamental problems with the revised approach. It begins, “Illinois has implemented new, higher standards for student learning in all schools across the state.” Let’s stop right there. I’m dubious that rewording the standards makes them “higher,” and from an English/language arts teacher perspective, the Common Core standards aren’t asking us to do anything different from what we’ve been doing since I started teaching in 1984. There’s an implied indictment in the opening sentence, suggesting that until now, the Common Core era, teachers haven’t been holding students to particularly high standards. I mean, logically, if there was space into which the standards could be raised, then they had to be lower before Common Core. It’s yet another iteration of the war-cry: Teachers, lazy dogs that they are, have been sandbagging all these years, and now they’re going to have to up their game — finally!
Then there’s the phrase “in all schools across the state,” that is, from the wealthiest Chicago suburb to the poorest downstate school district, and this idea gets at one of the biggest problems — if not the biggest — in education: grossly inequitable funding. We know that kids from well-to-do homes attending well-to-do schools do significantly better in school — and on assessments! — than kids who are battling poverty and all of its ill-effects. Teachers associations (aka, unions) have been among the many groups advocating to equalize school funding via changes to the tax code and other laws, but money buys power and powerful interests block funding reform again and again. So until the money being spent on every student’s education is the same, no assessment can hope to provide data that isn’t more about economic circumstances than student ability.
As if this disparity in funding weren’t problematic enough, school districts have been suffering cutbacks in state funding year after year, resulting in growing deficits, teacher layoffs (or non-replacement of retirees), and other direct hits to instruction.
According to the “The top 10 things” document, “[a] large number of Illinois educators have been involved in the development of the assessment.” I have no idea how large a “large number” is, but I know there’s a big difference between involvement and influence. From my experience over the last 31 years, it’s quite common for people to present proposals to school boards and the public clothed in the mantle of “teacher input,” but they fail to mention that the input was diametrically opposed to the proposal.
The very fact that the document says in talking point #1 that a large number of educators (who, by the way, are not necessarily the same as teachers) were involved in PARCC’s development tells us that PARCC was not developed by educators, and particularly not by classroom teachers. In other words, this reform movement was neither initiated nor orchestrated by educators. Some undefined number of undefined “educators” were brought on board, but there’s no guarantee that they had any substantive input into the assessment’s final form, or even endorsed it. I would hope that the teachers who were involved were vocal about the pointlessness of a revised assessment when the core problems (pun intended), like inadequate funding, are not being addressed. At all.
“The top 10 things” introduction ends with “Because teachers are at the center of these changes and directly contribute to student success, the Illinois State Board of Education has compiled a list of the ten most important things for teachers to know about the new tests.” In a better world, the sentence would be Because teachers are at the center of these changes and directly contribute to student success … the Illinois State Board of Education has tasked teachers with determining the best way to assess student performance. Instead, teachers are being given a two-page handout, which is heavy in snazzy graphics, two to three weeks before the start of the school year. In my district, we’ve had several inservices over the past two years regarding Common Core and PARCC, but our presenters had practically no concrete information to share with us because everything was in such a state of flux; as a consequence, we left meeting after meeting no better informed than we were after the previous one. Often the new possible developments revised or even replaced the old possible developments.
The second paragraph of the introduction claims that PARCC will “provide educators with reliable data that will help guide instruction … [more so] than the current tests required by the state.” I’ve already spoken to that so-called reliable data above, but a larger issue is that this statement assumes teachers are able to analyze all that data provided by previous tests in an attempt to guide instruction. It happens, and perhaps it happens in younger grades more so than in junior high and high school, but by and large teachers are so overwhelmed with the day-to-day — minute-to-minute! — demands of the job that there’s hardly time to pore through stacks of data and develop strategies based on what they appear to be saying about each student. Teachers generally have one prep or planning period per day, less than an hour in length. The rest of the time they’re up to their dry-erase boards in kids (25 to 30 or more per class is common). In that meager prep time and whatever time they can manage beyond that, they’re writing lesson plans; grading papers; developing worksheets, activities, tests, etc.; photocopying worksheets, activities, tests, etc.; contacting or responding to parents or administrators; filling out paperwork for students with IEPs or 504s; accommodating students’ individual needs, those with documented needs and those with undocumented ones; entering grades and updating their school websites; supervising hallways, cafeterias and parking lots; coaching, advising, sponsoring, chaperoning. . . .
Don’t get me wrong. I’m a scholar as well as a teacher. I believe in analyzing data. I’d love to have a better handle on what my students’ specific abilities are and how I might best deliver instruction to meet their needs. But the reality is that that isn’t a reasonable expectation given the traditional educational model — and it’s only getting worse in terms of time demands on teachers, with larger class sizes, ever-changing technology, and — now — allegedly higher standards.
Educational reformers are so light on classroom experience they haven’t a clue how demanding a teacher’s job is at its most fundamental level. In this regard I think education suffers from the fact that so many of its practitioners are so masterful at their job that their students and parents and board members and even administrators get the impression that it must be easy. Anyone who is excellent at what she or he does makes it look easy to the uninitiated observer.
I touched on ever-changing technology a moment ago; let me return to it. PARCC is intended to be an online assessment, but, as the document points out, having it online in all schools is unrealistic, and that “goal will take a few more years, as schools continue to update their equipment and infrastructure.” The goal of its being online is highly questionable in the first place. The more complicated one makes the assessment tool, the less cognitive processing space the student has to devote to the given question or task. Remember when you started driving a car? Just keeping the darn thing on the road was more than enough to think about. In those first few hours it was difficult to imagine that driving would become so effortless that one day you’d be able to drive, eat a cheeseburger, sing along with your favorite song, and argue with your cousin in the backseat, all simultaneously. At first, the demands of driving the car dominated your cognitive processing space. When students have to use an unfamiliar online environment to demonstrate their abilities to read, write, calculate and so on, how much will the online environment itself compromise the cognitive space they can devote to the reading, writing and calculating processes?
What is more, PARCC implies that schools, which are already financially strapped and overspending on technology (technology that has never been shown to improve student learning and may very well impede it), must channel dwindling resources — whether local, state or federal — to “update their equipment and infrastructure.” These are resources that could, if allowed, be used to lower class sizes, re-staff libraries and learning centers, and offer more diverse educational experiences to students via the fine arts and other non-core components of the curriculum. While PARCC may not require, per se, schools to spend money they don’t have on technology, it certainly encourages it.
What is even more, the online nature of PARCC introduces all kinds of variables into the testing situation that are greatly minimized by the paper-and-pencil tests it is supplanting. Students will need to take the test in computer labs, classrooms and other environments that may or may not be isolated and insulated from other parts of the school, or off-site setting. Granted, the sites of traditional testing have varied somewhat — you can’t make every setting precisely equal to every other setting — but it’s much, much easier to come much, much closer than when trying to do the test online. Desktop versus laptop computers (in myriad models), proximity to Wi-Fi, speed of connection (which may vary minute from minute), how much physical space can be inserted between test-takers — all of these are issues specific to online assessments, and they all will affect the results of the assessment.
So my beef comes down to this about PARCC and the Common Core: Hundreds of millions of dollars have been spent rewording standards and developing a new assessment that won’t actually help improve education. Here’s what would help teachers teach kids:
1. Equalize funding and increase it.
2. Lower class sizes, kindergarten through 12th grade, significantly — maximum fifteen per class, except for subjects that benefit from larger classes, like music courses.
3. Treat teachers better. Stop gunning for their jobs. Stop dismantling their unions. Stop driving them from the profession with onerous evaluation tools, low pay and benefits, underfunded pensions, too many students to teach to do their job well, and ridiculous mandates that make it harder to educate kids. Just stop it.
But these common sense suggestions will never fly because no one will make any money off of them, let alone get filthy rich, and education reform is big business — the test developers, textbook companies, technology companies, and high-priced consultants will make sure the gravy train of “reform” never gets derailed. In fact, the more they can make it look like kids are underachieving and teachers are underperforming, the more secure and more lucrative their scam is.
Thus PARCC and Common Core … let the good times roll.
Lowered teacher evaluations of Danielson Framework require special training
In an earlier post I analyzed the “Danielson Framework for Teacher Evaluation,” which has become the adopted model in numerous states, including Illinois, and I pointed out some of its many flaws. One of the aspects of Danielson that has been troubling to teachers from the beginning is its insistence that virtually no teacher is excellent (distinguished, outstanding). When the Framework was designed in 1996 it was intended to rate first-year teachers, so it made sense that very, very few would be rated in the top category. The Framework was revised three times (2007, 2011 and 2013) in an effort to be an evaluation tool for all educators and even non-classroom professionals (like librarians and school nurses). Nevertheless, the idea that virtually no teacher is capable of achieving the top echelon (however it may be labeled in a district’s specific evaluation instrument) has clung to the Framework.
In my district, we were told of the Danielson Framework a full two years before it was implemented, and from the start we were informed that it was all but impossible to achieve an “excellent” rating, even for teachers who have consistently been rated at the top level for several evaluation cycles (pre-Danielson era). After a full year of its being used, it seems that administrators’ predictions were true (or made to be true), and almost no one (or literally no one) received an excellent rating. We were encouraged to compile a substantial portfolio of evidence or artifacts to help insure that our assessment would be more comprehensive than the previous evaluation approach. I foolishly (in retrospect) spent approximately six hours pulling together my portfolio and writing a narrative to accompany it. A portfolio, as it turned out, we never discussed and could only have been glanced at given the timing of its being retrieved and the appointed hour of my conference.
As predicted, I was deemed “proficient.” It was a nearly surreal experience to be complimented again and again only to be informed at the end that I didn’t rate as “excellent” because the Danielson Framework makes it exceptionally difficult for a teacher to receive a top rating. There were literally no weaknesses noted — well, there were comments in the “weakness” areas of the domains, but they were phrased as “continue to …” In other words, I should improve by continuing to do what I’ve been doing all along. In fairness, I should note that the evaluator had numerous teachers to evaluate, therefore observations to record, portfolios to read, summative evaluations to write — so I’m certain the pressure of deadlines figured into the process. Nevertheless, it’s the system that’s in place, and my rating stands as a reflection of my merits as a teacher and my value to the district and the profession — there’s no recourse for appeal, nor, I suppose, purpose in it.
I was feeling a lot of things when I left my evaluation conference: angry, humiliated, defeated, underappreciated, naive, deceived (to list a few). And, moreover, I had zero respect for the Danielson Framework and (to be honest) little remained for my evaluator — though it seems that from the very beginning evaluators are trained (programmed) to give “proficient” as the top mark. After a year of pop-in observations in addition to the scheduled observation, the preparation of a portfolio based on the four domains, a conference, and the delivery of my official evaluation, I literally have no idea how to be a better teacher. Apparently, according to the Framework, I’m not excellent, and entering my fourth decade in the classroom I’m clueless how to be excellent in the World According to Charlotte Danielson (who, by the way, has very little classroom experience).
If the psychological strategy at work is that by denying veteran teachers a top rating, they will strive even harder to achieve the top next time around, it’s an inherently flawed concept, especially when there are no concrete directions for doing things differently. As I said in my previous post on Danielson, it would be like teachers telling their students that they should all strive for an “A” and do “A”-quality work — even though in the end the best they can get on their report card is a “B.” Or business owners telling their salespeople to strive for through-the-roof commissions, even though no matter how many sales they make, they’re all going to get the same modest paycheck. In the classroom, students would quickly realize that the person doing slightly above average work and the person doing exceptional work are both going to get a “B” … so there’s no point in doing exceptional work. On the job, salespeople would opt for the easiest path to the same result.
Under Danielson, it will take great personal and professional integrity to resist the common-sense urge to be the teacher that one’s evaluation says one is — to resist being merely proficient if that, in practice, is the best ranking that is available.
My experience regarding the Danielson Framework is not unique in my school, and clearly it’s not unique in Illinois as a whole. Each year administrators must participate in an Administrators Academy workshop, and one workshop being offered by the Sangamon County Regional Office of Education caught my eye in particular: “Communicating with Staff Regarding Performance Assessment,” presented by Dr. Susan Baker and Anita Plautz. The workshop description says,
“My rating has always been “excellent” [sic] and now it’s “basic”. [sic] Why are you doing this to me?” When a subordinate’s performance rating declines from the previous year, how do you prepare to deliver that difficult message? How do you effectively respond to a negative reaction from a staff member when they [sic] receive a lower performance rating? This course takes proven ideas from research and weaves them into practical activities that provide administrators with the tools needed to successfully communicate with others in difficult situations. (Sangamon Schools’ News, 11.3, spring 2014, p. 11; see here to download)
Apparently, then, school administrators are giving so many reduced ratings to teachers that they could benefit from special coaching on how to deliver the bad news so that the teacher doesn’t go postal right there in their office (I was tempted). In other words, the problem isn’t an instrument and an approach that consistently undervalues and humiliates experienced staff members; the problem, rather, is rhetorical — how do you structure the message to make it as palatable as possible?
While I’m at it, I have to point out the fallacious saw of citing “research,” and in this description even “proven ideas,” which is so common in education. The situation that this workshop speaks to, with its myriad dynamics, is unique and only recently a pervasive phenomenon. Therefore, if there have been studies that attempt to replicate the situation created by the Danielson Framework, they must be recent ones and could at best suggest some preliminary findings — they certainly couldn’t prove anything. If the research is older, it must be regarding some other communication situation which the workshop presenters are using to extrapolate strategies regarding the Danielson situation, and they shouldn’t be trying to pass it off as proof. As a literature person, I’m also amused by the word “weaves” in the description as it is often a metaphor for fanciful storytelling — and the contents of the alluded to research must be fanciful indeed. (By the way, I don’t mean to imply that Dr. Baker and Ms. Plautz are trying to deliberately mislead — they no doubt intend to offer a valuable experience to their participants.)
What is more, a lowered evaluation is not just a matter of hurting one’s pride. With recent changes in tenure and seniority laws in Illinois (and likely other states), evaluations could be manipulated to supersede seniority and remove more experienced teachers in favor of less experienced ones — which is why speaking out carries a certain amount of professional risk even for seasoned teachers.
My belief is that the Danielson Framework and the way that it’s being used are part of a calculated effort to cast teachers as expendable cogs in a broken wheel. Education reform is a billions-of-dollars-a-year industry — between textbook publishers, software and hardware developers, testing companies, and high-priced consultants (like Charlotte Danielson) — and how can cash-strapped states justify spending all those tax dollars on reform products if teachers are doing a damn fine job in the first place? It would make no sense.
It would make no sense.
The Loss of Intellect by Ted Morrissey
I appreciate NAR’s invitation to contribute to its blog.
My review of William H. Gass’s novel Middle C for NAR was a warm-up for a longer critical paper that I’ll present at the Louisville Conference on Literature and Culture Since 1900, and in preparing to write that paper I re-read several of Gass’s essays and interviews, including an interview from 1995 that was published in the Hungarian Journal of English and American Studies 3.1 (1997), and reprinted in Conversations with William H. Gass (2003), edited by Theodore G. Ammon.
The interviewer, Idiko Kaposi, asked Gass his view on emerging (mid-90s) technologies and how they would affect writing, reading, and ultimately, thinking. As a teacher, mainly of eighteen-year-olds, looking back at Gass’s remarks from nearly two decades ago, I find his insights disturbingly accurate. Gass, besides being an award-winning novelist and literary critic, was also a professor of philosophy at Washington University in St. Louis, since retired.
Gass suspected that the…
View original post 487 more words
Not speaking about Danielson Framework per se, but
Sir Ken Robinson has several TED Talks regarding education, and his “How to Escape Education’s Death Valley” is an especially appropriate follow-up to my last post about the Danielson Group’s Framework for Teaching Evaluation Instrument. Robinson, who is very funny and engaging, doesn’t reference Charlotte Danielson and her group per se, but he may as well. The Danielson Group’s Framework, which has been adopted as a teacher evaluation instrument in numerous states, including Illinois, is emblematic — in fact, the veritable flagship — of everything that’s wrong with education in America, according to Robinson.
Treat yourself to twenty minutes of Robinson’s wit and wisdom:
Fatal flaws of the Danielson Framework
The Danielson Group’s “Framework for Teaching Evaluation Instrument” has been sweeping the nation, including my home state of Illinois, in spite of the fact that the problems with the Group, the Framework, the Instrument, and even Ms. Danielson herself are as obvious as a Cardinals fan in the Wrigley Field bleachers. There have already been some thorough critiques of the Danielson Group, its figurehead, the Framework, and how it’s being used destructively rather than constructively. For example, Alan Singer’s article at the Huffington Post details some of the most glaring problems. I encourage you to read the article, but here are some of the highlights:
[N]obody … [has] demonstrated any positive correlation between teacher assessments based on the Danielson rubrics, good teaching, and the implementation of new higher academic standards for students under Common Core. A case demonstrating the relationship could have been made, if it actually exists.
[I]n a pretty comprehensive search on the Internet, I have had difficulty discovering who Charlotte Danielson really is and what her qualifications are for developing a teacher evaluation system … I can find no formal academic resume online … I am still not convinced she really exists as more than a front for the Danielson Group that is selling its teacher evaluation product. [In an article archived at the Danielson Group site, it describes the “crooked road” of her career, and I have little doubt that she’d be an interesting person with whom to have lunch — but in terms of practical classroom experience as a teacher, her CV, like most educational reformers’, is scant of information.]
The group’s services come at a cost, which is not a surprise, although you have to apply for their services to get an actual price quote. [Prices appear to range from $599 per person to attend a three-day workshop, $1,809 per person to participate in a companion four-week online class. For a Danielson Group consultant, the fee appears to be $4,000 per consultant/per day when three or more days are scheduled, and $4,500 per consultant/per day for one- to two-day consultations (plus travel, food and lodging costs). There are fees for keynote addresses, and several books are available for purchase.]
As I’ve stated, you should read Mr. Singer’s article in its entirety, and look into the Danielson Group and Charlotte Danielson yourself. The snake-oil core of their lucrative operation quickly becomes apparent. One of the chief purposes of the Danielson Framework, which allegedly works in conjunction with Common Core State Standards, is to turn students into critical readers who are able to dissect text, comprehending both its explicit and implicit meanings. What follows is my own dissection of the “Framework for Teaching Evaluation Instrument” (2013 edition). For now, I’m limiting my analysis to the not quite four-page Introduction, which, sadly, is the least problematic part of the Framework. The difficulties only increase as one reads farther and farther into the four Domains. (My citations refer to the PDF that is available at DanielsonGroup.org.)
First of all, the wrongheadedness of teacher evaluation
Before beginning my dissection in earnest, I should say that, rubrics aside, the basic idea of teacher evaluation is ludicrous — that sporadic observations, very often by superiors who aren’t themselves qualified to teach your subject, result in nothing especially accurate nor useful. As I’ve blogged before, other professionals — physicians, attorneys, business professionals, and so on — would never allow themselves to be assessed as teachers are. For one thing, and this is a good lead-in to my analysis, there are as many styles of teaching as there are of learning. There is no “best way” to teach, just as there is no “best way” to learn. Teachers have individual styles, just as tennis players do, and effective ones know how to adjust their style depending on their students’ needs.
But let us not sell learners short: adjusting to a teacher’s method of delivery is a human attribute — the one that allowed us to do things like wander away from the Savanna, learn to catch and eat meat, and survive the advance of glaciers — and it is well worth fine tuning before graduating from high school. I didn’t attend any college classes nor hold any jobs where the professor or the employer adjusted to fit me, at least not in any significant ways. Being successful in life (no matter how one chooses to define success) depends almost always on one’s ability to adjust to changing circumstances.
In essence, forcing teachers to adopt a very particular method of teaching tends to inhibit their natural pedagogical talents, and it’s also biased toward students who do, in fact, like the Danielsonesque approach, which places much of the responsibility for learning in the students’ lap. Worse than that, however, a homogenous approach — of any sort — gives students a very skewed sense of the world in which they’re expected to excel beyond graduation.
In fairness, “The Framework for Teaching Evaluation Instrument” begins with a quiet little disclaimer, saying in the second sentence, “While the Framework is not the only possible description of practice, these responsibilities seek to define what teachers should know and be able to do in the exercise of their profession” (3). That is, there are other ways to skin the pedagogical cat. It’s also worth noting that the Danielson Group is seek[ing] to define — it doesn’t claim to have found The Way, at least not explicitly. Nevertheless, that is how untold numbers of legislators, reformers, consultants and administrators have chosen to interpret the Framework. As the Introduction goes on to say, “The Framework quickly found wide acceptance by teachers, administrators, policymakers, and academics as a comprehensive description of good teaching …” (3).
Teachers, well, maybe … though I know very, very few who didn’t recognize it as bologna from the start. Administrators, well, maybe a few more of these, but I didn’t hear any that were loudly singing its praises once it appeared on the Prairie’s horizon. Academics … that’s pretty hard to imagine, too. I’ve been teaching high-school English for 31 years, and I’ve been an adjunct at both private and public universities for 18 years — and I can’t think of very many college folk who would embrace the Danielson Framework tactics. Policymakers (and the privateer consultants and the techno-industrialists who follow remora-like in their wake) … yes, the Framework fits snugly into their worldview.
Thus, the Group doesn’t claim the Framework is comprehensive, but they seem to be all right with others’ deluding themselves into believing it is.
The Framework in the beginning
The Introduction begins by explaining each incarnation of the Framework, starting with its 1996 inception as “an observation-based evaluation of first-year teachers used for the purpose of licensing” (3). The original 1996 edition, based on research compiled by Educational Testing Service (ETS), coined the performance-level labels of “unsatisfactory,” “basic,” “proficient,” and “distinguished” — labels which have clung tenaciously to the Framework through successive editions and adoptions by numerous state legislatures. In Illinois, the Danielson Group Framework of Teaching is the default evaluation instrument if school districts don’t modify it. Mine has … a little. The state mandates a four-part labeling structure, and evaluators have been trained (brainwashed?) to believe that “distinguished” teachers are as rare as four-leaf clovers … that have been hand-plucked and delivered to your doorstep by leprechauns.
In my school, it is virtually (if not literally) impossible to receive a “distinguished” rating, which leads to comments from evaluators like “I think you’re one of the best teachers in the state, but according to the rubric I can only give you a ‘proficient.'” It is the equivalent of teachers telling their students that they’re using the standard A-B-C-D scale, and they want them to do A-quality work and to strive for an A in the course, but, alas, virtually none of them are going to be found worthy and will have to settle for the B (“proficient”): Better luck next time, kids. Given the original purpose of the Framework — to evaluate first-year teachers — it made perfect sense to cast the top level of “distinguished” as all but unattainable, but it makes no sense to place that level beyond reach for high-performing, experienced educators. Quite honestly, it’s demeaning and demoralizing — it erodes morale as well as respect for the legitimacy of both the evaluator and the evaluation process.
Then came (some) differentiation
The 2007 edition of the Framework, according to the Introduction, was improved by providing modified evaluation instruments for “non-classroom specialist positions, such as school librarians, nurses, and counselors,” that is, people who “have very different responsibilities from those of classroom teachers”; and, as such, “they need their own frameworks, tailored to the details of their work” (3). There is no question that the differentiation is important. However, the problem is that it implies “classroom teacher” is a monolithic position, and nothing could be further from the truth. Thus, having one instrument that is to be used across grade levels, ability levels, not to mention for vocational, academic and fine arts courses is, simply, wrongheaded.
As any experienced teacher will tell you, each class (each gathering of students) has a personality of its own. On paper, you may have three sections of a given course, all with the same sort of students as far as age and ability; yet, in reality, each group is unique, and the lesson that works wonderfully for your 8 a.m. group may be doomed to fail with your 11 a.m. class, right before lunch, or your 1 p.m. after-lunch bunch — and on and on and on. So the Danielson-style approach, which is heavily student directed, may be quite workable for your early group, whereas something more teacher directed may be necessary at 11:00.
Therefore, according to the Danielson Group, I may be “distinguished” in the morning, but merely “proficient” by the middle of the day (and let us not speak of the last period). The evaluator can easily become like the blindman feeling the elephant: Depending on which piece he experiences, he can have very different impressions about what sort of thing, what sort of teacher, he has before him. Throw into the mix that evaluators, due to their training, have taken “distinguished” off the table from the start, and we have a very wobbly Framework indeed.
Enter Bill and Melinda Gates
The 2011 edition reflected revisions based on the Group’s 2009 encounter with the Bill and Melinda Gates Foundation and its Measures of Effective Teaching (MET) research project, which attempted “to determine which aspects of a teacher’s practice were most highly correlated with high levels of student progress” (4). Accordingly, the Danielson Group added more “[p]ossible examples for each level of performance for each component.” They make it clear, though, that “they should be regarded for what they are: possible examples. They are not intended to describe all the possible ways in which a certain level of performance might be demonstrated in the classroom.” Indeed, the “examples simply serve to illustrate what practice might look like in a range of settings” (4).
I would applaud this caveat if not for the fact that it’s embedded within an instrument whose overarching purpose is to make evaluation of a teacher appear easy. Regarding the 2011 revisions, the Group writes, “Practitioners found that the enhancements not only made it easier to determine the level of performance reflected in a classroom … but also contributed to judgments that are more accurate and more worthy of confidence” (4-5). Moreover, the Group says that changes in the rubric’s language helped to simplify the process: “While providing less detail, the component-level rubrics capture all the essential information from those at the element level and are far easier to use in evaluation than are those at the element level” (4).
I suspect it’s this ease-of-use selling point that has made the Framework so popular among policymakers, who are clueless as to the complexities of teaching and who want a nice, tidy way to assess teachers (especially one designed to find fault with educators and rate them as average to slightly above average). But it is disingenuous, on the part of Charlotte Danielson and the Group, to maintain that a highly complex and difficult activity can be easily evaluated and quantified. In a 2012 interview, Ms. Danielson said that her assessment techniques are “not like rocket science,” whereas “[t]eaching is rocket science. Teaching is really hard work. But doing that [describing what teaching “looks like in words”] isn’t that big a deal. Honestly, it’s not. But nobody had done it.”
It’s downright naive — or patently deceptive — to say that a highly complex process (and highly complex is a gross understatement) can be easily and simply evaluated — well, it can be done, but not with any accuracy or legitimacy.
Classic fallacy of begging the question
I want to touch on one other inherent flaw (or facet of deception) in the Danielson Framework and that is its bias toward “active, rather than passive, learning by students” (5). Speaking of the Framework’s alignment with the Common Core, the Group writes, “In all areas, they [CCSS] place a premium on deep conceptual understanding, thinking and reasoning, and the skill of argumentation (students taking a position and supporting it with logic and evidence).” On the one hand, I concur that these are worthy goals — ones I’ve had as an educator for more than three decades — but I don’t concur that they can be observed by someone popping into your classroom every so often, perhaps skimming through some bits of documentary evidence (so-called artifacts), and I certainly don’t concur that it can be done easily.
The Group’s reference to active learning, if one goes by the Domains themselves, seems to be the equivalent of students simply being active in an observable way (via small-group work, for example, or leading a class discussion), but learning happens in the brain and signs of it are rarely visible. Not to get too far afield here, but the Framework is intersecting at this point with introverted versus extroverted learning behaviors. Evaluators, perhaps reflecting a cultural bias, prefer extroverted learners because they can see them doing things, whereas introverted learners may very well be engaged in far deeper thinking, far deeper comprehension and analysis — which is, in fact, facilitated by their physical inactivity.
And speaking of “evidence,” the Introduction refers to “empirical research and theoretical research” (3), “analyses” and “stud[ies]” (4) and to “educational research” that “was fully described” in the appendix of the 2007 edition (3), but beyond this vague allusion (to data which must be getting close to a decade old) there are no citations whatsoever, so, in other words, the Danielson Group is making all sorts of fantastic claims void of any evidence, which I find the very definition of “unsatisfactory.” This tactic, of saying practices and policies are based on research (“Research shows …”), is common in education; yet citations, even vague ones, rarely follow — and when they do, the sources and/or methodologies are dubious, to put it politely.
I plan to look at the Danielson Framework Domains in subsequent posts, and I’m also planning a book about what’s really wrong in education, from a classroom teacher’s perspective.
’Tis the season–to traumatize young teachers
Illinois has many seasons–bow season, shotgun season … and every March is “traumatizing young teachers” season as school administrators across the state dismiss nontenured teachers, and they’re not even required to give a reason for the dismissal, hence, oftentimes they don’t. Teachers are left devastated, humiliated, and profoundly confused about whether they’ve chosen the right professional path after all.
A few years ago the Illinois legislature, in one of the opening salvos in its campaign to destroy and demoralize educators, expanded the length of time that teachers could be let go without cause to four years, which means that young professionals (or older ones entering the profession later in life) can be dedicated, hard-working teachers who are establishing themselves in their communities and developing collegial relationships for one, two, three and even four years when they’re blindsided by the administrator’s news that they won’t be coming back the following year.
Sometimes, of course, there have been issues raised, and the teacher has not corrected them to the administrator’s satisfaction; and sometimes the school district’s desperate financial situation has led to the dismissal. Too often, though, the young and developing educators are sacked without any warning whatsoever–they’ve fallen prey to the caprices of an administration that has no one to answer to, excerpt perhaps school board members, who tend to know only what administrators tell them since they rarely have direct contact with the teaching staff.
The situation has been exacerbated in the past year by the state’s mandate of a new model for evaluating teachers. It is more complicated and more labor intensive than the tools most district’s had been using. The increased complications and time commitments have not led to a better approach to evaluation, however. They’ve only opened the door for even more nebulous assessments of a teacher’s performance. Teacher evaluation is a rich subject in itself, too rich of a subject to discuss here–but the bottom line is that teaching is far too complex an endeavor to be reasonably evaluated by a single rubric that is used across grade levels, disciplines, and teaching assignments. In fact, it’s insulting to the profession that so many people believe such a model can be devised and successfully implemented. Physicians, attorneys, engineers, business professionals–and politicians!–would never allow themselves to be evaluated the way that a teacher’s worth is determined.
But no matter how simple or how complex the evaluation process is, its usefulness and fairness depend on the sagacity and integrity of the evaluating body. Unfortunately, sagacity and integrity are not prerequisites for becoming an administrator or a school board member. There are good administrators out there, of course, and well-meaning board members; but administrators and board members come in all stripes, just like the human population as a whole. Yet there is no check-and-balance built into the process. Young teachers who are dismissed unfairly, and the professional associations who represent their interests, have no recourse. No recourse at all.
In other words, there is no evaluation of the evaluator, whose sagacity and integrity, apparently, are assumed by the Illinois General Assembly … in all of its sagacity and integrity.
When there is an unfair and unwarranted dismissal, a shockwave goes through the faculty and the student body almost as palpable as an accidental death. Other nontenured teachers become like deer in hunting season and worry that they’ll be next–if not this spring maybe the next, or the next, or the next, or the next. Tenured teachers are angered, saddened and frustrated by the loss of a valuable colleague and trusted friend. It greatly diminishes their respect for their superiors and their good will in working with them. It disrupts students’ focus, and it teaches them a hard lesson about the perils of choosing a career in education. And once a district becomes known as one that mistreats young professionals, word spreads virulently and the best and brightest don’t bother to apply.
Who, in their right mind, would want to work for an administration and board that will dismiss them without reason after a year, or two, or three, or four of hard work and dedication? Who, in their right mind, would chance subjecting their spouse and possibly children to the trauma of a lost job beyond their control?
Young teachers have mainly debt (nowadays from colossal student loans) and very little savings. It’s frightening to be jobless, especially when it’s due to no fault of their own–at least, no fault they’ve been made aware of. Yet teachers must continue to teach for the remainder of the school year, while also looking for new employment. They are often–magnanimously–given the option of resigning instead of being dismissed, but it’s likely a thin disguise that fools no one in their search for another teaching job. They find themselves in very difficult situations when interviewing elsewhere because the question must come up “Why did you leave such-and-such school?” What, then, do they say that won’t compromise either their honesty or their chances of landing another job?
The fact that we as a society allow this devastating unfairness to be visited on our young teachers every spring is just another indication of how little we value education, educators and–for that matter–the children they’ve dedicated themselves to educating.
94 comments