I’ve reached a point after these past few years where I feel that I’ve spent way too much time critiquing poorly constructed arguments and shoddy analyses that seem to be playing far too large a role in influencing state and federal (especially federal) education policy. I find this frustrating not because I wish that my own work got more recognition. I actually think my own work gets too much recognition as well, simply because I’ve become more “media savvy” than some of my peers in recent years.
I find it frustrating because there are numerous exceptional scholars doing exceptional work in school finance and the economics of education whose entire body of rigorous disciplined research seems drowned out by a few prolific hacks with connections in the current policy debate.It may come as a surprise to readers of popular media, but individuals like Mike Petrilli, Eric Osberg, Rick Hess (all listed on the USDOE resource web site) or Bryan Hassel wouldn’t generally be considered credible scholars in school finance or economics of education. I’d perhaps have less concern – and be able to blow this off – if many of the assertions being made by these individuals – and others – weren’t so often completely unsupported by reasonable analysis and if those assertions didn’t lead to potentially dangerous and damaging policies.
This post is specifically about the body of methodologically flimsy research produced in recent years by Marguerite Roza, previously of the Center on Reinventing Public Education and currently an advisor to the Gates Foundation.
Why this post now? I’ve simply lost my patience.
This post is in part a response to the recent unveiling of the U.S. Dept. of Education web site on improving educational productivity http://www.ed.gov/oii-news/resources-framing-educational-productivity. Amazingly, this site lists primarily non-peer reviewed, shoddy work by Marguerite Roza and colleagues and bypasses entirely more serious research on educational productivity or methods for evaluating it. The quality of some of the examples on this site is particularly abysmal. Yet it is presented as “the work of leading thinkers in the field.” (interesting that “thinkers” is used in place of “researchers.”) Among the worst examples, this site lists as a credible resource the Center for American Progress Return on Investment analysis. (by Ulrich Boser, a great writer on the topic of art theft, but in this case, a bit out of field).
I don’t mind so much that this stuff exists. But it certainly doesn’t belong in a serious policy conversation, nor does it represent “the work of leading thinkers in the field.”
Let’s start with a few common attributes of the worst-of-the-worst types of policy research floating around out there and warping and misguiding the education policy debates in general and school finance debates in particular. For lack of a better term, let’s just call it “hack research.”
Perhaps most importantly, hack research fails to recognize all of the credible work that’s already been done on a topic, typically because the research hack who produced it lacks entirely the discipline to bother to understand that body of work and how to build on it in order to come to new, credible findings and conclusions.
Further, hack research displays little regard for the connection between rigorous analysis and conclusions that may be drawn from it. This stems in part from the lack of discipline to actually conduct rigorous analyses.
Particularly effective hacks will not just ignore the body of existing scholarship but will do so belligerently, proclaiming that no good work has ever been done, no credible methods of analysis do exist, and therefore the time is right for their own creative and new perspective! The hack research method substitute is usually some seemingly intuitive, completely shallow, poorly conceived back-of-the-napkin approach. In other words, the hack research motto is that we must think outside the box, because it’s just too much work to open and unpack that box!
Many of us start as hacks, but eventually grow out of it as we realize that there’s a lot of great stuff out there to read and exceptional scholars from which to learn. And, some non-hacky researchers will occasionally hack. Hack happens. It’s only really problematic when it’s a persistent pattern of hackyness or even gets worse over time.
The most dangerous hacks use their shtick to influence policy with catchy anecdotes, convincing policymakers and major players that they need look no further (at real research, for example) than their own hacky “research.” And the most effective hacks can spin findings that never were into pure urban legend – well-accepted myths turned realities – with serious policy implications!
Let’s take a look at a number of mythical findings from shoddy research produced by Marguerite Roza in recent years, including a few sources cited on the USDOE resources page.
Myth #1: States have largely solved between district funding disparities and within district disparities are the remaining problem of the day.
Sources of the myth: See references in Baker/Welner article (cited below)
A now common myth in school finance reiterated in numerous sources produced by the Education Trust, Center for American Progress and other DC think tanks and pundits is that states have largely resolved disparities in funding between districts and that persistent disparities are primarily within districts, between schools – a function of illogical district allocation formulas.
In a recent article Kevin Welner and I tackle this argument and dig deeply into the sources behind this argument, which invariably find their way back to Marguerite Roza, then of the Center on Inventing Research Findings – excuse me – Center for Reinventing Public Education (CRPE).
Kevin and I conclude in our article:
Two interlocking claims are being increasingly made around school finance: that states have largely met their obligations to resolve disparities between local public school districts and that the bulk of remaining disparities are those that persist within school districts. These local decisions are described as irrational and unfair school district practices in the allocation of resources between individual schools. In this article, we accept the basic contention of within-district inequities. But we offer a critique of the empirical basis for the claims that within-district gaps are the dominant form of persistent disparities in school finance, finding instead that claims to this effect are largely based on one or a handful of deeply flawed analyses.
Kevin Welner and I dissect in detail the problematic, “non-traditional” methods Roza and colleagues use for conducting their analyses (ignoring real methods used by real researchers in real publications), but perhaps more interesting are those cases where a narrow, measured finding pertaining to one specific estimate in one specific context becomes a national trend, a dominant reality soon thereafter. Op-Ed columns by Roza on the topic of within versus between district funding disparities include particularly egregious examples. Kevin Welner and I explain:
Following a state high court decision in New York mandating increased funding to New York City schools, Roza and Hill (2005) opined: “So, the real problem is not that New York City spends some $4,000 less per pupil than Westchester County, but that some schools in New York [City] spend $10,000 more per pupil than others in the same city.” That is, the state has fixed its end of the system enough.
This statement by Roza and Hill is even more problematic when one dissects it more carefully. What they are saying is that the average of per pupil spending in suburban districts is only $4,000 greater than spending per pupil in New York City but that the difference between maximum and minimum spending across schools in New York City is about $10,000 per pupil. Note the rather misleading apples-and-oranges issue. They are comparing the average in one case to the extremes in another.
In fact, among downstate suburban New York State districts, the range of between-district differences in 2005 was an astounding $50,000 per pupil (between the small, wealthy Bridgehampton district at $69,772 and Franklin Square at $13,979). In that same year, New York City as a district spent $16,616 per pupil, while nine downstate suburban districts spent more than $26,616 (that is, more than $10,000 beyond the average for New York City). Pocantico Hills and Greenburgh, both in Westchester County (the comparison County used by Roza and Hill), spent over $30,000 per pupil in 2005. These numbers dwarf even the purported $10,000 range within New York City (a range that we agree is presumptively problematic); our conclusion based on this cursory analysis is that the bigger problem likely remains the between-district disparity in funding.
For the full take down, see:
Baker, B. D., & Welner, K. G. (2010). “Premature celebrations: The persistence of interdistrict funding disparities” Educational Policy Analysis Archives, 18(9). Retrieved [date] from http://epaa.asu.edu/ojs/article/view/718
Myth #2: America’s public school system suffers from something called Baumol’s disease, therefore the only solutions must be found outside of public education
Source: Curing Baumol’s Disease: In Search of Productivity Gains in K–12 Schooling Paul Hill, Marguerite Roza
While I don’t think this one really ever caught on, it’s so absurd that it must be addressed. Further, it’s actually cited on the USDOE resources in educational productivity page despite the fact that it offers no useful guidance whatsoever on the topic.
The objective of this policy brief by Paul Hill and Marguerite Roza of CRPE is to explain how American public education suffers from Baumol’s disease, or “the tendency of labor-intensive organizations to become more expensive over time but not any more productive.” Hill and Roza’s attempt at empirical validation that American public education suffers from Baumol’s disease is presented in two oversimplified figures, a graph showing increased number of staff who are not core teachers (Figure 1) and a graph showing that student test scores on the National Assessment of Educational Progress have remained flat over time (Figure 2). The latter claim that we’ve seen no improvement in NAEP scores over time is contested. And the former claim, when aggregated nationally is not particularly meaningful. The authors provide no empirically rigorous link between the two.
Rather, the casual reader is simply to assume that public schools have added a lot of non-teaching staff and have, on average, nationally seen no yield for that increase costs. Hill and Roza posit:
“While these indicators clearly point to increased costs for education, efforts to quantify productivity changes have been hampered by measurement challenges on the outputs side of the equation. By most accounts, key indicators of outcomes have not shown comparable gains. A thirty-year look at NAEP performance for seventeen year-olds, for instance, suggests that test scores have changed very little.” (p. 3)
While this may, in fact, not be entirely untrue, the authors provide no rigorous validation that “Baumol’s Disease” is a persistent problem of American public schools.
However, without a disease with a catchy name, there would be little reason for their proposed cure. But the proposed cure is no more thoroughly vetted or precisely articulated than the disease. A central assumption in the Baumol’s disease policy brief is that American public education systems take on one single form, as represented by national averages in the TWO graphs provided, that there is little or no variation within the public education system in terms of resource use or outcomes achieved (e.g. that it all suffers Baumol’s disease), and that therefore the only possible cures are those that come from outside the public education system or at its fringes. That is, that we have nothing to learn from variation within the public education system itself, because there is no such variation. Instead, for example, the authors suggest a closer look at “home schooling, distance learning systems, foreign language learning, franchise tutoring programs, summer content camps, parent-paid instructional programs (music, swimming lessons, etc.), armed services training, industry training/development, apprentice programs, education systems abroad.” (p. 10)
Numerous more credible researchers have spent a great deal of time learning from the heterogeneity of how schools, school districts, and charter schools operate, as well as across states, including studying the relative efficiency of schools that either operate differently or change how they operate. The assumption that the only solutions must come from outside the system is patently absurd, when the “system” consists of 51 policy contexts, over 100,000 schools, 5,000 charter schools and about 15,000 public districts. And it’s just lazy, hack thinking.
While one might gain insights from other labor-intensive industries, or education at the fringes of the current public system, it would be foolish to ignore the extent of variation within the current American public education system, and across traditional public, magnet, charter and private schooling. Arguably, the authors present the view that there is little or nothing to learn from the current system specifically in order to avoid the need for conducting rigorous analysis of it. Further, while such policy briefs may be generously considered as useful conversation starters, we take serious issue with the U.S. Department of Education’s identification of sources of this type, which are purely speculative, and severely lacking in intellectual or empirical rigor, as “Key Readings on Educational Productivity.”
Myth #3: Poor, failing school districts have plenty of money but are squandering too much on Cheerleading and Ceramics when they need to be spending on basics!
Original Source of (the anecdote behind the) myth: “Now is a Great Time to Consider the Per Unit Cost of Everything in Education.”
As I explain in my recent conference paper:
Authors including Marguerite Roza and colleagues of the Center for Reinventing Public Education encourage public outrage that any school district not presently meeting state outcome standards would dare to allocate resources to courses like ceramics or activities like cheerleading. To support their argument, the authors provide anecdotes of per pupil expense on cheerleading being far greater than per pupil expense on core academic subjects like math or English.
- Imagine a high school that spends $328 per student for math courses and $1,348 per cheerleader for cheerleading activities. Or a school where the average per student cost of offering ceramics was $1,608; cosmetology, $1,997; and such core subjects as science, $739.1
These shocking anecdotes, however, are unhelpful for truly understanding resource allocation differences and reallocation options, and are an unfortunate and unnecessary distraction. For example, the major reason why cheerleading or ceramics expenses per pupil are seemingly high is the relatively small class sizes, compared to those in English or Math. In total, the funds allocated to either cheerleading of ceramics are unlikely to have much if any effect if redistributed to reading or math.
Now, this myth is a rather strange one, because the source from which it comes, which is authored by Marguerite alone, really isn’t totally unreasonable. It’s not useful in any way shape or form, but it’s not unreasonable either. This wacky anecdote about cheerleading and ceramics spending comes from a piece in which Roza is trying to explain the importance of comparing unit costs of providing specific programs/opportunities. This is a rather “no duh” idea, and the working paper and eventual book chapter is built on uninteresting anecdotes, at best. The original point of the paper is that if administrators look at the per unit cost of everything, they might find some things that stand out, and some things that might be reasonably reorganized to be offered at a lower unit cost (for example, the cost of cheerleading was reduced by moving it from a class period drawing on salaried time, to an after school activity, paid by small stipend).
But, the spin from this piece has been that this is all that low performing, poor urban districts need to do. They’ve all got enough. They themselves are responsible for the most persistent inequities – not the states. And they are the ones wasting way too much on things like cheerleading and ceramics. Given that this spin has had far more traction than the more reasonable paper behind it, one might assert that this is precisely what Roza intended.
In my paper, I conclude:
Rather, the emergent story from the data in both states was the contrast between high spending, high outcome districts, and low spending low outcome districts and their respective high schools. On average, high spending, high outcome districts were as one might expect much lower in student poverty concentration and low spending, low outcome districts much higher in poverty. That is, after applying thorough cost adjustment including adjustments for differences in student needs. Interestingly, the most striking differences between these groups of districts were not in the availability of assigned teachers or courses in the arts, but rather in the distribution of advanced versus basic course offerings in curricular areas such as math and physical science.
Note that to begin with, low spending, low outcome schools had fewer teacher main assignments and fewer course assignments per pupil. As such, they were, from the outset, more constrained in their allocation options. Further, there is at least some evidence that when evaluating district wide resource allocation, low resource, low outcome districts see greater necessity or feel greater pressure to allocate a larger overall share of resources to elementary classrooms (based on Illinois findings).
More thorough analyses of this issue see:
Baker, B.D. (2011) Cheerleading, Ceramics and the Non-Productive Use of Educational Resources in High Need Districts: Really? Paper presented at the Annual Meeting of the American Educational Research Association, New Orleans, LA 2011
Myth #4: High schools in Washington State pay math and science teachers less than other teacher despite public interest and state policies which encourage paying them more
Source: Washington State High Schools Pay Less for Math and Science Teachers than for Teachers in Other Subjects Jim Simpkins, Marguerite Roza, Cristina Sepe
This is one that suffers from both major issues identified at the beginning of this rant. First, the disconnect between the “study” and the press release:
The Press Release
The analysis finds that in twenty-five of the thirty largest districts, math and science teachers had fewer years of teaching experience due to higher turnover—an indication that labor market forces do indeed vary with subject matter expertise. The subject-neutral salary schedule works to ignore these differences.
That said, the lower teacher experience levels are indicative of greater turnover among the math and science teaching ranks, lending support to the hypothesis that math and science teachers may have access to more compelling non-teaching opportunities than do their peers. (p. 5)
That is, the conclusions of the study itself and the press release are, well, not consistent. But this pattern of behavior is entirely consistent for Roza and CRPE.
In a previous post I address just how ridiculous the methods in this analysis are, in which she compares STEM teacher salaries with non-STEM teacher salaries without any controls for other factors that affect salaries (on the argument that salaries shouldn’t be based on those things – experience and degree level – anyway).
All that Roza really found in this paper was that STEM teachers tend to be younger and as a result have lower average salaries than non-STEM teachers. From that, she spun the argument that because STEM teachers don’t earn more than other teachers, but STEM fields are more competitive, STEM teachers must be leaving teaching at a higher rate, leading to a less experienced pool and lower average salaries (a vicious cycle indeed! But one that’s never validated by the ridiculous analysis).
In my post, I actually evaluate several years of teacher level data on all teachers in Washington State, finding most of her conclusions to be flat out wrong. Here’s the figure on mean STEM and non-STEM teacher salaries over time: https://schoolfinance101.files.wordpress.com/2010/08/slide42.jpg
I also point out that credible researchers like Lori Taylor of Texas A&M have actually done better analyses of Washington teacher wages and addressed variations in labor market competitiveness by field:
Report on Taylor Study:
Somehow, not surprisingly, Roza was unaware of either this better research or the more credible methods used in this research.
For the full take down, see: https://schoolfinance101.wordpress.com/2010/08/20/new-from-the-center-on-inventing-research-findings/
Myth #5: With our handy-dandy basket of reformy fixes, we can cut significant funding from American public schools and dramatically increase productivity!
Source: Petrilli and Roza
Stretching the School Dollar (Brief)
In their policy brief on Stretching the School Dollar, Mike Petrilli of Thomas B. Fordham Institute and Marguerite Roza of the Gates Foundation provide a lengthy laundry list of strategies by which school districts and states might arguably increase their productivity at lower expense, or “stretch the dollar” so to speak. This policy brief is an extension of the Frederick Hess (American Enterprise Institute) and Eric Osberg (Fordham Institute) edited book by the same title. We highlight this source because of repeated specific references to this source in Secretary Duncan’s “New Normal” speeches during the Fall of 2010.
Because this policy brief and book specifically list strategies that are intended to improve productivity at comparable or lower expense, it would be particularly relevant for the book or brief to either provide directly or summarize from other sources, rigorous cost-effectiveness analysis of these options, or relative efficiency comparisons of schools and districts employing these options. But that is apparently asking way too much of Roza or Petrilli. I’ll cut Mike some slack here, because he isn’t the one actually presenting himself as a school finance expert/scholar. That’s Roza’s role in this partnership, therefore the burden falls on her. But after reading enough work by Roza and colleagues, I’m no-longer convinced that she is even aware that there is a body of research out there on Cost-effectiveness analysis or relative efficiency (more on this later). I certainly encourage her to go buy a copy of Hank Levin and Patrick McEwan’s book, not so subtly titled Cost-Effectiveness Analysis: Methods and Applications. It’s a relatively easy, non-academic read.
I’ll offer a primer on these methods and their application to these questions in a future post. There’s no need to beat a dead horse on this topic. I’ve taken down Roza and Petrilli’s reformy gift basket in two previous posts to which you can refer.
For the full take down, see:
Part 1 – Stretching the Truth, Not Dollars: School Finance in a Can: Unproven and Unsubstantiated Dollar-Stretching State Policies
Part 2 – Stretching the Truth, Not Dollars: Considering the Application of Cost-Benefit Analysis to Teacher Layoff Alternatives
13 thoughts on “School Finance through Roza-Tinted Glasses: 5 School Funding Myths from a single Misguided Source”
Hack and biased research not only has an adverse impact on and dilutes the reputation of quality educational research but also seems to have a disproportionate influence on state and federal education policy. Would you include Eric Hanushek within the group performing shoddy analyses?
You might be surprised that I do not include Eric Hanushek generally in this category. Much of his work is truly exceptional. He is a credible researcher with a long track record of high qualify empirical work. But, in my view, he has his moments where he constructs indefensible arguments and does so out of purely political interests. For my most thorough critique of one of Rick’s publications, see: http://www.tcrecord.org/content.asp?contentid=16106, where Kevin Welner and I critique his book in which he addresses the effects of school funding litigation. Note that this is a politically motivated book, and not his research. It’s a book co-authored with an attorney who assists in defending states against school funding litigation. I’ve also commented on other Hanushek musings on numerous occasions on this blog. I suspect a search on his name will reveal those comments. Perhaps my harshest critique of Hanushek’s work can be found in the concluding sections of this article (draft here): https://schoolfinance101.files.wordpress.com/2010/01/baker-missouri-peabody.pdf But, he clearly knows his stuff and produces some exceptional research, which is not something I can say of anything I’ve ever read by Marguerite Roza (or Bryan Hassel, for that matter).
You can critique the bad stuff forever.
Can you suggest some “good” research?
If you’re looking to get a start in the field of Education Finance/Economics of Education, I would spend some time digging through the Handbook of Research on Education Finance and Policy (linked in post above). It provides an exceptional summary of research across most of the major topics in the field. If you’re looking for other specific research – on specific topics – let me know and I’ll gladly list some good sources. There is plenty of good stuff out there… but it doesn’t get the media or policy play.
$200 for the book! Yikes! Bruce, there’s gotta be something cheaper or open source right?
It’s like the “great big book of everything” for school finance, so if you’re into it at all, it’s worth purchasing. I thought the paperback was cheaper. If there are specific topics that interest you, I might be able to point you toward some cheaper and/or open source stuff.
Here is something you might juxtapose against the Baumol’s disease argument: private prisons tend to be more expensive than state- or federally- run prisons.
The reality is this: if you have mostly private schools and these private (or quasi private charter) schools find ways to underpay teachers, eventually they are going to smarten up an unionize just like the prison guards did. Schools with poor outcomes will be hard pressed to hire teachers of quality. Teachers of quality will figure out their worth and won’t put up with this nonsense.
We are chasing a tiger by it’s tail if we think that teacher’s unions are the problem here. They are not. Great article, by the way.
Posted by a non-teacher, non-union member, moderate Republican.
Are there any studies in which cost equations (such as in Duncombe & Yinger 2004) have been empirically validated? As in, a random number of districts increased spending according to the terms of a cost equation study, and then their disadvantaged students in fact rose to the predicted level of performance, while the random districts chosen not to increase spending saw their disadvantaged students stay at a lower level?
You know full well that there aren’t randomized trials of school finance “remedies” so-to-speak. It would be interesting (though I suspect difficult to get through IRB). Indeed, the Duncombe and Yinger method attempts to extrapolate from the variations that exist across districts and over time in the system as it presently exists. Extrapolating from these variations is clearly no guarantee that if funding were raised to these levels in districts where it presently is not, that performance would also be raise to the estimated target. Certainly, extrapolating well outside the sample is far more problematic (e.g. projecting the costs of achieving outcome levels that no-one currently does). But it remains the most reasonable approach for estimating what would be the projected need toward achieving a specific measured outcome, given the existing underlying structures (and variations in them) for providing education (and it would be fun to estimate models across more varied structures/delivery systems).
I find it problematic to argue, however, that because we don’t know from experimental studies, whether increased funding would “work” or not, that we shouldn’t waste the additional money on districts that are projected to need it and we should instead leave the persistent disparities in place, and reshuffle students across alternative, comparably resource deprived alternatives (underfunded charters chasing private contributions, or low budget religious private schools). In many states, high need districts continue to have substantially lower resources than their surrounding more affluent neighbors (that national averages table on this point is meaningless: https://schoolfinance101.wordpress.com/2011/02/25/school-funding-equity-smokescreens-a-note-to-the-equity-commission/).
Further, as I note here: https://schoolfinance101.files.wordpress.com/2010/01/baker-missouri-peabody.pdf
“While cost functions may not necessarily produce definitive dollar predictions for each individual district in a state to achieve specific educational outcomes, they can provide guidance on how state aid might be better targeted to achieve greater equality of educational opportunity across varied settings. That is, cost functions may guide in reshaping or bending state school finance formulas. Cost functions may aid in identifying factors that influence costs, alternative proxy measures for cost factors and the relative sensitivity of education costs to alternative measures (See also Killeen and Baker, 2008).”
Allowing districts such as Riverview Gardens, Hazelwood and Caruthersville [Missouri] to languish with below state average funding when they clearly have greater than average needs is simply unfair and should be considered unconstitutional. It is morally repugnant to assert that legislators should not be compelled by courts or their own conscience to address this inequity because the education cost function does not produce estimates of “pure cost,” eradicating from those estimates 100% of the inefficiencies which we do not even know exist. Common sense tells us these districts and their children have greater needs and thus need more resources to achieve equal opportunity. The cost function provides us a reasonable empirical estimate of how much more.”
“Arguing that it is unreasonable to estimate “costs” of educational outcomes based on practices as they currently exist, on the untested assumption that a better and more efficient system exists somewhere out there, is unhelpful at best and destructive at worst. If one believes that a “better way” exists – a more efficient and productive use of educational resources – then one should try implementing that “better way” in a number of school districts across a state. Once this new approach takes hold in select settings across a state and produces its expected efficiency gains, updated cost function models on the state system should pick up these changes or at least identify schools implementing such highly efficient approaches as significant outliers which may then inform future policies.”
That aside, there is a sizable body of research on the effects of school finance reforms on the level and distribution of student outcomes. This research tends to show that school finance reforms that have altered the level of funding have subsequently altered the level of outcomes and school finance reforms that have altered the distribution of funding have altered the distribution of outcomes.
This article provides a thorough review of this research: http://www.tcrecord.org/content.asp?contentid=16106
Earlier draft here: https://schoolfinance101.files.wordpress.com/2010/01/doreformsmatter-baker-welner.pdf
The article also explains many of the conceptual and empirical flaws that exist in many analyses that purport to show that school finance reforms and school finance litigation have had no effect on outcomes. Certainly, when school finance litigation results in no actual substantive reform (even if a “win” for plaintiffs), we would not expect to see an effect on outcomes.
I suspect you have referenced it before, but the Think Tank Review Project is one organized attempt to address the junk science perpetrated by politically motivated outlets– And I love the fact that the Bunkum Awards have been given this year to the federal government for the crap they post on official websites.
You say, “Extrapolating from these variations is clearly no guarantee that if funding were raised to these levels in districts where it presently is not, that performance would also be raise to the estimated target.”
Well, that’s the issue I am thinking about. Where there are actual disparities, as you say, that’s one thing. But it’s another thing entirely to make the prediction that it costs 1.xx times as much to educate a poor student or an ELL to proficiency. Anyone who says that is making a very precise empirical prediction of what effect a multiplier will have at a particular level, but I can’t tell that there’s been any empirical validation of such predictions.
They are estimates… like estimates from any other statistical model, with error ranges and with the potential for bias in the estimate. The point is that they can provide clearer guidance, as I explain in that really long quote from a published paper, to “bend” school finance formulas in a more empirically based direction. Clearly in school finance nothing is ever perfect, perfectly rational or perfectly empirical, nor can it be. But, there are several state school finance systems out there that are sufficiently out of whack that they need some significant bending in a more reasonable and rational direction – empirically guided bending perhaps.
Comments are closed.