For the measures I am discussing, statistical significance is tough to generate even on a massive scale. A single district?
We collected 200,000 sheets of student work from 200 schools and barely made statistical significance on the measures we are performing. A doctoral student doesn't stand a chance. (The work stacks 80 feet high.)
Statistical significance is tough to pull off in education research, which is why they resort to meta-analysis to generate statistically significant results. (And don't even get me started on meta-analysis.)
We collected 200,000 sheets of student work from 200 schools and barely made statistical significance on the measures we are performing. A doctoral student doesn't stand a chance. (The work stacks 80 feet high.)
Statistical significance is tough to pull off in education research, which is why they resort to meta-analysis to generate statistically significant results. (And don't even get me started on meta-analysis.)
I fully disagree. Even if the state test was PERFECT in analyzing students' true proficiency, it does not satisfy the needs of "check" demanded of the Deming cycle because it is not formative. State tests do not check the process used to generate the results, they only test the outcomes.
So my math scores are low. Even if the state tests is perfect, that still doesn't point to a solution. Is it the instruction? Is it the curriculum? Is it the alignment of the curriculum? Is it the rigor of in-class questioning? State tests won't tell you.
In the Deming cycle, the check phase does not mean, "Sell the cars and see if you generate any customer complaints," which is directly analogous to the method of using state test scores to measure teaching effectiveness.
So my math scores are low. Even if the state tests is perfect, that still doesn't point to a solution. Is it the instruction? Is it the curriculum? Is it the alignment of the curriculum? Is it the rigor of in-class questioning? State tests won't tell you.
In the Deming cycle, the check phase does not mean, "Sell the cars and see if you generate any customer complaints," which is directly analogous to the method of using state test scores to measure teaching effectiveness.
These students don't know how to write because no one has ever taught them. Why? Because it was always someone else's responsibility.
Besides, what is it you are teaching and are subskill proficiencies getting in the way of assessing their proficiency on the content? That is one of the things I look for when observing classrooms. Can I measure that to national specifications? No. But it still needs measuring.