INTRO: Part I is here.
If you find this topic dry and hard to get your head around, then you are half way there. Because some of the most important things to know about government are dry and hard to get one's head around. And that makes it easy for politicians to bamboozle voters with falsehoods and misinformation.
So if you want to understand why ET's firing of civil servants (most of government employees) is a violation of law and various regulations, you'll have to buck up and read carefully. Even take notes.
This content is based on testimony I gave in a local discrimination case. So I had to pare it down to as simple an explanation as possible so that I didn't lose the jury. The attorney was nervous that his expert would talk over their heads, but when I was done he was relieved that I'd made it very easy to understand. And the jury said the local government was guilty.
So good luck. [I explained ET in the Intro to Part I, but it's not critical.]
From a February 19, 2018 post:
Graham v MOA #9: Exams 2 - Can You Explain These Terms: Merit Principles, Validity, And Reliability?
“Personnel policy and rules preserving the merit principle of employment.” AMC 3.30.041 and 3.30.044 explain examination types, content, and procedures consistent with these merit principles.Âs defined in the Anchorage Municipal Code Personnel Policies and Rules,
“Examination means objective evaluation of skills, experience, education and other characteristics demonstrating the ability of a person to perform the duties required of a class or position.” (AMC 3.30.005)
Jeff Graham's attorney made me boil this down to the most basic points to increase the likelihood I wouldn't put the jury to sleep. So bear with me and keep reading.
And, you can see an annotated index of all the posts at the Graham v MOA tab above or just link here.]
Basic Parts of Government In The United States
- The elected politicians who pass the laws and set the broad policy directions (legislature)
- The elected executive who carries out the laws.
- The administration is led by the elected executive - the president, the governor at the state level, and the mayor at the city level.
- Civil Service refers to the career government workers who actually carry out the policies. There are also appointed officials at the highest levels who are exempt from some or all of the civil service rules.
- Employees are chosen solely based on their skills, knowledge, and abilities (SKAs) that are directly related to their performance of the job.
- The purpose of this is to make government as as effective and efficient as possible by hiring people based on their job related qualities and nothing else.
- That also means other factors - political affiliation, race, color, nationality, marital status, age, and disability should not be considered in hiring or promotion. It also means that arbitrary actions and personal favoritism should not be involved.
- Selection and promotion criteria should be as objective as possible.
- Grounded on facts. Grounding your understanding or belief on something concrete, tangible. Something measurable that different people could 'see' and agree on.
- Unbiased. A second, implied meaning from the first, is that you make decisions neutrally, as free as you can be from bias, preconceived ideas. That’s not easy for most people to do, but there are ways to do it better.
To be valid, the selection or promotion test must be a good predictor of success on the job. People who score high on the exam, should perform the job better than those who score low. And people who score low should perform worse on the job than people who score high.
BUT, even if the test is intrinsically valid, the way it is administered could invalidate it. If the test is not also reliable (testing and grading is consistent enough that different test takers will get a very similar score regardless of when or where they take the test and regardless of who scores the test) the test will no longer be valid. This is because the scores will no longer be good predictors of who will do well on the job.
That should be enough but for people who want to know more about this, I'll give a bit more below.
-----------------------------------------------------------------------
Merit Principles
If you want to know more about merit principles I'd refer you to the Merit System Protection Board that was set up as part of the Merit System Reform Act of 1978.
A little more about reliability problems (because these are important to understand about the engineer promotion exam)
In the main part of this post I wrote that all the important (could affect the score) conditions of the test need to be the same no matter where or when or with whom a candidate takes the test. Here are some more details
- Location - If one location is less comfortable - temperature, noise, furniture, lighting, whatever - it could skew the scores of test takers there.
- Time - could be a problem in different ways.
- All candidates must have the same amount of time to take the test.
- Instructions - all instructions have to be identical
- Security of the test questions - if some applicants know the questions in advance and others do not, the test is not reliable.
The scoring, too, has to be consistent from grader to grader for each applicant.
And there are numerous ways that scoring a test can go wrong.
- Grader bias - conscious and unconscious. Raters who know the candidates may rate them differently than people who don’t know them at all.
- The Halo effect means if you have a positive view of the candidate, you’re likely to give him or her more slack. You think, 'I know they know this.'
- The Horn or Devil Effect is the opposite - If you already have a negative opinion about a candidate, you consciously or unconsciously give that a candidate less credit. These are well documented biases.
- Testing order bias affects graders and candidates.
- After three poor candidates, a mediocre candidate may look good to graders.
- Grading Standards - Is the grading scale clear and of a kind that the graders are familiar with?
- Are the expected answers and how to score them clear to the graders?
- Do the graders have enough time to calculate the scores consistently?
- Grader Training -
- If they aren't well trained, it could take a while to figure out how to use their scoring techniques, so they score different at the end from the beginning.
You have to construct standardized, objective rubrics and grading scales - this is critical, particularly for essay and oral exams.
On November 9, 2016 when the electoral college vote totals were tallied, everyone saw the same facts, the same results. But half the country thought the numbers were good and half thought they were bad. When evaluating the facts of a job or promotion candidate, the organization has to agree, before hand, what ‘good’ facts look like and what ‘bad’ facts look like. Good ones are valid ones - they are accurate predictors of who is more likely to be successful in the position. Good and bad are determined by the test maker, not by the graders. The graders merely test whether the performance matches the pre-determined standard of a good performance. |
What’s a rubric?
It’s where you describe in as much detail as possible what a good answer looks like. If you’re looking at content, you identify the key ideas in the answer, and possibly how many points a candidate should get if they mention each of those ideas. It has to be as objective as possible. The Fire Safety Instructor Training Manual has some examples, but even those aren't as strong as they could be.
Good rubrics take a lot of thought - but it's thought that helps you clarify and communicate what a good answer means so that different graders give the same answer the same score.
Here are some examples:
UC Berkeley Graduate Student Instructors Training
Society For Human Resource Management - This example doesn't explicitly tell graders what the scores (1,2, 3, 4, 5) look like, as the previous one does.
BARS - Behaviorally Anchored Rating Scales - This is an article on using BARS to grade Structured Interviews. Look particularly at Appendices A & B.
How Olympic Ice Skating is Scored - I couldn't find an actual scoring sheet, but this gives an overall explanation of the process.
My experience is that good rubrics force graders to ground their scores on something concrete, but they can also miss interesting and unexpected things. It's useful for graders to score each candidate independently, and then discuss why they gave the scores they did - particularly those whose scores vary from most of the scores. Individual graders may know more about the topic which gives their scores more value. Or may not have paid close attention. Ultimately, it comes down to an individual making a judgment. Otherwise we could just let machines grade. But the more precise the scoring rubric, the easier it is to detect bias in the graders.
Accountability
Q: What if a candidate thinks she got the answer right on a question, but it was scored wrong?
Everything in the test has to be documented. Candidates should be able to see what questions they missed and how they were scored. If the test key had an error, they should be able to challenge it.
Q: Are you saying everything needs to be documented?
If there is going to be any accountability each candidate’s test and each grader’s score sheets must be maintained so that if there are questions about whether a test was graded correctly and consistently from candidate to candidate, it can be checked.
In the case of an oral exam or interview, at least an audio (if not video) record should be kept so that reviewers can see what was actually said at the time by the candidate and the graders.
Q: Have you strayed a bit from the Merit Principles?
Not at all. This all goes back to the key Merit Principle - selecting and promoting the most qualified candidates for the job. There won’t be 100% accuracy. But in general, if the test is valid, a high score will correlate with a high job performance. But unless the test is also reliable, it won’t be valid. The more reliable the test, the more consistent the scores will be under different conditions and graders. The best way to make tests more reliable is to make them as objective as possible.
No comments:
Post a Comment
Comments will be reviewed, not for content (except ads), but for style. Comments with personal insults, rambling tirades, and significant repetition will be deleted. Ads disguised as comments, unless closely related to the post and of value to readers (my call) will be deleted. Click here to learn to put links in your comment.