Term
| If a teacher made statement that all students performed above average this year on a test, why is this false? |
|
Definition
there has to be students who made average -statistically not possible |
|
|
Term
| On criterion reference test, items will be more general or specific? |
|
Definition
|
|
Term
| On a criterion reference test, items will be of the same or varying difficulty? |
|
Definition
|
|
Term
| On criterion reference test, why are items more specific and of the same difficulty? |
|
Definition
| because objectives are written at certain levels, only vary 1 level at the most |
|
|
Term
| What is the idea behind criterion reference testing? |
|
Definition
|
|
Term
| Give example of a condition in instructional objective? |
|
Definition
| Given a map, calculator, globe, etc... |
|
|
Term
| What is the main function of a test blueprint? |
|
Definition
| determines if items match instructional objectives |
|
|
Term
| What are the components of a well-written objective? |
|
Definition
| conditions, criteria, behavior/performance |
|
|
Term
| What can we do with correlations? what can we NOT do? |
|
Definition
| make predictions, cannot tell cause and effect |
|
|
Term
| Which correlation is more useful for making predictions: large negative correlation or large positive correlation? |
|
Definition
|
|
Term
| what is the correlation between test scores and criterion called? |
|
Definition
|
|
Term
| If a teacher-made test has items that match her objectives, then the test should have high... |
|
Definition
|
|
Term
| If you compare a newly constructed rating scale that measured honesty to an older scale that also compared honesty, what validity are we looking for? |
|
Definition
concurrent validity -not construct, even though it deals with emotions |
|
|
Term
| Why should you be more confident in reliability of a test if it is given to the whole school, not just one grade level? |
|
Definition
| you ensure the variability with a heterogeneous (mixed) group |
|
|
Term
| What validity is it when SAT scores are correlated with GPA? |
|
Definition
|
|
Term
| In test-taker groups, the more similar the population, the less the... |
|
Definition
| reliability -doesn't show true population if everyone thinks the same |
|
|
Term
| what is a criterion referenced test? |
|
Definition
Compares to predetermined standards rather than to other students -measures mastery -standardized or teacher-made -shorter and narrower -reflects extent to which goals are being met |
|
|
Term
| How many options are available when making a checklist? |
|
Definition
Two -yes/no, observed/not observed, etc |
|
|
Term
| If you test the same people with the same test at a later time, it is what reliability? |
|
Definition
|
|
Term
| If you only have 1 test, what reliability can you calculate? |
|
Definition
Internal ONLY -item to item comparisons |
|
|
Term
| What kind of validity is most important for regular teachers? |
|
Definition
content validity -because what we test must match up to our standards |
|
|
Term
| What is construct validity? |
|
Definition
how well test scores represent behavior predicted by idea/theory/etc -unobservable traits (such as honesty, depression,etc) |
|
|
Term
| If a company selects a new hire based off a test, what kind of validity? |
|
Definition
|
|
Term
| What is a norm-referenced test? |
|
Definition
compare current test takers to past test takers (the norm group) -standardized test -results are "relative" to the norm group |
|
|
Term
| "Marie is above average" is norm or crit ref test? |
|
Definition
Norm reference compares her to the norm group |
|
|
Term
| "Marie is in top 25% of her class" is norm or crit ref test? |
|
Definition
|
|
Term
| "Marie got 97 out of 100 questions correct" is norm or crit ref test? |
|
Definition
criterion referenced test -measures mastery |
|
|
Term
| Does norm or criterion reference usually have more variability of scores? |
|
Definition
|
|
Term
| How many items per objective are usually on Norm-ref tests? |
|
Definition
1 or 2 per objective -covers many objs -purpose is to rank child |
|
|
Term
| what does a NON-overlapping band mean in band interpretation? |
|
Definition
-possibly a REAL difference in scores or ability -overlap would mean it is most likely from chance |
|
|
Term
| What is the difference between a true score and an obtained score? |
|
Definition
Error -obtained does not take error into account (true does) |
|
|
Term
| What is "Alternate Form" reliability? |
|
Definition
reliability b/w 2 diff forms of test given to the same person at 2 different times -indicatess equivalency of forms (such as test form A & B or retake exams) |
|
|
Term
| What are some sources of error? |
|
Definition
-test itself (errors) -test conditions (setting, etc) -test taker (sleepy, not studied) -scoring errors |
|
|
Term
|
Definition
-value representing a score that is free of error -impossible to find because all tests have errors -score that takes error into account |
|
|
Term
| What is content validity? |
|
Definition
how much a test matches/measures the teacher's objectives -can also apply to state standards |
|
|
Term
| what is the highest level of Bloom's taxonomy? |
|
Definition
|
|
Term
| What are the ranges of SD? |
|
Definition
1 SD = 68% 2 SD = 96% 3 SD = 99% |
|
|
Term
| What is test reliability? |
|
Definition
-measures dependability and consistency over time -should yield same measurement of same variable -3 kinds: test-retest, alternate forms, internal consistency |
|
|
Term
| Which measures level of mastery: Norm or Crit? |
|
Definition
|
|
Term
| Can we observe someone "defend"? |
|
Definition
|
|
Term
| How do you make a performance assessment more of a direct measure? |
|
Definition
| include a variety of mediums |
|
|
Term
| Is it possible to produce a test that is completely without error? |
|
Definition
No -there will always be SOME error |
|
|
Term
| Is norm or criterion referenced tests more useful in classroom decision making? |
|
Definition
| Criterion, because you look at mastery rather than comparison to other students |
|
|
Term
|
Definition
-Standard error of measurement -variability of error scores -NOT the difference between true score and obtained score -as reliability goes down, SEM goes up |
|
|
Term
| Jared has obtained score of 86, so we can be 68% sure his TRUE score lies between __ and ___ |
|
Definition
83 and 89
1 SD x 3 = 3 86 - 3 = 83 86 + 3 = 89 |
|
|
Term
| What kind of test will any test in the content area be? |
|
Definition
Achievement test -content validity |
|
|
Term
| "Reliability goes up, SEM goes down" is pos. or neg. correlation? |
|
Definition
negative -they go different directions |
|
|
Term
| What are the advantages of a performance based assessment? |
|
Definition
-Take different forms -Assess different kinds of skills -Easily worked in to lessons -Higher cog effect needed -active learning |
|
|
Term
| Jared has obtained score of 86, we can be 95% sure his true score lies between __ and ___ (SEM 3) |
|
Definition
| 80 and 92 (2 SD x 3 = 6 86 + 6 = 86 86 - 6 = 80) |
|
|
Term
| What are the 3 types of validity? |
|
Definition
Content validity Criterion-related validity Construct validity |
|
|
Term
| What is the most important validity for classroom teachers? |
|
Definition
|
|
Term
| What can you NOT tell from correlation between two variables? |
|
Definition
cause and effect (can only predict) |
|
|
Term
| What are you looking for when interpreting a band chart? |
|
Definition
| seeing which bars overlap and which do not |
|
|
Term
| What is test-retest reliability |
|
Definition
2 scores from same person on same test at diff times -indicates stability |
|
|
Term
| What are the 3 characteristics of a good obj? |
|
Definition
| specific, observable , measurable |
|
|
Term
| What is standard deviation? |
|
Definition
most accurate measure of variability -includes all scores in a distribution -estimate of variability |
|
|
Term
| How many items per obj does crit. ref usually have? |
|
Definition
3+ per obj -emphasis on figuring out where response is lacking -shows mastery |
|
|
Term
|
Definition
two variables move same at first, to a point, then show negative relationship
Ex: drinking and feeling good, to a point, then you feel sick so (+)alchohol = (+)feelings then (+)alchohol = (-)feelings |
|
|
Term
| What validity is it when you compare test items to objectives? |
|
Definition
|
|
Term
| What are the 3 different ways to grade a performance assessment? |
|
Definition
Holistic (don't need to know) -Checklist -Rating scales |
|
|
Term
| What is "obtained score"? |
|
Definition
= true score (+) and (-) error score
difference is error |
|
|
Term
| The less variability in a test, the lower the... |
|
Definition
reliability -shorter tests have lower reliability -not enough items to measure more accurately |
|
|
Term
| How do you get away from multiple choice tests? |
|
Definition
performance assessments (informal) |
|
|
Term
| What is criterion-related validity? |
|
Definition
how well tests scores correlate to other tests of the same thing -requires correlation coefficient to be computed |
|
|
Term
| What validity is it when SAT scores are correlated with GPA? |
|
Definition
|
|
Term
| What is negative correlation? |
|
Definition
2 variable move in opposite directions
-as A increases B decreases or vice versa -think of negative as BAD relationship leading to a breakup |
|
|
Term
| What can you tell if you calculate less error on one test than another? |
|
Definition
| less error = more accurate measure |
|
|
Term
| Is there a "one size fits all" test in regards to validity and reliability? |
|
Definition
No
-validity and reliability may only be appropriate for specific population, when administered by competent user |
|
|
Term
| What is positive correlation? |
|
Definition
both variables involved move in same direction
-as A increases, B increases
think of positive as a good relationship since they move together |
|
|
Term
| In band interpretation, how do you find if a difference in scores is related to chance? |
|
Definition
|
|
Term
| If my watch is on 2 oclock but is broken, then the watch is...? |
|
Definition
Reliable but not valid (always know it will be on 2oclock, but it is only true 2 times each day) |
|
|
Term
| Jared has obtained score of 86, we can be 99% sure his true score lies between ___ and ___ |
|
Definition
77 and 95
3SD x 3 = 9 86 + 9 = 95 86 - 9 = 77 |
|
|
Term
| Which is narrower? Norm or crit referencing |
|
Definition
Criterion ref
covers few objects because you want to know how well they mastered each one |
|
|
Term
| If you give a condition in an objective, where else must you give that condition? |
|
Definition
in the test -same condition, same materials, same accuracy |
|
|
Term
| What are some disadvantages of performance assessments? |
|
Definition
-unReliabile -subjective due to outside influences -time consuming |
|
|
Term
| What is the rationale behind performance assessments? |
|
Definition
| It provides a direct measure of abilities. |
|
|
Term
| Is "with 100% accuracy" a condition? |
|
Definition
| No, it does not provide a context like "with a map" or something. |
|
|
Term
| What type of referencing is most appropriate for broad objectives? |
|
Definition
|
|
Term
| If student has raw score of 80, and SEM is 3, and want to be 99% confident of Trues score range, what would that range be? |
|
Definition
| 89-71 (3 SD's, times 3, then plus and minus) |
|
|
Term
| When is it okay to put opinion in a True/False question? |
|
Definition
If you attribute it to a source. "according to so and so..." |
|
|
Term
| What should you avoid when writing a T/F question? |
|
Definition
-using always/never -opinion |
|
|
Term
| What is the best way to test organizational thinking skills? |
|
Definition
|
|
Term
| What kind of test item is best for assessing high level thinking skills? |
|
Definition
|
|
Term
| What kind of test item are students most likely to guess on? |
|
Definition
| True/False and Multiple choice |
|
|
Term
| Which test item is the easiest to score? |
|
Definition
|
|
Term
| What level is appropriate for restricted essay items? |
|
Definition
Anything lower than application -because they do not have free range on the topic, they are restricted |
|
|
Term
| How can you more objectively grade essays? |
|
Definition
| Grade one criteria at a time, make the essays anonymous |
|
|
Term
| What is the mean of 81, 83, 82? |
|
Definition
|
|
Term
| What is the median of 10, 12, 8, 9, 7? |
|
Definition
|
|
Term
| What are the measures of central tendency? |
|
Definition
|
|
Term
| Which measure of central tendancy is repeated the most? |
|
Definition
|
|
Term
| Which measure of CT divides a distribution in half? |
|
Definition
|
|
Term
| Which measure of CT is the average? |
|
Definition
|
|
Term
| Which measure of CT is the 50th percentile? |
|
Definition
|
|
Term
| Which measure of Central Tendency is the most stable? |
|
Definition
| Mean, because it takes every score into account |
|
|
Term
| Student scores 49 on vocab test, and the mean for the class is 40, with an SD is 3. What is the z-score and what percentage of the class scored higher? |
|
Definition
z-score = 3
less than 1% scored higher
z=x-m
SD
z=49-40
3
z=3 then look at bell curve |
|
|
Term
|
Definition
| How spread out the scores are in a distribution. |
|
|
Term
| Which measure of a variable is the most dependable? |
|
Definition
Standard Deviation -Takes all into consideration |
|
|
Term
| If a set of scores has a variance of 0, what can you conclude? |
|
Definition
| Everyone in the distribution has the same score |
|
|
Term
| What happens to SEM when you decrease SD? |
|
Definition
When you decrease reliability, you have a more accurate test -SD is variability between scores |
|
|
Term
| How can you change the level of a multiple choice item? |
|
Definition
Change the distractors -make them more plausible -in stem, have them choose "best" answer |
|
|
Term
| Should you grade one whole essay at a time (on the same test, by the same person) |
|
Definition
| No, grade by criteria for all essays |
|
|
Term
| What does the "scoring criteria" for short answer and essay items tell you? |
|
Definition
| how many points it is worth, and what you will accept as correct |
|
|
Term
| In norma dist, approx what percent of scores lies between T-scores of 40-80? |
|
Definition
|
|
Term
| NCLB requires assessment at what grades? |
|
Definition
|
|
Term
|
Definition
| that students be tested annually |
|
|
Term
| According to the text, what is the real argument against High Stakes Testing? |
|
Definition
Using one score to make high stakes decisions -that score is only one snapshot in time -it also is biased, produces narrow scores, and causes teachers to "teach to the test" |
|
|
Term
| What kind of discussions should teachers have with students about High Stakes Testing? |
|
Definition
|
|
Term
| What are the 12 conditions for a HST program according to the American Educational Research Association? |
|
Definition
1. don't use a single score for high stakes decisions 2. everyone should have same resources and learning opportunities 3. validation for each intended separate use (don't use same score to tell graduation, promotion, financing, etc) 4. tell users the possible negative consequences of HST programs 5. test and curriculum are aligned 6. validity of passing scores and achievement levels (what the scores mean) 7. remediation available to those who fail the HST 8. attention to language differences 9. attention to disabilities 10. stick to rules about who will and wont take test (don't tell low-performing students not to come to school that day) 11. sufficient reliability researched for each intended use 12. ongoing evaluation of intended and unintended effects of HST |
|
|
Term
|
Definition
| Moving teachers around or not renewing contracts because of test results |
|
|
Term
| What are some test-taking strategies you can teach students for HST? |
|
Definition
-sleep, breakfast, study -follow directions carefully -read each item, passages, information carefully -manage test-taking time -easier items first -eliminate options before answering -check answers after completing test |
|
|
Term
| What does a positively skewed distribution tell you about the scores? |
|
Definition
majority of scores fall below the middle of the score distribution -there are many low scores, but few high scores |
|
|
Term
[image]
positively or negatively skewed? |
|
Definition
[image]
positively skewed
-many low scores, few high scores
-most scores fall below the middle
-tail is toward the positive end of curve
|
|
|
Term
[image]
positively or negatively skewed? |
|
Definition
[image]
negatively skewed
-many high scores, few low scores
-scores lump above middle
-tail is toward neg. end of curve
|
|
|
Term
[image]
| | |
1. 2. 3.
what distribution, and label mean, median, mode |
|
Definition
[image]
| | |
1. 2. 3.
positively skewed
1. mode 2. median 3. mean |
|
|
Term
[image]
| | |
1. 2. 3.
what distribution, and label mean, median, mode |
|
Definition
[image]
| | |
1. 2. 3.
negatively skewed
1.mean 2. median 3. mode |
|
|
Term
| Which measure of CT is most frequently used? |
|
Definition
|
|
Term
| Which is not affected by extreme scores, the median or the mean? |
|
Definition
median -represents the middle better when scores are skewed |
|
|
Term
| What are 2 modes in a distribution called? |
|
Definition
|
|
Term
| what are 3 or more modes in a distribution called? |
|
Definition
|
|
Term
| If each score in a dist. occurs with equal frequency, what is the mode? |
|
Definition
|
|
Term
| What is the least stable measure of CT? |
|
Definition
mode (a few scores can influence significantly) |
|
|
Term
| In a normal distribution, which measure of CT has the most value? |
|
Definition
| none, they are all the same value (all reach the same highest points on the bell curve) |
|
|
Term
| If the mean is 47, the median is 54, an mode is 59, what is the shape of the distribution? |
|
Definition
negatively skewed goes in same order they are listed |
|
|
Term
| What does the semi-interquartile range do? |
|
Definition
prevents extreme scores from influencing the sensitivity of a range (such as if everyone scores in the 40s but one person scored a 90)
only the middle 50% is computed, top and bottom 25%s are left out |
|
|
Term
| What does each quartiles in SIQR mean? |
|
Definition
Q1 is the point below which 25% of scores lie Q2 is the median, or 50% of scores Q3 is the top 25% percent of scores |
|
|
Term
| What is the most commonly used estimate of variability? |
|
Definition
|
|
Term
| What is the most accurate measure of variability? |
|
Definition
Standard Deviation -includes all scores in a distribution |
|
|
Term
|
Definition
| how much a single estimated score actually describes all scores within the range |
|
|
Term
| What do large and small score values represent in SD? |
|
Definition
| large SD means more variability, smaller SD means less |
|
|
Term
| How can you tell the strength of a correlation? |
|
Definition
| how close the numbers are to -1.0 or +1.0 |
|
|
Term
| What does the sign (-/+) tell us about a correlation? |
|
Definition
| whether it is negative or positive |
|
|
Term
| What does a correlation of .00 mean? |
|
Definition
| there is no correlation at all, high and lows in one are associated with highs and lows in another |
|
|
Term
| How do you tell pos or neg correlation from scatterplots? |
|
Definition
the direction of the slope -from left to right is negative (pointing down) -from right to left is positive (up) |
|
|
Term
| "Does the test measure what it is supposed to test?" is asking about... |
|
Definition
|
|
Term
| "Does the test yield the same or similar score scores consistently?" is asking about... |
|
Definition
|
|
Term
| "Does the test score closely approximate an individuals true level of ability, skill or aptitude?" is asking about... |
|
Definition
|
|
Term
| Why must a test have validity? |
|
Definition
To show that it measures what it says it measures (such as a 3rd grade math test actually testing 5th grade math, or even a math test actually testing reading skills) |
|
|
Term
| If someone inspects test questions to see if they correspond to what should be covered by the test, they are looking at what kind of validity? |
|
Definition
|
|
Term
| What is the problem with content validity? |
|
Definition
-hard to tell if construct tests are valid because they are abstract -only tells if the test LOOKS valid but it might be measuring something else (such as guessing ability,reading skills, etc) |
|
|
Term
| What are the three main forms of validity? |
|
Definition
content validity criterion-related validity construct validity |
|
|
Term
| If matching test to the standards, what validity is it? |
|
Definition
content validity -making sure what is meant to be testing is actually being tested |
|
|
Term
| What is concurrent validity? |
|
Definition
| The correlation between two tests that measure the same thing when given to the same people |
|
|
Term
| What is the difference between Construct validity and other forms of validity? |
|
Definition
| Content, Concurrent, and predictive all measure against something else (standards, other tests, future scores) while Construct measure something that is hard to test or hasn't been tested before |
|
|
Term
What evidence is each of these compared to? Content validity Concurrent validity predictive validity |
|
Definition
Content: test compared to standards or objectives Concurrent: test compared to other tests of same thing Predictive: test scores compared to future scores or abilities |
|
|
Term
| Which type of validity is evidence for alignment with curriculum? |
|
Definition
content validity -making sure test matches the standards and curriculum |
|
|
Term
| What does reliability tell you? |
|
Definition
| consistency with which a test yields the same rank for individuals who take the test more than once (think broken watch example) |
|
|
Term
| What are the types of reliability? |
|
Definition
Test-Retest Alternate forms Internal consistency |
|
|