OIC, so may I know more about the meaning of this? Thanks a lot about your comments David.I don't exactly agree with chiyui about the methodology, but I think it is essentially similar (some % of an anchor that is based on top 5%)
I have highly confident that the pass methodology is based on a ratio of an anchor to the Top 5%. However, specifically:
- I am currently unclear on whether the anchor is the average of the top 5.0%; or simply the score of the candidate who fell on the 5.0% quantile. Note the analog to ES and VaR! if there are 1,000 candidates, the first method produces an "anchor" = average (top 50 scores); note as a conditional mean, this is analogous to ES; while the second method returns an "anchor" = the candidate who happens to earn the 51st-highest score (as a quantile, note this is analogous to VaR)
- I do not think percentage multiplier (e.g., 75%) is either constant or even ex ante specified: I think the ratio is ex post calibrated to manage an overall outcome vis-a-vis perceived difficulty. I thinks this value is X, in other words. FWIW, thanks,
So why doesn't GARP keep the exam difficulty level at a more stable state by means of a more careful evaluation of the multiple-choice questions, instead of lowering the threshold in order to push the passing rate to their desired level?Hi NewComer, before i go too far out on a limb: I cannot know GARP will be applying this methodology in May 2013, I spoke with GARP about their method in 2011 and again in late 2012, I have not even tried to ask them about it for the May 2013 exam. I cannot assume the 2011-2012 methodology persists ... but okay ...
... my understanding of the floating X was for one simple reason (I don't know about other factors like supply/demand; I don't even know that it has anything to do with managing a pass rate, although it is interesting that there is a fairly consistent pass rate, sans one outlier): due to the rapid syllabus changes, I had understood the floating X to be a "safety valve" in case the exam was tougher than expected, as evidenced by the distribution, such that the passing grade could be modestly lowered to compensate for a difficult exam (i.e., more difficult than intended). This sort of an outcome might not be evidenced by the top 5% (this group might be excellent under any test; in distributional terms, the upper tail may not be sufficiently descriptive-yes?). For example, maybe there is a target X = 70% or X = 75%, but maybe the exam is experienced as abnormally, unintentionally difficult such that 70%*top 5% [ES |VaR] admits too few candidates so it can be lowered. At some point, I got the impression the float's purpose was in this error direction (accepting additional candidates), not the other (i.e., not to tighten the noose). I doubt it totally floats, but it would make sense to me if there were a range given the way the syllabus changes so much ... it cannot be overstated: it would be almost impossible to write a sequence of exams with identical difficulty levels, over time, when the syllabus changes at its current rate but also (actually, the even bigger reason is... ) in conjunction with a significant degree of over-assignment. That is, the actual exam only tests a small % of the syllabus. Overshadowing even the rate of change is the exam sampling variation. (I don't agree with the rapid syllabus change nor with what i would call the "over-assignment" of AIMs, i think the combination of both is detrimental and counterproductive)
... so every year, naturally candidates would like to understand how many corrects might constitute a pass rate, but so far as i can tell, there has never been any evidentiary basis for throwing out any numbers at all. So far, it looks to be "grading on a curve," and if you like, to your point, the curve further has even two unknown parameters. Thanks,
So I think they would say the primary goal is not to generate stable multiple choice questions; and in their defense, the exam is not very amenable to traditional test taking techniques (I do mean that in a positive way). It could be argued, i guess, that the exam tries to be fair in the sense that everybody has the same experience, but not exactly fair in the sense that we know exactly what will be asked. I do think this methodology, in my humble opinion, as a byproduct, frustrates the right segment (i.e., those who prefer technique and rote memorization) and there is another segment (learners, practitioners) that it seem to give less pause. That's all devil's advocate, it's not even necessarily my position, it's too far theoretical for me now. Thanks,The FRM Exam is a practice oriented examination. Its questions are derived from a combination of theory, as set forth in the readings, and “real-world” work experience. [<<-- this is a key phrase: they have always claimed to source questions from practitioners in addition to the text assignments, a practice that challenges multiple choice stability] Candidates are expected to understand risk management concepts and approaches and how they would apply to a risk manager’s day-to-day activities