View original document

The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.

Key Excerpts from FCIC Staff Interview of Roger Stein
Roger Stein is a Group Managing Director at Moody's in charge of the Quantitative Research and
Analytics Group. Beginning in 2005, Mr. Stein led the group at Moody's that developed a new
model for rating subprime RMBS called "M3 Subprime".
On Wednesday, May 23, 2010, Ryan Bubb and Vic Cunicelli of the FCIC interviewed Mr. Stein of
Moody's at the New York office of Moody's outside counsel, Sullivan and Cromwell.
The key finding in the interview was that Moody's calibrated the M3 Sub prime model so that the
levels of losses that it predicted matched the views of Moody's mortgage analysts. This was done
by adjusting the parameters of the model that had been initially estimated based on historical
data.
To generate the credit enhancement level needed for a bond to be rated Aaa, the losses predicted
by the model were adjusted further. When a pool of loans was run through the model, the model
first generated a set of 1,250 economic scenarios; each scenario predicted the losses on the
mortgage pool (using the calibrated parameters described above). To calculate the credit
enhancement level needed for a bond issued by the pool to be rated Aaa, Moody's then took the
highest pool loss from among those 1,250 scenarios and multiplied it by a factor greater than two
to generate a "stress" scenario. This factor was determined during a calibration exercise
performed during the model's development so that the resulting losses in this worst scenario
generally matched Moody's analysts' expectations for how high a loss to consider for some set of
representative mortgage pools.
Below are the key passages from the transcript. The full transcript for the interview is
forthcoming.

Mr. Bubb: You talked about calibrating the model so it was consistent with experts on subprime pools in
the RMBS group. A natural way to do that I would think would be to calibrate it to the prior tool. But that
wasn't the approach?
Mr. Stein: No.
Mr. Bubb: Why? Do you have a sense?
Mr. Stein: No, I don't actually know why that tool wasn't one of the tools that people wished to use for
calibration. But I do know that we calibrated to a set of pools that a number of committees looked at
carefully and felt were representative and for which the partners of ours within the RMBS team felt they
had a good understanding.
Mr. Bubb: When you say calibrate, I think I understand what it means, but will you explain what it means?
1

Mr. Stein: Sure, it may be good to take a step back and just talk about models more generally. So I've been
building models for many, many years, and almost every model I can think of is used, let me say it more
precisely, I cannot think of a model in current use in finance that doesn't get calibrated in some way. Unless
three conditions hold. I think those conditions need to be that the data that one uses to build the model is
plentiful, descriptive, covers maybe several economic cycles. That one believes that one's model is in fact
complete, and functionally correct, a complete description of the entire problem. And thirdly, that one
believes the world will not change in any way materially from whatever the data and model development
period was. I don't think that we believed any of those three things to be true. So, in such a setting, it's
necessary to calibrate the model to something else. And one could choose many different things. Some
people have advocated using for example CDS spreads, which turned out to be very, very tight during the
period that we were developing the model. There are academic models that one could use. I suppose one
could look at the estimates of analysts on the Street. We have our own analysts who have expert views, so
we chose to calibrate the model to those views.Mr. Bubb: So what does that mean, to calibrate a model to
the expert views? I mean, describe that in more detail what that means.

2

Mr. Stein: So, the model output that comes out is, let's say there's an average loss in the pool, call it the
expected loss. If the loss comes out much lower than the analysts feel is appropriate, then what needs to be
done is to change some of the historical parameters that are based on historical data to increase the risk
such that the losses come out to higher levels that are consistent with what the analysts and our experts
believe the true risk of those pools are. That was particularly true in the tail.
Mr. Bubb: What were the historical parameters that were used for calibration?
Mr. Stein: There were many. So a number of things were done. The prepayment baselines and sensitivity
to home prices, to home price increases, was dramatically reduced. There were additional adjustments for
different types of loans that made those loans riskier in the model than they would have been historically in
the data. There were adjustments to the actual losses for a variety of different types of things, such as IfO
provisions and so forth, documentation type was made riskier, a bunch of things like that. You can imagine
a whole suite of these types of things. I don't recall all of them though.
Mr. Bubb: So I just want to make sure I understand this. So you've got a bunch of data which we're going to
get to eventually, actually just to talk about what the data was, that provided a set of parameters, for
example, you mentioned one of them was the prepayment sensitivity to increases in housing prices. The
historical data would tell you that based on that period from which the data came, on average, at this sort of
sensitivity, but experts in maybe the RMBS group, experts on subprime, thought that those sensitivities
were wrong, that they were insufficiently, the historical data reflected insufficient sensitivity ...
Mr. Stein: More generally, they felt that the losses the model was producing were not consistent with the
risks in the pools. They were too low.
Mr. Bubb: I see. So this was not based on expert views on this intermediate parameter but rather views on
the ...
Mr. Stein: In many cases it was.
Mr. Bubb: Okay.
Mr. Stein: But ultimately, if the analysts' theory about a particular parameter was X, and by setting the
parameter to X, the losses were not high enough, then we would make other adjustments as necessary for
subprime mortgages during that period. Our analysis suggested that prepayment is a big driver of pool
level losses, because if a loan prepays, it can't default And so if prepayment rates are very high, even very
risky borrowers leave the pool early. So by lowering prepayments we make the conditional probability of
default, the conditional loss rate of the pool, much higher.
Mr. Bubb: I see, but you increase the tweak here, you increase prepayment sensitivity.
Mr. Stein: Yes. We increase prepayment rates.
Mr. Bubb: Sorry. Is it the levels or the slope?
Mr. Stein: Both. The prepayment baseline was decreased.
3

Mr. Bubb: Was decreased?
Mr. Stein: Yes, decreased.
Mr. Bubb: I'm sorry. And, what about the slope?
Mr. Stein: The slope was made less sensitive to positive home price appreciation.
Mr. Bubb: I see. And was, forgive me if I just, if you said this and I just forgot, was this tweak because of
expert views on that prepayment sensitivity, or expert views on the ultimate answer in terms of expected
loss?
Mr. Stein: The calibration of the model was a iterative process.
Mr. Bubb: I bet.
Mr. Stein: With members of my team, the RMBS team, and other groups, so we would meet weekly to
discuss different approaches to doing calibration. I don't recall whether this specific tweak came from a
modeler or whether it came from an RMBS analyst. But there was a fair mix of both I would say.
Mr. Bubb: And, help me think about this calibration, so, urn, one view of what was going on here was that
your experts are wrong, that the data say what the data say about ultimate pool losses based on historical
experience and that these experts' views in fact are subject to various biases, we make mistakes, experts
make mistakes, and that you were over-riding what the data were saying based on, sort of, idiosyncratic
views. You know, why have a model at all, would go this view. What do you think?
Mr. Stein: I think several things, one is, I don't think ex post it's necessarily feasible to determine whether a
specific view on a probability is right or wrong. I think probabilities by construction, unless they're zero,
have some probability of happening. With a single draw it's impossible to say whether that probability was
correctly assessed or not. With that said, I think that history bears out the, let me say this differently. Had
we done as you suggest, had we done as you are recommending, for example, we would have had levels
that are much lower than what ultimately were assessed by the analysts. Historical data was not capable of
producing losses in the tail, particularly, that analysts felt were appropriate. More importantly, it couldn't
even produce losses at mean that were sufficiently risky. I've sometimes wondered if we had had more
historical data, whether that would have made a difference, but our recent experiments suggest that even
using this last period of data, one still could not produce losses that are sufficiently large in the tail based
solely on historical data.
Mr. Bubb: So do I infer from that, that essentially the losses that we experienced over the last several years
are unprecedented historically?
Mr. Stein: Certainly the depth of the home price decline on a national level is unprecedented since the
change of various banking regulations. At a national level that's true I think.

4

[48:30]

Mr. Stein: ... Maybe shocked is the wrong term, but it's certainly not the loss one would have gotten by
looking at historical data. It's a much higher loss than that. Then, after we've done that 1200, I'm sorry,
1250 times, we can then have a distribution in some states of the world, a dip in home prices, I'm sorry, for
some pools, a dip in home prices in Year Three might be bad, and for in others a dip in home prices in Year
Two might be bad. So whatever that last path is that generated the worst loss, that is then shocked again,
and that's pushed out into the tail to make it several times, a couple of times, some number of times worse
than what it would have shown up in the simulation by itself.
Mr. Bubb: I'm, forgive me, still a little confused. We've got, I was with you when we had our 1250 draws,
each of which had already, each of which represented an expected pool loss for that economic scenario.
And we shocked it, based on what you were calling before the shock to the historical simulation. And now,
I'm sorry, just explain it again, the second shock is applied to which of those draws?
Mr. Stein: So we take the worst case.
Mr. Bubb: The single worst case?
Mr. Stein: The single worst case under the simulation for that particular pool.
Mr. Bubb: Meaning, of those 1250 draws, this is the worst of 1250 draws.
Mr. Stein: Yes exactly.
Mr. Bubb: And that one we're going to ...
Mr. Stein: Shock again.
Mr. Bubb: I see, and that's used how?
Mr. Stein: That then becomes the basis for the triple-A enhancement level.
Mr. Bubb: And it's the case that the triple-A enhancement level is set so that under that scenario so
shocked out, losses just come to below the triple-A. Is that basically... This is for the model output I
understand that the committee has to do things.
Mr. Stein: There is, there ... The model isn't the same as the rating.
Mr. Bubb: Of course. Yep, yep, yep.
Mr. Stein: And so, there's a credit enhancement level that is for the collateral of the transaction. All of the
subprime transactions that I'm aware of had cash flow waterfalls as well. There's a separate analysis that
needs to be done on those cash flow waterfalls that is not part of this model. So this was a starting point for
the enhancement levels. There was additional analysis with stress testing and things that was then done,
using some of this output, to arrive at whatever the final enhancement levels would be for a transaction.
5

And that would also contemplate various types oflegal or structural issues as well as behavioral issues of
the servicer, and so forth.
Mr. Bubb: But is in general sort of the right way to think about this, that worst draw, the 1250, after being
subject again, this is another shock, that that outcome is modeled in the cash, whatever is being done to
model the cash flows to the actual bonds of the deal. And that the principle is, the triple-A ought to not be
hit in that scenario.
Mr. Stein: I believe a way to say it would be that the analysts would look at that output, make adjustments
to that output, and then do what you have described.
Mr. Bubb: The adjustments would be based on the legal issues ...
Mr. Stein: Those adjustments might be [inaudible] to the collateral-based adjustments, so they would
typically be based on thoughts about the collateral, and servicer, and these types of things.
Mr. Bubb: So how did you figure out how big to make that shock on the last, the worst draw?
Mr. Stein: Again this was a fairly long process. We made adjustments such that the levels were consistent
with what analysts and other experts in the RMBS team and more broadly within Moody's were
comfortable sufficiently represented the risks in the tail of the transaction, of the population of
transactions. It's important, I think, to note that the historical data provided no guidance in this regard, so
you might think of it as a Bayesian type exercise, if that works as a way of thinking about it.
Mr. Bubb: So, we're really using the analysts' current beliefs about the tail events to drive the ultimate
assumptions used in the model?
Mr. Stein: That's correct One of the, it wasn't just the analysts, again, I'm not sure how familiar you are
with Moody's processes, but these were not just opinions of analysts, they were long committees that
would discuss these issues and so forth for resolving them. And so I believe it was the case that these
represented the best views of the organization with respect to those tail losses primarily again because the
historical data provided no guidance in this regard.
Mr. Bubb: So, that's interesting. So the model in some sense is the tail, the analyst is the dog, the committee
is the long ... to use a metaphor.
Mr. Stein: I don't want to use that metaphor. I think we have a model that must contemplate events for
which there is no data. It's not clear how else one might calibrate that model, besides using expert
judgment.
Mr. Bubb: What was the point of having the model at that point? I mean, another way to go here would be,
look, we've got analysts who have debated these things for a long time, they're good Bayesians, they've
updated based on the latest greatest events in the housing market, they can just rate deals with their own
gestalt assessments of the state of things.

6

Mr. Stein: I think that there are actually, I mean there are several benefits for the model. One of those
benefits was transparency. The goal of the model was to provide market participants, investors, issuers,
others in the market, with a better understanding of the factors that drive Moody's ratings. In addition,
while, while the committee view that was used for this calibration may reflect a general view, on an
individual transaction, that committee's view is not available, typically. And so, for an individual
transaction, you need to be able to rank the different transactions that come in to be rated based on the
general expectation about losses in general on the market The expectation about losses in general is
what's used for the calibration. For individual analysis, an analyst will take to their committee an
individual transaction which may be very different than the general view, the general market So the
purpose of the model is two-fold: to provide transparency and to provide consistency in the ranking of the
different transactions.
Mr. Bubb: A related question. So in that view, one could have a transparent model, a model that performs
the role that you described, both the ranking and the transparency role, without using any historical data.
So you could literally have a set of assumed coefficients that would output risks for a loan that had been
tuned so that the ultimate losses of the pools matched expert judgment, but that didn't begin with what one
might characterize as a contrivance, of using historical data, and generate, and achieve the same two
functions, of transparency and ordinal ranking of pools that you described. What am I missing, or am I
missing ... ?
Mr. Stein: I don't think that this is in any way a contrivance. I believe that to do as you propose would be
extraordinarily difficult because of the interactions of the different relationships of prepayment, default,
and severity within the model. So I think that the historical data provides tremendous guidance on the
shapes, the relationships of different behaviors, but may not provide sufficient guidance on the levels of
those behaviors. So the shape of the prepayment curve, for example, is a complicated function of the
structure of the loan which is probably difficult to just come up with gestalt I think is the word you used.
On the other hand, moving that up or down will affect the levels of prepayments, and so while the historical
data may provide good guidance on the relationships and the shapes of things, it is probably given the
historical record going to provide optimistic levels of the model.
Mr. Bubb: On that note could you describe a bit the data that was used?
Mr. Stein: Certainly. At some point I'd like to take a break ...

4834-1368-3718, v. 1

7

•

•

•