-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Include *all* years matching a warming level #33
Comments
I report here from @NiklasSchwind
|
And my reply here. I agree in general. Having only the first-encountered 21-year can be a meaningful design choice if the intent is to apply to continuously warming temperature pathways. Now in our application (possibly in contrast to @byersiiasa's rime), at least for now, we don't limit the emulator to monotonously increasing scenarios. In any case, besides any testing that should be done obviously, it is a question we might want to discuss with Carl & CA, precisely because it depends on the range of desired application. From my perspective though, the emulator calibration base should be as broad as possible, to avoid being overconfident (overconfidence is the worst thing one want to avoid). This is especially true because of the focus on strong mitigation scenarios (like 1.5C) at CA and IIASA. In practical terms, in general, in term of making things optional or not, there should certainly be a simple way to try out various options in the code, this should be as generic and as little option-specific as possible (like a custom, user-defined path where calibration data should be stored), because we cannot possibly keep track of all variations and choices we have and will encounter. And on this particular issue, I believe we should meet a clear decision ASAP to move forward in a linear manner -- otherwise we just lose time, and I know what I am talking about (I do love exploring all options). |
here you are @carlschleussner |
Hi both, But tend to agree with your point Mahe, it would be actually useful to use all information we had. If we can make it optional (binary switch: Either only the first one, or all at that level, or only on the way 'down') that would be a nice feature, too. It would allow us to explore the differences in the way up/down in a systematic manner also - might be nice for the paper just to have that toggle-switch and show it in a figure. Is that what you had in mind also? |
From the code perspective, all options can be explored, and the functions are made as flexible and explicit as possible. And there is a config file that sets the defaults, where a new option can be introduced that controls this behavior. However, I think it's important to decide on a default you don't have to think about everytime you use the tool. I'd be in favor of including all points by default, but again that's a design choice. |
From this exchange, and to move forward, I suggest the following:
Let me know if you're OK with that, or if you want to keep the current default of not including all points. |
Thanks! Agree with that proposal. Just to make sure, you're implementing two options (default all, optional:first), right? As I see limited value in the 'first' only approach, an alternative version would be: Might require a bit more work though (because we'd need to know if and when peak warming had occurred in the underlying GMT time series). And I don't want to overcomplicate things :) |
Yeah that was my intention. I also don't see much interest in the "first-only" approach, except that it tries to minimize any effect a time-lag might have at constant warming -- that could be fine if we'd only use it with fast warming scenarios, but we don't. As far as peaking, I am not sure whether any issues that arise (I have not investigated it first hand but I do remember it being mentioned during Katja Frieler's work on pattern scaling) are related to the time lag or to other effects (thinking of negative atmosphere-ocean heat flux). It's not a big deal to flag any peaking and create warming table pre or post peak, but of course it would only apply to some scenarios and not other, and then there is the question of how to apply it. So probably you'd want to have two calibration tables (pre- and post-), and also diagnose pre- and post- peaking when using these tables on new temperature pathways to calculate. That is doable, but I wonder whether it's the smartest way to deal with that issue. On the one hand, it comes at a cost because it complicates the code (never a good thing for maintenance, bugs, need for testing etc), and on the other it's not clear to me how large the potential benefit is. I mean, is it really a binary thing ? (peaking, not peaking?) Maybe it depends more on the slope, or maybe on the integral of past warming (heat uptake)... And splitting the training data in two, with rather few scenarios providing training data for the post-peaking period, may introduce biases of its own. But I don't know. I'm just seeing the complications without having being confronted first-hand to the problems, and without a clear feeling that it would be a solution. My suggestion here would be to treat that as a separate issue on peaking and discuss that and try out various things when we want to dedicate time on this. |
In any case for now I'll have "selection_mode" or such parameter that can take different values that we can extend over time. |
Fair - not sure it’s so important for the CMIP6 badge - but might become in the future if we can train RIME actually on overshoot data. But that’s for a later day :) Good discussion, though.
On 17. Oct 2024, at 15:38, Mahé Perrette ***@***.***> wrote:
Yeah that was my intention. I also don't see much interest in the "first-only" approach, except that it tries to minimize any effect a time-lag might have at constant warming -- that could be fine if we'd only use it with fast warming scenarios, but we don't.
As far as peaking, I am not sure whether any issues that arise (I have not investigated it first hand but I do remember it being mentioned during Katja Frieler's work on pattern scaling) are related to the time lag or to other effects (thinking of negative atmosphere-ocean heat flux). It's not a big deal to flag any peaking and create warming table pre or post peak, but of course it would only apply to some scenarios and not other, and then there is the question of how to apply it. So probably you'd want to have two calibration tables (pre- and post-), and also diagnose pre- and post- peaking when using these tables on new temperature pathways to calculate. That is doable, but I wonder whether it's the smartest way to deal with that issue. On the one hand, it comes at a cost because it complicates the code (never a good thing for maintenance, bugs, need for testing etc), and on the other it's not clear to me how large the potential benefit is. I mean, is it really a binary thing ? (peaking, not peaking?) Maybe it depends more on the slope, or maybe on the integral of past warming (heat uptake)... And splitting the training data in two, with rather few scenarios providing training data for the post-peaking period, may introduce biases of its own. But I don't know. I'm just seeing the complications without having being confronted first-hand to the problems, and without a clear feeling that it would be a solution. My suggestion here would be to treat that as a separate issue on peaking and discuss that and try out various things when we want to dedicate time on this.
—
Reply to this email directly, view it on GitHub<https://protect.checkpoint.com/v2/r02/___https://github.com/iiasa/rime/issues/33%23issuecomment-2419576120___.YzJlOmlpYXNhOmM6bzpmMGYzYzZhNWI4ODc3NmMyYTY4MGRmYzEyY2NhNTkwOTo3OjkyNTg6N2ZlZGMyYzgwZWMwYjIwMTU3YzI0YTY5NzAwOGFmMmI5ZmU2YjRkYWZkOTRiNjdiNTYwNGNkZDA4NmEwMDliMzpoOlQ6Tg>, or unsubscribe<https://protect.checkpoint.com/v2/r02/___https://github.com/notifications/unsubscribe-auth/BKBMB6NF5YH5GK5I2QR3JJLZ364WBAVCNFSM6AAAAABQB6VTUKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDIMJZGU3TMMJSGA___.YzJlOmlpYXNhOmM6bzpmMGYzYzZhNWI4ODc3NmMyYTY4MGRmYzEyY2NhNTkwOTo3OjM3Mzk6OTI2N2RlMjc3Njk1ZTgxYTRmNjQ2YjRhMmUzZjRkODRlNmM3MzJiODRiM2E3YWU4MzZkYzM3ZmMyYWEyMThlYTpoOlQ6Tg>.
You are receiving this because you commented.Message ID: ***@***.***>
|
Great, thanks! |
Currently we include only the data points from a time-series when it first crosses a warming level. Why not include all data points that fall into the warming level?
The figure below (please ignore old CIE) shows the current rimeX emulator overlaid with the actual ISIMIP2 time-series. The black dots (not indicated in the legend) are the points we actually retain in the calibration. The 2.6 scenario for precipitation (top right) shows significant variability in the 21-yr running mean that is not captured by our emulator, because we simply don't use data points we could use without that self-imposed rule.
I am not sure where that rule came from in the first place. I understand this could be a conservative measure to avoid mixing data points from a "decline" scenario, but I'd argue they also contribute to the uncertainty. I rather think this was a "rule" useful in the previous methodology that is not useful here any more, since we can integrate that data naturally.
The text was updated successfully, but these errors were encountered: