[BUG] after the fix in this commit 6c69155, does the result in the paper still hold for Multivariate inference #463
Replies: 1 comment 2 replies
-
|
@FerociousCentaur Since this is not really a current bug, but more of a question, please use discussions for this. The results in the tech report are not affected by the issue because it primarily affected fine-tuning, as noted in the PR discussion. Could you clarify what you mean by the following and provide an MWE?
Using past_covariates as future is not the correct and expected behavior. However, even if it were, this is not really lookahead bias because past_covariates are only supposed to have data before the prediction start date. If it contains future data, then the data formatting is inaccurate. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Bug report checklist
Describe the bug
The bug fixed in the commit says the future and past covariates were not separate because of the way they were calculated. When I personally tried it, I found that even when i don't provide any future covariates, the past covariates keys were used as future covariates as well, so basically there was a data leak issue, which was giving the move look ahead bias. So I just want to confirm, that are the results in the paper still hold true ? Or if there are new results after the fix, can you please share them.
Expected behavior
Model shouldn't be able to see future data if it is not specified under future covariates and is specified as past-only covariate.
To reproduce
Use the model checkpoint/ version before this commit and put a breakpoint on line number 848 in _build_batch function in dataset.py. Do not pass future covarites, just pass past covariates.
Environment description
Operating system: ubuntu
Python version: 3.12
CUDA version: 13.0
PyTorch version: 2.9.1
HuggingFace transformers version:
HuggingFace accelerate version:
Pandas version:
Beta Was this translation helpful? Give feedback.
All reactions