Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why should the validation or testing dataset use the whole data? #3199

Open
arieffadhlan opened this issue Jun 20, 2024 · 0 comments
Open

Why should the validation or testing dataset use the whole data? #3199

arieffadhlan opened this issue Jun 20, 2024 · 0 comments
Labels
bug Something isn't working

Comments

@arieffadhlan
Copy link

arieffadhlan commented Jun 20, 2024

Description

I'm still new to using DeepAR model. I see AWS Sagemaker makes the following statement and it looks like GluonTS implements the same. "You can create training and test datasets that satisfy this criteria by using the entire dataset (the full length of all time series that are available) as a test set and removing the last prediction_length points from each time series for training."

Why should the testing dataset use the whole data?

For example, I have 1000 data and want to predict the next 30 days. The training data will be used from 1 - 970. Why does the testing data use 1 - 1000?

@arieffadhlan arieffadhlan added the bug Something isn't working label Jun 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant