Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Grid-like datasets are not scaled correctly in the DataSet #690

Closed
10 tasks done
wangcj05 opened this issue Jul 9, 2018 · 1 comment · Fixed by #730
Closed
10 tasks done

Grid-like datasets are not scaled correctly in the DataSet #690

wangcj05 opened this issue Jul 9, 2018 · 1 comment · Fixed by #730

Comments

@wangcj05
Copy link
Collaborator

wangcj05 commented Jul 9, 2018


Issue Description

When one restart the calculation, the dataset will be scaled before retrieving the realization from that dataset. If grid-like sampler is used, the scale factor could be zero, since the grid points for some variable could be constant, this will cause the failure of the restart capability as recently reported by one of our users.

What did you expect to see happen?

The grid sampler with restart should work.

What did you see instead?

From raven user:

so I started a test without the meshfile variable. Left it running for some time so I had a dataset of 34 restart points. Then cancled Raven and tried to restart from the previous results but again the same error and Raven ignored the restart points completely and started sampling the same points again that already exist in the restart data.
Do you have a suggested fix for the development team?

@mandd and I found we have an issue with the SCALE in the DataSet. The fix can be found in
https://github.com/idaholab/raven/tree/wangc/std_issue

Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.

For Change Control Board: Issue Review

This review should occur before any development is performed as a response to this issue.

  • 1. Is it tagged with a type: defect or improvement?
  • 2. Is it tagged with a priority: critical, normal or minor?
  • 3. If it will impact requirements or requirements tests, is it tagged with requirements?
  • 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
  • 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)

For Change Control Board: Issue Closure

This review should occur when the issue is imminently going to be closed.

  • 1. If the issue is a defect, is the defect fixed?
  • 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
  • 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
  • 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
  • 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
@alfoa
Copy link
Collaborator

alfoa commented Aug 26, 2018

closure check list passed...

No email to users is necessary since this issue can not cause wrong results but was causing a crash.

alfoa pushed a commit that referenced this issue Aug 26, 2018
)

* fix std in data comparison

* add test to test restart with constant variables, test issue #690

* rename folder name, since 'con' is not acceptable in windows
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants