-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Increasing CUAHSI WDC-related AOI search limit beyond 1500 km2 #2756
Comments
Don Setiawan (UW) has deployed the Wikiwatershed/MMW App locally on his laptop, for development and testing. We figured out where the CUAHSI WDC AOI limit of 1,500 km2 was set, and changed it to 5,000 km2. We then ran a test search on a squarish polygon search area that's 4,093 km2 (the actual area of the enclosing rectangular AOI issued to the CUAHSI WDC catalog API would most likely be larger), and was able to get a response (4,954 records). See screenshot below. This generally confirms my suggestion that increasing the AOI limit to 3,000 km2, if not larger, is most likely just fine; specially after ensuring |
Shares points with #2760 |
Is this on staging? I would love to test it! |
Staging deployments are currently failing due to a third party dependency failure, but I'll comment here as soon as it is ready. |
@aufdenkampe this is now on staging: Sorry for the delay. |
Thanks @rajadain. I tried it out, using the Brandywine-Christina HUC 8 (1,960 km2), and CUAHSI WDC search stops with an error icon (FYI, the CINERGI search also yields the error icon). For our reference, what's the new AOI size being used on staging? Based on your exchanges at #2784 (comment), it seems like it's 5,000 km2, but I'm not totally sure. @lsetiawan and I are able to run searches larger than the one I'm reporting on here, in terms of AOI HUC 8 polygon size, on his laptop deployment; I can't imagine that his laptop has more resources than your staging cloud environment. Anyway, we'll try to run this specific HUC 8 search later today, and report back. |
(Emilio here, masquerading as Don) We've run a similar AOI test on Don's laptop with his app deployment. HUC polygon selection is not enabled in his deployment, so we created an AOI using free-draw that roughly matched the Brandywine-Christina HUC 8, but was a bit larger (2,235 vs 1,960 km2). The WDC search took around a minute, but completed successfully, returning just short of 5,000 records. See screenshot. So, we don't know why it fails on the staging app. |
@emiliom our staging environments are running on smaller machines than production in an effort to keep hosting costs down. We just upped the staging VM from a We're also expanding our development VMs in #2803 to allow larger areas of interest to be run while in development. I had to increase my |
@rajadain thanks for the info and update. I do realize that the staging VM was bound to be less capable than the production VM, for the reasons you state. Out of curiosity (and for comparison), how much RAM is allocated to the production VM? The deployment we're using is just Don's mid-range newish laptop, as is. Obviously we have no intent to recreate or approximate a hardware environment that looks like the cloud-based staging or production allocations used for the Wikiwatershed app. Our goal is focused on the development work for adding the new Water Quality Portal catalog; the CUAHSI WDC tests have just been a side benefit and opportunity. |
I've investigated CUAHSI WDC search API performance issues, following up on our last Monitor MW call. Briefly, I:
Summary of my findings and recommendations
GetSeriesMetadataCountOrData
. Its response is consistently slower than the one we currently use,GetSeriesCatalogForBox2
. The only near-term potential use I can foresee is its capability to return only a count of the number of series records it will return; that response is extremely fast, and may be used to guide further client actions (including self pagination).GetSeriesCatalogForBox2
Detailed information and discussion of API tests and related previous Azavea assessments
CUAHSI_HISCentral_AOI_service_tests.ipynb
, can be accessed here. See the descriptions at the top of the notebook. This notebook was run once for each AOI listed in the table. (The specific results shown in the notebook snapshot (for the "1° N of the above PA/DRB point" AOI) differ from the ones listed in the table, because the data are dynamic and factors such as CUAHSI server loads and network latency are not constant. The results in the notebook were run today, Monday April 2 at 3:40pm PT, while results in the table above were run on Saturday March 24 (weekend server loads are probably lighter).). Each result is for a search based on a 1° x 1° square box ("square" in lat-lon coordinates) centered at the center point listed. Search requests were issued withsuds-jurko
. The last 3 columns show response times (including suds processing time) for 3 API's:GetSeriesCatalogForBox2
(currently used in the MMW portal)GetSeriesCatalogForBox3
GetSeriesMetadataCountOrData
(the newer API we're investigating)The API currently used in the portal,
GetSeriesCatalogForBox2
, clearly yields the fastest response times. I believe this is due simply to the fact that it handles a slimmer set of meatadata attributes, compared to the other two API's.I did not encounter any strict failures on either the WDC server end or the client (my laptop) side. Requests that returned more records (as many as 23K, close to the 25K limit at least for
GetSeriesMetadataCountOrData
) were simply slower, but never actually failed. Client Python SOAP processing and deserialization with "suds" never failed either, unlike what Azavea reported in Oct 2017. The only possible reason I can think for the failures reported by Azavea is if they were using the original, very old and unmantained suds package rather than its fork and more current replacement, suds-jurko, which I used.Notes:
cc @aufdenkampe
The text was updated successfully, but these errors were encountered: