2016
DOI: 10.1016/j.envsoft.2015.12.010
|View full text |Cite
|
Sign up to set email alerts
|

Using a data grid to automate data preparation pipelines required for regional-scale hydrologic modeling

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 27 publications
(21 citation statements)
references
References 22 publications
0
21
0
Order By: Relevance
“…Despite this availability of tools, Billah et al [14] recognized as a major challenge for data and modelling integration the lack of common file structures and semantics. This challenge has been addressed in the current work by taking inspiration from the Sensor Observation Service open standard (SOS) [15], which defines data models to represent and exchange observations derived by sensor measures.…”
Section: Data Integration Modulementioning
confidence: 99%
“…Despite this availability of tools, Billah et al [14] recognized as a major challenge for data and modelling integration the lack of common file structures and semantics. This challenge has been addressed in the current work by taking inspiration from the Sensor Observation Service open standard (SOS) [15], which defines data models to represent and exchange observations derived by sensor measures.…”
Section: Data Integration Modulementioning
confidence: 99%
“…A challenge concerns the availability of all of these data at every site selected in the modeling process; if these data are not available everywhere, then it raises the question of how best to learn about the model inputs for every required location from the available sparse data [33]. These data, when available, are normally also contained within multiple databases maintained by various data providers, with each dataset having unique data access protocols, file formats, and semantics [27,29,30]. This heterogeneity and the gaps in source data mean that to derive the input data needed to develop and run site-specific LF models, informatics and analytical tools that can combine empirical data discovery with data integration and estimation protocols will also be required [10,33,39].…”
Section: Lf Data Assemblymentioning
confidence: 99%
“…Learning parasite transmission models that take a fuller account of heterogeneous dynamics across a spatial domain is a difficult task, but the increasing availability of geolocated demographic, intervention, and disease data [23][24][25][26] together with growing advances made in computational science approaches to knowledge discovery, particularly in the areas of (1) high performance grid-based computing and programming [8,11], (2) data discovery, integration, and assembly [11,19,[27][28][29][30][31], and (3) datadriven approaches for inferring models from measurements [32][33][34][35][36][37][38], mean that simulating disease dynamics and responses to interventions effectively across heterogeneous spatially structured environments at large scales are now becoming increasingly feasible. Bayesian data-driven modeling frameworks have received considerable attention in this regard given their ability for not only facilitating the induction of a dynamical system from data, but also in the use of multiple data sources for constraining the parameters of a model to capture the local transmission features of a spatial setting [21,22,33,[39][40][41].…”
Section: Introductionmentioning
confidence: 99%
“…Thus, the general approach demonstrated here for MODFLOW-NWT could be expanded for any workflow that can be automated and that is compatible with Docker requirements. For example, in prior work we have constructed pre-and post-processing workflows for the Variable Infiltration Capacity (VIC) hydrologic model (Liang et al, 1996b) that could directly benefit from this method for curating, packaging, and sharing resources (Billah et al, 2016c;). These containers are efficient, lightweight, self-contained packages of computational experiments that can be nearly guaranteed to be repeated or reproduced regardless of deployment issues.…”
Section: Discussionmentioning
confidence: 99%
“…Figure 3.1 shows the data processing workflow used to generate the meteorological and land surface input datasets for a VIC model simulation. This workflow consists of a sequence of 15 data processing steps, each step requiring input datasets from different sources, and many of the datasets having unique data models (Billah et al, 2016b). These scripts are written with different programming languages including Fortran 77, C, and C++.…”
Section: Variable Infiltration Capacity (Vic) Model Pre-processing Womentioning
confidence: 99%