The special collection “Benchmarking Data-Driven Site Characterization” is available in the ASCE Library (https://ascelibrary.org/ajrua6/benchmarking_data_driven_characterization).
Data-centric geotechnics is an emerging field that advocates a “digital first” agenda (Phoon et al. 2022a). One aspect of this agenda is the development of methods that make sense of actual data. Perhaps more importantly, it calls for the industry to develop methods without imposing thresholds on quantity, quality, and/or other “ugly” data attributes as prerequisites for exploring data-informed decision-making. In short, the industry should engage in the spirit of offering our data an opportunity to speak for itself.
The end goal of any analysis is to improve decision-making in geotechnical engineering, which is always related to a project carried out at a specific site. It is a given that the role of engineers in the design process is indispensable because they take ultimate responsibility for the project. Improving an engineer’s site-specific knowledge of the ground conditions will surely enhance this role. It is natural for data-driven site characterization (DDSC) to attract the most attention in data-centric geotechnics (Phoon et al. 2022b).
One purpose of DDSC is to produce a three-dimensional (3D) stratigraphic map of the subsurface volume below a full-scale project site and to estimate the governing engineering properties and soil type at each spatial point based on site investigation data at the target site and relevant big indirect data (BID) from other sites, neighboring or otherwise. However, this is a significant challenge because the “ugly” attributes of actual site data are MUSIC-3X [multivariate, uncertain and unique, sparse, incomplete, and potentially corrupted with 3D spatial variability (3X)]. The widely held perception that data sparsity is a strong hindrance to DDSC is incomplete in terms of the scope of the challenge. In particular, site uniqueness is a fundamental feature of geotechnical engineering, but there is no fully satisfactory data-driven method to characterize this “U” attribute (also called intersite variability) in BID. For example, Clause 2.4.5.2(10) of Eurocode 7 (CEN 2004), “Characteristic Values of Geotechnical Parameters” highlights the importance of accounting for “U” but does not provide a quantitative method to do so: “If statistical methods are employed in the selection of characteristic values for ground properties, such methods should differentiate between local and regional sampling and should allow the use of a priori knowledge of comparable ground properties.” The jury is still out on many important questions pertaining to the kinds of data sets that will provide better support for decision-making at a specific site. Do we use the entire BID or a subset of “similar” sites (a quasi-regional cluster) (Phoon and Ching 2022)? Do we collect only high-quality data that are limited in quantity or only lower-quality data in larger quantity? Do we combine them? When should they not be combined?
DDSC will fill an existing gap in Building Information Modeling (BIM) where a digital model for the subsurface is largely missing. An exciting prospect is risk- and reliability-informed design at the systems level using BIM as a platform that may eventually evolve into a digital twin to support smart infrastructure life-cycle management. As such, research on DDSC methods has been active in recent years. Because site-specific data are incomplete and sparse, these DDSC methods—Bayesian machine learning being the most popular—are mainly probabilistic . However, there are other methods (Phoon and Zhang 2023). At present, it is not known which method is more appropriate and/or more efficient for given circumstances. The answer must largely depend on the attributes of the training data and the desired outcomes of DDSC. To promote the development of DDSC methods that are applicable to routine projects in a more purposeful way, one important step is to create reference geotechnical data sets and standard performance metrics to train and measure different DDSC methods on a uniform basis. This benchmark testing, or benchmarking, is widely used in machine learning (ML) to support unbiased and competitive evaluation of emerging ML methods (Thiyagalingam et al. 2022). In a relatively early attempt at establishing a benchmarking framework for DDSC, Phoon et al. (2022c) recommended the following considerations for benchmark examples: (1) they must be 3D and realistic in scale, stratigraphy, and properties; (2) they must cover a sufficient range of ground conditions; (3) they must restrict the training data set to data that engineers have at their disposal (MUSIC-3X or other data attributes); and (4) they must select validation data sets and performance metrics that showcase the value of DDSC to decision-makers [to meet the “fit for (and transform) practice” requirement in data-centric geotechnics].
Phoon and Zhang (2023) classified value according to “ML advantage” and “ML supremacy.” For the former, the ML method is effective but not indispensable. Hence, its value to practice is incremental. For the latter, the ML method can solve problems where conventional methods are found to be impractical or ineffective. Hence, its value to practice is significant. A simple “value” matrix is also proposed to guide future research: (1) Type 1 (incremental value) involving available data and existing conventional applications, (2) Type 2a (potentially high value) involving available data and new applications, (3) Type 2b (high value) involving new data and existing applications, and (4) Type 3 (disruptive value) involving new data and completely novel applications.
This special collection contains seven papers. Some of the research was presented at the second TC304 Forum “Data-Driven Modeling for Geotechnical Spatial Variability” organized by Tokyo City University and the Japanese Geotechnical Society. The Forum took place online on July 3, 2021. Its purpose was to compare the performance of five data-driven methods in site characterization using standard benchmark examples and performance metrics (including computational cost).
Phoon (2023) reviews reliability applications and real-world databases that are mature and already in use or promising new developments in data-centric geotechnics that may transform practice in the near future. The paper situates geotechnical reliability as one step toward a broader and deeper exploitation of data for decision-making: (1) at a real-time system-wide scale, (2) in hybrid intelligence (human, machine) mode, (3) for a specific project [e.g., “precision construction” in Phoon (2018)], and (4) to address complex design goals beyond safety and economy (e.g., sustainability, resilience). This state-of-the-art review can be regarded as a timely update to National Research Council (1995) and as reopening the conversation with a focus on what is useful for practice and what has changed in a material way that increases or decreases the need for probabilistic methods in practice. The National Research Council report, “Probabilistic Methods in Geotechnical Engineering,” has been influential, but its impact on the broader community has been mainly restricted to the load and resistance factor design (LRFD). Phoon (2023) argues that reliability or other uncertainty quantification methods have become more rather than less important since their inception in the 1960s because they exploit data in more systematic and comprehensive than deterministic ways, which are critically needed to address complex new challenges (ASCE 2019). Data infrastructure is now considered to be as important as physical infrastructure. This is not well appreciated by most engineers, although it is pivotal to digital transformation (Phoon et al. 2022a). Given the growing ubiquity of digital technologies, this state-of-the-art review concludes that the basis for decision-making will shift from a data-informed physics approach to a data-driven approach that may or may not be based on physics.
The following five papers present the benchmarking results of different DDSC methods based on the virtual ground examples in Phoon et al. (2022c). Lyu et al. (2023) apply a DDSC method based on Bayesian compressive sampling/sensing (BCS) for benchmarking and implement a user-friendly in-house software known as Analytics of Sparse Spatial Data (ASSD-BCS). A new benchmark example based on an actual cone penetration test (CPT) data set collected at a site in Adelaide, Australia (Jaksa 1995) is also proposed. Mavritsakis et al. (2023) propose a Bayesian site characterization framework (BaySiC) and report its benchmarking results. In BaySiC, a vector random field model containing two CPT parameters (tip resistance and sleeve friction) is used. The random field is assumed to be normally distributed. Cross-correlation between tip resistance and sleeve friction and spatial correlations with depth are considered. The hyperparameters of the model are estimated using the Hamiltonian Markov Chain Monte Carlo (MCMC) method. The authors formulate two models using BaySiC, aggregated model and split, and report that the former is effective in predicting tip resistance while the latter is effective in identifying material type. Tomizawa and Yoshida (2022) implement Gaussian process regression (GPR) using multiple Gaussian random fields (GPR-MR) for DDSC. GPR has recently attracted renewed attention and is being actively studied for DDSC (Yoshida et al. 2021; Ching et al. 2023). In GPR-MR, three autocorrelation functions (or kernel functions), Gaussian, Markovian, and Whittle-Mattern, are considered and the best one is selected based on the Akaike information criterion (AIC) and Bayesian information criterion (BIC). Ching and Yoshida (2023) compare the performance of two DDSC methods, sparse Bayesian learning (SBL) and GPR-MR, through benchmarking. Bayesian evidence is computed in the process of hyperparameter tuning and used to compare the performance of the two DDSC methods. The authors also propose a two-step algorithm that applies clustering (cluster analysis) for stratification before spatial interpolation. Shuku and Phoon (2023a) propose a new benchmark example, small virtual ground 5 (S-VG5) based on a more heterogeneous virtual ground than the four in Phoon et al. (2022c). The 3D virtual ground is created by a category-based approach that generates complex geological structures through iterations of simple mathematical rules such as a Markov chain. An alternate approach is to apply a graphical Markov random field model (Shuku and Phoon 2023b). In addition, the authors compare three DDSC methods, Glasso, Glasso-BFs (Glasso with basis functions), and GPR, from a Bayesian perspective. This allows the methodological similarities and differences to be explained in a theoretically consistent way.
Finally, Hu et al. (2023) present a framework to assess the effectiveness of a sampling layout in a site investigation program based on uncertainty reduction. The authors propose an index called expected reduced variance (ERV) for this purpose. If ERV is large, the investigation program is effective. Crude Monte Carlo and bootstrap methods are compared in evaluating the statistical uncertainty of ERV, and it is reported that the bootstrap method is computationally more efficient than crude Monte Carlo. Two design examples (shallow foundation and slope) are presented to demonstrate the proposed method.
The guest editors thank Professor Michael Beer, editor of the ASCE-ASME Journal of Risk and Uncertainty in Engineering Systems, for his encouragement and support to complete this special collection. The efforts of the authors and reviewers are also deeply appreciated.

References

ASCE. 2019. Future world vision: Infrastructure reimagined. Reston, VA: ASCE.
CEN (European Committee for Standardization). 2004. Geotechnical design—Part 1: General rules. Comité Européen de Normalisation. EN 1997-1. Brussels, Belgium: CEN.
Ching, J., and I. Yoshida. 2023. “Data-drive site characterization for benchmark examples: Spare Bayesian learning versus Gaussian process regression.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 9 (1). https://doi.org/10.1061/AJRUA6.RUENG-983.
Ching, J., I. Yoshida, and K. K. Phoon. 2023. “Comparison of trend models for geotechnical spatial variability: Sparse Bayesian learning vs. Gaussian process regression.” Gondwana Res. https://doi.org/10.1016/j.gr.2022.07.011.
Hu, J. Z., J. G. Zheng, J. Zhang, and H. W. Huang. 2023. “Bayesian framework for assessing effectiveness of geotechnical site investigation programs.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 9 (1). https://doi.org/10.1061/AJRUA6.0001278.
Jaksa, M. 1995. “The influence of spatial variability on the geotechnical design properties of a stiff, overconsolidated clay.” Ph.D. dissertation, Dept. of Civil and Environmental Engineering, Univ. of Adelaide.
Lyu, B., Y. Hu, and Y. Wang. 2023. “Data-driven development of three-dimensional subsurface models from sparse measurements using Bayesian compressive sampling: A benchmarking study.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 9 (2). https://doi.org/10.1061/AJRUA6.RUENG-935.
Mavritsakis, A., T. Schweckendiek, A. Teixeira, and E. Smyrniou. 2023. “Bayesian analysis of benchmark examples for data-driven site characterization.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 9 (2). https://doi.org/10.1061/AJRUA6.RUENG-975.
National Research Council. 1995. Probabilistic methods in geotechnical engineering. Washington, DC: National Academies Press.
Phoon, K. K. 2018. “Editorial for special collection on probabilistic site characterization.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 4 (4): 02018002. https://doi.org/10.1061/AJRUA6.0000992.
Phoon, K. K. 2023. “What geotechnical engineers want to know about reliability.” ASCE-ASME J. Risk Uncertainty Eng. Syst. Part A: Civ. Eng.
Phoon, K. K., and J. Ching. 2022. “Additional observations on the site recognition challenge.” J. GeoEng. 17 (4): 231–247. https://doi.org/10.6310/jog.202212_17(4).6.
Phoon, K. K., J. Ching, and Z. Cao. 2022a. “Unpacking data-centric geotechnics.” Underground Space 7 (6): 967–989. https://doi.org/10.1016/j.undsp.2022.04.001.
Phoon, K. K., J. Ching, and T. Shuku. 2022b. “Challenges in data-driven site characterization.” Georisk: Assess. Manage. Risk Eng. Syst. Geohazards 16 (1): 114–126. https://doi.org/10.1080/17499518.2021.1896005.
Phoon, K. K., T. Shuku, J. Ching, and I. Yoshida. 2022c. “Benchmark examples for data-driven site characterization.” Georisk: Assess. Manage. Risk Eng. Syst. Geohazards 16 (4): 599–621. https://doi.org/10.1080/17499518.2022.2025541.
Phoon, K. K., and W. Zhang. 2023. “Future of machine learning in geotechnics.” Georisk: Assess. Manage. Risk Eng. Syst. Geohazards 17 (1). https://doi.org/10.1080/17499518.2022.2087884.
Shuku, T., and K. K. Phoon. 2023a. “Comparison of data-driven site characterization methods through benchmarking: Methodological and application aspects.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 9 (2). https://doi.org/10.1061/AJRUA6.RUENG-977.
Shuku, T., and K. K. Phoon. 2023b. “Data-driven subsurface modeling using a Markov random field model.” Georisk: Assess. Manage. Risk Eng. Syst. Geohazards. https://doi.org/10.1080/17499518.2023.2181973.
Thiyagalingam, J., M. Shankar, G. Fox, and T. Hey. 2022. “Scientific machine learning benchmarks.” Nat. Rev. Phys. 4 (6): 413–420. https://doi.org/10.1038/s42254-022-00441-7.
Tomizawa, Y., and I. Yoshida. 2022. “Benchmarking of Gaussian process regression with multiple random fields for spatial variability estimation.” ASCE-ASME J. Risk Uncertainty Eng. Syst., Part A: Civ. Eng. 8 (4). https://doi.org/10.1061/AJRUA6.0001277.
Yoshida, I., Y. Tomizawa, and Y. Otake. 2021. “Estimation of trend and random components of conditional random field using Gaussian process regression.” Comput. Geotech. 136 (Aug): 104179. https://doi.org/10.1016/j.compgeo.2021.104179.

Information & Authors

Information

Published In

Go to ASCE-ASME Journal of Risk and Uncertainty in Engineering Systems, Part A: Civil Engineering
ASCE-ASME Journal of Risk and Uncertainty in Engineering Systems, Part A: Civil Engineering
Volume 9Issue 2June 2023

History

Received: Jan 14, 2023
Accepted: Jan 23, 2023
Published online: Mar 22, 2023
Published in print: Jun 1, 2023
Discussion open until: Aug 22, 2023

Permissions

Request permissions for this article.

Authors

Affiliations

Professor, Information Systems Technology and Design/Architecture and Sustainable Design, Singapore Univ. of Technology and Design, Singapore 487372 (corresponding author). ORCID: https://orcid.org/0000-0003-2577-8639. Email: [email protected]
Associate Professor, Dept. of Environmental Management, Okayama Univ., Kita-ku, Okayama 700-8530, Japan. ORCID: https://orcid.org/0000-0002-0745-1010. Email: [email protected]
Professor, Dept. of Civil Engineering, National Taiwan Univ., Taipei 10617, Taiwan. ORCID: https://orcid.org/0000-0001-6028-1674. Email: [email protected]
Professor, Dept. of Urban and Civil Engineering, Tokyo City Univ., Setagaya-ku, Tokyo 158-8557, Japan. ORCID: https://orcid.org/0000-0001-9770-2233. Email: [email protected]

Metrics & Citations

Metrics

Citations

Download citation

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited by

  • Future of Machine Learning in Geotechnics (FOMLIG), 5–6 Dec 2023, Okayama, Japan, Georisk: Assessment and Management of Risk for Engineered Systems and Geohazards, 10.1080/17499518.2024.2316882, 18, 1, (288-303), (2024).
  • Report for ISSMGE TC309/TC304/TC222 and ASCE Geo-Institute Risk Assessment and Management Committee Fourth Machine Learning in Geotechnics Dialogue on “Machine Learning Supremacy Projects”, Georisk: Assessment and Management of Risk for Engineered Systems and Geohazards, 10.1080/17499518.2024.2316879, 18, 1, (304-313), (2024).

View Options

Media

Figures

Other

Tables

Share

Share

Copy the content Link

Share with email

Email a colleague

Share