American Innovation and Competitiveness Act was adopted unanimously by the U.S. Congress and signed into law by president Obama in January 2017.
The law contains a section called Research Reproducibility and Replication, which asked the Director of the National Science Foundation in agreement with the National Research Council to prepare a report on issues related to research reproducibility and “to make recommendations for improving rigor and transparency in scientific research”.
To fulfill this requirement, a consensus report of the National Academies of Sciences, Engineering, and Medicine was published in 2019. The report is summarized in the special issue of Harvard Data Science Review in December 2020.
Among the recommendations:
All researchers should include a clear, specific, and complete description of how the reported results were reached. Reports should include details appropriate for the type of research, including:
- a clear description of all methods, instruments, materials, procedures, measurements, and other variables involved in the study;
- a clear description of the analysis of data and decisions for exclusion of some data or inclusion of other;
- for results that depend on statistical inference, a description of the analytic decisions and when these decisions were made and whether the study is exploratory or confirmatory;
- a discussion of the expected constraints on generality, such as which methodological features the authors think could be varied without affecting the result and which must remain constant;
- reporting of precision or statistical power; and
- discussion of the uncertainty of the measurements, results, and inferences.
Funding agencies and organizations should consider investing in research and development of open-source, usable tools and infrastructure that support reproducibility for a broad range of studies across different domains in a seamless fashion. Concurrently, investments would be helpful in outreach to inform and train researchers on best practices and how to use these tools.
Journals should consider ways to ensure computational reproducibility for publications that make claims based on computations, to the extent ethically and legally possible.
A new paper is added to the collection of reproducible documents: Full waveform inversion and joint migration inversion with an automatic directional total variation constraint
As full waveform inversion (FWI) is a non-unique and typically ill-posed inversion problem, it needs proper regularization to avoid cycle-skipping. To reduce the non-linearity of FWI, Joint Migration Inversion (JMI) is proposed as an alternative, explaining the reflection data with decoupled velocity and reflectivity parameters. However, the velocity update may also suffer from being trapped in local minima. To optimally include geologic information, we propose FWI/JMI with directional total variation as an L1-norm regularization on the velocity. We design the directional total variation operator based on the local dip field, instead of ignoring the local structural direction of the subsurface and only using horizontal- and vertical- gradients in the traditional TV. The local dip field is estimated using plane-wave destruction based on a raw reflectivity model, which is usually calculated from the initial velocity model. With two complex synthetic examples, based on the Marmousi model, we demonstrate that the proposed method is much more effective compared to both FWI/JMI without regularization and FWI/JMI with the conventional TV regularization. In the JMI-based example, we also show that L1 directional TV works better than L2 directional Laplacian smoothing. In addition, by comparing these two examples, it can be seen that the impact of regularization is larger for FWI than for JMI, because in JMI the velocity model only explains the propagation effects and, thereby, makes it less sensitive to the details in the velocity model.
A new paper is added to the collection of reproducible documents: Five-dimensional seismic data reconstruction using the optimally damped rank-reduction method
It is difficult to separate additive random noise from spatially coherent signal using a rank-reduction method that is based on the truncated singular value decomposition (TSVD) operation. This problem is due to the mixture of the signal and the noise subspaces after the TSVD operation. This drawback can be partially conquered using a damped rank reduction method, where the singular values corresponding to effective signals are adjusted via a carefully designed damping operator. The damping operator works most powerfully in the case of a small rank and a small damping factor. However, for complicated seismic data, e.g., multi-channel reflection seismic data containing highly curved events, the rank should be large enough to preserve the details in the data, which makes the damped rank reduction method less effective. In this paper, we develop an optimal damping strategy for adjusting the singular values when a large rank parameter is selected so that the estimated signal can best approximate the exact signal. We first weight the singular values using optimally calculated weights. The weights are theoretically derived by solving an optimization problem that minimizes the Frobenius-norm difference between the approximated signal components and the exact signal components. The damping operator is then derived based on the initial weighting operator to further reduce the residual noise after the optimal weighting. The resulted optimally damped rank reduction method is nearly an adaptive method, i.e., insensitive to the rank parameter. We demonstrate the performance of the proposed method on a group of synthetic and real five-dimensional seismic data.
A new paper is added to the collection of reproducible documents: Seismic signal enhancement based on the lowrank methods
Based on the fact that the Hankel matrix constructed by noise-free seismic data is lowrank (LR), LR approximation (or rank-reduction) methods have been widely used for removing noise from seismic data. Due to the linear-event assumption of the traditional LR approximation method, it is difficult to define a rank that optimally separates the data subspace into signal and noise subspaces. For preserving the most useful signal energy, a relatively large rank threshold is often chosen, which inevitably leaves residual noise. To reduce the energy of residual noise, we propose an optimally damped rank-reduction method. The optimal damping is applied via two steps. In the first step, a set of optimal damping weights is derived. In the second step, we derive an optimal singular-value damping operator. We review several traditional lowrank methods and compare their performance with the new one. We also compare these lowrank methods with two sparsity-promoting transform methods. Examples demonstrate that the proposed optimally damped rank-reduction method could get significantly cleaner denoised images compared with the state-of-the-art methods.
A new paper is added to the collection of reproducible documents: Simultaneous denoising and reconstruction of 5D seismic data via damped rank-reduction method
The Cadzow rank-reduction method can be effectively utilized in simultaneously denoising and reconstructing 5D seismic data that depends on four spatial dimensions. The classic version of Cadzow rank-reduction method arranges the 4D spatial data into a level-four block Hankel/Toeplitz matrix and then applies truncated singular value decomposition (TSVD) for rank-reduction. When the observed data is extremely noisy, which is often the feature of real seismic data, traditional TSVD cannot be adequate for attenuating the noise and reconstructing the signals. The reconstructed data tends to contain a significant amount of residual noise using the traditional TSVD method, which can be explained by the fact that the reconstructed data space is a mixture of both signal subspace and noise subspace. In order to better decompose the block Hankel matrix into signal and noise components, we introduced a damping operator into the traditional TSVD formula, which we call the damped rank-reduction method. The damped rank-reduction method can obtain a perfect reconstruction performance even when the observed data has extremely low signal-to-noise ratio (SNR). The feasibility of the improved 5D seismic data reconstruction method was validated via both 5D synthetic and field data examples. We presented comprehensive analysis of the data examples and obtained valuable experience and guidelines in better utilizing the proposed method in practice. Since the proposed method is convenient to implement and can achieve immediate improvement, we suggest its wide application in the industry.
A new paper is added to the collection of reproducible documents: Separation and imaging of seismic diffractions using a localized rank-reduction method with adaptively selected ranks
Seismic diffractions are some weak seismic events hidden within the more dominant reflection events in a seismic profile. Separating diffraction energy from the post-stack seismic profiles can help infer the subsurface discontinuities that generate the diffraction events. The separated seismic diffractions can be migrated with a traditional seismic imaging method or a specifically designed migration method to highlight the diffractors, i.e., the diffraction image. Traditional diffraction separation methods based on the the underlying plane-wave assumption are limited by either the inaccurate slope estimation or the plane-wave assumption of the PWD filter, and thus will cause reflection leakage into the separated diffraction profile. The leaked reflection energy will deteriorate the resolution of the subsequent diffraction imaging result. Here, we propose a new diffraction separation method based on a localized rank-reduction method. The localized rank-reduction method assumes the reflection events to be locally low-rank and the diffraction energy can be separated by a rank-reduction operation. Compared with the global rank-reduction method, the localized rank-reduction method is more constrained in selecting the rank and is free of separation artifacts. We use a carefully designed synthetic example to demonstrate that the localized rank-reduction method can help separate the diffraction energy from a post-stack seismic profile with both kinematically and dynamically accurate performance.
A new paper is added to the collection of reproducible documents: Data-driven time-frequency analysis of seismic data using non-stationary Prony method
The empirical mode decomposition aims to decompose the input signal into a small number of components named intrinsic mode functions with slowly varying amplitudes and frequencies. In spite of its simplicity and usefulness, however, the empirical mode decomposition lack solid mathematical foundation. In this paper, we describe a method to extract the intrinsic mode functions of the input signal using non-stationary Prony method. The proposed method captures the philosophy of the empirical mode decomposition, but use a different method to compute the intrinsic mode functions. Having the intrinsic mode functions obtained, we then compute the spectrum of the input signal using Hilbert transform. Synthetic and field data validate the proposed method can correctly compute the spectrum of the input signal, and could be used in seismic data analysis to facilitate interpretation.
A new paper is added to the collection of reproducible documents: Enhancing seismic reflections using empirical mode decomposition in the flattened domain
Due to different reasons, the seismic reflections are not continuous even when no faults or no discontinuities exist. We propose a novel approach for enhancing the amplitude of seismic reflections and making the seismic reflections continuous. We use plane-wave flattening technique to provide horizontal events for the following empirical mode decomposition (EMD) based smoothing in the flattened domain. The inverse plane-wave flattening can be used to obtain original curved events. The plane-wave flattening process requires a precise local slope estimation, which is provided by the plane-wave destruction (PWD) algorithm. The EMD based smoothing filter is a non-parametric and adaptive filter, thus can be conveniently used. Both pre-stack and post-stack field data examples show tremendous improvement for the data quality, which makes the following interpretation easier and more reliable.
A new paper is added to the collection of reproducible documents: Application of principal component analysis in weighted stacking of seismic data
Optimal stacking of multiple datasets plays a significant role in many scientific domains. The quality of stacking will affect the signal-to-noise ratio (SNR) and amplitude fidelity of the stacked image. In seismic data processing, the similarity-weighted stacking makes use of the local similarity between each trace and a reference trace as the weight to stack the flattened prestack seismic data after normal moveout (NMO) correction. The traditional reference trace is an approximated zero-offset trace that is calculated from a direct arithmetic mean of the data matrix along the spatial direction. However, in the case that the data matrix contains abnormal mis-aligned trace, erratic and non-gaussian random noise, the accuracy of the approximated zero-offset trace would be greatly affected, thereby further influence the quality of stacking. We propose a novel weighted stacking method that is based on principal component analysis (PCA). The principal components of the data matrix, namely the useful signals, are extracted based on a low-rank decomposition method by solving an optimization problem with a low-rank constraint. The optimization problem is solved via a common singular value decomposition algorithm. The low-rank decomposition of the data matrix will alleviate the influence of abnormal trace, erratic and non-gaussian random noise, thus will be more robust than the traditional alternatives. We use both synthetic and field data examples to show the successful performance of the proposed approach.
A new paper is added to the collection of reproducible documents: Velocity analysis of simultaneous-source data using high-resolution semblance – coping with the strong noise
Direct imaging of simultaneous-source (or blended) data, without the need of deblending, requires a precise subsurface velocity model. In this paper, we focus on the velocity analysis of simultaneous-source data using the NMO-based velocity picking approach. We demonstrate that it is possible to obtain a precise velocity model directly from the blended data in the common-midpoint (CMP) domain. The similarity-weighted semblance can help us obtain much better velocity spectrum with higher resolution and higher reliability compared with the traditional semblance. The similarity-weighted semblance enforces an inherent noise attenuation solely in the semblance calculation stage, thus is not sensitive to the intense interference. We use both simulated synthetic and field data examples to demonstrate the performance of the similarity-weighted semblance in obtaining reliable subsurface velocity model for direct migration of simultaneous-source data. The migrated image of blended field data using prestack kirchhoff time migration (PSKTM) approach based on the picked velocity from the similarity-weighted semblance is very close to the migrated image of unblended data.