290 likes | 441 Vues
Speeding Up Seismic Imaging with Reduced-Order Modeling. Victor Pereyra Energy Resources Engineering Stanford University Scientific Computing and Matrix Computations Seminar U.C. Berkeley 2/13/13. Seismic Oil Exploration.
E N D
Speeding Up Seismic Imaging with Reduced-Order Modeling Victor Pereyra Energy Resources Engineering Stanford University Scientific Computing and Matrix Computations Seminar U.C. Berkeley 2/13/13
Seismic Oil Exploration • Seismic oil exploration and reservoir imaging require the processing of very large scale data sets. Data is collected in the time domain but maps of the underground are required in the space domain. The physics involved is wave propagation in complex media. • One of the most sophisticated data processing employed is Depth Migration and until recently only approximate solutions to the wave equation were used, such as the one way or parabolic approximation. This has limitations in complex areas. • Current technology is called Reverse Time Migration that uses full wave simulation from the sources and backward propagation of the seismic data from the receivers in order to create depth migrated images from seismic data by cross-correlation. • Full waveform tomography is another technique used to improve velocity models that requires large computing resources. All these techniques require massive numbers of simulations. • In 3D it is not unusual to manipulate many terabytes of data in large clusters. These problems are still a challenge today even with large dedicated clusters, both because of computing, storage requirements and network traffic.
BP builds world’s largest supercomputing complex for exploration applications • BP has started to build in Houston what it says will be the largest commercial supercomputing complex in the world. • The High-Performance Computing center is scheduled to open in mid-2013.The HPC will be the worldwide hub for processing and managing geologic and seismic data across all of BP’s assets. • The HPC will be equipped with 67,000 CPUs and is expected to process data at a rate of as much as two petaflops. The total memory will be 536 terabytes and the disk space will total 23.5
Reverse Time Migration (RTM) Overturned Rays Courtesy of Antoine Guitton and Bruno Kaelin
Reduced-order modeling • It would be of great interest to be able to run these multiple simulations faster and using less data. • Model order reduction is a technique that has been used in many application domains to do just that. • Since there is a paucity of work for applications to the wave equation in the time domain, we are performing some feasibility studies in order to see if the resulting approximate solutions are sufficiently accurate and cost effective for some of these problems. Fast wave propagation by model order reduction. V. Pereyra and B. Kaelin. ETNA 30, 2008. Wave equation simulation in two-dimensions using a compressed modeler. V. Pereyra. Submitted for publication in ETNA, 2012.
Acoustic Wave Equation • Consider the wave equation in inhomogeneous media, discretized in space : 2w/t2 = Aw + Bu(t), v = Cw, Where xRl,w(t)Rn and A (discrete Laplacian),B (source term),C (output) are appropriate matrices, u(t) is an input forcing function and v is the desired output. n will usually be very large.
Basic Idea • The dynamics is obtained by integrating this very large system of Ordinary Differential Equations (method of lines). • The procedure we propose attempts to obtain an approximate solution by reducing drastically the size of this system of ODE’s. • This is specially important and useful when rapid response or repeated simulations are required. • It is assumed that we have the ability to perform a few full size (high fidelity) solves.
Proper Orthogonal Decomposition (POD) • From one or a few full fidelity simulations we extract a certain number of snapshots: i.e., spatial states of the field at different times, fi, i=1,…,m. • We form a matrix F (nxm), where n is the number of degrees of freedom of the original spatial discretization (veeery large) and m is the number of snapshots (relatively small). • Then we proceed to calculate its QR Decomposition: F = Q R. • The columns of the matrix Q form an orthogonal basis for the space of columns of F, the snapshots, and R is upper triangular. • We actually use a progressive QR, in which a new snapshot is reduced on the fly and if its angle with the existing basis is not large enough it is skipped. In this way we incorporate some adaptivity in the snapshot selection and end up with a well conditioned basis.
Reduced System • Now we use Galerkin collocation on the original equations, by proposing the Ansatz: w = Qa(t). • The key here is to introduce the time dependence through the weights a(t). • Replacing in the wave equation we obtain the reduced system (of size m, the number of snapshots): att = (QTAQ) a(t) + (QTB) u(t) Observe that we still can input a source (it does not have to be the same source wavelet or even be applied at the same location as in the problem that generated the snapshots); we can also change the material properties or change the output in this reduced equation. i.e., we can change A, B (not drastically, obviously).
High Fidelity and MOR Codes • As a full fidelity code we use an 8th order finite difference approximation in space and a Runge-Kutta-Fehlberg fourth order integrator in time. Absorbing boundary conditions are included (R. Kosloff and D. Kosloff, Absorbing boundaries for wave propagation problems. J. Comp. Physics 63:363-376 (1986).) • The second order acoustic wave equation is written in first order form (as required by RKF45), so that doubles the number of ODE’s for the time integration. • The MOR code implements the procedure indicated above, using snapshots of previously run HF simulations.
Vel4: a Large Size 2D Model • In this example we consider an inhomogeneous model with the following specs: • dx=dz=6.25m; dt=0.05sec (snapshot spacing) • nx=2001, nz=2001 (total # space mesh n = 4004001). • Nine shots are placed on the free surface, starting at x=5931.25 m, spaced by 25 m for a total distance of 200 m. • We take 200 snapshots from the first and last shot simulations, for a total of 400 snapshots. With a threshold of the angle between a candidate snapshot and the current basis of 5o, we select m = 363 basis snapshots. • We use RKF45, a standard Runge-Kutta 4th order solver to integrate the 8008002 ODE’s in time. RKF45 adapts its integration step to preserve stability and achieve a prescribed accuracy (in this case 0.001).
A Complex, Large Scale 2D Model.Vel4, a Salt Body Courtesy of BP Velocity model. 2001 x 2001 mesh. Salt velocity is very high 13,000 ft/sec
Seismic Section: A number of seismograms plotted side by side
Performance • Observe that if we consider the preprocessing time and prorate it over the seven internal shots, then the gain in time is much smaller, but still significant: 5.7. • HF is Open MP optimized while MOR is not. This was run in a 12 processors machine and HF had parallel efficiency of 80%. • This factor would be much larger in 3D or if we consider more internal shots.
Full waveform inversion • In this application we have seismic data collected (say) at the free surface on a number of receivers: Sio(tj) and a parameterized velocity model v(a). • Starting from an initial guess v(a0) we use the HF and MOR codes to generate the calculated response at the receivers and use a least squares data fit to improve upon the values of the velocity parameters. • For this purpose we employ a Levenberg-Marquardt code that approximates gradients with finite differences (LMDIF from MINPAK).
Regularization/Convexification/ContinuationRCC • For difficult inverse problems is recommended to use regularization, i.e., add to the misfit functional a term of the form: mina (1-l)||So – Sc(a)||22 + l ||a – a0||22, that penalizes large changes in the unknown parameters from some reference value. • We observe that the penalty term is a convex function. • Also, this is the form for homotopy continuation, where we start from a known solution (a0, for l =1) and move l incrementally, towards 0, the problem we really want to solve.
Multimodal Response Function • We consider a simple 3 constant layers model and vary only the velocity in the first layer. We plot the response function for a range of values of v1 and of l.
Some Preliminary Inversion Results • We consider the same 2D, 3 homogeneous layer problem. Model size: 3500x3500. • The data is generated by the HF code and corresponds to a shot at x=1600. There are 200 receivers spaced by 10m, starting at x=700 (400 samples per receiver). l was set to 0.05. • The purpose is to recover the (constant) velocity in the layers, when starting away from the values that generated the synthetic data. • This results in a 80000 x 3 nonlinear least squares problem. MOR uses 108 snapshots from the first HF simulation at a0.
Numerical results Speed up = 5025/397.75 = 12.63 Using snapshots at initial guess.
A Larger Problem: Model vel4pc • Space: 2001x2001 • 7 parameters • No regularization • 12 threads (mainly for High Fidelity; efficiency 90%) • HF Time/fcn = 5580” • MOR Time/fcn = 186”; Pre-process Time: 3349”; rank=122 • Ratio: 11.83 • Ratio for 30 iters. Including pre-processing: 7.4
The real interest: 3D • Currently, parallelization in a distributed setting is done per shot. In 3D, simulating a survey may require 50,000 shots. Each shot is simulated in a multi-core, large memory node and many nodes are used to simulate the whole survey. • Can MOR help? Yes and no. • What we did in 2D shows that the method works and will provide savings as long as the proportion of MOR and HF+pre-processing is adequate. This would in principle work even better in 3D. • The problem with 3D is that those problems get too big to fit in current machinery.
3D: sizes of interest • The local models associated with a shot can be as large as: nx=1000, ny=1000, nz=5000, so that snapshots will have dimension N=5x109, i.e., 20 Giga-bites in single precision. • As the frequency, resolution and complexity of the models are increased, more snapshots are necessary. • Say we consider 1000 snapshots. That means that the size of the matrix of snapshots is 20 Tera-bites, which is about 200 times larger than current large memory single nodes. • With Michael Mahoney we are starting to look into randomized algorithms to see if those can do the trick.
Proper Slanted Decomposition (PSD) • The only place where orthogonality was required in POD is when eliminating U from the left hand side. • Let us consider what happens if we use the original matrix of snapshots S instead. For this we assume that the solution is of the form: u(t)=S a(t). Thus: S att= A S a + B u(t) – 2 D(e) S at. Now, we multiply by ST to get att =(STS)-1 ST A S a + ST B … The coefficient of the first term in the right hand side X can be obtained by solving the matrix LS problem: S X = A S
LSRN: A parallel iterative solver • In a recent still unpublished work, Meng, Saunders and Mahoney describe a code for solving large, strongly over determined systems by least squares. Moreover, there is a publicly available code LSNR. • The method uses random projections to generate a pre-conditioner that then makes iterative methods such as conjugate gradients or Chebyshev run very efficiently. • We are investigating if this, or similar approaches in randomized linear algebra, can help in cracking this problem.
Conclusions • At this evaluation stage one can monitor the true error by comparing the MOR and the HF solutions as one walks away from the problems that generated the basis snapshots. We have done that and as we showed in one of the examples, that there are increasing errors as we solve problems farther away. The errors, however, seem to be concentrated on the amplitudes, while the phases of the events are quite accurate. This is good, since the phases are more important than the amplitudes in migration algorithms. • It is important not to loose sight of the final objective of the simulations (whatever that might be). We are planning to test this technique for more applications in seismic data processing, to see how the approximation error affects the final result. • The more MOR we can use for the same basis, the more efficient the overall approach will be. Very large problems are still out of reach.