Mostrar el registro sencillo del ítem

dc.creatorDuque L.
dc.creatorGutiérrez G.
dc.creatorArias C.
dc.creatorRüger A.
dc.creatorJaramillo H.
dc.date2019
dc.date.accessioned2020-04-29T14:53:58Z
dc.date.available2020-04-29T14:53:58Z
dc.identifier.isbn9789462822894
dc.identifier.urihttp://hdl.handle.net/11407/5775
dc.descriptionWe propose a novel method for velocity estimation that leverages the newest advances in Deep Learning (DL) technology. This method is fully automatic and maps seismic shot-domain data to corresponding depth-domain velocity fields via two neural networks. Our new method is conceptually different from conventional methods such as seismic tomography or Full Waveform Inversion (FWI) that minimize a fixed objective function. Here, a system of neural networks automatically and continuously learns an objective function while training the seismic-to-velocity mapping. The newly introduced method avoids many of the drawbacks of conventional velocity estimation techniques, such as dependence on initial models or cycle-skipping. It uses the full seismic wavefield and avoids picking of first-arrival traveltimes. The system needs to be trained with hundreds or thousands of examples of seismic data paired with their corresponding velocity models relevant for the current project. Training the system is the main computationally demanding step and produces a mapping function that contains the seismic know-how for the presumed geologic setting. The computational cost of the subsequent estimation of velocity from new seismic data is negligent. Our first tests on complex two-dimensional synthetic data produce impressive results, underlining the potential of DL for velocity analysis. © 81st EAGE Conference and Exhibition 2019. All rights reserved.
dc.language.isoeng
dc.publisherEAGE Publishing BV
dc.relation.isversionofhttps://www.scopus.com/inward/record.uri?eid=2-s2.0-85073599084&partnerID=40&md5=f022ea7847272e7a984047f35b0b574a
dc.source81st EAGE Conference and Exhibition 2019
dc.titleAutomated velocity estimation by deep learning based seismic-to-velocity mapping
dc.typeConference Papereng
dc.rights.accessrightsinfo:eu-repo/semantics/restrictedAccess
dc.publisher.programFacultad de Ciencias Básicas
dc.publisher.facultyFacultad de Ciencias Básicas
dc.affiliationDuque, L., Instituto Tecnológico Metropolitano, Colombia; Gutiérrez, G., MOONBLOCK; Arias, C., Instituto Tecnológico Metropolitano, Colombia; Rüger, A., Colorado School of Mines, United States; Jaramillo, H., University of Medellín, Colombia
dc.relation.referencesAminzadeh, F., Jean, B., Kunz, T., 3-D salt and overthrust models (1997) Society of Exploration Geophysicists
dc.relation.referencesAraya-Polo, M., Jennings, J., Adler, A., Dahlke, T., Deep-learning tomography (2018) The Leading Edge, 37 (1), pp. 58-66. , https://doi.org/10.1190/tle37010058.1
dc.relation.referencesHuang, Z., Shimeld, J., Williamson, M., Katsube, J., Permeability prediction with artificial neural network modeling in the Venture gas field, offshore eastern Canada (1996) Geophysics, 61 (2), pp. 422-436. , https://doi.org/10.1190/1.1443970
dc.relation.referencesIsola, P., Zhu, J.Y., Zhou, T., Efros, A.A., (2016) Image-to-Image Translation with Conditional Adversarial Networks, , arXiv preprint
dc.relation.referencesMichelsanti, D., Tan, Z.H., (2017) Conditional Generative Adversarial Networks for Speech Enhancement and Noise-Robust Speaker Verification, , arXiv preprint
dc.relation.referencesMurat, M.E., Rudman, A.J., Automated first arrival picking: A neural network approach (1992) Geophysical Prospecting, 40 (6), pp. 587-604. , https://doi.org/10.1111/j.13652478.1992.tb00543.x
dc.relation.referencesRöth, G., Tarantola, A., Neural networks and inversion of seismic data (1994) Journal of Geophysical Research: Solid Earth, 99 (B4), pp. 6753-6768. , https://doi.org/10.1029/93JB01563
dc.type.versioninfo:eu-repo/semantics/publishedVersion
dc.type.driverinfo:eu-repo/semantics/article


Ficheros en el ítem

FicherosTamañoFormatoVer

No hay ficheros asociados a este ítem.

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem