Machine Learning Techniques for Noise Attenuation in Geophysical Surveys
Techniques are disclosed relating to machine learning in the context of noise filters for sensor data, e.g., as produced by geophysical surveys. In some embodiments, one or more filters are applied to sensor data, such a harsh filter determined to cause a threshold level of distortion in measured reflections, a mild filter determined to leave a threshold level of remaining noise signals, or an acceptable filter. In some embodiments, the system trains a machine learning classifier based on outputs of the filtering procedures and uses the classifier to determine whether other filtered sensor data from the same survey exhibits acceptable filtering. This may improve accuracy or performance in detecting unacceptable filtering, in some embodiments.
This application claims the benefit of U.S. Provisional Application No. 62/687,506, filed on Jun. 20, 2018, and of U.S. Provisional Application No. 62/776,854, filed on Dec. 7, 2018. Each of these applications is hereby incorporated by reference as if entirely set forth herein.
BACKGROUNDGeophysical surveys are often used for oil and gas exploration in geological formations, which may be located below marine environments. Seismic surveys, for example, are based on the use of acoustic waves. In seismic surveys, a survey vessel may tow one or more signal sources (e.g., an air gun) and a plurality of streamers along which a number of acoustic sensors (e.g., hydrophones and/or geophones) are located. Acoustic waves generated by the source(s) are transmitted into the earth's crust and then reflected back and captured at the sensors. Data collected during a marine geophysical survey is analyzed to locate hydrocarbon-bearing geological formations, and thus determine where deposits of oil and natural gas may be located.
Typically, sensor data from a geophysical survey exhibits noise. It may be desirable to filter this noise in order to distinguish the desired signals (signals originating from the survey sources that are reflected and measured by sensors). Examples of noise include ambient vibrations, ocean currents, shipping noise, noise from ocean lifeforms, etc. Sometimes, noise filtration may be performed during a survey (e.g., on-board a survey vessel). Traditionally, filtration quality has been checked manually, e.g., to detect if filtration is too mild (leaving substantial noise in the signal) or too harsh (causing distortions in the desired signals).
In survey system 100, survey vessel 10 is shown towing five signal sources 37A-37E (referred to collectively as “sources 37” or “signal sources 37”) using source cables 18. Note that, in some embodiments, sources may be towed in various patterns (e.g., square or circular patches) using various cable configurations for multi-dimensional data acquisition. In various embodiments, survey vessel 10 may tow any appropriate number of signal sources, including as few as none (e.g., when sources are towed by another vessel) or as many as six or more. In some embodiments, one or more of sources 37 may be impulsive sources such as air guns. In some embodiments, one or more of signal sources 37 may be vibratory signal sources that may be configured to be driven according to a given function by control equipment 12. For example, in various embodiments, one or more of signal sources 37 may be vibratory signal sources configured to be driven according to a given function or digital code.
Survey vessel 10 includes equipment, shown generally at 12 and, for convenience, collectively referred to as “control equipment.” Control equipment 12 may include devices such as a data recording unit (not shown separately) for making a record of signals generated by various geophysical sensors 22 in the system 100. Control equipment 12 may also include navigation equipment (not shown separately), which may be configured to control, determine, and record, at selected times, the geodetic positions of: survey vessel 10, each of a plurality of geophysical sensors 22 disposed at locations on streamers 20, and signal sources 37. Geodetic position may be determined using various devices, including global navigation satellite systems such as the global positioning system (GPS), for example. In the illustrated embodiment, survey vessel 10 includes geodetic positioning device 12A. Additional positioning devices may be placed at various locations on streamers 20. In some embodiments, control equipment 12 is configured to control sources 37, e.g., to control when the sources 37 activate, where the sources 37 are positioned, the manner in which the sources 37 are activated, etc. Note that, although control equipment 12 is shown on survey vessel 10, this embodiment is provided merely as an example and is not intended to limit the scope of the present disclosure. In various embodiments, components of control equipment 12, or the entirety of control equipment 12, may alternatively be located on a separate vessel (not shown) or at a remote location as desired.
Control equipment 12, in various embodiments, includes a computing system (an example embodiment of which is discussed below with reference to
In
In various embodiments, streamers 20 may include any of various appropriate modules in addition to geophysical sensors 22. In geophysical survey systems that include a plurality of laterally spaced-apart streamers, such as system 100, streamers 20 are typically coupled to towing equipment that secures the forward end of each of streamers 20 at a selected lateral position with respect to adjacent streamers and with respect to survey vessel 10. For example, as shown in
Seismic imaging of the sub-surface is very important for hydrocarbon prospection. The acquired seismic data typically flows through a complex chain of compute-intensive processing stages in order to produce the final image of the sub-surface geology.
One particular stage in this processing chain is the attenuation of noise. “Noise,” in this context, refers to any seismic energy that does not represent a reflection of signals (originating from a survey source) from a geological layer. Sources of noise are both environmental and man-made. Noise degrades the quality of the sub-surface image and renders interpretation of the geological layers less accurate. Hence, it is important to remove noise, but this needs to be done in a controlled way in order not to attenuate the desired signal.
Quality control (QC) assessment is often performed at each noise attenuation step to ensure that key seismic image properties such as bandwidth, resolution, and signal to noise ratio (SNR) are not compromised. The QC step traditionally requires a good deal of human interaction and represents a considerable part of the project's time line. With the continuous increases in the volume of the acquired seismic data, automating at least a portion of the QC step may be desirable to reduce QC time. In disclosed embodiments, automation is performed using machine learning, where a machine learning classifier is trained based on a sub-set of the data and is then used to generalize QC decisions on a larger dataset. Example techniques for training and using such a machine learning classifier are discussed in detail below with reference to
In various embodiments, sensor measurements may be filtered to remove noise. Filtering parameters may be determined for a survey using a subset of the survey data and filter outputs may be monitored to determine whether filtering meets one or more quality control criteria.
For example, in the illustrated embodiment, the system processes survey data subset 210 using mild filter 212, harsh filter 214, and acceptable filter 216. In some embodiments, a geophysicist may examine initial filter results for subset 210 to identify mild and harsh filters. The geophysicist (or an automated module) may adjust filter parameters or apply a number of filtering procedures to ensure that both harsh and mild filtering occurs for subset 210 (or more generally, to ensure that filtering occurs that does not visually pass quality control requirements, for use in training). In some embodiments, examples of harsh, mild, and acceptable filtering are determined during a parameter testing phase of the filtering process. Non-limiting examples of filters that may be utilized, with various input parameters, include: Singular spectrum analysis (SSA), reduced-rank filtering, prediction error filtering, time-frequency peak filtering (TFPF), wavelet analysis, independent component analysis (ICA), etc. In various embodiments, one or more of these filtering techniques correspond to various means for applying a filtering procedure to sensor data.
The outputs of the filters are stored as training data 220, in the illustrated embodiment, which includes data with residual noise 222 due to mild filtering, data with signal distortion 224 due to harsh filtering, and acceptable data 226.
Attribute determination module 230, in the illustrated embodiment, determines attributes for the training data. In some embodiments, attribute values are determined for each ensemble of data (e.g., group of traces) such as each shot point, e.g., for embodiments with impulsive sources such as airguns. In embodiments with other types of sources such as vibratory sources, attributes may be determined for each ensemble of data such as receiver-gather, shot-gather, or any other collection of seismic traces on which the filtering or the QC is to be performed. In some instances, filtering is performed within a given ensemble of data (e.g. common depth point (CDP)), but the attribute computation for the purpose of automatic QC is performed on a different ensemble of data (e.g., SHOT). Examples of techniques for determining attributes are discussed in further detail below and include, for example, cross-correlation, mutual information, mean lambda, and rank correlation.
In various embodiments, attribute determination module 230 is configured to compute attribute points for each ensemble of data based on degree of similarity between filter outputs and the residual, which is the difference between input data and the filter output. For example, referring briefly to
In some embodiments, because neighboring seismic ensembles (such as successive shot gathers) will likely have the same classification, the attribute vector for each ensemble may be augmented with attributes of neighboring ensembles (and the number of neighboring ensembles considered for this procedure may be user-adjustable). The new vector of attributes may be referred to as an augmented attribute vector.
Referring again to
Training module 245 uses the feature extraction results to train classifier 250, in some embodiments. For example, referring briefly to
In some embodiments, to determine the classification of a new ensemble of seismic data that was not used in building the classification system, first its corresponding vector of attributes is computed, then the vector of features is obtained using the already derived feature mapping. The vector of features is determined in the feature space (e.g., as plotted in
In various embodiments, training a classifier on a per-survey or per-survey-portion basis may improve automatic quality control, given that desirable filtering parameters may vary between surveys and/or survey portions.
Note that although attribute determination, feature extraction, and decision spaces are discussed herein for purposes of illustration, these techniques are not intended to limit the scope of the present disclosure. In other embodiments, any of various training techniques may be implemented to train a machine learning classifier based on different sets of training data with different filtering characteristics. Further, although the example techniques discussed herein train a classifier using harsh, mild, and acceptable sets of training data, in other embodiments, the disclosed techniques may use larger or smaller numbers of classifications. For example, the classifier may be trained to differentiate between harsh and acceptable filtering only or between mild and acceptable filtering only. As another example, the classifier may be trained to differentiate four or more classes (e.g., with multiple classes of acceptable filtering, mild filtering, and/or harsh filtering to indicate severity of filtering problems, which may inform filter parameter adjustment).
In various embodiments, the operations performed by elements 212, 214, 216, 230, 240, and 250 of
In various embodiments, the disclosed techniques may advantageously improve accuracy in detecting undesirable filtering of a set of survey data, relative to traditional manual techniques. The disclosed techniques may also improve performance, reducing time required for quality control procedures. In some embodiments, the disclosed techniques may provide a new way to perform quality control that was previously performed manually by geophysicists. In some embodiments, the disclosed techniques may be performed on-board a survey vessel that tows sources or sensors for the survey, which may be difficult using traditional techniques. Speaking generally, the disclosed techniques are directed to improvements in the technical field of quality control for filtering of sensor data, which may be particularly relevant in geophysical surveying.
Example Techniques for Generating AttributesThe following discussion provides examples for determining attributes based on filter output and a difference between the filter output and the input data, according to some embodiments. In some embodiments and situations, these attributes can be very informative for distinguishing between different types of filtering. These examples are not intended to limit the scope of the present disclosure. Any of various other attributes may be used in addition to or in place of the disclosed attributes. In some embodiments, various attributes may be used in combination to generate points in an attribute space (e.g., an N-dimensional space when generating N attributes).
A first example technique for generating an attribute is cross correlation, e.g., Pearson's cross correlation. This may utilize the following equation, for example, in which xi represents the output seismic samples after filtering, di represents the difference seismic samples, and N is the number of samples:
A second example technique for generating an attribute is determining mean lambda. This technique may utilize the following equation, for example, in which yi represents the input to the filter and x and d represent output samples and difference samples as described above:
A third example technique for generating an attribute is a mutual information function, e.g., as defined by the following equation:
in which p (x, d) is a joint probability density function of the seismic samples of the output (x) and the difference (d) and p(x) and p(d) are, respectively, the individual probability density functions for the seismic samples of the output and for the difference. Empirical estimation of PDF is a well-known topic in statistical analysis.
A fourth example technique for generating an attribute is correlation, e.g., rank correlation such as Kendall rank correlation or Markov rank correlation. Kendall rank correlation may utilize the following procedure. Draw randomly two samples for X and D: (xi, di) and (xj, dj). Set a1=sign(xi-xj) and a2=sign(di-dj). If (a1==a2), then the pairs are concordant. If not, then the pairs are discordant. After processing a number of pairs, the Kendall rank correlation is:
Markov rank correlation may utilize the following procedure. Draw randomly one sample for X and D: (xi, di). Set a1=sign(xi-x(i-L)) and a2=sign(di-d(i-L)), where L is an integer that indicates the distance between neighboring samples. If (a1==a2), then the pairs are concordant. If not, then the pairs are discordant. In some embodiments, a relatively large number of random pair drawings are used for Kendall and Markov correlation to produce reliable statistics. After processing a number of pairs, the Markov rank correlation is:
Turning now to
At 610, in the illustrated embodiment, a computing system accesses sensor data. The sensor data may be obtained by a geophysical survey, for example, where the sensor data is based on sensor measurements of subsurface reflections of signals emitted by one or more survey sources (e.g., seismic sources, electromagnetic sources, etc.). The accessing may include reading from one or more computer-readable media to access sensor data recorded by survey sensors. In various embodiments, element 610 alone, in combination with the other operations of
At 620, in the illustrated embodiment, a one or more filtering procedures are applied to the sensor data, an output of the one or more filtering procedures is determined to have a threshold level of distortion to the sensor data (e.g., due to a harsh filter) or leave a threshold level of remaining noise signals in the sensor data (e.g., due to a mild filter). Note that the harsh and mild filtering procedures may implement the same type of filter (e.g., with different input parameters) or may implement different types of filters.
At 630, in the illustrated embodiment, the computing system trains a classification engine based on outputs of the first and second filtering procedures. Generally, the training may be based on various combinations of training data, such as harsh/normal, mild/normal, harsh/mild/normal, very harsh/harsh/normal, etc. As discussed above, the classification engine may implement various machine learning techniques, alone or in combination (e.g., using model stacking).
The training may include generating respective sets of attributes in an attribute space for the one or more filtering procedures. This may be determined based on similarity between: outputs of the one or more filtering procedures and differences between the outputs of the one or more filtering procedures and the sensor data. The training may include performing feature extraction to generate respective sets of features for the one or more filtering procedures in a feature space.
At 640, in the illustrated embodiment, the computing system classifies, using the trained classification engine, other filtered sensor data to determine whether the other filtered sensor data exhibits acceptable filtering. In some embodiments, sensor data that exhibits acceptable filtering is stored on a non-transitory computer-readable medium, thereby completing the manufacture of a geophysical data product. The training and classifying of elements 630 and 640 may be performed during the same geophysical survey. The classification may generate attributes and features for the other sensor data and determine which portions of the feature space correspond to the generated features.
Example Computing DeviceTurning now to
In the illustrated embodiment, computing device 710 further includes computer-readable medium 714 as a possibly distinct element from storage subsystem 712. For example, computer-readable medium 714 may include non-transitory, persistent, tangible storage such as tape reels, hard drives, CDs, DVDs, flash memory, optical media, holographic media, or other suitable types of storage. In some embodiments, computer-readable medium 714 may be physically separable from computing device 710 to facilitate transport. In some embodiments, computer-readable medium 714 may be used to manufacture a geophysical data product. For example, in some embodiments, seismic data (generated and recorded according to any one of various disclosed embodiments), or further processed versions of such seismic data, may be stored on computer-readable medium 714, thereby completing manufacture of a geophysical data product. Although shown to be distinct from storage subsystem 712, in some embodiments, computer-readable medium 714 may be integrated within storage subsystem 712.
In various embodiments, processing unit 750 includes one or more processors. In some embodiments, processing unit 750 includes one or more coprocessor units. In some embodiments, multiple instances of processing unit 750 may be coupled to interconnect 760. Processing unit 750 (or each processor within 750) may contain a cache or other form of on-board memory. In some embodiments, processing unit 750 may be implemented as a general-purpose processing unit, and in other embodiments it may be implemented as a special purpose processing unit (e.g., an ASIC). In general, computing device 710 is not limited to any particular type of processing unit or processor subsystem.
As used herein, the terms “processing unit” or “processing element” refer to circuitry configured to perform operations. Accordingly, a processing unit may be implemented as a hardware circuit implemented in a variety of ways. The hardware circuit may include, for example, custom very-large-scale integration (VLSI) circuits or gate arrays, off-the-shelf semiconductors such as logic chips, transistors, or other discrete components. A processing unit may also be implemented in programmable hardware devices such as field programmable gate arrays, programmable array logic, programmable logic devices, or the like. A processing unit may also be configured to execute program instructions from any suitable form of non-transitory computer-readable media to perform specified operations.
Storage subsystem 712 is usable by processing unit 750 (e.g., to store instructions executable by and data used by processing unit 750). Storage subsystem 712 may be implemented by any suitable type of physical memory media, including hard disk storage, floppy disk storage, removable disk storage, flash memory, random access memory (RAM-SRAM, EDO RAM, SDRAM, DDR SDRAM, RDRAM, etc.), ROM (PROM, EEPROM, etc.), and so on. Storage subsystem 712 may consist solely of volatile memory in one embodiment. Storage subsystem 712 may store program instructions executable by computing device 710 using processing unit 750, including program instructions executable to cause computing device 710 to implement the various techniques disclosed herein.
I/O interface 730 may represent one or more interfaces and may be any of various types of interfaces configured to couple to and communicate with other devices, according to various embodiments. In one embodiment, I/O interface 730 is a bridge chip from a front-side to one or more back-side buses. I/O interface 730 may be coupled to one or more I/O devices 740 via one or more corresponding buses or other interfaces. Examples of I/O devices include storage devices (hard disk, optical drive, removable flash drive, storage array, SAN, or an associated controller), network interface devices, user interface devices or other devices (e.g., graphics, sound, etc.).
Various articles of manufacture that store instructions (and, optionally, data) executable by a computing system to implement techniques disclosed herein are also contemplated. These articles of manufacture include non-transitory computer-readable memory media. The contemplated non-transitory computer-readable memory media include portions of a memory subsystem of a computing device as well as storage media or memory media such as magnetic media (e.g., disk) or optical media (e.g., CD, DVD, and related technologies, etc.). The non-transitory computer-readable media may be either volatile or nonvolatile memory.
This disclosure includes references to “one embodiment,” “a particular embodiment,” “some embodiments,” “various embodiments,” “an embodiment,” etc. The appearances of these phrases do not necessarily refer to the same embodiment. Particular features, structures, or characteristics may be combined in any suitable manner consistent with this disclosure.
Within this disclosure, different entities (which may variously be referred to as “units,” “circuits,” other components, etc.) may be described or claimed as “configured” to perform one or more tasks or operations. This formulation—[entity] configured to [perform one or more tasks]—is used herein to refer to structure (i.e., something physical, such as an electronic circuit). More specifically, this formulation is used to indicate that this structure is arranged to perform the one or more tasks during operation. A structure can be said to be “configured to” perform some task even if the structure is not currently being operated. For example, a “filter configured to process sensor data” is intended to cover, for example, equipment that has a module or circuitry that performs this function during operation, even if the circuitry in question is not currently being used (e.g., a power supply is not connected to it). Thus, an entity described or recited as “configured to” perform some task refers to something physical, such as a device, circuit, memory storing program instructions executable to implement the task, etc. This phrase is not used herein to refer to something intangible. The term “configured to” is not intended to mean “configurable to.” An unprogrammed FPGA, for example, would not be considered to be “configured to” perform some specific function, although it may be “configurable to” perform that function after programming.
Reciting in the appended claims that a structure is “configured to” perform one or more tasks is expressly intended not to invoke 35 U.S.C. § 112(f) for that claim element. Accordingly, none of the claims in this application as filed are intended to be interpreted as having means-plus-function elements. Should Applicant wish to invoke Section 112(f) during prosecution, it will recite claim elements using the “means for” [performing a function] construct.
It is to be understood that the present disclosure is not limited to particular devices or methods, which may, of course, vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting. As used herein, the singular forms “a”, “an”, and “the” include singular and plural referents unless the context clearly dictates otherwise. Furthermore, the words “can” and “may” are used throughout this application in a permissive sense (i.e., having the potential to, being able to), not in a mandatory sense (i.e., must). The term “include,” “comprise,” and derivations thereof, mean “including, but not limited to.” The term “coupled” means directly or indirectly connected.
As used herein, the term “based on” is used to describe one or more factors that affect a determination. This term does not foreclose the possibility that additional factors may affect the determination. That is, a determination may be solely based on specified factors or based on the specified factors as well as other, unspecified factors. Consider the phrase “determine A based on B.” This phrase specifies that B is a factor used to determine A or that affects the determination of A. This phrase does not foreclose that the determination of A may also be based on some other factor, such as C. This phrase is also intended to cover an embodiment in which A is determined based solely on B. As used herein, the phrase “based on” is synonymous with the phrase “based at least in part on.”
As used herein, the phrase “in response to” describes one or more factors that trigger an effect. This phrase does not foreclose the possibility that additional factors may affect or otherwise trigger the effect. That is, an effect may be solely in response to those factors, or may be in response to the specified factors as well as other, unspecified factors. Consider the phrase “perform A in response to B.” This phrase specifies that B is a factor that triggers the performance of A. This phrase does not foreclose that performing A may also be in response to some other factor, such as C. This phrase is also intended to cover an embodiment in which A is performed solely in response to B.
As used herein, the terms “first,” “second,” etc. are used as labels for nouns that they precede, and do not imply any type of ordering (e.g., spatial, temporal, logical, etc.), unless stated otherwise. When used in the claims, the term “or” is used as an inclusive or and not as an exclusive or. For example, the phrase “at least one of x, y, or z” means any one of x, y, and z, as well as any combination thereof (e.g., x and y, but not z).
Although specific embodiments have been described above, these embodiments are not intended to limit the scope of the present disclosure, even where only a single embodiment is described with respect to a particular feature. Examples of features provided in the disclosure are intended to be illustrative rather than restrictive unless stated otherwise. The above description is intended to cover such alternatives, modifications, and equivalents as would be apparent to a person skilled in the art having the benefit of this disclosure.
The scope of the present disclosure includes any feature or combination of features disclosed herein (either explicitly or implicitly), or any generalization thereof, whether or not it mitigates any or all of the problems addressed herein. Accordingly, new claims may be formulated during prosecution of this application (or an application claiming priority thereto) to any such combination of features. In particular, with reference to the appended claims, features from dependent claims may be combined with those of the independent claims and features from respective independent claims may be combined in any appropriate manner and not merely in the specific combinations enumerated in the appended claims.
Claims
1. A method, comprising:
- accessing, by a computing system, sensor data;
- applying one or more filtering procedures to the sensor data, wherein an output of the one or more filtering procedures is determined to have a threshold level of distortion to the sensor data or leave a threshold level of remaining noise signals in the sensor data;
- training, by the computing system, a classification engine based on the output of the one or more filtering procedures; and
- classifying, by the computing system using the trained classification engine, other filtered sensor data to determine whether the other filtered sensor data exhibits acceptable filtering.
2. The method of claim 1,
- wherein the one or more filtering procedures include: a first filtering procedure, wherein an output of the first filtering procedure is determined to have a threshold level of distortion to the sensor data; a second filtering procedure, wherein an output of the second filtering procedure is determined to leave a threshold level of remaining noise signals in the sensor data; and a third filtering procedure, wherein an output of the third filtering procedure is determined to provide acceptable filtering outputs.
3. The method of claim 1,
- wherein the training comprises: generating respective sets of attributes in an attribute space for the one or more filtering procedures based on determined similarity between: outputs of the one or more filtering procedures; and differences between the outputs of the one or more filtering procedures and the sensor data; and performing feature extraction to generate respective sets of features for the one or more filtering procedures in a feature space; and
- wherein the classifying generates attributes and features for the other sensor data and determines which portions of the feature space correspond to the generated features.
4. The method of claim 3, wherein the generating the respective sets of attributes include determining one or more of:
- cross-correlation;
- mean lambda;
- rank correlation; or
- Markov correlation.
5. The method of claim 4, wherein the feature space has a smaller number of dimensions than the attribute space.
6. The method of claim 1, further comprising:
- performing additional filtering in response to a classification that indicates that filtering of the other sensor data left noise signals in filtered data.
7. The method of claim 1, further comprising:
- performing another filtering procedure with different parameters in response to a classification that indicates that filtering of the other sensor data caused distortion.
8. The method of claim 1, further comprising:
- storing sensor data that is classified as acceptable on one or more non-transitory computer readable media to generate a geophysical data product.
9. The method of claim 1, wherein the classification engine is a neural network or a support vector machine.
10. The method of claim 1, wherein the sensor data is obtained by a geophysical survey based on sensor measurements of subsurface reflections of signals emitted by one or more survey sources.
11. The method of claim 10, wherein the training and classifying are performed during the same geophysical survey.
12. A non-transitory computer-readable medium having instructions stored thereon, where the instructions are executable by one or more processors to perform operations comprising:
- accessing sensor data;
- applying one or more filtering procedures to the sensor data, wherein an output of the one or more filtering procedures is determined to have a threshold level of distortion to the sensor data or leave a threshold level of remaining noise signals in the sensor data;
- training a classification engine based on the output of the one or more filtering procedures; and
- classifying, using the trained classification engine, other filtered sensor data to determine whether the other filtered sensor data exhibits acceptable filtering.
13. The non-transitory computer-readable medium of claim 12,
- wherein the one or more filtering procedures include: a first filtering procedure, wherein an output of the first filtering procedure is determined to have a threshold level of distortion to the sensor data; a second filtering procedure, wherein an output of the second filtering procedure is determined to leave a threshold level of remaining noise signals in the sensor data; and a third filtering procedure, wherein an output of the third filtering procedure is determined to provide acceptable filtering outputs.
14. The non-transitory computer-readable medium of claim 12,
- wherein the training comprises: generating respective sets of attributes in an attribute space for the one or more filtering procedures based on determined similarity between: outputs of the one or more filtering procedures; and differences between the outputs of the one or more filtering procedures and the sensor data; and performing feature extraction to generate respective sets of features for the one or more filtering procedures in a feature space; and
- wherein the classifying generates attributes and features for the other sensor data and determines which portions of the feature space correspond to the generated features.
15. The non-transitory computer-readable medium of claim 12, wherein the operations further comprise:
- performing additional filtering in response to a classification that indicates that filtering of the other sensor data left noise signals in filtered data.
16. The non-transitory computer-readable medium of claim 12, wherein the operations further comprise:
- performing another filtering procedure with different parameters in response to a classification that indicates that filtering of the other sensor data caused distortion.
17. An apparatus for classifying sensor processed data from a marine geophysical survey, comprising:
- means for accessing sensor data;
- means for applying one or more filtering procedures to the sensor data, wherein an output of the one or more filtering procedures is determined to have a threshold level of distortion to the sensor data or leave a threshold level of remaining noise signals in the sensor data;
- means for training a classification engine based on the output of the one or more filtering procedures; and
- means for classifying, using the trained classification engine, other filtered sensor data to determine whether the other filtered sensor data exhibits acceptable filtering.
18. The apparatus of claim 17, wherein the training includes:
- generating respective sets of attributes in an attribute space for the one or more filtering procedures based on determined similarity between: outputs of the one or more filtering procedures; and differences between the outputs of the one or more filtering procedures and the sensor data; and
- performing feature extraction to generate respective sets of features for the one or more filtering procedures in a feature space; and
- wherein the classifying generates attributes and features for the other sensor data and determines which portions of the feature space correspond to the generated features.
19. In a technological method of filtering survey sensor measurements that includes: the specific technological improvement comprising:
- accessing, by a computing system, sensor data;
- applying one or more filtering procedures to the sensor data, wherein an output of the one or more filtering procedures is determined to have a threshold level of distortion to the sensor data or leave a threshold level of remaining noise signals in the sensor data;
- training, by the computing system, a classification engine based on the output of the one or more filtering procedures; and
- classifying, by the computing system using the trained classification engine, other filtered sensor data to determine whether the other filtered sensor data exhibits acceptable filtering, thereby improving automated filtering quality control accuracy for the sensor data.
20. The method of claim 19,
- wherein the training comprises: generating respective sets of attributes in an attribute space for the one or more filtering procedures based on determined similarity between: outputs of the one or more filtering procedures; and differences between the outputs of the one or more filtering procedures and the sensor data; and performing feature extraction to generate respective sets of features for the one or more filtering procedures in a feature space; and
- wherein the classifying generates attributes and features for the other sensor data and determines which portions of the feature space correspond to the generated features.
Type: Application
Filed: Jun 10, 2019
Publication Date: Dec 26, 2019
Inventor: Maiza Bekara (Kingston Upon Thames)
Application Number: 16/436,211