Seismic data processing workbook(ProMax)

Embed Size (px)

Citation preview

  • 8/19/2019 Seismic data processing workbook(ProMax)

    1/49

    ASSIGNMENT COVER SHEET

    Student Name: __________________________________________

    Student ID: __________________________________________

    Unit Name: __________________________________________

    Lecturer’s Name: __________________________________________

    Due Date: __________________________________________

    Date Submitted: __________________________________________

    DECLARATION

    I have read and understood Curtin’s policy on plagiarism, and, except where indicated, thisassignment is my own work and has not been submitted for assessment in another unit orcourse. I have given appropriate references where ideas have been taken from the published orunpublished work of others, and clearly acknowledge where blocks of text have been taken fromother sources.

    I have retained a copy of the assignment for my own records.

    ________________________________________

    [Signature of student]

    For Lecturer’s Use Only:

    Overall Mark: ________ out of a tot al of _________ Percentage:

    Lecturer’s Comments:

    Lecturer’s Name: Date Returned:

    Sutthisrisaarng Pholpark

    17682974

    Geophysical Data Processing 612 (Petroleum)

    Sasha S.

    7 November 2014

    7 November 2014

    Sutthisrisaarng Pholpark

  • 8/19/2019 Seismic data processing workbook(ProMax)

    2/49

    Lab 1. Unix basics, ProMAX user interface.Part 1. Unix basicsPromax basic command

    Command Responseman Call manual, a page must be specified.man --helpman man Call a manual of manual.man ls Call a manual of ls.ls Show files in a current directory.

    [T17682974@egplnxc1 ~]$ lsconfig-local-gp312 mycalendar.txt part1.sgy run wd_nov2008_shots.sgy

    ls Show files in a current directory in a long detail format.ls -a Show files in a current directory (short forma6t), included hidden files .pwd Show a current directory.

    [T17682974@egplnxc1 ~]$ pwd/share/data/home/students/T17682974

    who List a current user in the same network.whoami Show a current username.cp file1 file2 Copy a file.

    [T17682974@egplnxc1 ~]$ lsconfig-local-gp312 mycalendar.txt part1.sgy run wd_nov2008_shots.sgy[T17682974@egplnxc1 ~]$ cp part1.sgy copy.sgy[T17682974@egplnxc1 ~]$ lsconfig-local-gp312 copy.sgy mycalendar.txt part1.sgy runwd_nov2008_shots.sgy

    rm Remove a file

    [T17682974@egplnxc1 ~]$ rm copy.sgy[T17682974@egplnxc1 ~]$ lsconfig-local-gp312 mycalendar.txt part1.sgy run wd_nov2008_shots.sgy

    rmdir dir1 Remove an empty directory.mkdir dir1 Create a new directory.

    [T17682974@egplnxc1 ~]$ mkdir testdirect[T17682974@egplnxc1 ~]$ lsconfig-local-gp312 mycalendar.txt part1.sgy run testdirectwd_nov2008_shots.sgy

    chmod Changes the permissions of file.cal Call a calendar.cat Print a text file on the screentop Display all current tasks.ps Show a snap shot of the current operation.ps -e To see every process on the system using standard syntaxkillxterm Strart a new panel, the old panel is unable to operate until the new panel is

    closed.xterm& Strart a new panel, the old panel is still able to operate.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    3/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    4/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    5/49

    Land seismic- Answer the following questions:a. What are the main parameters of file, i.e. sampling interval, number of samples (see thelog file)?

    # Traces per Ensemble .............. = 156# Auxiliary Traces per Ensemble .... = 51614

    Sample interval (micro sec) ........ = 1000Recording sample interval .......... = 0# samples per trace ................ = 3001# recording samples per trace ...... = 1250Data sample format ................. = 4 Byte IBM floating point

    b. Which trace headers have some defined values?b. Which trace headers have some defined values?FFID: Field File Identification Number , increaseswith #of shots

    CHAN: #of channelOFFSETAOFFSETCDP: Common Depth PointCDP_X,CDP_YLINE_NOREC_X,REC_XDFILE_NOSEQNOSOURCE

    SOU_X,SOU_XDSOU_H2ODTFULL_ETFULL_STLIVE_ETLIVE_STRACENOTRC_TYPETR_FOLD

    c. Identify main components of the wave field.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    6/49

    Lab 2: GeometryObjective

    Learn to set geometry for raw seismic data and perform quality control after geometry isassigned.

    FLOWS

    1/5 Load seismic data to the databaseFlow: 010 – SEGY Input

    SEG-Y Input - read all traces from ‘’part1.sgy ” Disk Data Output – Stored data in the database (at the current flow) name “raw data”

    The loaded data can be viewed by clicking‘datasets’ tab, then use MB2 to click ‘raw data’.

    If use MB3, Promax will show history of the file.

    2/5 Creating database files Flow: 020 – Extract DB FilesThe main function of this flow is to extract headers from ‘raw data’ in order to generate Promaxdatabase entries.

    Since trace headers are only used for DB files

    generation, click ‘YES’ at ‘Process trace heade rsonly’.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    7/49

    3/5 Assigning geometry Flow: 030 – GeometryThe only one function of this flow is to generate ‘2D Marine Geometry Spreadsheet’ .The picture below has been taken from GP312, lecture2 geometry.

    Number of channels in the streamer: 230Minimum inline offset: 130 mCrossline source offset: 0 mReceiver group interval 12.5 mSource interval: 50 mStreamer towing depth: 8 mSource towing depth: 6 mNominal sail line azimuth: 0°Use ‘matching pattern’ midpointsassignment method.

    After flow ‘030 – Geometry’ has beenexecuted, the geometry spreadsheet will showup.

    Geometry assignment sequence (menu items) as provided in the lab instructions are:1. Setup – enter parameters2. Sources – coordinates of sources3. Pattern – define pattern of recievers4. Bin – assign midpoints, bin and finalize database5. QC the results – show survey pattern

    Setup Sources

    Patterns

  • 8/19/2019 Seismic data processing workbook(ProMax)

    8/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    9/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    10/49

    After project the direct wave pick to all ensembles, all of them follow the projected lines whichindicates that assigned geometry works properly.

    4. Resort data in CDP:AOFFSET order, explain changes in number of traces per ensemble in respectto SOURCE:AOFFSET.SOURCE:AOFFSET - #trace = 230 traces per ensemble (constant number thorough all of SIN)CDP:AOFFSET- #trace = traces per ensemble increasing as the number of fold increasing as in the

    fold diagram below. Number of traces start from 1 at CDP1, increasing up to 29 and then decreasingto 1 afterwards.Hence, in the case of CDP:AOFFSET sorting, #trace in each CDP ensemble depends on #fold whileSOURCE:AOFFSET sorting #trace in SIN ensemble is a constant number 230 traces per ensemble.

    #traces increses with #fold

    Fold diagram

  • 8/19/2019 Seismic data processing workbook(ProMax)

    11/49

    5. Plot only nearest channel, all shots. Find direct wave, bottom reflection, water bottom multiple.

    SIN:AOFFSET

    CHAN:SIN

  • 8/19/2019 Seismic data processing workbook(ProMax)

    12/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    13/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    14/49

    Do all the traces have similar quality and if not answer to following question. How many ‘bad’ traces are there in the dataset? Is it just a faulty receiver or a number of receivers or a

    whole shot? Identify ‘bad’ data and show proofs.

    In CHAN:FFID sorting, not all the traces have similar quality. There are bad recording reoccur at FFID 187 (onerecording totally fail).

    In FFID:OFFSET sorting, as the picture below we can clearly see that FFID187 failed to record the data.

    1.3 Resort data in CDP: AOFFSET order;- explain changes in number of traces per ensemble in respect to SOURCE: AOFFSET.In CDP: AOFFSET, number of traces per ensemble increasing regarding to #fold. SIN:AOFFSET sorting, however, has aconstant #traces per ensemble through all the SIN. The differences are shown in the picture below.

    CDP: AOFFSET - #traces start from 1 and increase towards, maximum at 29 traces per ensemble and the decrease

    afterwards to 1. (As show in the fold diagram in the next page)

    SIN:AOFFSET - #traces = 230 through all SIN

  • 8/19/2019 Seismic data processing workbook(ProMax)

    15/49

    - Compute expected fold from the survey parameters (Lecture 2),

    =230∗ 12.5

    2 ∗ 50 = 28.5

    - Compare it to the fold you could estimate visually. Why number of traces per ensemble on first few ensembles issmaller than on the CDP ensembles from the central part?

    The calculated fold number is not equal to the true fold in each CDP. In CDP:AOFFSET sorting, number of tracesper ensemble depends on #fold at a particular CDP. Since first few ensembles have small fold number, they have lower#fold than CDPs at the central as shown in the fold diagram.

    1.4 Explain the following sorting order

    Sorting order - CDP:OFFSET1-50(5):1-500For CDP: taking trace 1 - 50 every 5 traces e.g. 1,6,11,… and for OFFSET: taking 1 - 500 every traces./ means ‘and’ (use to separate sorting order, but merging all of them together)50-1000(10):500-1000For CDP: taking trace 50 - 1000 every 10 traces e.g. 50,60,70… and for OFFSET: taking 500 -100 every traces.

    Can this sorting order be used- for the marine dataset ‘raw data’ - for the marine dataset ‘raw data wg’ and why? This sorting order cannot be used for the raw data because it has not been assigned CPD, however, it can be used forthe raw data with geometry because its CDPs has been assigned in bin process.

    2. Interactive analysis of travel time curves1. Sort data in SOURCE: AOFFSETorder, display each 50 th source position.Sorting order is 1-500(50):*/2. Using tool estimate apparent velocity of direct wave.

    The picture below shown the estimate Va from SIN1, 51, 101, 151 and 201.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    16/49

  • 8/19/2019 Seismic data processing workbook(ProMax)

    17/49

    3. Create or copy flow ‘ TRACE DISPLAY’, read and display first 5 ensembles from ‘land raw data’ dataset usingSOURCE: CHAN sorting.

    4. Identify direct wave, reflected wave and ground roll . Estimate parameters of travel time curves of these waves.Each wave type is shown in the picture below. The average velocity of the direct wave is 1800 ft/s and the averagevelocity of the reflected is 2040 ft/s.

    3. Interactive spectral analysis – marine data1. Create a flow SPECTRAL ANALYSISwith Disk Data input and Interactive Spectral Analysis routines . Data must besorted in SOURCE: CHAN order.2. Select parameters of the ISA as shown on the figure below:

  • 8/19/2019 Seismic data processing workbook(ProMax)

    18/49

    3. Compute power spectrum of direct wave, water bottom reflection, and ambient noise recorded before the firstarrivals . Explain the results.

    Ambient noise- Energy scatteringthroughout all channels- Energy peak at 3 Hz (lowfrequency)

    - Normal distribution ofthe energy centred at 35Hz.

    Reflected wave

    - Peak energy at 50 Hz- Distribution of energy isin a range between 0 – 70Hz.- Energy spread throughall of the channels.

    Direct wave- Energy peak at 8 Hz anddistribute between 0-60Hz.- Energy concentrates atCH 0-20.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    19/49

    4. Try different windows (small, big, narrow or wide) for power spectrum calculation. Analyse the result.The window dictates the area of calculation. The wider area the more information, however, some of information isirrelevant to a particular analysis objective. Hence, we need to set our analysis target and select window size andlocation according to the target e.g. for direct wave spectral analysis, only select the area covered direct wave and trynot to select others feature.

    5. Display Average Power spectrum for ambient noise. Change display type from Percent Power to dB Power. Go toView -> Power Scaling -> Display Type. Analyse the difference.How do the values on the Percent power spectrum relate to those on dB power spectrum?dB power is computed from energy in each signal frequency and display without weighting (power dB = 10logP(f)). Onthe other hand, %power display is obtained from P(f)=(abs F(f))^2 with weighting effect depends on how muchfrequency lying at a particular frequency. As the display below, energy highly concentrate at 0 – 10 Hz, dB powerdisplay showing true dB power at the range of frequency, but %power weights this energy with power of energy forwhole frequency content. Overall, %power can clearly show us which frequency contains the maximum and theminimum energy.

    6. Change paremeters of phase visualisation. Go to View -> Phase Spectrum ->Phase shift type – >LinearThis function shows us how the phase spectrum will look like after linear shift (according to input value). We canchange it to sample shift or time shift, which is shifting the phase to a particular sample/time.Phase wrapping -> Unwrapping

    Unwrapping the phase spectrum removes the discontinuities at +/- 180 degrees in phase spectrum. So it will makephase spectrum smoother by average individually unwrapping phase.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    20/49

    3a. Interactive spectral analysis – land dataRepeat same procedure for land data example. What is amplitude spectrum of direct wave, ground roll, reflections?

    Central frequency of the direct wave is 40 Hz. There is a relatively high energy spectral at 2-3 Hz, which is very lowfrequency that may attribute from background noise.

    Amplitude spectrum of ground roll concentrated at 20 Hz, can be considered as central frequency of the ground roll.

    There is a peak energy spectral at 3 Hz, which is very low frequency that may attribute from background noise. Thecentral frequency of the reflected wave is 25 Hz.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    21/49

    Workshop 4. Pre-processing and signal processing

    1. Trace KillSome traces can contain no signal (due to hardware malfunction) or be polluted with large amount of noise (acoustic orelectric). These are to be removed from processing as they are not going to improve results of processing.

    Create processing flow 050 – Pre-processing with Disk Data Input (DDI, from raw data with geometry) and Trace Display(TD). Data should be sorted in SIN:CHAN order, display should be in grayscale mode, ~20 ensembles per screen.

    Scroll though the data and pick bad traces (Picking/Bad traces, see the figure below). You’ll need to create new pick,select CHAN as a secondary trace header. Save the pick (File->Save Picks). Exit TD.

    After picking the traces are stored in : ‘bad trace ’.

    After add ‘Trace Kill’, select to remove bad trace from the pick and execute , the bad trace is removed from the data.

    2. MutingTop/bottom muting can be used to remove coherent noises recorded above/below certain time. In the training marinedataset the only type of noise which can be removed in that way is direct wave. In fact there’s not much reason to muteit as it is not interfering with the signal. However in order to learn how to use muting pick top mute which will eliminate

    direct wave as a function of (a)offset and try applying it in the pre-processing flow using Trace Muting routine .

    It might be convenient to change parameters of DDI and TD in the pre-processing flow in order to adjust parametersusing limited amount of ensembles (keep the same sorting, but input every 20-th common-source gather ).

    The flow will be applied to the whole dataset later on.Muting parameters should be picked in common source gathers:

  • 8/19/2019 Seismic data processing workbook(ProMax)

    22/49

    1. Picking- >Top Mute, create new table (‘Top Mute’ ), use AOFFSET as a secondary key . 2. Pick top mute on one of the ensembles and project to others (MB3, Project):3. Save the table (File->Save picks).

    4. Add Trace Muting to processing flow:

    3. Bandpass filtering

    The data contains low-frequency random noise. To remove it, add Bandpass Filter to processing flow. Select optimumparameters and explain your approach in the lab book.

    Usable seismic reflection energy – 10-70 Hz with 30 Hz – peak frequency (Lecture note 4). The approaches of filterdesigns are to keep the usable signal as much as possible and screen noise out. Interactive spectral analysis is used todetermine frequency content in the signal.

    Ormsby bandpass filter 5-10-65-75 is selected. The result is shown below.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    23/49

    4. Amplitude decay compensation, Parameter testTwo different approaches can be applied to amplitude decay compensation.

    3.1 True Amplitude RecoveryTry using True Amplitude Recovery to apply ‘Time raised to a power’ correction with different parameters.

    Note: This setting can suppress the presence of the direct wave in early time and enhance reflection waves in later time.

    It is convenient to experiment with TAR parameters in common offset domain, looking to nearest channel i.e. resort thedata by changing parameters of DDI to read only first channel (but all shots):

    Note: CHAN:SIN sorting shows effect of TAR clearer

  • 8/19/2019 Seismic data processing workbook(ProMax)

    24/49

    Using Parameter Test routine create display with time-power TAR constant being 0.5, 1.0, 1.5, 2.0, 3.0. Select bestoption:3. Change parameters of the TAR as shown below (99999 is a special value for the parameter being tested):

    4. Run the flow.

    5. Select the best result (explain why), change parameter at TAR from 99999 to the parameter which produced it.

    Power 2.0 is selected because it yields the optimum results in both early time and late time reflections as shown in thepicture below. From power=0.5 and 1.0, TAR fail to attenuate direct wave. Moreover, they are unable to recoveramplitudes of late time reflection. Power = 1.5 is acceptable, however, there still clearly see the presence of the directwave. Power = 3.0 yields the best late time reflection, on the other hand, the early time reflections are suppressed toomuch as some reflectors are hardly recognize.

    6. Remove Parameter Test from the flow (or make it inactive).

    3.2 Automatic gain control (AGC)Try using AGC instead of TAR. Use Parameter Test to vary AGC operator length from 30 to 1000 ms (4-5 values).Select one of the approaches (TAR or AGC).

    Parameter test options

    In AGC, time window 500 ms is selected due to it yields the optimum result. We can clearly see reflector details afterAGC (with window500 ms) , however, coherence noise is also boosted as shown in the picture below.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    25/49

    Comparing TAR power 2.0 against AGC time window 500ms, TAR gives better subsurface reflector image. In addition, amajority of coherence noise is eliminated while in TAR coherence noise is boosted. Hence TAR power 2.0 is selectedmethod in amplitude recovery process.

    Finalise the pre-processing flow. It should contain:DDI

    Trace Kill(Muting)Bandpass filterTAR or AGC

    My finalise pre-process flow.

    5. Apply the pre-processing flow to the whole datasetAdd Disk Data Output routine to the 050 – Pre-processing flow. Adjust parameters of DDI to apply the flow to wholedata range. Execute the flow.

    Compare raw and pre-processed data using the TRACE DISPLAYflow.The comparisons are shown in the picture below.

    Note: processing summary1. Trace kill: remove bad trace/bad recording(2. Trace muting: remove unwanted part of the traces )3. Bandpass filter: Remove unwanted frequency and preserve usable signal4. Amplitude recovery (TAR/AGC): Recover amplitude at later time. (TAR also suppresses amplitude of ea rly time. It’svery useful for direct wave removal since it appear in the early time record.)

  • 8/19/2019 Seismic data processing workbook(ProMax)

    26/49

    Workshop 5. Deconvolution

    This workshop involves operations with synthetic data only.In your area in ProMAX create line ‘Deconvolution’.

    Deterministic deconvolution1. In a separate processing flow generate synthetic data using routine ‘Synthetic Trc Generation’ and save thedata to a dataset (A).

    List of synthetic events:2,100,10,1,300,0,20/2,200,10,-1,300,0,20/2,230,10,0.5,300,0,20/2,400,10,1,300,0,20/ 2,500,10,-2,300,0,20/2,800,10,1,300,0,20/Last parameter in the line is the central frequency of (minimum phase equivalent of) Ricker wavelet.

    2. In the same flow create dataset (B) with same events using spikes instead of Ricker wavelet. Comparedatasets A and B.

    The first dataset simulates earth response filtered with the wavelet. Find corresponding events, explain yourobservations.The corresponding events are that both dataset occur at the same time and their amplitude at a particular timeare match e.g. at 500 ms we can clearly see the biggest Ricker wavelet and the biggest impulse with amplitude -2. Ricker wavelet is a simulate earth response convolved with the transmitted wavelet, on the other hand, spikewavelet represents the earth response.

    3. Create a dedicated flow to derive an inverse filter.First, create single trace containing the same impulse you used for synthetic generation.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    27/49

    Create inverse filter and save it to separate dataset:

    4. Compare amplitude spectra of the impulse and the inverse filter.

    In inverse filter spectra analysis, it has thecentral frequency at 59 Hz. Its energymostly lying between 45 - 65 Hz. After 50Hz, it has low energy spectral-20 dB throughout the entire energyspectral.

    In synthetic impulse spectral analysis,however, has central frequency at 20 Hz.Majority of the energy concentrated at 0 – 40 Hz.

    From the observation, synthetic impulsehas lower frequency than its inverse sinceits central frequency concentrated at thelower frequency.

    5. In a new flow apply inverse filter to the dataset. Explain the results.

    Purpose of this flow is to convolve Ricker wavelet (wavelet A) with its inverse filter. After the convolutionprocess, the result is spikier than wavelet A. The result has some resemblances to impulse train, except itsspikiness, it has high amplitude in a very short time at the central frequency of wavelet A events.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    28/49

    6. Add 1% of random noise to the dataset and repeat step 4. Explain the results.

    Adding noise to wavelet A resulting in a spikierconvolution result. Comparing convolution results, thequiet zone in ‘no noise’ result clearly appears to bespikier ‘noise added’ result. According to ideal impulsecharacteristic, adding noise in to wavelet resulting inits convolution result (convolution with its inverse) isfar different from an ideal impulse train which weexpected.

    Compare amplitude spectra of theresults : The results from ‘noise free’and ‘noise added’ are very similar.The example of a small differencecan be seen at the arrow point.Since noise is random, it containsenergy in every frequency evenlyand its energy in a particularfrequency is relatively low comparedto the signal. Hence, adding noise

    into the signal before deconvolution,the effect in changing amplitudespectra in the deconvolution resultis very low.

    7. Repeat steps 3-6 varying length of the inverse filter (try values 50, 100, 200).

    8. Create display comparing datasets , and result of deconvolution, explore changes in the amplitude spectrum

    (using Interactive Spectral Analysis).

    Applying the filter with different lengths (50, 100, 200 ms) to “A” dataset with noise free .From the display, we can clear see that convolution wavelet A with a longer inverse filter yields the betterresult e.g. more similar to impulse train which we expected.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    29/49

    Applying the filter with different lengths (50, 100, 200 ms) to “A” dataset with noise added .From the display, the longer inverse filter length yields the better noise suppression in convolution process.

    Spectral analysis

    The longer inverse filter length gives the better noise suppression result. With inverse filter length 200ms, noise almost disappears from energy spectra.

    The longer inverse filter length, the less range of energy distribution in convolution result. With inversefilter length 200 ms, energy tends to concentrate between 0 – 50 Hz, while with filter length 100 ms,energy spectral distribute between 0 – 80 Hz.

    Since energy is conserve, concentrate energy at the useful frequency should enhance the desire signal.

    Predictive deconvolutionCreate separate flow for the predictive deconvolution exercise. It should consist of following routines:Note: Additive Noise and Spikes routine will be used to add up to 5% of random noise:

    1. Using ‘Synthetic Trc Generation’ generate several (~30) traces with model of primary reflections :

  • 8/19/2019 Seismic data processing workbook(ProMax)

    30/49

    Events list: 2,100,10,0.5,300,0,60/ 2,400,10,-1,300,0,60/ 2,500,10,1,300,0,60/ 2,700,10,1,300,0,60/Save it to a dedicated dataset.

    2. Generate synthetic data containing model of primary reflections plus model of peg-leg multiples. To do thischange events list to:2,100,10,0.5,300,0,60/ 2,140,10, 0.25 ,300,0,60 /2,400,10,-1,300,0,60 / 2,440,10,-0.5,300,0,60 /2,500,10,1,300,0,60/ 2,540,10,0.5,300,0,60 / 2,700,10,1,300,0,60/ 2,740,10,0.5,300,0,60 /

    I.e. add model of multiples as half-amplitude primaries, shifted down by 40 ms (left display – primaries, right – primaries and multiples):

    3. Perform predictive deconvolution using known(40 ms) prediction interval and varying operatorlength (10, 20, 40, 60, 80, 160 ms):

    Determine optimum operator length and compare results of predictive deconvolution to primaries-onlysynthetic dataset.The operator length 40 ms is selected since it is optimum parameter to suppress multiples while in othersoperator length, the presence of multiples is obvious even though they are shifted to the later time as show inthe figure below.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    31/49

    Evaluate influence of white noise level parameter. Explain your observations.

    White noise level influences the ability of deconvolution to suppress multiples. As seen in the noise freesimulation, after deconvolution, we can scarcely see the presence of multiples. Noise added simulation,however, we still can clearly see the presence of multiples (in the black rectangulars frame). In summary, thelevel of noise decreases the ability of deconvolution to suppress multiples.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    32/49

    Workshops 6 and 7. Velocity analysis, NMO and stacking(Given information)The CDP is a powerful technique for suppression of different noises on reflection seismic data, both random andcoherent, having kinematic characteristics being different from those of primary reflected waves. In order to apply thistechnique one should:1. Combine traces in CDP gathers2. Apply normal move out correction (NMO), i.e. stretch each trace in order to convert it to zero-offset using the factwhat travel time curves for reflected waves can be approximated with hyperbola:

    ( ,) = √2 + 22 (), ≈ ( 1 3. Stack all traces in each CDP gather.

    Stacking velocity = (, ̅) can vary with depth (however it is more convenient to parameterise itusing -time of arrival of reflected wave on a zero-offset trace) and CDP coordinates and can be estimated from theactual data using Velocity analysis .

    Next several workshops are aimed to show how this approach works and will consist of three steps:1. NMO + Brut stack. Here we will sort input data in CDP:OFFSET order, apply NMO with just one velocity function andstack them to produce so called brute stack (i.e. preliminary, very draft stack) .

    2. Stacking velocity analysis.3. NMO(using estimated velocities) + stack

    Part 1. NMO+Brut stack1. Create a new flow, 060 – Brute stack and apply NMO with single velocity to several ensembles and analyse the effect ofNMO to data.

    In DDI sort data in CDP: AOFFSET order, include several CDP ensembles from full-fold interval, for instance, use each 200 thCDP gather from 500 to 1500. Trace display should be in WT/VA mode. Initially make NMO routine inactive. Execute theflow.

    Trace display

    Make NMO routine active, edit parameters as shown in the figure below. It will apply NMO with velocity as in water andwithout stretch mute.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    33/49

    (CDP1:T1-V1,T2,V2/)

    Execute the flow, explain your observations.What happened to: 1) water bottom reflection, 2) sub bottom reflections, 3) water-bottom multiples?All of the 1, 2 and 3 are stretched. However the levels of stretching are different depend on a velocity of a particularwave. Water bottom reflection and water-bottom multiples have velocity approximately 1500 m/s, hence it almostperfectly stretched after NMO correction. Sub-bottom reflections, on the other hand, have various velocity (>1500 m/s) ,hence they are under stretched after NMO correction.

    Execute the same flow with single velocity being equal to 1300 m/s, 1400 m/s, 1550 m/s, 1600 m/s, explain observations.

    Overcorrection, estimated V NMO is too low.

    Overcorrection, estimated V NMO is too low.

    Undercorrection, estimated V NMO is too high.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    34/49

    Undercorrection, estimated V NMO is too high.

    Observations: In NMO correction, it uses estimate velocity of a particular reflected wave to calculate its travel time.When V NMO is lower than the real reflected wave velocity, the result in NMO correction appears to be overcorrected whilewhen V NMO is higher than the real reflected wave velocity, the result in NMO correction appears to be undercorrected.

    2. Derive single depth/time varying velocity function . To do this in the same processing flow make NMO routine inactive,execute it and approximate several primary reflections with hyperbolas as it is shown in the figure below. This will allowyou to obtain several pairs T 0- VRMS .

    (T0 - VRMS); 1:0-1500,1392-1500,2112-1762,2556-1941,10000-2000/Note that we could not see any primary reflections below multiples clearly.

    Apply NMO with single time-variable velocity function and no stretch mute.

    How the result is different from those cases then just single value of velocity was used for all T 0 values? Find how signal isbeing stretched on large offsets.

    Using single velocity in NMO correction is unable stretch all the events with different velocity perfectly (mostlyover/understretched). While using a proper velocity for a particular event (as shown in the figure below) improves abilityof NMO correction to stretch that particular event specifically. For an optimum NMO correction, all of the target eventsneed to be assigned correct Tn and Vn in order to increase they S/N in stacking process.

    The signals at large offset increase in their frequency. They appear to have longer wavelengths than before NMOstretching

    Change stretch mute to 30% and execute the flow again. What has changed?

  • 8/19/2019 Seismic data processing workbook(ProMax)

    35/49

    Comment: NMO stretching contributes to frequency distortion in shallow features at far offsets (events are shifted to

    lower frequency). Stretching is quantified that ∆ = ∆ . Stretch wave form at large offsets can damage shallowfeatures in the data. However the problem can be solved by muting the stretched zones in the gather.

    Observations: No stretch allows all the data to be appear on the screen. 30% stretching allows only ∆ < 30% to beshown while 80% stretching allows only ∆ < 80% to be shown. Since ∆ increases with offset, the more percentage ofmuting is the more data to be shown.3. Apply NMO to whole pre-processed dataset and produce brut stackChange parameters of DDI to bring all CDP gathers to the processing flow; add routine CDP/Ensemble Stack and DDO

    after NMO:

    Trace display

    Compare brut stack with single (first) channel record (use pre-processed data) .

  • 8/19/2019 Seismic data processing workbook(ProMax)

    36/49

    Are there any improvements in S/N?After stacking, the display shows that there is a significant improvement in S/N compared to pre-processed data. We cansee stronger reflection events after the stacking.

    Recompute brut stack (but do not save it to disk) using just single velocity value (1500 m/s).What is the difference between these two brut stacks?In brute stack 1500 m/s, CDP gather shows strong reflector event only at the reflected waves V~1500 m/s as circle in thepicture below. Brute stack, however, shows strong reflector events at various reflected wave velocity e.g. 1500, 1762 and1941. Expectedly, if we input the right velocity for a particular reflected wave, its stacking subsequently improved in S/N

    and showed as stronger reflector CDP gathers display.

    Part 2. Stacking velocity analysisStacking velocity analysis (VA) is a parameterisation of travel time curves of reflected waves using usually two parameters (some additional parameters can be used in case of presence of anisotropy).Traditional (sometimes called ‘vertical’) velocity analysis allows to estimate = ( ̅) in a several discreetpoints along the line. To apply NMO for each trace stacking velocity is needed for the CDP gather to which this tracebelongs. If no stacking velocity function is defined for it, it will be generated by interpolation using nearest CDPs for whichvelocity analysis was performed. I.e. you should select velocity analysis step along the line small enough to describelateral velocity variation .

    Quite often signal/noise ratio for some weak reflection (especially in a lower part of the section) is poor. In order toobtain robust estimation of stacking velocities several dedicated workflow improvements are being used.

    1. Seismic data can be prepared separately for the purposes of velocity analysis and for obtaining optimum stackedsections. In first case the main almost only signal/noise ratio if the key factor while in the second we might want toachieve some specific goals like improve vertical resolution or preserve amplitudes. This means that to run the VA ismight be useful to create separate dataset with some routines which can improve S/N in exchange for resolution (narrow

    bandpass filter) and true amplitudes (AGC) applied.2. Signal/noise ratio can be improved further if we use several neighbouring CDPs together to form so calledsupergathers . This is relatively seldom method while producing stacked section as it affects lateral resolution (it smearsstack), but for purposes of the VA it is the default approach.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    37/49

    Finally, computation of velocity spectra and constant velocity stacks is a relatively slow process. In order to make thevelocity analysis being a user-friendly routine in ProMAX velocity analysis requires precompute step.

    1. Stacking velocity analysis1. Create separate flow to run velocity analysis ( 070 – Velocity analysis ). Flow should consist of following routines:

    Parameters should be defined similar to examples shown below.Super gather:

    Velocity analysis:

    VA precompute:

    Pick velocity functions for each super gather; store them to a dedicated velocity table.

    Note: We don’t need to pick multiples velocity because we want to remove it in stacking process. Hence Vrms isextrapolated in the time where there are the presences of multiples.

    2. NMO (using estimated velocities) + stackIn a new flow, similar to the one used to compute the brute stack,1. Apply NMO using to several CDP gathers using derived velocity table.What happened to: 1) water bottom reflection, 2) sub bottom reflections, 3) water-bottom multiples?

    Both water bottom reflection and sub bottom reflections are almost perfectly stretch, however, water-bottommultiples are undercorrected. Since we extrapolate the velocities after the last sub-bottom reflection by assumingthat the velocity of reflected wave increases with depth but in case of multiples it do not follow this assumption,resulting in undercorrected as shown in the figure below.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    38/49

    2. Compute stack section using velocity table obtained on previous stage, and compare this stack to the brute stack. S/N of water bottom reflection is almost the same compared to brute stack 1500 m/s due to we use the proper

    velocity for water bottom reflection in NMO correction before staking. However, if we use higher or lowervelocity than the real water bottom reflection velocity in NMO correction, S/N in water bottom reflection of thatbrute stack should be significantly low comparing to NMO correction using estimated velocity + stacking.

    S/N of sub-bottom reflections are clearly improved owing to we use estimated velocities in NMO correctionbefore stacking.

    Water bottom multiples are re moved after stacking because we didn’t estimate their T0 and Vrms in NMO

    correction. So when stacking, these signals are cancelled. The results are shown below.

    3. Velocity analysis - QCCreate new flow; name it ‘080 – Velocity viewer’. This flow shou ld consist of only one routine:Velocity Viewer/Point Editor

    After correction4. Velocity manipulationProMAX routine Velocity Manipulation can be used to perform various transformations with velocity tables, such astransformation of stacking velocity to interval velocities, smoothing of the velocity field, etc.Use this routine to generate single average RMS velocity function.

    This function can be a good guide function for next iterations of the velocity analysis.

    Comment: After execute the flow, this function will compute Vrms of each layer, resulting in constant velocity in eachparticular layer. In addition, it also smoothen the velocity curve as shown below.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    39/49

    Workshops 8: FK filter application

    We continue processing of the real 2D marine line. Here we assume that pre-processed data aswell as initial results of velocity analysis were prepared during previous workshops.The goal is to compare two different ways of multiples suppression: radon and FK filtering.- Perform FK analysis

    - Define mute polygon

    - Stack data with FK filter applied

    - Compare with stack data with radon filter applied.

    1. Velocity manipulationUse this routine to generate single RMS velocity function which will contain average table withstacking velocities decreased by 25%.As input velocity table use the best possible velocities obtained during the last velocity analysis

    2. FK analysisCreate dedicated flow to perform FK analysis ( 100 – FK analysis ).We use common source gathers.Read and display pre-processed source gathers, apply NMO with single decreased RMS velocityfunction v1_75. Zero stretch mute should be used.Introduce routine FK Analysis to the flow:

    Execute the analysis; pick the mute function.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    40/49

    3. FK filter applicationChange the flow as shown below:

  • 8/19/2019 Seismic data processing workbook(ProMax)

    41/49

    4. CDP stack after FK filteringCompute stack after FK filter , compare it with previously computed stack sections after velocityanalysis and radon filtering.

    Flow for staking data after F-K filtering

    Comment: After using FK-filter before stacking, the frequencies inside the selected triangle in FK domain areremoved. Hence when we stack the data after F-K domain, we are unable to see that frequency contents in thedata display. In addition, the removed frequencies are belong to multiples, so that multiples can be removed byF-K filter.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    42/49

    Workshops 9: Demultiple in Tau-P domain

    One of standard demultiple approaches is the one based on parabolic τ− p transform.Generic workflow for it provided below:1. Estimate stacking velocities V NMO(x, T0)2. Decrease V NMO velocities by several percent. New stacking velocities should overcorrect primary reflections, butundercorrect multiples.3. Using parabolic tau-p transform model multiples and subtract them from initial wavefield.

    4. Repeat velocity analysis using results of de-multiple workflow and update stack section.

    We continue processing of the real 2D marine line. Here we assume that pre-processed data as well as initial results ofvelocity analysis were prepared during previous workshops.

    1. Velocity manipulationProMAX routine Velocity Manipulation can be used to perform various transformations with velocity tables, such astransformation of stacking velocity to interval velocities, smoothing of the velocity field, etc.Use this routine to generate single RMS velocity function which will contain average table with stacking velocitiesdecreased by 5-10%.

    2. Interactive tau-pi analysisCreate dedicated flow to perform tau-p demultiple ( 100 – Tau-P demultiple ).

    Initially we use it to do interactive tau-p analysis and define parameters of the muting in parabolic tau-p domain whichwill separate primaries and multiples. It is possible to do tau-p demultiple using different gathers, this time we usecommon source gathers.

    Read and display each 40-th pre-processed source gather, apply NMO with single average RMS velocity function andsingle function with decreased velocities (prepared at the previous step). Compare the results.Note: Since the velocities are decreased by 10%, the primary reflections are overcorrected while multiples are remainundercorrected.

    Note what in the demultiple flow purposes we should avoid using stretch mute :

  • 8/19/2019 Seismic data processing workbook(ProMax)

    43/49

    Introduce routine Interactive Radon/Tau-P Analysis to the flow:Parameters of the Tau-P analysis routine:

    Note: mute (in this case – top mute) should use two header values, the primary should be SIN, the secondary must beuser-defined header Moveout .Execute the analysis; pick the top mute function for all selected common source gathers. Comment: After picking and muting primaries, the program performs inverse of radon for multiples into T-X domain, andthen multiples are subtracted in T-X domain resulting in multiples are removed from T-X domain display as shown below.

    3. Application of Tau-P demultipleChange the flow as shown below:Parameters of the Radon Filter should mimic those of Interactive Radon/Tau-P Analysis :

    Execute the flow, compare results of tau-p demultiple to the input data.Comment: After tau-p demultiple is computed, multiples are suppressed. The presence of multiples is hardly recognizedwhile multiples in an input data are obviously shown.

    5. Velocity analysis, NMO and stack1. Repeat velocity analysis using result of the demultiple procedure as an input. Velocity analysis step should be 50 CDPs ,use CDP range of 200-1600.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    44/49

    2. Compare velocity spectrum obtained after demultiple to the velocity spectrum before the procedure.Explain your observations.

    Observation: In the velocity spectra of input data (beforeRadon filter is applied), there is a discontinuity of velocitybetween primaries and multiples while the velocity spectraafter using Radon filter shows a continuity throughout all ofthe spectra. However, the velocity spectra after Radonfilter in the late time are scattered, unlike the input datawhich multiple velocities are concentrated in one line.

    3. QC velocity field, repeat velocity analysis if necessary.

    Comment: There are discontinuities in velocities of Tau-P data demultiple data compared to processed data. To make itsmoother (more geological possible), its velocities are adjusted in velocity editor flow. The result is shown below.

    4. Compute stack after tau-p demultiple , compare to previously computed stack sections.Comment: Even though stacking by usingvelocity analysis provides a good multiplessuppression, tau-p demultiples and F-Kfilter show superior ability in multiplessuppression as show in the display. It’squite hard to distinguish the final resultsbetween stacking data after tau-P and F-Kfilter, however, when looking closely tau-pdata shows stronger primaries than F-Kfilter data. In addition, tau-p performsgood multiples reduction while preservingfrequencies of primaries. In my opinion, I

    would select tau-p in multiples removal forthis dataset.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    45/49

    Workshop 10: Migration

    Migration is the process of placing seismic reflection energy in its proper subsurface location. It can beconsidered as a dip-distortion correction made by moving reflection points away from their positions onvertical lines, onto inclined lines that correspond to travel paths. Migration operators can be applied in both thetime and space domains. There are many formulations of Time migration:- Stolt migration,

    - Kirchhoff migration,- Finite-difference migration.

    We continue processing of the real 2D marine line. Here we assume that we got the best possible stack as wellas the best possible velocity field during previous workshops.The best possible stack in our case means the stack with multiples suppressed .The best possible velocity field means the smooth velocities obtained after multiples suppression.

    Memory Stolt FK MigrationThis algorithm uses Stolt’s stretching technique to account for vertical and lateral velocity variations. It iscomputationally efficient and is very accurate for constant velocities , but has difficulties imaging steep dips inareas where there are large vertical or lateral velocity variations.

    We use this routine to migrate stacked data.1. Create dedicated flow to perform Time Migration 200 – Migration.

    We first migrate the data with the slightly incorrect velocities - over and underestimated. We analyse results.Then we improve the velocity field by using the real velocities instead of over and underestimated ones.The purpose of this exercise is to compare the results and tell the difference between them. How velocitiesaffect the data?

    2. First, obtain single rms velocity function from the best possible velocities. Use Velocity manipulation routine.It provides a lateral average of velocities from your best velocity field. Save it in a new velocity table “single velfor migration”. It will be used in Memory Stolt FK Migration tool

    3. Migrate data with underestimated by 10 and 20%, correct and overestimated by 10 and 20% velocities .Percent velocity scale factor can be changed in Memory Stolt FK Migration routine.Parameters for Migration should be as follows:

    Obtain 5 migrated stacks changing “Percent velocity scale factor” from 80% to 120% with a step of 10%.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    46/49

    Display the migrated stacks in Greyscale mode . Compare it to the stack before migration . In your case it will bestack after radon filter. Note what’s happened with diffractions . Describe the difference.Which one gives the best migration results?

    In Radon stack, we canclearly see diffractions inCDP gathers while afterstolt F-K migration using100% Vrms, diffractionsare reduced. Howeversome of reflectors lookwavy which mayregardless of geology butfrom the mirgrationitselves.

    In Stolt F-K migration withunderestimatedvelocities, we can still seethe traces of diffractions.However 10%underestimated velocityhas a better ability indiffractions removal

    compared to 20%underestimated velocity.

    With overestimatedvelocity migrations,diffractions are almost

    gone, however, thereflectors are too wavythan they should be.

    Comment: In my opinion, Stolt F-K migration with 100% Vrms gives the best result because it removesdiffractions and damages reflectors appearance in a lower degree compared to those overestimated velocities.The reason that Stolt F-K does not give a very good migrations result in this data set may attributes to variationof the velocity and geological complexity. We can see that most of wavy reflectors occur at areas near the fault.In migration process, the accuracy of velocity directly related to migration result, the best possible velocitiesshould be used in order to get optimum migration result.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    47/49

    Kirchhoff Time MigrationThis type of migration performs a time migration by way of summing of traces over hyperbola calculated from avertically and laterally variant RMS velocity field in time. This type of migration is generally used for smoothlyvarying velocity fields with moderately complex structure.1. Introduce Poststack Kirchhoff 2D Time Migration routine to the flow.

    Note: If repeat running is required, ‘Run type’ should be changed to ‘Normal (overwrite)’.

    2. Execute the flow and compare the result with the previous one (Memory Stolt FK Migration - used with singlevelocity function obtained from real data ).

    Comment: Kirchhoff migration significantly improves diffractions around the fault (in the red circle), so we cansee the shape of the fault clearer than in Stolt F-K migration. In addition, since Kirchhoff migration allowsvelocity variations and moderate geological complexity, the wavy reflectors as seen in Stolt F-K migrations areimproved to be more geological possible. Despite the edged of Kirchhoff migration result is stretched, it doesnot affect appearance the dataset. However, there are some traces of diffractions as pointing by yellow arrows.

    Finite-Difference Time MigrationThe finite-difference migration can be calculated using Fast Explicit FD Time Migration . It migrates stackeddata using a modification of explicit finite-difference extrapolators . FD migration can handle fully variableinterval velocity fields in time , and likewise handles well moderate dips .1. Introduce Fast Explicit FD Time Migration routine to the flow.Note that we are using interval velocities in time!2. Use Velocity Manipulation tool to convert RMS velocities to the interval velocities in time. We are usingsingle average velocity table.

  • 8/19/2019 Seismic data processing workbook(ProMax)

    48/49

    3. Execute flow and compare the result with the stack before migration and with the migrated stacks obtainedwith Memory Stolt FK Migration and Poststack Kirchhoff 2D Time Migration.Comment: In the stack before migration, we can clearly see diffraction events. From the display below, FD timemigration perform better diffractions removal over Kirchhoff 2D time migration and FD time migration, hencewe can see reflectors and the presence of the fault clearer after FD timmigration without interferences ofdiffractions. However, with the same dataset, FD time migration takes the longest computing time.

    Finite-Difference Depth MigrationFinite-difference Migration can also be performed in the depth domain using Explicit FD Depth Migration . It isalso good at handling vertically and laterally variant velocities and relatively steep dips.1. Introduce Explicit FD Depth Migration routine to the flow. Parameters of Explicit FD Depth Migration:Note that this time we are using interval velocities in DEPTH !

    2. Use Velocity Manipulation tool to convert RMS velocities to the interval velocities in depth (describedabove).

    3. Execute flow and analyse the result. Compare the result with the stack before migration and with thepreviously migrated stacks . What is the main difference? Can we actually compare these stacks?The main different between the result after FD depth migration and others is that the domain of data (FD depthmigration is in depth domain, others are in time domain). Hence we are unable to compare the results indifferent domains.

    (CDP Gathers in Depth domain)

  • 8/19/2019 Seismic data processing workbook(ProMax)

    49/49

    4. In order to be able to compare the dataset obtained in depth domain to the ones obtained in time domainwe need to perform depth to time conversion.Create dedicated flow 300 – Time/Depth Conversion.Introduce Time/Depth Conversion routine to the flow. Time/Depth Conversion parameters:

    5. Execute flow. Compare the result with the stack before migration and with the migrated stacks obtained withMemory Stolt FK Migration , Poststack Kirchhoff 2D Time Migration and Fast Explicit FD Time Migration.

    Comment: In Radon stack data (before migration), diffractions obscure the presence of the fault. Aftermigration, diffraction events are suppressed, so that we can see the fault and reflectors clearer. Stolt FKmigration caused wavy reflectors in this dataset which is undesirable effect. Kirchhoff 2D has moderately goodperformance in diffractions removal, however, some diffractions still remain. FD time migration gives betterresult in diffractions removal compared to those mentioned method (Stolt FK and Kirchhoff). However FD depthmigration, perform the best diffraction removal while preserving reflector plains.

    In order to obtain good migration result, the best possible velocities are required. Moreover we have tohave some ideas about geological complexity and velocity variations to select the most appropriate migrationmethod according its constrain.