Multivariate Analyses using Geologic Information from Cores (MAGIC).CONTENTSMAGIC PROJECT OVERVIEW
|
![]() | Discriminant Analysis |
![]() | Principle Components |
![]() | Cluster Analysis |
The literature provides several examples of discriminant analysis (Crowe & Dodman, 1989), principal components (Mulhern, et. al., 1988) and cluster analysis (Anxionnaz, et. al., 1990). All of these procedures for lithofacies identification will give roughly similar results given similar input data.
The Multivariate Cluster analysis program known as MAGIC has been applied to 364 wells in Southwest Alabama as part of a regional study of the Smackover Formation. MAGIC uses the data from wireline logs (360,000 feet of neutron, density, sonic, gamma ray and deep resistivity data) along with core data (47,000 feet of lithologically described and analyzed cores) to predict lithologies. The availability of modern vintage (i.e., post 1970) cored and logged wells in the study area is exceptional from a data standpoint. The density of this core and log information has provided experimental verification of the multivariate technique.
MAGIC is written using Statistical Analysis System (SAS) language. The statistical results of MAGIC log analysis are saved to external SAS datasets for further analysis and to provide input to other software packages for stratigraphic study and mapping.
Detailed predicted lithologies derived from MAGIC in the SAS-maintained datasets are used to identify favorable stratigraphic reservoir trends in complex carbonate and evaporite sequences.
The following benefits are realized from using the MAGIC processing technique:
![]() | The combination of log-derived lithology, core descriptions and eight composite log curves on each permanent hard copy computer-generated color geophysical log simplifies lithofacies identifications. The detail in which the representation of lithologic data is reproduced greatly improves difficult stratigraphic correlations. | ||||||
![]() | Statistical examination of log information does the following: | ||||||
![]() | Identifies bad hole and other inaccurate curve values before assignment of lithology using multivariate analysis. | ||||||
![]() | Generates a missing curve or regenerates bad data in a curve by statistical inferences
to a digital log database. For example, the following list represents the program's
capability:
| ||||||
![]() | The prediction of porosity and permeability values from log data are related to core analyses by multivariate statistical regressions. | ||||||
![]() | Multiple well processing is supported by maintaining a database of all log and core data including porosity and permeability with lithologic descriptions, and locations of all wells analyzed. |
![]() | Simplifies a combined stratigraphic analysis of both subsurface geological rock information and geophysical seismic information: |
![]() | Supports seismic modeling by generating a sonic log for synthetic seismograms and stratigraphic simulations if one was not run in a key well. |
![]() | Improves the identification of lithologies to interpret seismic information for search of subtle stratigraphic traps. |
![]() | Digital geophysical log data are stored in a database for mapping and future advances in
processing technology. |
Color MAGIC logs have been used to develop time-correlative sequence events in an area with no distinctive fossils or marker bed boundaries. The MAGIC logs are further integrated with core thin sections, field and production data, and map analysis
Multivariate log and core statistical analysis and seismic data provided a comprehensive dictionary of Smackover data in Mississippi and Alabama. Since much of these data are digital, many tentative (what if ?) correlations can be made in short order.
The traditional method of using chart book cross-plots is not suitable for determining correct rock petrology in areas of complex lithology. Statistical treatment of well log information improves lithologic predictions.
Procedures based on discriminant analysis use a truth table of log responses and core lithology. Unfortunately, dolomites and compacting shales have such a wide range of log values which makes the precise lithofacies identification of these rock types hard to perform.
Schlumberger Faciolog and Geocolumn (Mulhern, et. al., 1986) procedures are based on principal component analysis. Principal components are manually grouped into local facies nodes which determine lithofacies identifications. These are incorporated into their CLIFF (Core-Like Interpretation of Facies Features, Anxionnaz, 1990) experimental system. The results from these analyses cannot easily do multiwell population studies or provide quantitative exploration mapping of lithofacies distributions.
![]() | Requires the use of Schlumberger's ELAN (Elemental Log Analysis) minicomputer system or RockClass software. |
![]() | Cannot easily do population studies or exploration mapping of lithofacies distributions. |
MAGIC is based on SAS code. As such it is an integrated system which runs on any system (PC, VAX, Main Frame, UNIX, and APPLE computers) and talks to many types of databases (Oracle, M204, VSAM, DB2, dBase, Sybase, Access, and others). The built in graphics, statistics, utilities, and database input/output allow rapid end user access to data analysis.
The Smackover Regional project began January, 1988 to investigate the use of well log derived lithology data integrated with core information. Very little previous work existed in a form that allowed us to add additional lithologic data in a quantified manner. An accurate depiction of complex carbonate lithofacies is required for detection of subtle hydrocarbon traps in the Smackover Formation.
A cursory examination of all well data resources in Alabama shows that:
![]() | 95% of all log data (1,200 wells) in the area was obtained after 1970. A large majority of these wells have modern log suites. |
![]() | There are about 600 wells which cored the Smackover and Norphlet formations. |
The large percentage of core and modern log information has provided the statistical base needed to prove the validity of multivariate lithology identification technique. Chevron has about 230 wells with extensive log suites of neutron, density, sonic data in the area. These data are available in technical well files.
A very important part of our study is based on the existence of thousands of thin sections archived at the Alabama State Geological Survey. The State of Alabama requires all core and logs taken from wells drilled within state borders be deposited at the Survey after a period of time. These core and core-chip specimens are available for public use. The only operational requirement is that a copy of the study results are given to the State. Many thin-sections made from core materials have accumulated over the years from all different sources, both industry and academic, and may be examined at the Alabama Geological Survey.
We have micro-photographed about 800 thin-sections from the Alabama collection. These photomicrographs are used to determine lithological facies and make interpretations on original environments of deposition. MAGIC logs helped us to determine what the rocks are now, the thin-sections help us to determine what the rocks were during deposition.
After analyzing the results from 364 wells, 190 of them with cores, 12 of them by personal description of the cores, and 35 of them with extensive thin-sections; Chevron now has a Smackover database containing 360,000 feet of normalized, depth adjusted induction, gamma ray, spontaneous potential, neutron porosity, density, sonic and caliper curves with 47,000 feet of core description plus core porosity and permeability data.
This use of statistical log and core analyses has resulted in a significant improvement in understanding Smackover depositional processes as compared to prior in-house studies and other reports. This improvement in our understanding of Smackover depositional environments has allowed selection of many additional areas for hydrocarbon prospecting in Mississippi and SW Alabama.
Figure 1. Old MAGIC Results Using All Core Descriptions.
CORE DESCRIPTIONS MOSTLY FROM COMMERCIAL LABS
CORE LITHOLOGIES | ||||||||
|
ANH | DOL | LS | SH | SS | SALT | TOTAL | % AGREE |
ANH | 1,450 | 178 | 66 | 6 | 4 | . | 1,704 | 85.1 |
DOL | 229 | 5,396 | 1,855 | 65 | 384 | . | 7,929 | 68.1 |
LS | 170 | 1192 | 4,877 | 17 | 160 | . | 6,416 | 76.0 |
SALT | . | 1 | . | . | 9 | 12 | 22 | 54.5 |
SH | 11 | 66 | 13 | 29 | 38 | . | 157 | 18.5 |
SS | 7 | 145 | 320 | 13 | 1,987 | . | 2,472 | 80.4 |
TOTAL | 1,867 | 6,978 | 7,131 | 130 | 2,582 | 12 | 18,700 | |
% AGREE | 77.7 | 77.3 | 67.8 | 22.3 | 77.0 | 100.0 |
OVERALL AGREEMENT LOG TO CORE LITHOLOGY = 73.5%
Figure 2. Old MAGIC Results Using Experienced Geologist Core Descriptions.
CORE LITHOLOGIES | ||||||||
|
ANH | DOL | LS | SH | SS | SALT | TOTAL | % AGREE |
ANH | 152 | 42 | 18 | 5 | . | . | 217 | 70.0 |
DOL | 7 | 537 | 116 | . | 69 | . | 729 | 73.7 |
LS | 24 | 251 | 803 | . | 34 2 | . . | 1,112 | 72.2 |
SALT | 6 | 6 | 0.0 | |||||
SH | . | 5 | 1 | . | 11 | . | 17 | 0.0 |
SS | . | 24 | 4 | . | 566 | . | 594 | 95.3 |
TOTAL | 183 | 859 | 942 | 5 | 686 | . | 2,675 | |
% AGREE | 83.1 . | 62.5 | 85.2 | 0.0 | 82.5 |
OVERALL AGREEMENT LOG TO CORE LITHOLOGY = 76.9%
MAGIC log data processing consists of reading log and core data files from datasets containing the project files. SAS reads in project file log data and carries out the following steps:
![]() | Finds and statistically reconstructs bad or missing log curve data. |
![]() | Performs multivariate analysis using clustering techniques. |
![]() | Merges clusters with cluster-lithology key table. |
![]() | Writes output needed for MAGIC log. |
![]() | Saves log and core data with statistics to a SAS database. This resultant database is
used for: Multiwell processing of log, core and lithology information for Exploration mapping. Performing core to log regressions for porosity and permeability. Provides statistical database for restoring 'bad' log data. |
Usually, much of the 'bad' log data are identified using cut-off lines on neutron-density, neutron-sonic, and sonic-density cross-plot tables. Cut-offs for 'bad' log curve data recognized during this study are as follows:
ACCEPTED DATA RANGES FOR CURVES USED IN CLUSTERING
CURVE | Low | High | Units | |
Deep Resistivity | .02 | to | 2000. | MOMH/M |
Neutron Porosity | -5.0 | to | 60. | PHI UNITS |
Sonic Travel Time | 40.0 | to | 140. | MICROSEC/FT |
Bulk Density | 1.74 | to | 3.1 | GM/CC |
Gamma Ray | 1.0 | to | 300. | API UNITS |
CROSS-PLOT DATA CUT-OFFS
CURVES | Cross Plot Cut-off Values |
SONIC - DENSITY | |
High DRHO<(-.01400*DT)+3.230 | |
Low DRHO>(-.01330*DT)+3.770 | |
DENSITY - NEUTRON | |
High DRHO<(-0.01640*NPHI)+2.480 | |
Low(1) DRHO>(-0.00857*NPHI)+3.100 | |
Low(2) DRHO>(-0.02140*NPHI)+3.330 | |
SONIC - NEUTRON | |
High DT>(1.3300*NPHI)+65 | |
Low(1) DT<(1.8330*NPHI)+4.2 | |
Low(2) DT<(0.3750*NPHI)+40 |
The caliper curve was not used to determine 'bad' data. Many MAGIC logs have
demonstrated that hole size along with hole rugosity has a great effect on density pad
contact with the borehole surface. This generally shows up as low values of bulk density
readings along with curve correction values in the high
minus range. In this particular case density curve values may be invalid (i.e. 'bad') even
though the sonic, neutron and induction curves show valid rock values. By using heuristic
logic developed during the Smackover Study, offending curve(s) are set to null values for
statistical curve reconstruction.
SAS code member DEADTR executes a statistical reconstruction of all 'bad' or missing data by separating missing data into three temporary SAS data files; no-density, no-sonic, and no-neutron. For example, the no-sonic file has valid values for density, neutron, and deep induction curves with a missing sonic value. The no-density and the no-neutron files have valid values for other curves.
Statistical reconstruction of missing data starts by going to the SAS database containing log digits of all logs for all wells previously processed by MAGIC. From this multiwell log file, only 'good' (i.e., no bad or missing log curves) log data is selected. To further window the data to manageable size, only those 'good' data from identical formations in the missing data files are selected for statistical analysis.
For each SAS data file with missing curves, (no-density, no-neutron, and no-sonic cases) data are retrieved from the 'good' log data set by selecting only those values occurring in the same modal class as the 'missing' log data by using an identification (ID) variable:
![]() | Sonic curve ID variable = FORMATION + NPHI + DRHO+ log10(ILD) |
![]() | Density curve ID variable = FORMATION + NPHI + DT + log10(ILD) |
![]() | Neutron curve ID variable = FORMATION + DRHO + DT + log10(ILD) |
The data in the 'good' log data are further reduced by calculation of the means, sums, and frequencies of all data in each good-log-data-set by using a ID variable. This reduces the population of 'good' log data to 1,000-to-10,000 points. This makes the DEADTR statistical routines which determine the estimated value of missing log curves cost effective. Since the ID variable is identical in both the 'bad' and the 'good' data, all good-log-data is merged with all "bad" log-data. The missing curve value is supplied by the 'good' dataset. Insufficient data in any one of the 'good' log sets nullify the DEADTR algorithm.
Two statistical procedures are executed for each case, no-sonic, no-density, and no-neutron:
![]() | SAS procedure General Liner Models (GLM) estimates a missing value by using linear interpolation from the 'good' log data to the 'bad' curve values. Predicted curve results using GLM tend to have the lowest standard deviation. |
![]() | Another estimate is made from the modal value created by using PROC MEANS on the curve's ID variable. This is estimated by using the 'good' log-data-modes for seed values in a FASTCLUS execution. By using no-curve dataset as input to FASTCLUS and 'good' log data modes as cluster seeds, this prediction represents a modal, or most common, value by formation. This prediction is 'lumpy' when compared to most log curves and has a higher standard deviation. |
The value for DEADTR no-curve estimate is calculated by averaging the GLM value with the FASTCLUS modal value to smooth out 'lumpy' modal values and make a more natural appearing curve.
Prior to using multivariate statistics, replacement DEADTR data for bad or missing curve values are merged with original log curve data. The use of member BADLOGS finds all 'bad' data points and sets those values to null values. This avoids clustering of data outliers and limits 'good' log data to the coverage of seed data values in 5 dimensional space. The set up calculation for all the clustering variables is then performed by member MSPEQ.
Member MSPEQ contains log based equations to calculate sonic-density, density-neutron, neutron-sonic xplot porosity and sonic and density matrix values. There are optional NPHI environmental calculations if needed.
All default values for log constants are as follows:
![]() | ![]() |
![]() | ![]() |
![]() | ![]() |
![]() | ![]() |
![]() | ![]() |
![]() | ![]() |
![]() | ![]() |
![]() | ![]() |
- IF
THEN DO;
- PORDA=(RHOM-4.0)/(RHOM-RHOF);
- PORNA=0.76-10**(-(5*NPHI+0.16));
- END;
- IF
DO;
- PORDA=1.0; PORNA=-(2.06*NPHI+1.17)+10**(-(16*NPHI+0.4));
- END;
- NDXPHI=(PORDA*NPHI-DPHI*PORNA)/(PORDA-PORNA);
- IF ABS(NDXPHI)¬>=1.0 THEN RHOMA=(DRHO-ABS(NDXPHI)*RHOF)/(1-ABS(NDXPHI));
- ELSE RHOMA=1.0;
- NPHI=NPHI*100;
- DPHI=DPHI*100;
- NDXPHI=NDXPHI*100;
![]() | ![]() |
![]() | ![]() |
- IF NPHI>SPHI THEN DO;
- PORSA=-.146;
- PORNA=0.5-10**(-(5*NPHI+0.3));
- END;
- IF NPHI<=SPHI THEN DO;
- PORSA=0.5; PORNA=-(.62*NPHI+0.36)+10**(-(18*NPHI+0.92));
- END;
- NSXPHI=(PORSA*NPHI-SPHI*PORNA)/(PORSA-PORNA);
- IF ABS(NSXPHI)¬>=1.0 THEN DTMA=(DT-ABS(NSXPHI)*DELTF)/(1-ABS(NSXPHI));
- NPHI=NPHI*100;
- SPHI=SPHI*100;
- NSXPHI=NSXPHI*100;
- X=((RHOF-RHOM)*(DELTSH-DELTM))-((DELTF-DELTM)*(RHOSH-RHOM));
- SDXPHI=(((DRHO-RHOM)*(DELTSH-DELTM))-((DT-DELTM)* (RHOSH-RHOM)))/X;
- SDXPHI=SDXPHI*100;
This allows a prediction of lithology using the best possible input data to the FASTCLUS procedure.
Data preparation for clustering log data includes a calculation of the actual variables used in the clustering algorithms. The variables derived from log data are as follows:
CLUSTER VARIABLE | LOG CURVE(S) | REASON FOR CALCULATION |
RHOMA | DRHO,DT,NPHI | Apparent grain density is directly related to lithology. |
DTMA | DRHO,DT,NPHI | Apparent sonic travel time is directly related to lithology. |
LOGILD | LOG10(ILD) | The logarithms of the deep resistivity values are normally distributed. |
LOGGAMMA | LOG10(GAMMA) | The logarithms of the gamma ray values are normally distributed. |
MEAN | NPHI,DPHI,SPHI | Mean values of sonic, neutron, and density porosity values are related to rock matrix and porosity. |
STD | NPHI,DPHI,SPHI | Standard deviations of sonic, neutron, and density porosity values are related to the rock matrix. |
The curve weighting factors used to generate clustering variables for FASTCLUS are calculated as follows:
|
|
|
|
||
GD | = | RHOMA | x | 10.00 | 4 |
MT | = | DTMA | x | 0.50 | 5 |
M | = | MEAN | x | 0.04 | 1 |
S | = | STD | x | 0.13 | 2 |
G | = | LOGGAMMA | x | 0.86 | 1 |
I | = | LOGILD | x | .55 | 2 |
* Weighting factor is calculated by dividing relative lithologic contribution by population values of the FASTCLUS variable at the 98% cumulative distribution point minus the value at the 2% cumulative distribution point. RHOMA and DTMA are weighted the most.
The remaining 'good' data are clustered using a disjoint clustering SAS procedure called FASTCLUS. FASTCLUS is used to find 2628 preliminary clusters. These clusters and the output data set are sorted in preparation for later merges with the cluster lithology key table. Seed values are generic for each cluster and have been chosen by using cumulative distribution frequencies of all log curve data.
All 'good' log data are merged with the cluster-lithology key table by cluster. The cluster-lithology key table (With all non-parametric lithologic associations) contains cluster numbers and associated lithology assignments which have been determined by comparison to previous core intervals. The following items pertain to construction and use of the key table:
![]() | Cluster numbers and associated lithologies are the heart and soul of the multivariate technique. Care used in construction of this table will determine the validity of log to core comparisons. |
![]() | Assignments of cluster lithology should be made using core lithology as ground truth. This ground truth assumes correct lithologic identification by geologic personnel. This insures that lithologies identified using log derived clusters are as accurate as possible. Other lithologic data such as sidewall cores, sample logs or ditch cuttings should only be used if conventional core descriptions are unavailable. |
![]() | Log and core data were assembled primarily from a three state area of Mississippi, Alabama, and Florida. The data covers wells drilled in Jurassic carbonate terrain with additional information on sandstones of the Jurassic Norphlet Formation. Additional log and core data from other carbonate terranes in Texas, New Mexico, Wyoming, North Dakota, California, Alaska and North Africa have been analyzed. These results are being analyzed for validity. |
Lithology is assigned by using key petrologic terms which are recognizable by later steps of the MAGIC analysis. These key terms are combined into a single variable called LITH. LITH is composed of a major and (if needed) a minor lithology. For example, a limey dolomite is identified to MAGIC as DOL,LS. Using major-minor pairs of lithologies, it is possible to pair up 92 separate lithologic types using the following basic rock types:
MAJOR LITHOLOGY * | MINOR LITHOLOGY MODIFIER | ||
Anhydrite | (ANH) | Anhydritic | (ANH) |
Basement | (BSMT) ** | ||
Chert | (CHERT) | Cherty | (CHERT) |
Coal | (CARB) | Carbonaceous | (CARB) |
Dolomite | (DOL) | Dolomitic | (DOL) |
Limestone | (LS) | Limey | (LS) |
Salt | (SALT) | Salty | (SALT) |
Sandstone | (SS) | Sandy | (SS) |
Sandy Limestone | (LSSS) | ||
Shale | (SH) | Shaly | (SH) |
Shaly Sandstone | (SSSH) | ||
Shaly Limestone | (LSSH) | ||
Shaly Dolomite | (DOLSH) | ||
Dense | (DENSE) DOL | ||
Good Reservoir | (GOOD) Med Xtal Dol | ||
Moldic Porosity | (MOLD) LS & DOL | ||
Vuggy Porosity | (VUGS) LS & DOL |
* Note - These lithology types are based on lithology symbols available to the plotting package which is used. Additional symbols could be used if more lithology symbols are available.
** If a basement formation call is made in the proprietary marker files or a BSMT call is made in the core files then all lithologies are changed to BSMT.
The major-minor lithologies from logs match major-minor nomenclature in core files.
These major-minor pairs are also identified in member SYMTABLE which contain the symbol and color instructions for preparing log displays with log and core lithofacies.
SAS writes data needed to create a color log copy. The procedure is described as follows:
![]() | Log data are read from flat files and written to the log plotting package as trace data strings. |
![]() | Core data are read and described lithologies are written as trace data strings using binary code. Core files have a provision to allow presentation of grain size if such information is available. |
![]() | Multivariate cluster lithology assignments are written as trace data strings. |
![]() | Color and symbol information are written to the output files. The coding is keyed-off log and core lithologies using SAS members SYMTABLE, CLUSLITH, CORELITH and LEGEND. |
![]() | Well log header information is read from an external database. |
SAS and MSP code books in the References Cited section are needed to get the most out of this documentation. The first assumption made is that the reader is somewhat familiar with SAS.
SAS documentation is especially needed since file maintenance is done using interactive SAS. When using combined SAS file LOG data to do multiple well processing, much information can be gained by simple SAS procedures. For example:
![]() | PROC RSQUARE is needed to generate log to core regressions for both permeability and porosity from a merge of log and core data. The resultant regression is used to write core to log relations for core porosity and permeability. |
![]() | PROC GPLOT is an interactive graphic plot of information in the SAS log files. Plots are needed to see cross-plot relationships of log data in the SAS data set. |
![]() | PROC PRINT can be used to generate lists of well data for future reference. |
MAGIC uses a process of statistically analyzing information contained in data files of logs and cores. There are many external files (SAS MAGIC source code, Project Data, Utility, and MSP JCL) used by MAGIC which need user maintenance. These are:
![]() | SAS code used by MAGIC to read, analyze, and save log, core, and statistics needed to obtain a MAGIC log plot. |
![]() | Input log and core data from project files. |
![]() | Statistically manipulate log and core data. |
![]() | Create and save data needed to plot a MAGIC log. |
![]() | Save log, core, and statistical information to a SAS database for MAGIC missing curve estimation, multiwell processing, and other AD HOC usage. These SAS databases are active and are maintained by the MAGIC program by creating a MAGIC log. |
![]() | SAS code needed by the user to interactively modify and maintain both the SAS data and external project files. |
![]() | SAS code needed by the user to interactively analyze the SAS log and core data. |
Project log data are copied from MSP TRACES or WELLLOG data into a flat file (Eg. EXL.R$MAJIC.AASSCCC.LOGS where AA=state, SS=state code, and CCC=county code) for MAGIC to read.
Project core data are entered from core reports and verbal core descriptions into a flat file (Eg. EXL.R$CORE.MAJIC) for MAGIC to read.
Utility SAS datafiles (Eg. EXL.R$SAS.UTILITY) used by MAGIC for marker name translations and state-county names from API numbers. The general utility data does not change and is not maintained by MAGIC.
Marker date maintained by the user are important since they are used in reconstructing DEADTR intervals and annotation on track 1 of the MAGIC logs. Formation calls are used in MAGIC to annotate the MAGIC log and tag LOG digits for use in multiwell processing and DEADTR reconstructions.
The data is entered using PIDMS or GWDS proprietary NA card records. The study name are the only marker calls recognized by MAGIC. This name has the value of the SAS macro variable &STUDY. This value of the 5 character formation call nmonic is maintained downward from the point of first occurrence until another formation call is encountered. Formation data is downward propagating by depth in the log data.
SAS active data files are SAS datasets maintained by MAGIC. These datasets form a basis for multiwell processing of data. Since they are in SAS format it is necessary to learn certain SAS PROCs such as PROC FSEDIT to do any file maintenance.
![]() | CORES - Data CORES are porosity and permeability data from project flat files. They are replaced every time a MAGIC log is created. This means they are always current as to depth adjustment presented on the latest run of the MAGIC log. | ||||||
![]() | INVENT - Data INVENT (for INVENTORY) contains basic well data recovered from the PIDMS retrieval and all subsequent information on whether or not specific log information is present in project core and log files. | ||||||
![]() | LOGS - Data LOGS are the raw log curve information from project flat files along with statistical results of cluster lithology, bad or missing data, cluster numbers, and formation. Additional data can be created by algorithm if needed (e.g. Sonic - Density Cross Plot Porosity). | ||||||
![]() | DEADTR - Data DEADTR contains information generated by the missing or bad data curve
algorithm are composed of the following datasets:
| ||||||
![]() | SEED - Data SEED is the control file required when using SAS PROC FASTCLUS. It was originally calculated using FAKECLUS (Appendix C) and is based on the cumulative distribution of all log curves used in the multivariate analysis. SEED should not be touched unless level 3 maintenance of SAS code is needed. |
So far, in discussions of MAGIC programs, areas used in the examples are limited to the Haynesville, Smackover, and Norphlet in Mississippi and Alabama. While most of the multivariate observations loaded into member CLUSMKV are in these formations, there is evidence that a significant portion of these observations have world-wide significance. For example, anhydrite precipitated from a Permian sea is identical to anhydrite precipitated from a Jurassic or Cretaceous sea (given similar depths of burial).
Petrologic differences in lithologic rock types of a similar composition may be due to differences in local usage. Truly different rock types should be recognizable as unidentified clusters on a MAGIC log. Currently only half of the 2832 clusters in the FASTCLUS seed table are correlated to sandstone, evaporite, limestone, and dolomite lithologies found in the Jurassic of SW Alabama and Mississippi. The other half of all missing or unidentified clusters probably represent:
![]() | Lithologies yet-to-be identified in cores or sample logs. |
![]() | Non-existent lithologies caused by the introduction of curve errors during logging or digitizing which are not identified as 'bad' log data. |
![]() | Bed boundaries where a boundary separates rocks of different compositions and logs are reading composite curve values from both rock A and rock B. |
Because of the large number of clusters used in FASTCLUS, it may be impractical to start out with a clean cluster key table. Perhaps a better way would be to start out with a copy of CLUSMKV and change cluster assignments in the copy of the cluster-lithology key table as new log and core data show differences. Many of those cluster-lithology identifications now used for the Jurassic of SW Alabama in CLUSMKV may be valid for many areas in the United States.
To examine the feasibility of using MAGIC in new areas, the following need to be determined.
![]() | How many cores have been taken from the subject formation(s) in the area of interest and how many have good porosity, permeability, and lithologic information? |
![]() | Lacking cores, how many wells have sidewall cores or good sample logs? |
![]() | How many wells have modern sonic, neutron, and density geophysical data in the cored
interval? |
Loading of log and core data forms the project database for the MAGIC program. You should be consistent in the internal layout of the project file. It is suggested that naming conventions on the project log files be followed as used in the Smackover Regional project.
All data used in this study were digitized from hardcopy paper records because extensive digital files do not exist in SW Alabama. As a result, not all environmental corrections were applied to log data. Only simple log transformations were performed such as deep resistivity from conductivity, bulk density from density porosity, & etc.
These logs were initially examined by us and each log curve was color coded for digitizing. Color coded logs from each well were digitized and loaded into a WELLLOG database without any corrections, i.e. verbatim copies of each log:
![]() | Spontaneous Potential (SP) Correlation curve in millivolts |
![]() | Gamma Ray (* GAMMA) * API Gamma Ray Units |
![]() | Caliper (CALIP) Hole Size in inches |
![]() | Deep Resistivity (* ILD) * Induction, Laterolog, or Conductivity |
![]() | Density Porosity (*.DPHI) * Limestone Porosity Units |
![]() | Bulk Density (*.RHOB) * Grams/CC (if DPHI not present) |
![]() | Neutron Porosity (* NPHI) Neutron Porosity Units |
![]() | Sonic Travel Time (* DT) Microseconds/foot |
Digital log data generally come from two sources:
![]() | Read from a logging company digital tape made at the time the well is logged. | ||||
![]() | Digitized from hardcopy paper logs using:
|
After log curves are digitized then a replot of the curves are compared with the hardcopy records for quality control. The digital curve data is then loaded as recorded into the WELLLOG database for safekeeping and preservation of raw data.
The project file is loaded from the WELLLOG database by merging the following curves by depth:
GAMMA | Gamma Ray in API Units. |
NPHI | Neutron Porosity in Limestone Phi Units (%). Both Conventional and Sidewall Neutron Curves can be used. |
ILD | Deep Induction or Laterolog in OHM M/M2 |
DT | Sonic Transit Time in MICROSEC/FT |
DRHO | Bulk Density in GM/CC |
DPHI | Density Porosity in Limestone Matrix Phi Units (%) |
CALIP | Caliper in inches. |
SP | Spontaneous Potential in Relative MICROVOLTS |
These project files become the basis for creating MAGIC logs. Curves are normalized and depth adjusted as needed by inspection of the curve log plots. We have found that the following curve adjustments are often necessary. They are (in order of decreasing occurrence) as follows:
![]() | Depth adjustment of one or more curves to the Gamma Ray.
| ||||||
![]() | Amplitude adjustment to the Gamma Ray. Using known lithologies, the gamma ray value is shifted and/or multiplied by a known factor to match other wells. In the Smackover Project this means having a value of 6 to 14 API units in evaporites or clean carbonates and a value of 50 to 90 API units in the eolian Norphlet Formation. An occasional Gamma Ray log in Radium Units needs conversion to API Units. | ||||||
![]() | Environmental correction to the neutron curve. Using known lithologies, the Neutron value is shifted so that it reads 0 in tite, clean limestone, 0 to -0.5 in anhydrite, and -2 in salt. | ||||||
![]() | Adjustment shift in sonic curve. Using known lithologies, the sonic value is shifted so
that it reads no less than:
|
Since the MAGIC program only recognizes log data having a valid reference numbers and depths, the user may write notes in the top of each file in the project file to record changes made to data in each well for future reference.
Core data was handled by keypunch from core reports found in wellfiles or obtained by scouting sources. Core descriptions were added to keypunch files if good proprietary data were found, otherwise contractor core descriptions were used. Keypunch files were concatenated together and contain the following data:
![]() | Reference Number |
![]() | Core depths Depth on core report |
![]() | Depth Adjustment for Core Correction to match core to log |
![]() | Measured Permeability (PERM1) Millidarcy (plug, horizontal) |
![]() | Measured Porosity (PORO) Core Porosity |
![]() | Lithologic Code (MAJOR,MINOR) Standard Lithology Codes |
![]() | Grain Size for Core log Texture |
For the Smackover Regional project, core data was entered by local key punch or individual geologist. Core descriptions were obtained from in-house proprietary documents, personal inspections of core at the Alabama Geological Survey, published geologic literature, or lastly descriptions accompanying the commercial core report. Porosity and permeability values come from the commercial report.
We have found that the following core file adjustments are often necessary. They are (in order of decreasing occurrence) as follows:
![]() | Depth Adjustment. All cores needed depth adjusting to match core depths with log depths.
The measured core porosity is displayed as a curve in track 3 and can be compared directly
to porosity sensitive curves (![]() ![]() ![]() |
![]() | Core Descriptions. Commercial core descriptions as they appear on the core report are
often in error. This information is used to to generate core lithology patterns on the
MAGIC log and to calibrate clusters in the cluster lithology table. Use the best core
description you have and only believe in trustworthy sources. |
Anxionnoz, H., Delfiner, P., and Delhomme, J.P., 1990: Computer-Generated Corelike Descriptions from Open-Hole Logs, AAPG Bull, V. 74, No. 4,p. 375-393
Mulhern, M. E., Laing, J. E., Widdicombe, R. E., Isselhardt, C., and Bowersox, J. R., 1986: Electrofacies Identification of Lithology and Stratigraphic Trap, in GEOBYTE, V. 1, No. 4, p. 48-56
SAS, 1979: Statistical Analysis System (SAS) Users Guide, Version 79.2, SAS, Institute Inc., Raliegh, NC.
SAS, 1982: Statistical Analysis System (SAS) Users Guide-Basics, Version, 82.4, SAS Institute Inc., Raliegh, NC.
SAS, 1985a: Statistical Analysis System (SAS) Users Guide-Basics, Version, 85.0, SAS Institute Inc., Raliegh, NC.
SAS, 1985b: Statistical Analysis System (SAS) Users Guide-Statistics, Version, 85.0, SAS Institute Inc., Raliegh, NC.
SAS, 1985c: Statistical Analysis System (SAS/GRAPH) Users Guide-Graphics, .Version 85.0, SAS Institute Inc., Raliegh, NC.
SAS, 1985d: Statistical Analysis System (SAS/FSP) Users Guide, Version 85.0, SAS Institute Inc., Raliegh, NC.
© 1999, 2000
|