Seismic waves are the vibrations from earthquakes that travel through the Earth; they are recorded on instruments called seismographs. Seismographs record a zig-zag trace that shows the varying amplitude of ground oscillations beneath the instrument. Sensitive seismographs, which greatly magnify these ground motions, can detect strong earthquakes from sources anywhere in the world. The time, locations, and magnitude of an earthquake can be determined from the data recorded by seismograph stations.
The Richter Scale
The Richter magnitude scale was developed in 1935 by Charles F. Richter of the California Institute of Technology as a mathematical device to compare the size of earthquakes. The magnitude of an earthquake is determined from the logarithm of the amplitude of waves recorded by seismographs. Adjustments are included for the variation in the distance between the various seismographs and the epicenter of the earthquakes. On the Richter Scale, magnitude is expressed in whole numbers and decimal fractions. For example, a magnitude 5.3 might be computed for a moderate earthquake, and a strong earthquake might be rated as magnitude 6.3. Because of the logarithmic basis of the scale, each whole number increase in magnitude represents a tenfold increase in measured amplitude; as an estimate of energy, each whole number step in the magnitude scale corresponds to the release of about 31 times more energy than the amount associated with the preceding whole number value.
At first, the Richter Scale could be applied only to the records from instruments of identical manufacture. Now, instruments are carefully calibrated with respect to each other. Thus, magnitude can be computed from the record of any calibrated seismograph.
Earthquakes with magnitude of about 2.0 or less are usually called microearthquakes; they are not commonly felt by people and are generally recorded only on local seismographs. Events with magnitudes of about 4.5 or greater - there are several thousand such shocks annually - are strong enough to be recorded by sensitive seismographs all over the world. Great earthquakes, such as the 1964 Good Friday earthquake in Alaska, have magnitudes of 8.0 or higher. On the average, one earthquake of such size occurs somewhere in the world each year.
The Richter Scale is not commonly used anymore, as it has been replaced by another scale called the moment magnitude scale which is a more accurate measure of the earthquake size.
Modern seismographic systems precisely amplify and record ground motion (typically at periods of between 0.1 and 100 seconds) as a function of time. This amplification and recording as a function of time is the source of instrumental amplitude and arrival-time data on near and distant earthquakes. Although similar seismographs have existed since the 1890's, it was only in the 1930's that Charles F. Richter, a California seismologist, introduced the concept of earthquake magnitude. His original definition held only for California earthquakes occurring within 600 km of a particular type of seismograph (the Woods-Anderson torsion instrument). His basic idea was quite simple: by knowing the distance from a seismograph to an earthquake and observing the maximum signal amplitude recorded on the seismograph, an empirical quantitative ranking of the earthquake's inherent size or strength could be made. Most California earthquakes occur within the top 16 km of the crust; to a first approximation, corrections for variations in earthquake focal depth were, therefore, unnecessary.
Richter's original magnitude scale (ML) was then extended to observations of earthquakes of any distance and of focal depths ranging between 0 and 700 km. Because earthquakes excite both body waves, which travel into and through the Earth, and surface waves, which are constrained to follow the natural wave guide of the Earth's uppermost layers, two magnitude scales evolved - the mb and MS scales.
The standard body-wave magnitude formula is
mb = log10(A/T) + Q(D,h) ,
where A is the amplitude of ground motion (in microns); T is the corresponding period (in seconds); and Q(D,h) is a correction factor that is a function of distance, D (degrees), between epicenter and station and focal depth, h (in kilometers), of the earthquake. The standard surface-wave formula is
MS = log10 (A/T) + 1.66 log10 (D) + 3.30 .
There are many variations of these formulas that take into account effects of specific geographic regions, so that the final computed magnitude is reasonably consistent with Richter's original definition of ML. Negative magnitude values are permissible.
A rough idea of frequency of occurrence of large
earthquakes is given by the following table:
|MS||Earthquakes per year|
|8.5 - 8.9||0.3|
|8.0 - 8.4||1.1|
|7.5 - 7.9||3.1|
|7.0 - 7.4||15|
|6.5 - 6.9||56|
|6.0 - 6.4||210|
This table is based on data for a recent 47 year period. Perhaps the rates of earthquake occurrence are highly variable and some other 47 year period could give quite different results.
The original mb scale utilized compressional body P-wave amplitudes with periods of 4-5 s, but recent observations are generally of 1 s-period P waves. The MS scale has consistently used Rayleigh surface waves in the period range from 18 to 22 s.
When initially developed, these magnitude scales were considered to be equivalent; in other words, earthquakes of all sizes were thought to radiate fixed proportions of energy at different periods. But it turns out that larger earthquakes, which have larger rupture surfaces, systematically radiate more long-period energy. Thus, for very large earthquakes, body-wave magnitudes badly underestimate true earthquake size; the maximum body-wave magnitudes are about 6.5 - 6.8. In fact, the surface-wave magnitudes underestimate the size of very large earthquakes; the maximum observed values are about 8.3 - 8.7. Some investigators have suggested that the 100 s mantle Love waves (a type of surface wave) should be used to estimate magnitude of great earthquakes. However, even this approach ignores the fact that damage to structure is often caused by energy at shorter periods. Thus, modern seismologists are increasingly turning to two separate parameters to describe the physical effects of an earthquake: seismic moment and radiated energy.
Fault Geometry and Seismic Moment, MO
The orientation of the fault, direction of fault movement, and size of an earthquake can be described by the fault geometry and seismic moment. These parameters are determined from waveform analysis of the seismograms produced by an earthquake. The differing shapes and directions of motion of the waveforms recorded at different distances and azimuths from the earthquake are used to determine the fault geometry, and the wave amplitudes are used to compute moment. The seismic moment is related to fundamental parameters of the faulting process.
MO = µSd ,
where µ is the shear strength of the faulted rock, S is the area of the fault, and <d> is the average displacement on the fault. Because fault geometry and observer azimuth are a part of the computation, moment is a more consistent measure of earthquake size than is magnitude, and more importantly, moment does not have an intrinsic upper bound. These factors have led to the definition of a new magnitude scale MW, based on seismic moment, where
MW = 2/3 log10(MO) - 10.7 .
The two largest reported moments are 2.5 X 1030 dyn·cm (dyne·centimeters) for the 1960 Chile earthquake (MS 8.5; MW 9.6) and 7.5 X 1029 dyn·cm for the 1964 Alaska earthquake (MS 8.3; MW 9.2). MS approaches its maximum value at a moment between 1028 and 1029 dyn·cm.
The amount of energy radiated by an earthquake is a measure of the potential for damage to man-made structures. Theoretically, its computation requires summing the energy flux over a broad suite of frequencies generated by an earthquake as it ruptures a fault. Because of instrumental limitations, most estimates of energy have historically relied on the empirical relationship developed by Beno Gutenberg and Charles Richter:
log10E = 11.8 + 1.5MS
where energy, E, is expressed in ergs. The drawback of this method is that MS is computed from an bandwidth between approximately 18 to 22 s. It is now known that the energy radiated by an earthquake is concentrated over a different bandwidth and at higher frequencies. With the worldwide deployment of modern digitally recording seismograph with broad bandwidth response, computerized methods are now able to make accurate and explicit estimates of energy on a routine basis for all major earthquakes. A magnitude based on energy radiated by an earthquake, Me, can now be defined,
Me = 2/3 log10E - 2.9.
For every increase in magnitude by 1 unit, the associated seismic energy increases by about 32 times.
Although Mw and Me are both magnitudes, they describe different physical properites of the earthquake. Mw, computed from low-frequency seismic data, is a measure of the area ruptured by an earthquake. Me, computed from high frequency seismic data, is a measure of seismic potential for damage. Consequently, Mw and Me often do not have the same numerical value.
The increase in the degree of surface shaking (intensity) for each unit increase of magnitude of a shallow crustal earthquake is unknown. Intensity is based on an earthquake's local accelerations and how long these persist. Intensity and magnitude thus both depend on many variables that include exactly how rock breaks and how energy travels from an earthquake to a receiver. These factors make it difficult for engineers and others who use earthquake intensity and magnitude data to evaluate the error bounds that may exist for their particular applications.
An example of how local soil conditions can greatly influence local intensity is given by catastrophic damage in Mexico City from the 1985, MS 8.1 Mexico earthquake centered some 300 km away. Resonances of the soil-filled basin under parts of Mexico City amplified ground motions for periods of 2 seconds by a factor of 75 times. This shaking led to selective damage to buildings 15 - 25 stories high (same resonant period), resulting in losses to buildings of about $4.0 billion and at least 8,000 fatalities.
The occurrence of an earthquake is a complex physical process. When an earthquake occurs, much of the available local stress is used to power the earthquake fracture growth to produce heat rather than to generate seismic waves. Of an earthquake system's total energy, perhaps 10 percent to less that 1 percent is ultimately radiated as seismic energy. So the degree to which an earthquake lowers the Earth's available potential energy is only fractionally observed as radiated seismic energy.
Determining the Depth of an Earthquake
Earthquakes can occur anywhere between the Earth's surface and about 700 kilometers below the surface. For scientific purposes, this earthquake depth range of 0 - 700 km is divided into three zones: shallow, intermediate, and deep.
Shallow earthquakes are between 0 and 70 km deep; intermediate earthquakes, 70 - 300 km deep; and deep earthquakes, 300 - 700 km deep. In general, the term "deep-focus earthquakes" is applied to earthquakes deeper than 70 km. All earthquakes deeper than 70 km are localized within great slabs of shallow lithosphere that are sinking into the Earth's mantle.
The evidence for deep-focus earthquakes was discovered in 1922 by H.H. Turner of Oxford, England. Previously, all earthquakes were considered to have shallow focal depths. The existence of deep-focus earthquakes was confirmed in 1931 from studies of the seismograms of several earthquakes, which in turn led to the construction of travel-time curves for intermediate and deep earthquakes.
The most obvious indication on a seismogram that a large earthquake has a deep focus is the small amplitude, or height, of the recorded surface waves and the uncomplicated character of the P and S waves. Although the surface-wave pattern does generally indicate that an earthquake is either shallow or may have some depth, the most accurate method of determining the focal depth of an earthquake is to read a depth phase recorded on the seismogram.
The most characteristic depth phase is pP. This is the P wave that is reflected from the surface of the Earth at a point relatively near the epicenter. At distant seismograph stations, the pP follows the P wave by a time interval that changes slowly with distance but rapidly with depth. This time interval, pP-P (pP minus P), is used to compute depth-of-focus tables. Using the time difference of pP-P as read from the seismogram and the distance between the epicenter and the seismograph station, the depth of the earthquake can be determined from published travel-time curves or depth tables.
Another seismic wave used to determine focal depth is the sP phase - an S wave reflected as a P wave from the Earth's surface at a point near the epicenter. This wave is recorded after the pP by about one-half of the pP-P time interval. The depth of an earthquake can be determined from the sP phase in the same manner as the pP phase by using the appropriate travel-time curves or depth tables for sP.
If the pP and sP waves can be identified on the seismogram, an accurate focal depth can be determined.
by William Spence, Stuart A. Sipkin, and George L. Choy
Earthquakes and Volcanoes
Volume 21, Number 1, 1989