In astronomy, luminosity is the total amount of energy emitted by a star, galaxy, or other astronomical object per unit time. It is related to the brightness, which is the luminosity of an object in a given spectral region.
In SI units luminosity is measured in joules per second or watts. Values for luminosity are often given in the terms of the luminosity of the Sun, L⊙. Luminosity can also be given in terms of magnitude: the absolute bolometric magnitude (Mbol) of an object is a logarithmic measure of its total energy emission rate.
In astronomy, luminosity is the amount of electromagnetic energy a body radiates per unit of time. When not qualified, the term "luminosity" means bolometric luminosity, which is measured either in the SI units, watts, or in terms of solar luminosities (L☉). A bolometer is the instrument used to measure radiant energy over a wide band by absorption and measurement of heating. A star also radiates neutrinos, which carry off some energy (about 2% in the case of our Sun), contributing to the star's total luminosity. The IAU has defined a nominal solar luminosity of 3.828×1026 W to promote publication of consistent and comparable values in units of the solar luminosity.
While bolometers do exist, they cannot be used to measure even the apparent brightness of a star because they are insufficiently sensitive across the electromagnetic spectrum and because most wavelengths do not reach the surface of the Earth. In practice bolometric magnitudes are measured by taking measurements at certain wavelengths and constructing a model of the total spectrum that is most likely to match those measurements. In some cases, the process of estimation is extreme, with luminosities being calculated when less than 1% of the energy output is observed, for example with a hot Wolf-Rayet star observed only in the infra-red. Bolometric luminosities can also be calculated using a bolometric correction to a luminosity in a particular passband.
The term luminosity is also used in relation to particular passbands such as a visual luminosity of K-band luminosity. These are not generally luminosities in the strict sense of an absolute measure of radiated power, but absolute magnitudes defined for a given filter in a photometric system. Several different photometric systems exist. Some such as the UBV or Johnson system are defined against photometric standard stars, while others such as the AB system are defined in terms of a spectral flux density.
A star's luminosity can be determined from two stellar characteristics: size and effective temperature. The former is typically represented in terms of solar radii, R⊙, while the latter is represented in kelvins, but in most cases neither can be measured directly. To determine a star's radius, two other metrics are needed: the star's angular diameter and its distance from Earth, often calculated using parallax. Both can be measured with great accuracy in certain cases, with cool supergiants often having large angular diameters, and some cool evolved stars having masers in their atmospheres that can be used to measure the parallax using VLBI. However, for most stars the angular diameter or parallax, or both, are far below our ability to measure with any certainty. Since the effective temperature is merely a number that represents the temperature of a black body that would reproduce the luminosity, it obviously cannot be measured directly, but it can be estimated from the spectrum.
An alternative way to measure stellar luminosity is to measure the star's apparent brightness and distance. A third component needed to derive the luminosity is the degree of interstellar extinction that is present, a condition that usually arises because of gas and dust present in the interstellar medium (ISM), the Earth's atmosphere, and circumstellar matter. Consequently, one of astronomy's central challenges in determining a star's luminosity is to derive accurate measurements for each of these components, without which an accurate luminosity figure remains elusive. Extinction can only be measured directly if the actual and observed luminosities are both known, but it can be estimated from the observed colour of a star, using models of the expected level of reddening from the interstellar medium.
In the current system of stellar classification, stars are grouped according to temperature, with the massive, very young and energetic Class O stars boasting temperatures in excess of 30,000 K while the less massive, typically older Class M stars exhibit temperatures less than 3,500 K. Because luminosity is proportional to temperature to the fourth power, the large variation in stellar temperatures produces an even vaster variation in stellar luminosity. Because the luminosity depends on a high power of the stellar mass, high mass luminous stars have much shorter lifetimes. The most luminous stars are always young stars, no more than a few million years for the most extreme. In the Hertzsprung–Russell diagram, the x-axis represents temperature or spectral type while the y-axis represents luminosity or magnitude. The vast majority of stars are found along the main sequence with blue Class 0 stars found at the top left of the chart while red Class M stars fall to the bottom right. Certain stars like Deneb and Betelgeuse are found above and to the right of the main sequence, more luminous or cooler than their equivalents on the main sequence. Increased luminosity at the same temperature, or alternatively cooler temperature at the same luminosity, indicates that these stars are larger than those on the main sequence and they are called giants or supergiants.
Blue and white supergiants are high luminosity stars somewhat cooler than the most luminous main sequence stars. A star like Deneb, for example, has a luminosity around 200,000 L⊙, a spectral type of A2, and an effective temperature around 8,500 K, meaning it has a radius around 203 R⊙. For comparison, the red supergiant Betelgeuse has a luminosity around 100,000 L⊙, a spectral type of M2, and a temperature around 3,500 K, meaning its radius is about 1,000 R⊙. Red supergiants are the largest type of star, but the most luminous are much smaller and hotter, with temperatures up to 50,000 K and more and luminosities of several million L⊙, meaning their radii are just a few tens of R⊙. An example is R136a1, over 50,000 K and shining at over 8,000,000 L⊙ (mostly in the UV), it is only 35 R⊙.
The luminosity of a radio source is measured in W Hz−1, to avoid having to specify a bandwidth over which it is measured. The observed strength, or flux density, of a radio source is measured in Jansky where 1 Jy = 10−26 W m−2 Hz−1.
For example, consider a 10W transmitter at a distance of 1 million metres, radiating over a bandwidth of 1 MHz. By the time that power has reached the observer, the power is spread over the surface of a sphere with area 4πr2 or about 1.26×1013 m2, so its flux density is 10 / 106 / 1.26×1013 W m−2 Hz−1 = 108 Jy.
More generally, for sources at cosmological distances, a k-correction must be made for the spectral index α of the source, and a relativistic correction must be made for the fact that the frequency scale in the emitted rest frame is different from that in the observer's rest frame. So the full expression for radio luminosity, assuming isotropic emission, is
where Lν is the luminosity in W Hz−1, Sobs is the observed flux density in W m−2 Hz−1, DL is the luminosity distance in metres, z is the redshift, α is the spectral index (in the sense , and in radio astronomy, assuming thermal emission the spectral index is typically equal to 2.)
For example, consider a 1 Jy signal from a radio source at a redshift of 1, at a frequency of 1.4 GHz. Ned Wright's cosmology calculator calculates a luminosity distance for a redshift of 1 to be 6701 Mpc = 2×1026 m giving a radio luminosity of 10−26 × 4π(2×1026)2 / (1+1)(1+2) = 6×1026 W Hz−1.
To calculate the total radio power, this luminosity must be integrated over the bandwidth of the emission. A common assumption is to set the bandwidth to the observing frequency, which effectively assumes the power radiated has uniform intensity from zero frequency up to the observing frequency. In the case above, the total power is 4×1027 × 1.4×109 = 5.7×1036 W. This is sometimes expressed in terms of the total (i.e. integrated over all wavelengths) luminosity of the Sun which is 3.86×1026 W, giving a radio power of 1.5×1010 L⊙.
|First magnitude||< 1.5||Vega||0.03|
|Second magnitude||1.5 to 2.5||Denebola||2.14|
|Third magnitude||2.5 to 3.5||Rastaban||2.79|
|Fourth magnitude||3.5 to 4.5||Sadalpheretz||3.96|
|Fifth magnitude||4.5 to 5.5||Pleione||5.05|
|Sixth magnitude||5.5 to 6.5||54 Piscium||5.88|
|Seventh magnitude||6.5 to 7.5||HD 40307||7.17|
|Eighth magnitude||7.5 to 8.5||HD 113766||7.56|
|Ninth magnitude||8.5 to 9.5||HD 149382||8.94|
|Tenth magnitude||9.5 to 10.5||HIP 13044||9.98|
Luminosity is an intrinsic measurable property of a star independent of distance. The concept of magnitude, on the other hand, incorporates distance. First conceived by the Greek astronomer Hipparchus in the second century BC, the original concept of magnitude grouped stars into six discrete categories depending on how bright they appeared. The brightest first magnitude stars were twice as bright as the next brightest stars, which were second magnitude; second was twice as bright as third, third twice as bright as fourth and so on down to the faintest stars, which Hipparchus categorized as sixth magnitude. The system was but a simple delineation of stellar brightness into six distinct groups and made no allowance for the variations in brightness within a group. With the invention of the telescope at the beginning of the seventeenth century, researchers soon realized that there were subtle variations among stars and millions fainter than the sixth magnitude—hence the need for a more sophisticated system to describe a continuous range of values beyond what the naked eye could see.
In 1856 Norman Pogson, noticing that photometric measurements had established first magnitude stars as being about 100 times brighter than sixth magnitude stars, formalized the Hipparchus system by creating a logarithmic scale, with every interval of one magnitude equating to a variation in brightness of 1001/5 or roughly 2.512 times. Consequently, a first magnitude star is about 2.5 times brighter than a second magnitude star, 2.52 brighter than a third magnitude star, 2.53 brighter than a fourth magnitude star, et cetera. Based on this continuous scale, any star with a magnitude between 5.5 and 6.5 is now considered to be sixth magnitude, a star with a magnitude between 4.5 and 5.5 is fifth magnitude and so on. With this new mathematical rigor, a first magnitude star should then have a magnitude in the range 0.5 to 1.5, thus excluding the nine brightest stars with magnitudes lower than 0.5, as well as the four brightest with negative values. It is customary therefore to extend the definition of a first magnitude star to any star with a magnitude less than 0.5, as can be seen in accompanying table.
The Pogson logarithmic scale is used to measure both apparent and absolute magnitudes, the latter corresponding to the brightness of a star or other celestial body as seen if it would be located at an interstellar distance of 10 parsecs. The apparent magnitude is a measure of the diminishing flux of light as a result of distance according to the inverse-square law. In addition to this brightness decrease from increased distance, there is an extra decrease of brightness due to extinction from intervening interstellar dust.
By measuring the width of certain absorption lines in the stellar spectrum, it is often possible to assign a certain luminosity class to a star without knowing its distance. Thus a fair measure of its absolute magnitude can be determined without knowing its distance nor the interstellar extinction, allowing astronomers to estimate a star's distance and extinction without parallax calculations. Since the stellar parallax is usually too small to be measured for many distant stars, this is a common method of determining such distances.
To conceptualize the range of magnitudes in our own galaxy, the smallest star to be identified has about 8% of the Sun’s mass and glows feebly at absolute magnitude +19. Compared to the Sun, which has an absolute of +4.8, this faint star is 14 magnitudes or 400,000 times dimmer than our Sun. Our galaxy's most massive stars begin their lives with masses of roughly 100 times solar, radiating at upwards of absolute magnitude –8, over 160,000 times the solar luminosity. The total range of stellar luminosities, then, occupies a range of 27 magnitudes, or a factor of 60 billion.
In measuring star brightnesses, absolute magnitude, apparent magnitude, and distance are interrelated parameters—if two are known, the third can be determined. Since the Sun's luminosity is the standard, comparing these parameters with the Sun's apparent magnitude and distance is the easiest way to remember how to convert between them.
Imagine a point source of light of luminosity that radiates equally in all directions. A hollow sphere centered on the point would have its entire interior surface illuminated. As the radius increases, the surface area will also increase, and the constant luminosity has more surface area to illuminate, leading to a decrease in observed brightness.
The surface area of a sphere with radius r is , so for stars and other point sources of light:
where is the distance from the observer to the light source.
where and are the radius and temperature of the Sun, respectively.
The magnitude of a star, a unitless measure, is a logarithmic scale of observed visible brightness. The apparent magnitude is the observed visible brightness from Earth which depends on the distance of the object. The absolute magnitude is the apparent magnitude at a distance of 10 parsecs, therefore the bolometric absolute magnitude is a logarithmic measure of the bolometric luminosity.
The difference in bolometric magnitude between two objects is related to their luminosity ratio according to:
This can be used to derive a luminosity in solar units:
which makes by inversion:
Although the absolute bolometric magnitude of the sun is approximately 4.7554, the zero point of the absolute magnitude scale is actually defined as a fixed luminosity of 3.0128×1028 W. Therefore the absolute magnitude can be calculated from a luminosity in watts:
and the luminosity in watts can be calculated from an absolute magnitude (although absolute magnitudes are often not measured relative to an absolute flux):
2014 CODATA recommended values