In statistics and its applications, the root mean square (abbreviated RMS or rms) is defined as the square root of the mean square (the arithmetic mean of the squares of a set of numbers). The RMS is also known as the quadratic mean and is a particular case of the generalized mean with exponent 2. RMS can also be defined for a continuously varying function in terms of an integral of the squares of the instantaneous values during a cycle.
The RMS value of a set of values (or a continuous-time waveform) is the square root of the arithmetic mean of the squares of the values, or the square of the function that defines the continuous waveform. In physics, the RMS current is the "value of the direct current that dissipates power in a resistor."
In the case of a set of n values , the RMS is
The corresponding formula for a continuous function (or waveform) f(t) defined over the interval is
and the RMS for a function over all time is
The RMS over all time of a periodic function is equal to the RMS of one period of the function. The RMS value of a continuous function or signal can be approximated by taking the RMS of a sequence of equally spaced samples. Additionally, the RMS value of various waveforms can also be determined without calculus, as shown by Cartwright.
If the waveform is a pure sine wave, the relationships between amplitudes (peak-to-peak, peak) and RMS are fixed and known, as they are for any continuous periodic wave. However, this is not true for an arbitrary waveform, which may not be periodic or continuous. For a zero-mean sine wave, the relationship between RMS and peak-to-peak amplitude is:
For other waveforms the relationships are not the same as they are for sine waves. For example, for either a triangular or sawtooth wave
|DC-shifted square wave|
|Modified sine wave|
Waveforms made by summing known simple waveforms have an RMS that is the root of the sum of squares of the component RMS values, if the component waveforms are orthogonal (that is, if the average of the product of one simple waveform with another is zero for all pairs other than a waveform times itself).
(If the waveforms are in phase, then their RMS amplitudes sum directly.)
A special case of RMS of waveform combinations is:
Electrical engineers often need to know the power, P, dissipated by an electrical resistance, R. It is easy to do the calculation when there is a constant current, I, through the resistance. For a load of R ohms, power is defined simply as:
However, if the current is a time-varying function, I(t), this formula must be extended to reflect the fact that the current (and thus the instantaneous power) is varying over time. If the function is periodic (such as household AC power), it is still meaningful to discuss the average power dissipated over time, which is calculated by taking the average power dissipation:
|(where denotes the mean of a function)|
|(as R does not vary over time, it can be factored out)|
|(by definition of RMS)|
So, the RMS value, IRMS, of the function I(t) is the constant current that yields the same power dissipation as the time-averaged power dissipation of the current I(t).
Average power can also be found using the same method that in the case of a time-varying voltage, V(t), with RMS value VRMS,
By taking the square root of both these equations and multiplying them together, the power is found to be:
Both derivations depend on voltage and current being proportional (i.e., the load, R, is purely resistive). Reactive loads (i.e., loads capable of not just dissipating energy but also storing it) are discussed under the topic of AC power.
In the common case of alternating current when I(t) is a sinusoidal current, as is approximately true for mains power, the RMS value is easy to calculate from the continuous case equation above. If Ip is defined to be the peak current, then:
where t is time and ω is the angular frequency (ω = 2π/T, where T is the period of the wave).
Since Ip is a positive constant:
Using a trigonometric identity to eliminate squaring of trig function:
but since the interval is a whole number of complete cycles (per definition of RMS), the sin terms will cancel out, leaving:
A similar analysis leads to the analogous equation for sinusoidal voltage:
Where IP represents the peak current and VP represents the peak voltage.
Because of their usefulness in carrying out power calculations, listed voltages for power outlets (e.g., 120 V in the USA, or 230 V in Europe) are almost always quoted in RMS values, and not peak values. Peak values can be calculated from RMS values from the above formula, which implies VP = VRMS × √, assuming the source is a pure sine wave. Thus the peak value of the mains voltage in the USA is about 120 × √, or about 170 volts. The peak-to-peak voltage, being double this, is about 340 volts. A similar calculation indicates that the peak mains voltage in Europe is about 325 volts, and the peak-to-peak mains voltage, about 650 volts.
RMS quantities such as electric current are usually calculated over one cycle. However, for some purposes the RMS current over a longer period is required when calculating transmission power losses. The same principle applies, and (for example) a current of 10 amps used for 12 hours each day represents an RMS current of 5 amps in the long term.
The term "RMS power" is sometimes erroneously used in the audio industry as a synonym for "mean power" or "average power" (it is proportional to the square of the RMS voltage or RMS current in a resistive load). For a discussion of audio power measurements and their shortcomings, see Audio power.
where R represents the ideal gas constant, 8.314 J/(mol·K), T is the temperature of the gas in kelvins, and M is the molar mass of the gas in kilograms per mole. The generally accepted terminology for speed as compared to velocity is that the former is the scalar magnitude of the latter. Therefore, although the average speed is between zero and the RMS speed, the average velocity for a stationary gas is zero.
When two data sets—one set from theoretical prediction and the other from actual measurement of some physical variable, for instance—are compared, the RMS of the pairwise differences of the two data sets can serve as a measure how far on average the error is from 0. The mean of the pairwise differences does not measure the variability of the difference, and the variability as indicated by the standard deviation is around the mean instead of 0. Therefore, the RMS of the differences is a meaningful measure of the error.
The RMS can be computed in the frequency domain, using Parseval's theorem. For a sampled signal , where is the sampling period,
where and N is number of samples and FFT coefficients.
In this case, the RMS computed in the time domain is the same as in the frequency domain:
From this it is clear that the RMS value is always greater than or equal to the average, in that the RMS includes the "error" / square deviation as well.
Physical scientists often use the term "root mean square" as a synonym for standard deviation when it can be assumed the input signal has zero mean, i.e., referring to the square root of the mean squared deviation of a signal from a given baseline or fit. This is useful for electrical engineers in calculating the "AC only" RMS of a signal. Standard deviation being the root mean square of a signal's variation about the mean, rather than about 0, the DC component is removed (i.e. RMS(signal) = Stdev(signal) if the mean signal is 0).
Alternating current (AC) is an electric current which periodically reverses direction, in contrast to direct current (DC) which flows only in one direction. Alternating current is the form in which electric power is delivered to businesses and residences, and it is the form of electrical energy that consumers typically use when they plug kitchen appliances, televisions, fans and electric lamps into a wall socket. A common source of DC power is a battery cell in a flashlight. The abbreviations AC and DC are often used to mean simply alternating and direct, as when they modify current or voltage.The usual waveform of alternating current in most electric power circuits is a sine wave, whose positive half-period corresponds with positive direction of the current and vice versa. In certain applications, different waveforms are used, such as triangular or square waves. Audio and radio signals carried on electrical wires are also examples of alternating current. These types of alternating current carry information such as sound (audio) or images (video) sometimes carried by modulation of an AC carrier signal. These currents typically alternate at higher frequencies than those used in power transmission.Amplitude
The amplitude of a periodic variable is a measure of its change over a single period (such as time or spatial period). There are various definitions of amplitude (see below), which are all functions of the magnitude of the difference between the variable's extreme values. In older texts the phase is sometimes called the amplitude.Average rectified value
In electrical engineering, the average rectified value (ARV) of the quantity is the average of its absolute value.
The average of a symmetric alternating value is zero and it is therefore not useful to characterize it. Thus the easiest way to determine a quantitative measurement size is to use the average rectified value. The average rectified value is mainly used to characterize alternating voltage and alternating current. It can be computed by averaging the absolute value of a waveform over one full period of the waveform.
While conceptually similar to the root mean square (RMS), ARV will differ from it whenever a function's absolute value varies locally, as the former then increases disproportionately. The difference is expressed by the form factor
In the military science of ballistics, circular error probable (CEP) (also circular error probability or circle of equal probability) is a measure of a weapon system's precision. It is defined as the radius of a circle, centered on the mean, whose boundary is expected to include the landing points of 50% of the rounds; said otherwise, it is the median error radius. That is, if a given bomb design has a CEP of 100 m, when 100 are targeted at the same point, 50 will fall within a 100 m circle around their average impact point. (The distance between the target point and the average impact point is referred to as bias.)
There are associated concepts, such as the DRMS (distance root mean square), which is the square root of the average squared distance error, and R95, which is the radius of the circle where 95% of the values would fall in.
The concept of CEP also plays a role when measuring the accuracy of a position obtained by a navigation system, such as GPS or older systems such as LORAN and Loran-C.Confirmatory factor analysis
In statistics, confirmatory factor analysis (CFA) is a special form of factor analysis, most commonly used in social research. It is used to test whether measures of a construct are consistent with a researcher's understanding of the nature of that construct (or factor). As such, the objective of confirmatory factor analysis is to test whether the data fit a hypothesized measurement model. This hypothesized model is based on theory and/or previous analytic research. CFA was first developed by Jöreskog and has built upon and replaced older methods of analyzing construct validity such as the MTMM Matrix as described in Campbell & Fiske (1959).In confirmatory factor analysis, the researcher first develops a hypothesis about what factors they believe are underlying the measures used (e.g., "Depression" being the factor underlying the Beck Depression Inventory and the Hamilton Rating Scale for Depression) and may impose constraints on the model based on these a priori hypotheses. By imposing these constraints, the researcher is forcing the model to be consistent with their theory. For example, if it is posited that there are two factors accounting for the covariance in the measures, and that these factors are unrelated to one another, the researcher can create a model where the correlation between factor A and factor B is constrained to zero. Model fit measures could then be obtained to assess how well the proposed model captured the covariance between all the items or measures in the model. If the constraints the researcher has imposed on the model are inconsistent with the sample data, then the results of statistical tests of model fit will indicate a poor fit, and the model will be rejected. If the fit is poor, it may be due to some items measuring multiple factors. It might also be that some items within a factor are more related to each other than others.
For some applications, the requirement of "zero loadings" (for indicators not supposed to load on a certain factor) has been regarded as too strict. A newly developed analysis method, "exploratory structural equation modeling", specifies hypotheses about the relation between observed indicators and their supposed primary latent factors while allowing for estimation of loadings with other latent factors as well.Cross-validation (statistics)
Cross-validation, sometimes called rotation estimation, or out-of-sample testing is any of various similar model validation techniques for assessing how the results of a statistical analysis will generalize to an independent data set. It is mainly used in settings where the goal is prediction, and one wants to estimate how accurately a predictive model will perform in practice. In a prediction problem, a model is usually given a dataset of known data on which training is run (training dataset), and a dataset of unknown data (or first seen data) against which the model is tested (called the validation dataset or testing set). The goal of cross-validation is to test the model's ability to predict new data that was not used in estimating it, in order to flag problems like overfitting or selection bias and to give an insight on how the model will generalize to an independent dataset (i.e., an unknown dataset, for instance from a real problem).
One round of cross-validation involves partitioning a sample of data into complementary subsets, performing the analysis on one subset (called the training set), and validating the analysis on the other subset (called the validation set or testing set). To reduce variability, in most methods multiple rounds of cross-validation are performed using different partitions, and the validation results are combined (e.g. averaged) over the rounds to give an estimate of the model's predictive performance.
In summary, cross-validation combines (averages) measures of fitness in prediction to derive a more accurate estimate of model prediction performance.IEC 60038
International Standard IEC 60038:1983 defines a set of standard voltages for use in low voltage and high voltage AC electricity supply systems.Kabsch algorithm
The Kabsch algorithm, named after Wolfgang Kabsch, is a method for calculating the optimal rotation matrix that minimizes the RMSD (root mean squared deviation) between two paired sets of points. It is useful in graphics, cheminformatics to compare molecular structures, and also bioinformatics for comparing protein structures (in particular, see root-mean-square deviation (bioinformatics)).
The algorithm only computes the rotation matrix, but it also requires the computation of a translation vector. When both the translation and rotation are actually performed, the algorithm is sometimes called partial Procrustes superimposition (see also orthogonal Procrustes problem).Maxwell–Boltzmann distribution
In physics (in particular in statistical mechanics), the Maxwell–Boltzmann distribution is a particular probability distribution named after James Clerk Maxwell and Ludwig Boltzmann.
It was first defined and used for describing particle speeds in idealized gases, where the particles move freely inside a stationary container without interacting with one another, except for very brief collisions in which they exchange energy and momentum with each other or with their thermal environment. The term "particle" in this context refers to gaseous particles only (atoms or molecules), and the system of particles is assumed to have reached thermodynamic equilibrium. The energies of such particles follow what is known as Maxwell-Boltzmann statistics, and the statistical distribution of speeds is derived by equating particle energies with kinetic energy.
Mathematically, the Maxwell–Boltzmann distribution is the chi distribution with three degrees of freedom (the components of the velocity vector in Euclidean space), with a scale parameter measuring speeds in units proportional to the square root of (the ratio of temperature and particle mass).
The Maxwell–Boltzmann distribution is a result of the kinetic theory of gases, which provides a simplified explanation of many fundamental gaseous properties, including pressure and diffusion. The Maxwell–Boltzmann distribution applies fundamentally to particle velocities in three dimensions, but turns out to depend only on the speed (the magnitude of the velocity) of the particles. A particle speed probability distribution indicates which speeds are more likely: a particle will have a speed selected randomly from the distribution, and is more likely to be within one range of speeds than another. The kinetic theory of gases applies to the classical ideal gas, which is an idealization of real gases. In real gases, there are various effects (e.g., van der Waals interactions, vortical flow, relativistic speed limits, and quantum exchange interactions) that can make their speed distribution different from the Maxwell–Boltzmann form. However, rarefied gases at ordinary temperatures behave very nearly like an ideal gas and the Maxwell speed distribution is an excellent approximation for such gases. Ideal plasmas, which are ionized gases of sufficiently low density, frequently also have particle distributions that are partially or entirely Maxwellian.
The distribution was first derived by Maxwell in 1860 on heuristic grounds. Boltzmann later, in the 1870s, carried out significant investigations into the physical origins of this distribution.
The distribution can be derived on the ground that it maximizes the entropy of the system. A list of derivations are:Mean squared displacement
In statistical mechanics, the mean squared displacement (MSD, also mean square displacement, average squared displacement, or mean square fluctuation) is a measure of the deviation of the position of a particle with respect to a reference position over time. It is the most common measure of the spatial extent of random motion, and can be thought of as measuring the portion of the system "explored" by the random walker. In the realm of biophysics and environmental engineering, the Mean Squared Displacement is measured over time to determine if a particle is spreading solely due to diffusion, or if an advective force is also contributing. Another relevant concept, the Variance-Related Diameter (VRD, which is twice the square root of MSD), is also used in studying the transportation and mixing phenomena in the realm of environmental engineering. It prominently appears in the Debye–Waller factor (describing vibrations within the solid state) and in the Langevin equation (describing diffusion of a Brownian particle). The MSD is defined as
where N is the number of particles to be averaged, is the reference position of each particle, is the position of each particle in determined time t.Mean squared error
In statistics, the mean squared error (MSE) or mean squared deviation (MSD) of an estimator (of a procedure for estimating an unobserved quantity) measures the average of the squares of the errors—that is, the average squared difference between the estimated values and what is estimated. MSE is a risk function, corresponding to the expected value of the squared error loss. The fact that MSE is almost always strictly positive (and not zero) is because of randomness or because the estimator does not account for information that could produce a more accurate estimate.The MSE is a measure of the quality of an estimator—it is always non-negative, and values closer to zero are better.
The MSE is the second moment (about the origin) of the error, and thus incorporates both the variance of the estimator (how widely spread the estimates are from one data sample to another) and its bias (how far off the average estimated value is from the truth). For an unbiased estimator, the MSE is the variance of the estimator. Like the variance, MSE has the same units of measurement as the square of the quantity being estimated. In an analogy to standard deviation, taking the square root of MSE yields the root-mean-square error or root-mean-square deviation (RMSE or RMSD), which has the same units as the quantity being estimated; for an unbiased estimator, the RMSE is the square root of the variance, known as the standard error.Percentage point
A percentage point or percent point is the unit for the arithmetic difference of two percentages. For example, moving up from 40% to 44% is a 4 percentage point increase, but is a 10 percent increase in what is being measured. In the literature, the percentage point unit is usually either written out, or abbreviated as pp or p.p. to avoid ambiguity. After the first occurrence, some writers abbreviate by using just "point" or "points".
Consider the following hypothetical example: In 1980, 50 percent of the population smoked, and in 1990 only 40 percent smoked. One can thus say that from 1980 to 1990, the prevalence of smoking decreased by 10 percentage points although smoking did not decrease by 10 percent (it decreased by 20 percent) – percentages indicate ratios, not differences.
Percentage-point differences are one way to express a risk or probability. Consider a drug that cures a given disease in 70 percent of all cases, while without the drug, the disease heals spontaneously in only 50 percent of cases. The drug reduces absolute risk by 20 percentage points. Alternatives may be more meaningful to consumers of statistics, such as the reciprocal, also known as the number needed to treat (NNT). In this case, the reciprocal transform of the percentage-point difference would be 1/(20pp) = 1/0.20 = 5. Thus if 5 patients are treated with the drug, one could expect to heal one more case of the disease than would have occurred in the absence of the drug.
For measurements involving percentages as a unit, such as, growth, yield, or ejection fraction, statistical deviations and related descriptive statistics, including the standard deviation and root-mean-square error, the result should be expressed in units of percentage points instead of percentage. Mistakenly using percentage as the unit for the standard deviation is confusing, since percentage is also used as a unit for the relative standard deviation, i.e. standard deviation divided by average value (coefficient of variation).Radius of gyration
Radius of gyration or gyradius of a body about an axis of rotation is defined as the radial distance of a point from the axis of rotation at which, if whole mass of the body is assumed to be concentrated, its moment of inertia about the given axis would be the same as with its actual distribution of mass. It is denoted by .
Mathematically the radius of gyration is the root mean square distance of the object's parts from either its center of mass or a given axis, depending on the relevant application. It is actually the perpendicular distance from point mass to the axis of rotation.
Suppose a body consists of particles each of mass . Let be their perpendicular distances from the axis of rotation. Then, the moment of inertia of the body about the axis of rotation is
If all the masses are the same (), then the moment of inertia is .
Since ( being the total mass of the body),
From the above equations, we have
Radius of gyration is the root mean square distance of particles from axis formula
Therefore, the radius of gyration of a body about a given axis may also be defined as the root mean square distance of the various particles of the body from the axis of rotation.
The root-mean-square deviation (RMSD) or root-mean-square error (RMSE) (or sometimes root-mean-squared error) is a frequently used measure of the differences between values (sample or population values) predicted by a model or an estimator and the values observed. The RMSD represents the square root of the second sample moment of the differences between predicted values and observed values or the quadratic mean of these differences. These deviations are called residuals when the calculations are performed over the data sample that was used for estimation and are called errors (or prediction errors) when computed out-of-sample. The RMSD serves to aggregate the magnitudes of the errors in predictions for various times into a single measure of predictive power. RMSD is a measure of accuracy, to compare forecasting errors of different models for a particular dataset and not between datasets, as it is scale-dependent.RMSD is always non-negative, and a value of 0 (almost never achieved in practice) would indicate a perfect fit to the data. In general, a lower RMSD is better than a higher one. However, comparisons across different types of data would be invalid because the measure is dependent on the scale of the numbers used.
RMSD is the square root of the average of squared errors. The effect of each error on RMSD is proportional to the size of the squared error; thus larger errors have a disproportionately large effect on RMSD. Consequently, RMSD is sensitive to outliers.Root-mean-square deviation of atomic positions
In bioinformatics, the root-mean-square deviation of atomic positions (or simply root-mean-square deviation, RMSD) is the measure of the average distance between the atoms (usually the backbone atoms) of superimposed proteins. Note that RMSD calculation can be applied to other, non-protein molecules, such as small organic molecules. In the study of globular protein conformations, one customarily measures the similarity in three-dimensional structure by the RMSD of the Cα atomic coordinates after optimal rigid body superposition.
When a dynamical system fluctuates about some well-defined average position, the RMSD from the average over time can be referred to as the RMSF or root mean square fluctuation. The size of this fluctuation can be measured, for example using Mössbauer spectroscopy or nuclear magnetic resonance, and can provide important physical information. The Lindemann index is a method of placing the RMSF in the context of the parameters of the system.
A widely used way to compare the structures of biomolecules or solid bodies is to translate and rotate one structure with respect to the other to minimize the RMSD. Coutsias, et al. presented a simple derivation, based on quaternions, for the optimal solid body transformation (rotation-translation) that minimizes the RMSD between two sets of vectors. They proved that the quaternion method is equivalent to the well-known Kabsch algorithm. The solution given by Kabsch is an instance of the solution of the d-dimensional problem, introduced by Hurley and Cattell. The quaternion solution to compute the optimal rotation was published in the appendix of a paper of Petitjean. This quaternion solution and the calculation of the optimal isometry in the d-dimensional case were both extended to infinite sets and to the continuous case in the appendix A of an other paper of Petitjean.Spectral width
In telecommunications, spectral width is the wavelength interval over which the magnitude of all spectral components is equal to or greater than a specified fraction of the magnitude of the component having the maximum value.
In optical communications applications, the usual method of specifying spectral width is the full width at half maximum. This is the same convention used in bandwidth, defined as the frequency range where power drops by less than half (at most −3 dB).
The FWHM method may be difficult to apply when the spectrum has a complex shape. Another method of specifying spectral width is a special case of root-mean-square deviation where the independent variable is wavelength, λ, and f (λ) is a suitable radiometric quantity.
The relative spectral width, Δλ/λ, is frequently used where Δλ is obtained according to note 1, and λ is the center wavelength.Time–frequency representation
A time–frequency representation (TFR) is a view of a signal (taken to be a function of time) represented over both time and frequency. Time–frequency analysis means analysis into the time–frequency domain provided by a TFR. This is achieved by using a formulation often called "Time–Frequency Distribution", abbreviated as TFD.
TFRs are often complex-valued fields over time and frequency, where the modulus of the field represents either amplitude or "energy density" (the concentration of the root mean square over time and frequency), and the argument of the field represents phase.Two-point equidistant projection
The two-point equidistant projection is a map projection first described by Hans Maurer in 1919. It is a generalization of the much simpler azimuthal equidistant projection. In this two-point form, two locus points are chosen by the mapmaker to configure the projection. Distances from the two loci to any other point on the map are correct: that is, they scale to the distances of the same points on the sphere.
The projection has been used for all maps of the Asian continent by the National Geographic Society atlases since 1959, though its purpose in that case was to reduce distortion throughout Asia rather than to measure from the two loci. The projection sometimes appears in maps of air routes. The Chamberlin trimetric projection is a logical extension of the two-point idea to three points, but the three-point case only yields a sort of minimum error for distances from the three loci, rather than yielding correct distances. Tobler extended this idea to arbitrarily large number of loci by using automated root-mean-square minimization techniques rather than using closed-form formulae.Volt-ampere
A volt-ampere (VA) is the unit used for the apparent power in an electrical circuit, equal to the product of root-mean-square (RMS) voltage and RMS current. In direct current (DC) circuits, this product is equal to the real power (active power) in watts. Volt-amperes are useful only in the context of alternating current (AC) circuits (sinusoidal voltages and currents of the same frequency).
With a purely resistive load, the apparent power is equal to the real power. Where a reactive (capacitive or inductive) component is present in the load, the apparent power is greater than the real power as voltage and current are no longer in phase. In the limiting case of a purely reactive load, current is drawn but no power is dissipated in the load.
Some devices, including uninterruptible power supplies (UPSs), have ratings both for maximum volt-amperes and maximum watts. The VA rating is limited by the maximum permissible current, and the watt rating by the power-handling capacity of the device. When a UPS powers equipment which presents a reactive load with a low power factor, neither limit may safely be exceeded. For example, a (large) UPS system rated to deliver 400,000 volt-amperes at 220 volts can deliver a current of 1818 amperes.
VA ratings are also often used for transformers; maximum output current is then VA rating divided by nominal output voltage. Transformers with the same sized core usually have the same VA rating.
The convention of using the volt-ampere to distinguish apparent power from real power is allowed by the SI standard.