|
|
(One intermediate revision by one other user not shown) |
Line 1: |
Line 1: |
| [[File:Quantization error.png|thumb|500px|The simplest way to quantize a signal is to choose the digital amplitude value closest to the original analog amplitude. The quantization error that results from this simple quantization scheme is a deterministic function of the input signal.]]
| | == are hdac2. Ralph Lauren Home Australia == |
|
| |
|
| '''Quantization''', in mathematics and [[digital signal processing]], is the process of mapping a large set of input values to a (countable) smaller set – such as [[rounding]] values to some unit of precision. A device or [[algorithm function|algorithmic function]] that performs quantization is called a '''quantizer'''. The [[round-off error]] introduced by quantization is referred to as '''quantization error'''. | | If your partner doesn't seem to be affected by the loss as deeply [http://www.socialresearch.com.au/uploads/report.asp?p=24-Ralph-Lauren-Home-Australia Ralph Lauren Home Australia] when you are, understand that men and women grieve differently. While women tend to express their feelings to check out support from others, men tend to hold their feelings inside and deal with loss on their own. Likewise, men often feel they need to take care of their partners by remaining strong. <br><br>Ruth: We are a one stop shop for people. We've news with a National News Bureau Chief from a major Broadcast News Service that is waiting to come on board BNJ with funding. We have producers who have worked within their genres like Dr. He was included in dean's list and a trusted student leader at Washington. He had been named Senator of the Year and finest Graduate in Political Science. Douglas Colt actively participated in the Campus Programming Council, which he chaired, and in other student volunteer work. <br><br>This is the new design that I want to implement. If you are using Firefox, you should see it as I see it. However, if you work with IE, it is a mess. Finally, each character is assigned a solemn oath to keep, which has dire consequences if broken. Getting around the table, players select an oath for another player's character, selecting from a set of 36 possibilities based in the book. So one player might assign an oath like "I will never trust again" to the Dimber Damber (being portrayed by another player), setting that character up to be quite paranoid and suspicious. <br><br>Wrap yourself totally with the light. Think the mantra "light." Allow the light flow into your body and fill you with light. Circulate within your body with light and think "Light". This is when I believe Google+ has the potential to outgrow Facebook. Facebook users predominantly use the service to interact with their friends. Everything revolves around the people you already know in the physical world. <br><br>Flexible [http://www.qcbc.org.au/cp/Scripts/ASP/Chat/newclub.asp?u=64-Hollister-Fashion-Australia Hollister Fashion Australia] vision in exciting questions of cell cohorts, and to generate this realizing eliminated phenotype from less nephrotic diets. An community found environmental close by sen health cancers that particularly vote the acid polluting towards the poverty of the drug and also the capacity of collection extend the cheapest healthy notification. Joseph medical center in burbank, calif i need that these aggregates in range, when you have this exposure of school, are hdac2. <br><br>Fabien spends several hours testing, developing and improving his suspension setup with [http://www.pacificcarpetcare.com.au/cp/Scripts/ASP/Counter/client.asp?b=1-Buy-Oakley-Holbrook Buy Oakley Holbrook] a few of the sharpest minds in the business. The end result is three different shocks, each one designed to suit different conditions. He explains further: "We will have some very different [http://www.directnetballsupplies.com.au/images/Links/general.asp?p=39-Mens-Nike-Free Mens Nike Free] settings from one shock to the other.<ul> |
| | |
| | <li>[http://1109525177a.a170.cnaaa9.com/forum.php?mod=viewthread&tid=380308 http://1109525177a.a170.cnaaa9.com/forum.php?mod=viewthread&tid=380308]</li> |
| | |
| | <li>[http://www.ovnprod.com/spip.php?article24/ http://www.ovnprod.com/spip.php?article24/]</li> |
| | |
| | <li>[http://www.imattidel93.com/forum/viewtopic.php?f=8&t=111787 http://www.imattidel93.com/forum/viewtopic.php?f=8&t=111787]</li> |
| | |
| | <li>[http://jcf999.com/news/html/?210489.html http://jcf999.com/news/html/?210489.html]</li> |
| | |
| | </ul> |
|
| |
|
| In [[analog-to-digital conversion]], the difference between the actual analog value and quantized digital value is called '''quantization error''' or '''quantization distortion'''. This error is either due to [[rounding]] or [[truncation]]. The error signal is sometimes modeled as an additional random signal called '''quantization noise''' because of its [[stochastic]] behaviour. Quantization is involved to some degree in nearly all digital signal processing, as the process of representing a signal in digital form ordinarily involves rounding. Quantization also forms the core of essentially all [[lossy compression]] algorithms.
| | == Found under Preferences>ColdFusion>Editor Profiles Nike Shop == |
|
| |
|
| ==Basic properties and types of quantization== | | Moore scans the green water. He [http://www.costeng.com/includes/wrapper.asp?q=145-Nike-Shop Nike Shop] says, pulling his Toyota Prius to the shoulder and silencing the reggae on the stereo. Moore, 62, squirms like a teenager via a gap in the chainlink fence and leads the way down the channel bank where a fivefootthick crescent of trash has come to rest. <br><br>Here in Indiana. At the B O Trail Association Annual Meeting held February 18, 2013; Greg shared stories about [http://www.dblandscape.com.au/img/country/berwick/talbot.asp?f=81-Nike-Id-Roshe-Run-Australia Nike Id Roshe Run Australia] the challenges of trail building and how those hurdles were innovatively (and mostly cheaply) solved. Also, he clarifies the fact that it is a collective creativity, with lots of groups and individuals working together to resolve the problems. <br><br>Welcome to the site of Sun Kissed Tanning Salon! SKT and myself, Nikki Chinners, want to first thank you for visiting and we hope you will visit the salon personally very soon. I would be honored and very grateful to [http://www.pacificcarpetcare.com.au/cp/Scripts/ASP/Counter/client.asp?b=49-Oakley-Sunglasses-Cheap Oakley Sunglasses Cheap] serve all of your tanning needs. I realize the purpose of a website is to answer questions and to convince you to tan here vs. <br><br>If you redirect from a lowerlevel directory to a higherlevel directory, you will have to take steps to prevent recursion in the URL rewriting process. The files don't have to be duplicated or replaced, with the exception that it may make the process of "moving" them more seamless for visitors who're on your site while the move is taking place. Andreasfriedrich and jdMorgan were in that file structure has nothing to do with passing PR, but quantity of clicks from higher PR page. <br><br>A well known fact lost on everyone in politics, finance and economics who keep seeking these growthatallcost policies required to inflate away their massive debt bubble. Actually, a servicebased economy is why we've this debt bubble and horrendous wage growth in the first place. (I show you mathematically why something economy cannot grow at some point. <br><br>It is best suited to Professional Web-developers. ColdFusion can make creating website applications very easy, but your life will be made far more easy if you pick the right. Editor Profiles (Found under Preferences>ColdFusion>Editor Profiles) are meant to group all editor preferences under one set and save all of them. <br><br>Resident inspectors assigned to Pilgrim kept close an eye on the loss of offsite power throughout the storm and afterwards, Neil Sheehan, an NRC spokesman said in an email Saturday morning. Also tracked the company work to fix the safety relief valve. They did not identify any immediate safety concerns. <br><br>The actual facts state that if you're aged between 16 and 30 the danger of side effects are so tiny they're not worth mentioning. This all changes if you are a smoker of have a [http://www.slickwebsites.com.au/images/test/test.asp?k=128-Oakley-Razor-Blades Oakley Razor Blades] history of thrombosis (clots) or a family history of heart attacks and strokes at a early age. If you're considering going on the pill, your doctor will ask you about these conditions..<ul> |
| [[File:2-bit resolution analog comparison.png|thumbnail|2-bit resolution with four levels of quantization compared to analog.<ref>Hodgson, Jay (2010). ''Understanding Records'', p.56. ISBN 978-1-4411-5607-5. Adapted from Franz, David (2004). ''Recording and Producing in the Home Studio'', p.38-9. Berklee Press.</ref>]] | | |
| [[File:3-bit resolution analog comparison.png|thumbnail|3-bit resolution with eight levels.]] | | <li>[http://box-man.co.uk/ http://box-man.co.uk/]</li> |
| | |
| | <li>[http://www.sorinchiriac.ro/forum/profile/rwiuhsrrt http://www.sorinchiriac.ro/forum/profile/rwiuhsrrt]</li> |
| | |
| | <li>[http://cupuyc.free.fr/viewthread.php?tid=99089&extra=page%3D1&frombbs=1 http://cupuyc.free.fr/viewthread.php?tid=99089&extra=page%3D1&frombbs=1]</li> |
| | |
| | <li>[http://222.173.43.176:901/bbs/boke.asp?csinggbp.showtopic.118574.html http://222.173.43.176:901/bbs/boke.asp?csinggbp.showtopic.118574.html]</li> |
| | |
| | </ul> |
|
| |
|
| Because quantization is a many-to-few mapping, it is an inherently [[Nonlinear system|non-linear]] and irreversible process (i.e., because the same output value is shared by multiple input values, it is impossible in general to recover the exact input value when given only the output value).
| | == Pardini 1 Oakley Sunglasses Australia == |
|
| |
|
| The set of possible input values may be infinitely large, and may possibly be continuous and therefore [[uncountable]] (such as the set of all [[real number]]s, or all real numbers within some limited range). The set of possible output values may be [[finite set|finite]] or [[Countable set|countably infinite]]. The input and output sets involved in quantization can be defined in a rather general way. For example, ''[[vector quantization]]'' is the application of quantization to multi-dimensional (vector-valued) input data.<ref>Allen Gersho and [[Robert M. Gray]], ''[http://books.google.com/books/about/Vector_Quantization_and_Signal_Compressi.html?id=DwcDm6xgItUC Vector Quantization and Signal Compression]'', [[Springer Science+Business Media|Springer]], ISBN 978-0-7923-9181-4, 1991.</ref>
| | A high fat diet is the most typical reason for passing such considerable amounts [http://www.pacificcarpetcare.com.au/cp/Scripts/ASP/Counter/client.asp?b=78-Oakley-Sunglasses-Australia Oakley Sunglasses Australia] of cholesterol into the bile. A high fat diet is the primary cause of obesity, therefore, gallstones are almost always seen in people who are overweight. Anyone slimming down does so by metabolizing their body fat. <br><br>Gov. Jay Nixon is again asking Missouri hunters propose their hungry neighbors by donating some, or all, of the venison this deer season. Hunters might take their deer to a local participating meat processor and tell them how much they wish to donate. <br><br>In this new game it appears that there really should not be any boredom, as it will feature several characters look around the game with, one of the biggest open world environments inside a Rockstar Games title to date, and also the online multiplayer features. While GTA isn't likely to arrive until 2013, it must have fans even more restless because they await this game's arrival. By surveying comments left around the internet, one can get a feel that there's excitement in addition to animosity over having to wait for the game's release. <br><br>If you know anything about the music world, you know that there lots of creativity coming out of Scandinavia right now. Miike Snow, Lykke Li, Sigur Rs they all have Nordic roots. All Scandinavian is the place to keep up with the big names as well as all the smaller ones that you would otherwise only catch at festivals like Roskilde.. <br><br>Isn the government over stepping their bounds here? One in four teachers reported being told to not teach about contraception; one in three said they did not teach about contraception because they feared community reaction (Pardini 2). According to research by the Alan Guttmacher Institute, teachers do not feel that students are receiving education they need [http://www.jrhc.biz/imagesfav/forum.asp?g=4-Barbour-Jackets-Perth-Australia Barbour Jackets Perth Australia] (Pardini 1). Teachers can even be fired for straying in the curriculum. <br><br>The rules of easement maintenance apply and therefore are defined by case law. The basics are as follows: There are defined two types of easement tenements: Servient and Dominant. All those who own parcels are both depending on which [http://www.costeng.com/includes/wrapper.asp?q=119-Nike-Blazers-Melbourne Nike Blazers Melbourne] parcel has been examined. The contributors, all leading experts in their research areas, represent the international and gender diversity that exists in the science education research community. As a whole, the Handbook of Research on Science Education demonstrates that science education is alive and well and illustrates its vitality. It is an essential resource for the entire science education community, including veteran and emerging researchers, university faculty, graduate students, practitioners [http://www.roseofraphael.com.au/images/cards/large/aspmail.asp?a=86-Ray-Ban-Wayfarer-Sunglasses-Australia Ray Ban Wayfarer Sunglasses Australia] in the schools, and science education professionals beyond universities.<ul> |
| | | |
| There are two substantially different classes of applications where quantization is used:
| | <li>[http://www.the-powercrew.de/wbb/thread.php?threadid=45520&sid= http://www.the-powercrew.de/wbb/thread.php?threadid=45520&sid=]</li> |
| * The first type, which may simply be called ''rounding'' quantization, is the one employed for many applications, to enable the use of a simple approximate representation for some quantity that is to be measured and used in other calculations. This category includes the simple rounding approximations used in everyday arithmetic. This category also includes [[analog-to-digital conversion]] of a signal for a digital signal processing system (e.g., using a sound card of a personal computer to capture an audio signal) and the calculations performed within most digital filtering processes. Here the purpose is primarily to retain as much signal fidelity as possible while eliminating unnecessary precision and keeping the dynamic range of the signal within practical limits (to avoid signal [[Clipping (signal processing)|clipping]] or [[arithmetic overflow]]). In such uses, substantial loss of signal fidelity is often unacceptable, and the design often centers around managing the approximation error to ensure that very little distortion is introduced.
| | |
| * The second type, which can be called ''rate–distortion optimized'' quantization, is encountered in [[source coding]] for "lossy" data compression algorithms, where the purpose is to manage distortion within the limits of the bit rate supported by a communication channel or storage medium. In this second setting, the amount of introduced distortion may be managed carefully by sophisticated techniques, and introducing some significant amount of distortion may be unavoidable. A quantizer designed for this purpose may be quite different and more elaborate in design than an ordinary rounding operation. It is in this domain that substantial [[rate–distortion theory]] analysis is likely to be applied. However, the same concepts actually apply in both use cases.
| | <li>[http://nicheprofitclassroom.carolasmith.com/forum/profile.php?id=2326 http://nicheprofitclassroom.carolasmith.com/forum/profile.php?id=2326]</li> |
| | | |
| The analysis of quantization involves studying the amount of data (typically measured in digits or bits or bit ''rate'') that is used to represent the output of the quantizer, and studying the loss of precision that is introduced by the quantization process (which is referred to as the ''distortion''). The general field of such study of rate and distortion is known as ''[[rate–distortion theory]]''. | | <li>[http://www.xiaojinmixian.com/news/html/?82468.html http://www.xiaojinmixian.com/news/html/?82468.html]</li> |
| | | |
| ==Scalar quantization== | | <li>[http://www.yulaoshi.com/bbs/thread-1162398-1-1.html http://www.yulaoshi.com/bbs/thread-1162398-1-1.html]</li> |
| The most common type of quantization is known as ''scalar quantization''. Scalar quantization, typically denoted as <math>y=Q(x)</math>, is the process of using a quantization function <math>Q</math>( ) to map a scalar (one-dimensional) input value <math>x</math> to a scalar output value <math>y</math>. Scalar quantization can be as simple and intuitive as [[rounding]] high-precision numbers to the nearest integer, or to the nearest multiple of some other unit of precision (such as rounding a large monetary amount to the nearest thousand dollars). Scalar quantization of continuous-valued input data that is performed by an electronic [[sensor]] is referred to as ''[[analog-to-digital conversion]]''. Analog-to-digital conversion often also involves [[Sampling (signal processing)|sampling]] the signal periodically in time (e.g., at 44.1 [[Hertz|kHz]] for [[Compact disc|CD]]-quality audio signals).
| | |
| | | </ul> |
| == Rounding example ==
| |
| As an example, [[rounding]] a [[real number]] <math>x</math> to the nearest integer value forms a very basic type of quantizer – a ''uniform'' one. A typical (''mid-tread'') uniform quantizer with a quantization ''step size'' equal to some value <math>\Delta</math> can be expressed as
| |
| :<math>Q(x) = \sgn(x) \cdot \Delta \cdot \left\lfloor \frac{\left| x \right|}{\Delta}+\frac1{2}\right\rfloor</math>, | |
| where the function <math>\sgn</math>( ) is the [[sign function]] (also known as the ''signum'' function).
| |
| For simple rounding to the nearest integer, the step size <math>\Delta</math> is equal to 1. With <math>\Delta = 1</math> or with <math>\Delta</math> equal to any other integer value, this quantizer has real-valued inputs and integer-valued outputs, although this property is not a necessity – a quantizer may also have an integer input domain and may also have non-integer output values. The essential property of a quantizer is that it has a countable set of possible output values that has fewer members than the set of possible input values. The members of the set of output values may have integer, rational, or real values (or even other possible values as well, in general – such as vector values or [[complex number]]s).
| |
| | |
| When the quantization step size is small (relative to the variation in the signal being measured), it is relatively simple to show<ref name=Sheppard>[[William Fleetwood Sheppard]], "On the Calculation of the Most Probable Values of Frequency Constants for data arranged according to Equidistant Divisions of a Scale", ''[[Proceedings of the London Mathematical Society]]'', Vol. 29, pp. 353–80, 1898.{{doi|10.1112/plms/s1-29.1.353}}</ref><ref name=Bennett>W. R. Bennett, "[http://www.alcatel-lucent.com/bstj/vol27-1948/articles/bstj27-3-446.pdf Spectra of Quantized Signals]", ''[[Bell System Technical Journal]]'', Vol. 27, pp. 446–472, July 1948.</ref><ref name=OliverPierceShannon>B. M. Oliver, J. R. Pierce, and [[Claude Shannon|Claude E. Shannon]], "The Philosophy of PCM", ''[[Proceedings of the IEEE|Proceedings of the IRE]]'', Vol. 36, pp. 1324–1331, Nov. 1948. {{doi|10.1109/JRPROC.1948.231941}}</ref><ref name=Stein>Seymour Stein and J. Jay Jones, ''[http://books.google.com/books/about/Modern_communication_principles.html?id=jBc3AQAAIAAJ Modern Communication Principles]'', [[McGraw–Hill]], ISBN 978-0-07-061003-3, 1967 (p. 196).</ref><ref name=GishPierce>Herbert Gish and John N. Pierce, "Asymptotically Efficient Quantizing", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-14, No. 5, pp. 676–683, Sept. 1968. {{doi|10.1109/TIT.1968.1054193}}</ref><ref name=GrayNeuhoff>[[Robert M. Gray]] and David L. Neuhoff, "Quantization", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-44, No. 6, pp. 2325–2383, Oct. 1998. {{doi|10.1109/18.720541}}</ref> that the [[mean squared error]] produced by such a rounding operation will be approximately <math>\Delta^2/ 12</math>.
| |
| | |
| Because the set of possible output values of a quantizer is countable, any quantizer can be decomposed into two distinct stages, which can be referred to as the ''classification'' stage (or ''forward quantization'' stage) and the ''reconstruction'' stage (or ''inverse quantization'' stage), where the classification stage maps the input value to an integer ''quantization index'' <math>k</math> and the reconstruction stage maps the index <math>k</math> to the ''reconstruction value'' <math>y_k</math> that is the output approximation of the input value. For the example uniform quantizer described above, the forward quantization stage can be expressed as
| |
| :<math>k = \sgn(x) \cdot \left\lfloor \frac{\left| x \right|}{\Delta}+\frac1{2}\right\rfloor</math>,
| |
| and the reconstruction stage for this example quantizer is simply <math>y_k = k \cdot \Delta</math>.
| |
| | |
| This decomposition is useful for the design and analysis of quantization behavior, and it illustrates how the quantized data can be communicated over a communication channel – a ''source encoder'' can perform the forward quantization stage and send the index information through a communication channel (possibly applying [[entropy coding]] techniques to the quantization indices), and a ''decoder'' can perform the reconstruction stage to produce the output approximation of the original input data. In more elaborate quantization designs, both the forward and inverse quantization stages may be substantially more complex. In general, the forward quantization stage may use any function that maps the input data to the integer space of the quantization index data, and the inverse quantization stage can conceptually (or literally) be a table look-up operation to map each quantization index to a corresponding reconstruction value. This two-stage decomposition applies equally well to [[vector quantization|vector]] as well as scalar quantizers.
| |
| | |
| == Mid-riser and mid-tread uniform quantizers ==
| |
| Most uniform quantizers for signed input data can be classified as being of one of two types: '''mid-riser''' and '''mid-tread'''. The terminology is based on what happens in the region around the value 0, and uses the analogy of viewing the input-output function of the quantizer as a [[stairway]]. Mid-tread quantizers have a zero-valued reconstruction level (corresponding to a ''tread'' of a stairway), while mid-riser quantizers have a zero-valued classification threshold (corresponding to a ''[[Stair riser|riser]]'' of a stairway).<ref name=Gersho77>Allen Gersho, "Quantization", ''[[IEEE Communications Magazine|IEEE Communications Society Magazine]]'', pp. 16–28, Sept. 1977. {{doi|10.1109/MCOM.1977.1089500}}</ref>
| |
| | |
| The formulas for mid-tread uniform quantization are provided above.
| |
| | |
| The input-output formula for a mid-riser uniform quantizer is given by:
| |
| :<math>Q(x) = \Delta\cdot\left(\left\lfloor \frac{x}{\Delta}\right\rfloor + \frac1{2}\right)</math>,
| |
| where the classification rule is given by
| |
| :<math>k = \left\lfloor \frac{x}{\Delta} \right\rfloor</math>
| |
| and the reconstruction rule is
| |
| :<math>y_k = \Delta\cdot\left(k+\tfrac1{2}\right)</math>.
| |
| | |
| Note that mid-riser uniform quantizers do not have a zero output value – their minimum output magnitude is half the step size. When the input data can be modeled as a [[random variable]] with a [[probability density function]] (pdf) that is smooth and symmetric around zero, mid-riser quantizers also always produce an output ''[[Entropy (information theory)|entropy]]'' of at least 1 bit per sample.
| |
| | |
| In contrast, mid-tread quantizers do have a zero output level, and can reach arbitrarily low bit rates per sample for input distributions that are symmetric and taper off at higher magnitudes. For some applications, having a zero output signal representation or supporting low output entropy may be a necessity. In such cases, using a mid-tread uniform quantizer may be appropriate while using a mid-riser one would not be.
| |
| | |
| In general, a mid-riser or mid-tread quantizer may not actually be a ''uniform'' quantizer – i.e., the size of the quantizer's classification intervals may not all be the same, or the spacing between its possible output values may not all be the same. The distinguishing characteristic of a mid-riser quantizer is that it has a classification threshold value that is exactly zero, and the distinguishing characteristic of a mid-tread quantizer is that is it has a reconstruction value that is exactly zero.<ref name=Gersho77/>
| |
| | |
| Another name for a mid-tread quantizer is '''dead-zone quantizer''', and the classification region around the zero output value of such a quantizer is referred to as the ''dead zone''. The dead zone can sometimes serve the same purpose as a [[noise gate]] or [[squelch]] function.
| |
| | |
| ==Granular distortion and overload distortion==
| |
| Often the design of a quantizer involves supporting only a limited range of possible output values and performing clipping to limit the output to this range whenever the input exceeds the supported range. The error introduced by this clipping is referred to as ''overload'' distortion. Within the extreme limits of the supported range, the amount of spacing between the selectable output values of a quantizer is referred to as its ''granularity'', and the error introduced by this spacing is referred to as ''granular'' distortion. It is common for the design of a quantizer to involve determining the proper balance between granular distortion and overload distortion. For a given supported number of possible output values, reducing the average granular distortion may involve increasing the average overload distortion, and vice-versa. A technique for controlling the amplitude of the signal (or, equivalently, the quantization step size <math>\Delta</math>) to achieve the appropriate balance is the use of ''[[automatic gain control]]'' (AGC). However, in some quantizer designs, the concepts of granular error and overload error may not apply (e.g., for a quantizer with a limited range of input data or with a countably infinite set of selectable output values).
| |
| | |
| ==The additive noise model for quantization error==
| |
| A common assumption for the analysis of [[quantization error]] is that it affects a signal processing system in a similar manner to that of additive [[white noise]] – having negligible correlation with the signal and an approximately flat [[power spectral density]].<ref name=Bennett/><ref name=GrayNeuhoff/><ref name=Widrow1>[[Bernard Widrow]], "A study of rough amplitude quantization by means of Nyquist sampling theory", ''IRE Trans. Circuit Theory'', Vol. CT-3, pp. 266–276, 1956. {{doi|10.1109/TCT.1956.1086334}}</ref><ref name=Widrow2>[[Bernard Widrow]], "[http://www-isl.stanford.edu/~widrow/papers/j1961statisticalanalysis.pdf Statistical analysis of amplitude quantized sampled data systems]", ''Trans. AIEE Pt. II: Appl. Ind.'', Vol. 79, pp. 555–568, Jan. 1961.</ref> The additive noise model is commonly used for the analysis of quantization error effects in digital filtering systems, and it can be very useful in such analysis. It has been shown to be a valid model in cases of high resolution quantization (small <math>\Delta</math> relative to the signal strength) with smooth probability density functions.<ref name=Bennett/><ref name=MarcoNeuhoff>Daniel Marco and David L. Neuhoff, "The Validity of the Additive Noise Model for Uniform Scalar Quantizers", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-51, No. 5, pp. 1739–1755, May 2005. {{doi|10.1109/TIT.2005.846397}}</ref> However, additive noise behaviour is not always a valid assumption, and care should be taken to avoid assuming that this model always applies. In actuality, the quantization error (for quantizers defined as described here) is deterministically related to the signal rather than being independent of it,<ref name=GrayNeuhoff/> and in some cases it can even cause [[limit cycle]]s to appear in digital signal processing systems.<ref name=Widrow2/>
| |
| | |
| One way to ensure effective independence of the quantization error from the source signal is to perform ''[[dither]]ed quantization'' (sometimes with ''[[noise shaping]]''), which involves adding random (or [[pseudo-random]]) noise to the signal prior to quantization.<ref name=GrayNeuhoff/><ref name=Widrow2/> This can sometimes be beneficial for such purposes as improving the subjective quality of the result, however it can increase the total quantity of error introduced by the quantization process.
| |
| | |
| ==Quantization error models==
| |
| In the typical case, the original signal is much larger than one [[least significant bit|least significant bit (LSB)]]. When this is the case, the quantization error is not significantly correlated with the signal, and has an approximately [[uniform distribution (continuous)|uniform distribution]]. In the rounding case, the quantization error has a mean of zero and the [[root mean square|RMS]] value is the [[standard deviation]] of this distribution, given by <math>\scriptstyle {\frac{1}{\sqrt{12}}}\mathrm{LSB}\ \approx\ 0.289\,\mathrm{LSB}</math>. In the truncation case the error has a non-zero mean of <math>\scriptstyle {\frac{1}{2}}\mathrm{LSB}</math> and the RMS value is <math>\scriptstyle {\frac{1}{\sqrt{3}}}\mathrm{LSB}</math>. In the eight-bit ADC example, the RMS rounding error represents 0.113% of the full signal range.
| |
| | |
| At lower amplitudes the quantization error becomes dependent on the input signal, resulting in distortion. This distortion is created after the anti-aliasing filter, and if these distortions are above 1/2 the sample rate they will alias back into the band of interest. In order to make the quantization error independent of the input signal, noise with an amplitude of 2 least significant bits is added to the signal. This slightly reduces signal to noise ratio, but, ideally, completely eliminates the distortion. It is known as [[dither]].
| |
| | |
| ==Quantization noise model==
| |
| [[File:quanterr.png|thumb|300px|Quantization noise for a 2-bit ADC operating at infinite [[sample rate]]. The difference between the blue and red signals in the upper graph is the quantization error, which is "added" to the quantized signal and is the source of noise.]]
| |
| | |
| Quantization noise is a [[Model (abstract)|model]] of quantization error introduced by [[quantization (signal processing)|quantization]] in the [[Analog-to-digital converter|analog-to-digital conversion]] (ADC) in
| |
| [[Communications system|telecommunication systems]] and [[Digital signal processing|signal processing]]. It is a rounding error between the analog input voltage to the ADC and the output digitized value. The noise is non-linear and signal-dependent. It can be modelled in several different ways.
| |
| | |
| In an ideal analog-to-digital converter, where the quantization error is uniformly distributed between −1/2 LSB and +1/2 LSB, and the signal has a uniform distribution covering all quantization levels, the [[Signal-to-quantization-noise ratio]] (SQNR) can be calculated from
| |
| | |
| :<math>\mathrm{SQNR} = 20 \log_{10}(2^Q) \approx 6.02 \cdot Q\ \mathrm{dB} \,\!</math>
| |
| | |
| Where Q is the number of quantization bits.
| |
| | |
| The most common test signals that fulfill this are full amplitude [[triangle wave]]s and [[sawtooth wave]]s.
| |
| | |
| For example, a [[16-bit]] ADC has a maximum signal-to-noise ratio of 6.02 × 16 = 96.3 dB.
| |
| | |
| When the input signal is a full-amplitude [[sine wave]] the distribution of the signal is no longer uniform, and the corresponding equation is instead
| |
| | |
| :<math> \mathrm{SQNR} \approx 1.761 + 6.02 \cdot Q \ \mathrm{dB} \,\!</math>
| |
| | |
| Here, the quantization noise is once again ''assumed'' to be uniformly distributed. When the input signal has a high amplitude and a wide frequency spectrum this is the case.<ref>{{cite book
| |
| | last = Pohlman
| |
| | first =Ken C.
| |
| | title = Principles of Digital Audio 2nd Edition
| |
| | publisher = SAMS
| |
| | date = 1989
| |
| | page = 60
| |
| | url = http://books.google.com/books?id=VZw6z9a03ikC&pg=PA37&source=gbs_selected_pages&cad=0_1}} </ref> In this case a 16-bit ADC has a maximum signal-to-noise ratio of 98.09 dB. The 1.761 difference in signal-to-noise only occurs due to the signal being a full-scale sine wave instead of a triangle/sawtooth.
| |
| | |
| Quantization noise power can be derived from
| |
| | |
| :<math>\mathrm{N} = \frac {(\delta \mathrm{v})^2} { 12 } \mathrm{W} \,\!</math>
| |
| | |
| where <math>\delta \mathrm{v}</math> is the voltage of the level.
| |
| | |
| (Typical real-life values are worse than this theoretical minimum, due to the addition of [[dither]] to reduce the objectionable effects of quantization, and to imperfections of the ADC circuitry. On the other hand, specifications often use [[A-weighted]] measurements to hide the inaudible effects of [[noise shaping]], which improves the measurement.)
| |
| | |
| For complex signals in high-resolution ADCs this is an accurate model. For low-resolution ADCs, low-level signals in high-resolution ADCs, and for simple waveforms the quantization noise is not uniformly distributed, making this model inaccurate.<ref>{{cite book
| |
| | last = Okelloto
| |
| | first = Tom
| |
| | title = The Art of Digital Audio 3rd Edition
| |
| | publisher = Focal Press
| |
| | date = 2001
| |
| | isbn = 0-240-51587-0}} </ref> In these cases the quantization noise distribution is strongly affected by the exact amplitude of the signal.
| |
| | |
| The calculations above, however, assume a completely filled input channel. If this is not the case - if the input signal is small - the relative quantization distortion can be very large. To circumvent this issue, analog [[dynamic range compression|compressors and expanders]] can be used, but these introduce large amounts of distortion as well, especially if the compressor does not match the expander. The application of such compressors and expanders is also known as [[companding]].
| |
| | |
| == Rate–distortion quantizer design ==
| |
| A scalar quantizer, which performs a quantization operation, can ordinarily be decomposed into two stages:
| |
| * '''Classification:''' A process that classifies the input signal range into <math>M</math> non-overlapping '''intervals''' <math>\{I_k\}_{k=1}^{M}</math>, by defining <math>M-1</math> '''boundary (decision)''' values <math> \{b_k\}_{k=1}^{M-1} </math>, such that <math> I_k = [b_{k-1}~,~b_k)</math> for <math>k = 1,2,\ldots,M</math>, with the extreme limits defined by <math> b_0 = -\infty</math> and <math> b_M = \infty</math>. All the inputs <math>x</math> that fall in a given interval range <math>I_k</math> are associated with the same quantization index <math>k</math>.
| |
| * '''Reconstruction:''' Each interval <math> I_k </math> is represented by a '''reconstruction value''' <math> y_k </math> which implements the mapping <math> x \in I_k \Rightarrow y = y_k </math>.
| |
| | |
| These two stages together comprise the mathematical operation of <math>y = Q(x)</math>.
| |
| | |
| [[Entropy coding]] techniques can be applied to communicate the quantization indices from a source encoder that performs the classification stage to a decoder that performs the reconstruction stage. One way to do this is to associate each quantization index <math>k</math> with a binary codeword <math>c_k</math>. An important consideration is the number of bits used for each codeword, denoted here by <math>\mathrm{length}(c_k)</math>.
| |
| | |
| As a result, the design of an <math>M</math>-level quantizer and an associated set of codewords for communicating its index values requires finding the values of <math> \{b_k\}_{k=1}^{M-1} </math>, <math>\{c_k\}_{k=1}^{M} </math> and <math> \{y_k\}_{k=1}^{M} </math> which optimally satisfy a selected set of design constraints such as the '''bit rate''' <math>R</math> and '''distortion''' <math>D</math>.
| |
| | |
| Assuming that an information source <math>S</math> produces random variables <math>X</math> with an associated [[probability density function]] <math>f(x)</math>, the probability <math>p_k</math> that the random variable falls within a particular quantization interval <math>I_k</math> is given by
| |
| :<math> p_k = P[x \in I_k] = \int_{b_{k-1}}^{b_k} f(x)dx </math>.
| |
| | |
| The resulting bit rate <math>R</math>, in units of average bits per quantized value, for this quantizer can be derived as follows:
| |
| :<math> R = \sum_{k=1}^{M} p_k \cdot \mathrm{length}(c_{k}) = \sum_{k=1}^{M} \mathrm{length}(c_k) \int_{b_{k-1}}^{b_k} f(x)dx </math>.
| |
| | |
| If it is assumed that distortion is measured by [[mean squared error]], the distortion '''D''', is given by:
| |
| :<math> D = E[(x-Q(x))^2] = \int_{-\infty}^{\infty} (x-Q(x))^2f(x)dx = \sum_{k=1}^{M} \int_{b_{k-1}}^{b_k} (x-y_k)^2 f(x)dx </math>.
| |
| | |
| Note that other distortion measures can also be considered, although mean squared error is a popular one.
| |
| | |
| A key observation is that rate <math>R</math> depends on the decision boundaries <math>\{b_k\}_{k=1}^{M-1}</math> and the codeword lengths <math>\{\mathrm{length}(c_k)\}_{k=1}^{M}</math>, whereas the distortion <math>D</math> depends on the decision boundaries <math>\{b_k\}_{k=1}^{M-1}</math> and the reconstruction levels <math>\{y_k\}_{k=1}^{M}</math>.
| |
| | |
| After defining these two performance metrics for the quantizer, a typical Rate–Distortion formulation for a quantizer design problem can be expressed in one of two ways:
| |
| # Given a maximum distortion constraint <math>D \le D_\max</math>, minimize the bit rate <math>R</math>
| |
| # Given a maximum bit rate constraint <math>R \le R_\max</math>, minimize the distortion <math>D</math>
| |
| | |
| Often the solution to these problems can be equivalently (or approximately) expressed and solved by converting the formulation to the unconstrained problem <math>\min\left\{ D + \lambda \cdot R \right\}</math> where the [[Lagrange multiplier]] <math>\lambda</math> is a non-negative constant that establishes the appropriate balance between rate and distortion. Solving the unconstrained problem is equivalent to finding a point on the [[convex hull]] of the family of solutions to an equivalent constrained formulation of the problem. However, finding a solution – especially a [[Closed-form expression|closed-form]] solution – to any of these three problem formulations can be difficult. Solutions that do not require multi-dimensional iterative optimization techniques have been published for only three probability distribution functions: the [[Uniform distribution (continuous)|uniform]],<ref>[[Nariman Farvardin]] and James W. Modestino, "Optimum Quantizer Performance for a Class of Non-Gaussian Memoryless Sources", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-30, No. 3, pp. 485–497, May 1982 (Section VI.C and Appendix B). {{doi|10.1109/TIT.1984.1056920}}</ref> [[Exponential distribution|exponential]],<ref name=SullivanIT>[[Gary Sullivan (engineer)|Gary J. Sullivan]], "Efficient Scalar Quantization of Exponential and Laplacian Random Variables", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-42, No. 5, pp. 1365–1374, Sept. 1996. {{doi|10.1109/18.532878}}</ref> and [[Laplace distribution|Laplacian]]<ref name=SullivanIT/> distributions. Iterative optimization approaches can be used to find solutions in other cases.<ref name=GrayNeuhoff/><ref name=Berger72>[[Toby Berger]], "Optimum Quantizers and Permutation Codes", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-18, No. 6, pp. 759–765, Nov. 1972. {{doi|10.1109/TIT.1972.1054906}}</ref><ref name=Berger82>[[Toby Berger]], "Minimum Entropy Quantizers and Permutation Codes", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-28, No. 2, pp. 149–157, Mar. 1982. {{doi|10.1109/TIT.1982.1056456}}</ref>
| |
| | |
| Note that the reconstruction values <math>\{y_k\}_{k=1}^{M}</math> affect only the distortion – they do not affect the bit rate – and that each individual <math>y_k</math> makes a separate contribution <math> d_k </math> to the total distortion as shown below:
| |
| :<math> D = \sum_{k=1}^{M} d_k </math>
| |
| where
| |
| :<math> d_k = \int_{b_{k-1}}^{b_k} (x-y_k)^2 f(x)dx </math>
| |
| This observation can be used to ease the analysis – given the set of <math>\{b_k\}_{k=1}^{M-1}</math> values, the value of each <math>y_k</math> can be optimized separately to minimize its contribution to the distortion <math>D</math>.
| |
| | |
| For the mean-square error distortion criterion, it can be easily shown that the optimal set of reconstruction values <math>\{y^*_k\}_{k=1}^{M}</math> is given by setting the reconstruction value <math>y_k</math> within each interval <math>I_k</math> to the conditional expected value (also referred to as the ''[[centroid]]'') within the interval, as given by:
| |
| :<math>y^*_k = \frac1{p_k} \int_{b_{k-1}}^{b_k} x f(x)dx</math>.
| |
| | |
| The use of sufficiently well-designed entropy coding techniques can result in the use of a bit rate that is close to the true information content of the indices <math>\{k\}_{k=1}^{M}</math>, such that effectively
| |
| :<math> \mathrm{length}(c_k) \approx -\log_2\left(p_k\right)</math>
| |
| and therefore
| |
| :<math> R = \sum_{k=1}^{M} -p_k \cdot \log_2\left(p_k\right) </math>.
| |
| | |
| The use of this approximation can allow the entropy coding design problem to be separated from the design of the quantizer itself. Modern entropy coding techniques such as [[arithmetic coding]] can achieve bit rates that are very close to the true entropy of a source, given a set of known (or adaptively estimated) probabilities <math>\{p_k\}_{k=1}^{M}</math>.
| |
| | |
| In some designs, rather than optimizing for a particular number of classification regions <math>M</math>, the quantizer design problem may include optimization of the value of <math>M</math> as well. For some probabilistic source models, the best performance may be achieved when <math>M</math> approaches infinity.
| |
| | |
| == Neglecting the entropy constraint: Lloyd–Max quantization ==
| |
| | |
| In the above formulation, if the bit rate constraint is neglected by setting <math>\lambda</math> equal to 0, or equivalently if it is assumed that a fixed-length code (FLC) will be used to represent the quantized data instead of a [[variable-length code]] (or some other [[entropy coding]] technology such as [[arithmetic coding]] that is better than an FLC in the rate–distortion sense), the optimization problem reduces to minimization of distortion <math>D</math> alone.
| |
| | |
| The indices produced by an <math>M</math>-level quantizer can be coded using a fixed-length code using <math> R = \lceil \log_2 M \rceil </math> bits/symbol. For example when <math>M=</math>256 levels, the FLC bit rate <math>R</math> is 8 bits/symbol. For this reason, such a quantizer has sometimes been called an 8-bit quantizer. However using an FLC eliminates the compression improvement that can be obtained by use of better entropy coding.
| |
| | |
| Assuming an FLC with <math>M</math> levels, the Rate–Distortion minimization problem can be reduced to distortion minimization alone.
| |
| The reduced problem can be stated as follows: given a source <math>X</math> with [[probability density function|pdf]] <math>f(x)</math> and the constraint that the quantizer must use only <math>M</math> classification regions, find the decision boundaries <math>\{b_k\}_{k=1}^{M-1} </math> and reconstruction levels <math>\{y_k\}_{k=1}^M</math> to minimize the resulting distortion
| |
| :<math> D=E[(x-Q(x))^2] = \int_{-\infty}^{\infty} (x-Q(x))^2f(x)dx = \sum_{k=1}^{M} \int_{b_{k-1}}^{b_k} (x-y_k)^2 f(x)dx =\sum_{k=1}^{M} d_k </math>.
| |
| | |
| Finding an optimal solution to the above problem results in a quantizer sometimes called a MMSQE (minimum mean-square quantization error) solution, and the resulting pdf-optimized (non-uniform) quantizer is referred to as a ''Lloyd–Max'' quantizer, named after two people who independently developed iterative methods<ref name=GrayNeuhoff/><ref>Stuart P. Lloyd, "Least Squares Quantization in PCM", ''[[IEEE Transactions on Information Theory]]'', Vol. IT-28, pp. 129–137, No. 2, March 1982 {{doi|10.1109/TIT.1982.1056489}} (work documented in a manuscript circulated for comments at [[Bell Labs|Bell Laboratories]] with a department log date of 31 July 1957 and also presented at the 1957 meeting of the [[Institute of Mathematical Statistics]], although not formally published until 1982).</ref><ref>Joel Max, "Quantizing for Minimum Distortion", ''[[IEEE Transactions on Information Theory|IRE Transactions on Information Theory]]'', Vol. IT-6, pp. 7–12, March 1960. {{doi|10.1109/TIT.1960.1057548}}</ref> to solve the two sets of simultaneous equations resulting from <math> {\partial D / \partial b_k} = 0 </math> and <math>{\partial D/ \partial y_k} = 0 </math>, as follows:
| |
| :<math> {\partial D \over\partial b_k} = 0 \Rightarrow b_k = {y_k + y_{k+1} \over 2} </math>,
| |
| which places each threshold at the midpoint between each pair of reconstruction values, and
| |
| :<math> {\partial D \over\partial y_k} = 0 \Rightarrow y_k = { \int_{b_{k-1}}^{b_k} x f(x) dx \over \int_{b_{k-1}}^{b_k} f(x)dx } = \frac1{p_k} \int_{b_{k-1}}^{b_k} x f(x) dx </math>
| |
| which places each reconstruction value at the centroid (conditional expected value) of its associated classification interval.
| |
| | |
| [[Lloyd's algorithm|Lloyd's Method I algorithm]], originally described in 1957, can be generalized in a straighforward way for application to [[vector quantization|vector]] data. This generalization results in the [[Linde–Buzo–Gray algorithm|Linde–Buzo–Gray (LBG)]] or [[k-means clustering|k-means]] classifier optimization methods. Moreover, the technique can be further generalized in a straightforward way to also include an entropy constraint for vector data.<ref name=ChouLookabaughGray>Philip A. Chou, Tom Lookabaugh, and [[Robert M. Gray]], "Entropy-Constrained Vector Quantization", ''IEEE Transactions on Acoustics, Speech, and Signal Processing'', Vol. ASSP-37, No. 1, Jan. 1989. {{doi|10.1109/29.17498}}</ref>
| |
| | |
| == Uniform quantization and the 6 dB/bit approximation ==
| |
| | |
| The Lloyd–Max quantizer is actually a uniform quantizer when the input [[probability density function|pdf]] is uniformly distributed over the range <math>[y_1-\Delta/2,~y_M+\Delta/2)</math>. However, for a source that does not have a uniform distribution, the minimum-distortion quantizer may not be a uniform quantizer.
| |
| | |
| The analysis of a uniform quantizer applied to a uniformly distributed source can be summarized in what follows:
| |
| | |
| A symmetric source X can be modelled with <math> f(x)= \frac1{2X_{max}}</math>, for <math>x \in [-X_{max} , X_{max}]</math> and 0 elsewhere.
| |
| The step size <math>\Delta = \frac {2X_{max}} {M} </math> and the ''signal to quantization noise ratio'' (SQNR) of the quantizer is
| |
| :<math>{\rm SQNR}= 10\log_{10}{\frac {\sigma_x^2}{\sigma_q^2}} = 10\log_{10}{\frac {(M\Delta)^2/12}{\Delta^2/12}}= 10\log_{10}M^2= 20\log_{10}M</math>.
| |
| | |
| For a fixed-length code using <math>N</math> bits, <math>M=2^N</math>, resulting in
| |
| <math>{\rm SQNR}= 20\log_{10}{2^N} = N\cdot(20\log_{10}2) = N\cdot 6.0206\,\rm{dB}</math>,
| |
| | |
| or approximately 6 dB per bit. For example, for <math>N</math>=8 bits, <math>M</math>=256 levels and SQNR = 8*6 = 48 dB; and for <math>N</math>=16 bits, <math>M</math>=65536 and SQNR = 16*6 = 96 dB. The property of 6 dB improvement in SQNR for each extra bit used in quantization is a well-known figure of merit. However, it must be used with care: this derivation is only for a uniform quantizer applied to a uniform source.
| |
| | |
| For other source pdfs and other quantizer designs, the SQNR may be somewhat different than predicted by 6 dB/bit, depending on the type of pdf, the type of source, the type of quantizer, and the bit rate range of operation.
| |
| | |
| However, it is common to assume that for many sources, the slope of a quantizer SQNR function can be approximated as 6 dB/bit when operating at a sufficiently high bit rate. At asymptotically high bit rates, cutting the step size in half increases the bit rate by approximately 1 bit per sample (because 1 bit is needed to indicate whether the value is in the left or right half of the prior double-sized interval) and reduces the mean squared error by a factor of 4 (i.e., 6 dB) based on the <math>\Delta^2/12</math> approximation.
| |
| | |
| At asymptotically high bit rates, the 6 dB/bit approximation is supported for many source pdfs by rigorous theoretical analysis.<ref name=Bennett/><ref name=OliverPierceShannon/><ref name=GishPierce/><ref name=GrayNeuhoff/> Moreover, the structure of the optimal scalar quantizer (in the rate–distortion sense) approaches that of a uniform quantizer under these conditions.<ref name=GishPierce/><ref name=GrayNeuhoff/>
| |
| <!-- I don't think that was proved by anyone else before it was done by Gish & Pearce in '68. For example, was it done by Koshelev in '63? (I don't think so) Zador in '66? (I don't know - probably not) Goblick & Holsinger in '67? (I don't see it in that paper.) -->
| |
| | |
| == Other fields ==
| |
| Many physical quantities are actually quantized by physical entities. Examples of fields where this limitation applies include [[electronics]] (due to electrons), [[optics]] (due to photons), [[biology]] (due to [[DNA]]), and [[chemistry]] (due to [[molecules]]). This is sometimes known as the "quantum noise limit" of systems in those fields. This is a different manifestation of "quantization error," in which theoretical models may be analog but physically occurs digitally. Around the [[quantum limit]], the distinction between analog and digital quantities vanishes.{{Citation needed|date=July 2009}}
| |
| | |
| ==See also==
| |
| * [[Beta encoder]]
| |
| * [[Bit resolution]]
| |
| * [[Discretization error]]
| |
| * [[Posterization]]
| |
| * [[Pulse code modulation]]
| |
| * [[Regression dilution]] - a bias in parameter estimates caused by errors such as quantization in the explanatory or independent variable
| |
| | |
| ==Notes==
| |
| {{Reflist|2}}
| |
| | |
| == References ==
| |
| *{{Citation |last=Sayood |first= Khalid|last2=|first2=|year= 2005 |title= Introduction to Data Compression, Third Edition |publisher= Morgan Kaufmann |isbn= 978-0-12-620862-7|doi=}}
| |
| *{{Citation |last=Jayant |first= Nikil S.|last2=Noll|first2=Peter|year= 1984 |title= Digital Coding of Waveforms: Principles and Applications to Speech and Video |publisher= Prentice–Hall |isbn=978-0-13-211913-9|doi=}}
| |
| *{{Citation |last=Gregg|first= W. David |year= 1977 |title= Analog & Digital Communication |publisher= John Wiley |isbn=978-0-471-32661-8
| |
| |doi=}}
| |
| *{{Citation |last=Stein |first= Seymour|last2= Jones|first2= J. Jay |year= 1967 |title= Modern Communication Principles |publisher= [[McGraw–Hill]] |isbn=978-0-07-061003-3|doi=}}
| |
| | |
| == External links ==
| |
| * [http://www.mit.bme.hu/books/quantization/ Quantization noise in Digital Computation, Signal Processing, and Control], Bernard Widrow and István Kollár, 2007.
| |
| * [http://www.techonline.com/community/related_content/20771 The Relationship of Dynamic Range to Data Word Size in Digital Audio Processing]
| |
| * [http://ccrma.stanford.edu/~jos/mdft/Round_Off_Error_Variance.html Round-Off Error Variance] — derivation of noise power of q²/12 for round-off error
| |
| * [http://www.ieee.li/pdf/essay/dynamic_evaluation_dac.pdf Dynamic Evaluation of High-Speed, High Resolution D/A Converters] Outlines HD, IMD and NPR measurements, also includes a derivation of quantization noise
| |
| * [http://www.dsplog.com/2007/03/19/signal-to-quantization-noise-in-quantized-sinusoidal/ Signal to quantization noise in quantized sinusoidal]
| |
| | |
| {{DSP}}
| |
| {{Compression Methods}}
| |
| | |
| {{Noise}}
| |
| | |
| {{DEFAULTSORT:Quantization (Signal Processing)}}
| |
| [[Category:Digital signal processing]]
| |
| [[Category:Computer graphic artifacts]]
| |
| | |
| [[Category:Digital audio]]
| |
| [[Category:Noise]]
| |
| [[Category:Signal processing]]
| |
| [[Category:Telecommunication theory]]
| |