Has anyone been able to define well or measure differences between vinyl and digital?


It’s obvious right? They sound different, and I’m sure they measure differently. Well we know the dynamic range of cd’s is larger than vinyl.

But do we have an agreed description or agreed measurements of the differences between vinyl and digital?

I know this is a hot topic so I am asking not for trouble but for well reasoned and detailed replies, if possible. And courtesy among us. Please.

I’ve always wondered why vinyl sounds more open, airy and transparent in the mid range. And of cd’s and most digital sounds quieter and yet lifeless than compared with vinyl. YMMV of course, I am looking for the reasons, and appreciation of one another’s experience.

128x128johnread57

Showing 16 responses by fair

@thespeakerdude

I see one major flaw in your logic. CD and two channel DSD is just that, two channels. When I am in a room, out in the wilds, or anywhere, there could be an infinite number of sound sources, that all contribute to that data you mention. When I am at home, there is only 2 sound sources. They may bounce off the walls, the floor, the windows, but there is only 2 sources. In another thread we are talking about ATMOS with 9, 11 or more speaker which still only simulates all that we can hear.

Use that 11 speaker example at CD data rates. The rate is 7.8 mbits / second. 11 speakers is not enough. 24? Now 16.8 mbits/second. Well beyond your 3.5 - 4 mbits/second.

I don't think you can correlate the data rate for the cochlea with the brain, which I suspect is a WAG, from sound information that comes from all directions, with what comes out of 2 speakers.

How much have you experimented with binaural recordings? Done right, at 24/192, they provide convincing illusion of being there. Until listener turns his or her head.

Turning the head, moving it, standing up and moving body around, going to an adjacent room, and so on. Those of course break the illusion.

Yet this is an orthogonal consideration. Naturally, physical movement and physical action may change what the listener hears, with all else staying same.

What I was discussing wasn't Complete Illusion of Being There. That would heavily depend on the degrees of freedom the listener possesses.

For instance, let's restrain the listener to only rotating the head 60 degrees left and 60 degrees right. Then we'd need to increase the amount of information 121x, in a brute force approach.

121 variants of binaural recording made for this particular listener with the rotation resolution of one degree would maintain convincing illusion of still being there, as his or her head rotation is tracked.

There are ingenious compression schemes cutting down the amount of information that needs to be recorded in such case, yet, as with any lossy format, one must carefully think about compression artifacts elimination. 

What I was discussing is rather different: the amount of audio information that needs to reach each ear, every second, such that further increase of this amount can't change what the listener perceives.

If there isn't enough information - because it just can't be encoded in a given format - then there is a possibility of the listener noticing the artifacts beyond those inherent in the audio setup. The artifacts may ruin music enjoyment.

Yet another relevant consideration is that high amount of information may not even be contained in a second of a specific piece of music. "A girl and a guitar" and "full symphonic orchestra" have quite different information-generating capabilities.

In this context, I claimed that CD format is insufficient for capturing full information inherently transmittable by stereo setup, whereas stereo DSD128 and PCM 24/192 formats are sufficient.

So, discussion of Analog vs Digital ought to take into consideration what is meant by "Digital". It is true regarding "Analog" as well of course, yet the context of this discussion was clear on that, the Analog being stereo LP format.

@cleeds

We know that CD has a greater potential dynamic range than LP but in practice, the opposite is often the case.  Just look at the DR database.

"Dynamic range" has multiple definitions even in professional sense. For starters, for a given recording one can derive large number of dynamic ranges based on even a single parameter of averaging over time.

Then, if we go deeper, to psychoacoustics of music perception, we may start discerning different large sets of dynamic ranges for large number of frequency ranges, for "standard ear".

Going deeper yet, we ought to take into account individual hearing systems differences. A typical teenager, for instance, may discern a wider dynamic range at 15KHz compared to a typical retiree.

The standardized procedures are indeed useful, as they give general idea about the dynamic range of a given recording, allowing to compare different recordings in this regard, yet those are gross simplifications.

Depending on the method of dynamic range measurement, most suitable in a certain sense for a given music piece and listener, ether CD or LP may be "proven" to have a greater potential dynamic range.

@tomcy6

A blind analog/digital test recently came to light which caused a great deal of consternation among the analog contingent of our hobby and brought doubt.to some of their claims of what they can hear.

It was recently discovered that MoFi Labs had a digital step in the mastering - pressing chain of their Lps going back at least to 2011 and maybe even further back.

MoFi found that record labels were often not willing to loan master tapes out to them, so they put together a portable Studer tape deck that they could take to the record label vaults to make copies of the master tapes that they then used to make their MoFi Lps, including the very expensive one-steps.

MoFi started with the analog master tapes but they were recording them to DSD, plain old DSD in some cases but 4x DSD in most cases. Audiophiles bought these Lps for over a decade and loved them. There was the rare voice here and there that didn’t like them, but no more than with any album no matter how pure its lineage. Michael Fremer had a number of them on his 100 best sounding Lps list.

Thousands and thousands of analog listeners could not tell that the MoFi Lps had been produced from a digital source even after many listens over a period of years on their own systems.

So, MoFi definitely should have been upfront about the source for its Lps, but they weren’t, and no one could tell. I’m not saying that there aren’t differences between analog and digital, but there may be factors other than sound quality involved for those who find digital fundamentally flawed, In My Humble Opinion, YMMV..

Interesting story indeed. Thank you for sharing.

Plain DSD format, otherwise known as DSD64, uses 1 bit at 2.8224 MHz. Thus the density of recorded information is 2.8224 MegaBits per second.

CD uses 16 bits at 44.1 KHz. The corresponding density is 16 x 44.1 =  705.6 KiloBits per second = 0.7056 MegaBits per second.

2.8224 / 0.7056 = 4.0. Thus, DSD encodes 4x amount of information per second compared to CD.

DSD 4x, otherwise known as DSD256, encodes 16x the amount of information per second compared to CD. 11.2896 MegaBits per second. That's quite a difference.

Why is that important? It is important because research of human hearing system resulted in understanding that the rate of flow of information from cochlea to brain is between 3.5 MegaBits per second and 4.0 MegaBits per second.

Auditory processing circuits in the brain drastically compress this flow of information: this explains why lossy encoding works so well. Still, if some part of the original 3.5. to 4.0 MegaBits per second flow is arbitrarily removed, artifacts may occur.

As we can see, CD falls about 5.7x short of the target of complete digital transparency. DSD64 falls about 1.4x short.

DSD128 encodes more information, by factor of 1.4x, than the nerves running between cochlea and brain can transport. DSD256 exceeds the sufficiency threshold by factor of 2.8x, and thus shall be considered far more than enough.

How does this compare to other common PCM formats? Let's see.

24 bits x 48 KHz = 1.152 MegaBits per second. Falls short of sufficient 4.0 by factor of 2.88x.

24 bits x 96 KHz = 2.304 MegaBits per second. Falls short of the sufficient 4.0 MegaBits per second by factor of 1.44x. About the same in this regard as DSD64.

And finally, 24 bits x 192 KHz = 4.608 MegaBits per second. This exceeds the sufficiency threshold. Fittingly, this is the predominant format professional studios use for mixing the most complex recordings.

So, two of the most commonly used formats, DSD128 and PCM 24/192, can be considered as transparent to human ear as any analog format can possibly be.

The real tragedy of the 20th century is that so much music was published in the insufficient for full transparency format: CD of 16 bits x 44.1 KHz.

 

@teo_audio

The place it counts is in the micro expression of transients and micro transients and the differences in level and timing between them.

This sums it up well.

This is where digital and class d falls apart. Those are the points of greatest distortion, in digital and class D.

I would agree, if "digital and class d" meant "mass market digital and mass market class d". Highest-end digital and class d are much harder to differentiate from highest-end analog.

In science, things are supposed to correlate to the situation at hand. Do you understand the question? Is the measurement relevant to the question at hand? If not, go back to the start and have a go at it again. Even when done, keep questioning the results and facts don’t exist..so that all finalized things can be gone over again and altered according to new results on the complex investigation of it all. That’s science.

Exactly. That's what I was pointing out to certain ASR folks. If a theory doesn't fit facts, keep working on the theory, instead of rejecting facts out of hand.

Engineering is specifically NOT exploration, engineering is designed for building things that work and use scientific theories turned into scientific law. Law...Which is a falsehood built for the engineering trade and training within it, for linear minds which are principally dogmatic in form and function.

I see it differently. Not a falsehood, but a model simple enough to be applicable in economical way to a day-to-day engineering.

In audio, the measurement and the analysis is wrong, just plain wrong. Too many engineering minds on the job, trying to play it safe and keep things ordered & black and white.

Measurements are measurements. If they are done competently, with calibrated instruments, and only what is actually measured is claimed, I'm happy to use them.

Analysis is a different story. Analysis always presupposes a theory, or at least a paradigm. And this I consider too rigid in the current mainstream audio.

This is why the audiophile conundrum has existed for about 50 years. The ignorance of projection in the pundits that surround the engineering trade and ideals that are involved in the audio world. Interference (engineers from other areas) from outside audio (even more ignorant!!) helps keep the insanity frothing along nicely.

There are other reasons for relative ignorance of the hearing system fundamental properties among practicing engineers. One of them is that not all relevant knowledge is even discovered yet. Another is that some very relevant knowledge was discovered relatively recently, and practicing engineers weren't taught it.

To clarify, an engineer is not trained to commit to the scientific method or invention, they are trained to follow the books, as that is why they are engineers, not scientists who explore and change things as required when required.

Agree. People like me, trained as scientists, are often perceived as "irreverent" in regard to dozens of audio engineering handbooks published over several past decades. Most engineers (not all) take doubting certain things written in these handbooks as a manifestation of sheer stupidity.

Meanwhile, a whole parallel world of peer-reviewed audio science publications exists. It is instrumental to observe how drastically the theories changed over the past 50 years, prompted by more and more sophisticated experiments, and breakthrough discoveries in the field of mammalian audio system physiology. 

If you want to explore in formal sense, go back to school and get trained to see all as theories, which are subject to change from/on new data, tests and proofing, correlation, etc. Get trained as a scientist.

Not practical for most practicing engineers. The change will only occur gradually, as older generations retire and new ones are taking their place.

When this mess erupts into fully blown projections in insanity of following the dogmatic rule books of engineering, we end up with things like ASR.

I like pretty much all ASR measurements. What I don't agree with is some of the analysis they derive from the measurements. ASR crowd is very uneven: there are bone fide luminaries posting there, and also folks who keep scoring points for slighting others. Guess who ends up with more points?

The longer a problem sits unsolved, unresolved.. the more fundamental the error in the formulation of the question.

Agree. As an example, Ptolemaic System was generally believed to be true for about 14 centuries.

Thus, the audiophile conundrum is deeper than this surface level stuff that people generally think it is. It’s deep in the minds involved, regarding how they explore reality.

Indeed.

As long as dogmatic minds try to figure out what is wrong in audiophiles vs measurements, without moving to true and proper scientific method...the longer they’ll be spinning around and getting no real correlating clarity in any of it.

I'd say the truly dogmatic minds don't even try to figure out what is wrong. They just reject the facts as aberrations, just like later-centuries Ptolemaic scholars ignored observed deviations in planets movements not explainable by their preferred theory.

Let's dissect thinking behind ignoring one of such facts in audio: certain types of music, for instance classical symphonies and gamelan, tend to not sound right when published in CD format.

What is usually offered as grounds for rejecting such statement? The Sampling Theorem and one of the ways to calculate dynamic range of a digital format.

 

The Sampling Theorem (https://en.wikipedia.org/wiki/Nyquist–Shannon_sampling_theorem) reads in its original edition:

If a function x(t) contains no frequencies higher than B hertz, it is completely determined by giving its ordinates at a series of points spaced 1/(2B) seconds apart.

This theorem is often taken as "proof" that sampling frequency of 44.1KHz is sufficient for encoding any meaningful music signal. Because, "obviously", everything above 20 KHz can't be heard by humans, and thus is not worth encoding.

Let's look closely. What does "contains no frequencies higher than B hertz" actually mean? It means, using formulation in same Wikipedia article, that "Strictly speaking, the theorem only applies to a class of mathematical functions having a Fourier transform that is zero outside of a finite region of frequencies."

Do analog signals corresponding to practical music pieces have Fourier transform that is zero outside of a finite region of frequencies? Absolutely not! Because, as another theorem from Fourier analysis proves, only functions of infinite duration can have such Fourier transform.

Let it slowly sink in. The Sampling Theorem, strictly speaking, is not applicable to analog signals corresponding to practical music pieces. But, obviously, some form of Fourier transform is widely used in audio digital signal processing. What's going on here?

What is actually being used, in discrete form, are variations of Short-Term Fourier Transform.

A fragment of a signal, let's say with a duration of 20-25 milliseconds, is taken, then multiplied by a so-called "smoothing window". The resulting function of time is guaranteed to smoothly start as 0, and smoothly end as 0.

Then the signal is mathematically virtually replicated infinite number of times. Since it is now of infinite duration, the Fourier Transform result has limited range of frequencies.

Then the process repeats with another fragment of the signal, starting at 10-12.5 milliseconds later than the previous piece. For the purposes of digital filters, this process sometimes virtually repeats with shift of just one digital sample duration.

So, in practical applications, digital signal processing uses an approximation of Fourier Transform. Correspondingly, the Sampling Theorem only works approximately. Most of the time more than well enough. Sometimes not at all.

 

Now let's consider the issue of sufficient dynamic range. It is oft-cited that CD format has dynamic range of 96 dB. Let's see, approximately, how one could come to such conclusion. 1 bit corresponds to 6 dB SPL. So, "obviously", 16 bits correspond to 6 dB x 16 = 96 dB.

According to https://hub.yamaha.com/audio/music/what-is-dynamic-range-and-why-does-it-matter/:

As a group, classical recordings have the widest dynamic range of any genre. The same study cited above found that recorded classical music typically offers between about 20 dB and 32 dB of dynamic range. While that might seem like a lot, it’s still quite a bit smaller than that of a live symphony orchestra performance, which can be as large as 90 dB.

Technically, those are good news, aren't they? Live symphony orchestra dynamic range is 90 dB. CD dynamic range is presumably 96 dB. 90 < 96. So, CD should be able to reproduce the whole dynamic range of a symphony orchestra, right?

In practice though, we have those presumably stupid music producers and audio engineers, who fell to the Dark Side during the Loudness Wars, and who will only record classical music CDs with 20 dB to 32 dB of dynamic range.

What would happen if they attempted the 90 dB? They would need to allocate 90 / 6 = 15 bits to the dynamic range encoding. For the quietest sound, they'd only be left with 1 bit for encoding it. Wait, what?

Yes, imagine a quiet passage in a symphony, nevertheless involving a dozen of instruments, each with a complex multi-overtone spectrum. With frequency slides, amplitude rides, tremolos etc. All of this would need to be recorded with just one bit at 44,100 Hz!

This is exact equivalent of DSD encoding, only its frequency is 64 times lower. Correspondingly, the highest frequency that we can hope to encode with similar fidelity as DSD will be 44,100 / 2 / 64 = 344.5 Hz. Say goodbye to the "micro expression of transients and micro transients"!

How much different is what the audio engineers are actually doing? Let's say they decided to limit the dynamic range to 30 dB. This corresponds to 30 / 6 = 5 bits. This leaves 11 bits to encode the quietest part of the symphony.

How good are 11 bits? 2 to the power of 11 is 2,048. 1/2,048 = 0.00049. An average digitization error would be half of that, which is 0.025%. Interesting, it is just below the widely accepted threshold of THD defining a hi-fi power amplifier, which is 0.03%.

This is not a coincidence. If they'd allocated less bits for the quietest parts of the signal, they would hear noise and distortions in them, similarly to how they'd be able to hear noise and distortions introduced by a low-sound-quality power amplifier.

If they's wanted to go audiophile quality for the quietest passages, they'd need to up the ante 3 bits more, to 14 bits. so that digitization noise and distortions would be approximately equal to that of a high-quality DAC, and thus would be likely unnoticeable even on a high-quality professional headphones.

So, for faithful reproduction of a symphony we would need 90 / 6 = 15 bits for encoding the dynamic range, and 14 bits for encoding the shape of the signal. 15 + 14 = 29 bits. Uh-oh, but professional ADC and DAC only encode 24 bits? How could they manage to effectively push to 29?

And here we come to understanding of why arguably overkill digital formats are desirable. The seemingly excessive amount of information per second inherent in 24/192, DSD128, and especially in 24/384 and DSD256 can be divided between encoding the dynamic range, encoding the shape of the quietest signal, and sampling the signal frequently enough to capture its evolution over shorter periods of time.

How all of the above relates to the current thread theme? By the virtue of analog recording and reproduction system, which in principle doesn't place fundamental limits, other than noise and maximum acceleration of mechanical parts, on either effective bit depth or sampling frequency.

It is commonly accepted that the best analog systems have about 70 dB of dynamic range. Which would roughly correspond to 70 / 6 = 12 bits. This gives an excuse for proponents of CD superiority over LP to claim that this must be so because obviously 16 > 12.

However, in order for the quietest signal to be still distinguishable, it only needs to be 6 db, or 1 bit, above the noise floor. This leaves an equivalent of 11 bits for dynamic range, which is more than twice of the 5 bits of the usable CD dynamic range.

Instead of the last 1 bit with which to encode the shape of the quiet signal, a high-quality analog system has many more. Actual number depends on the analog media granularity and its speed, yet the most important fact is that there is no hard stop similar to the one a digital system would have.

So, the analog system would reproduce the quiet passages in higher fidelity signal-shape wise, superimposed with noticeable noise of course. Yet the human hearing system is capable of filtering out this noise at higher levels of processing in the brain, and enjoying the quiet passage hidden underneath.

Viewed from this perspective, LP has twice as wide usable dynamic range in comparison with CD. But higher noise and distortions. For classical music especially, this could be a desirable compromise. For some other genres, for instance, extremely-narrow-dynamic-range very-simple-signal-shape electronic dance music, CD could be preferable.

I would expect a classical recording made by multiple microphones sampled at 24/384, or even at 32/384, and delivered in DSD256 after careful mixing and mastering, to be the ultimate one for the time being. As I recall, they produce such recordings in Europe.

 

@thespeakerdude

... These theories are tested day in and day out ...

... They do what they do accurately, understanding their limitations ...

I agree with these two statements of yours. What I was referring to are approaches where limitations of the classic theory were disregarded. Let me try to explain from a different angle what I meant, using a concrete example.

Imagine an audio signal which is a sinusoid of frequency 12 KHz, with amplitude described as piecewise function of two segments linear on dB scale. First segment goes from 0 to 100 dB SPL during first half cycle of the the sinusoid. Second segment goes from 100 db SPL to below quantization noise during the next four cycles of the sinusoid.

Try to sample it with 16/44.1. Then try to reconstruct the signal from the samples. Then shift capture time of the first sample by 1/8 of the sinusoid period. Repeat the exercise.

What you’ll find is that, first, reconstruction will be pretty rough, and second, that it will be wildly changing with even small shifts of the first sample capture time.

From Fourier Analysis viewpoint, this is an example of a signal with spectrum extending significantly beyond 20 KHz, which makes sampling at 44.1 KHz untenable, and result of reverse transform unpredictable.

Yet from human hearing system standpoint, such a signal is perfectly valid, and will result in physiological reactions inside several inner hair cells. Most likely, if it manages to evoke a sensation of pitch in a particular individual, perceived pitch frequency will be close to the intended 12 KHz.

An analog system doesn’t care about the sampling frequency, and at what precise moment of time the first sample happens to be taken, and would capture this signal fully, with some distortions of course, yet nevertheless it will capture the shape definitively. And it will be reconstructed definitively as well.

Imagine further, that some short time later, another signal comes in, which is exact reversal of the first one.

Depending on the time difference of the signals start, the sampled values of the second signal may range from exact opposites of the first set of sampled values to something seemingly unrelated.

Once again, human hearing system, with its half-wave rectification capability, will react to the second signal in a similar way it reacted to the first. And once again, the analog system, not restrained by sampling and time shift considerations, will capture the second signal fully.

If, on the other hand, we significantly increase the piece-wise linear segments duration: let’s say first segment goes up for 100 cycles, and the second one goes down for 1,000 cycles, then the 16/44.1 sampling with consequent reconstruction will produce much more agreeable result.

So, I gave an example of a signal which is meaningful and definitive both from the hearing systems and analog recording standpoints, yet non-definitive from the digital sampling standpoint.

Also, an example of a signal with the same general shape, yet with different duration of its characteristic segments. Which happens to be both meaningful and definitive from all three standpoints.

Which illustrates the limitations of digital sampling and classic Fourier-analysis-based DSP: they work well enough in most practically encountered cases, yet not always.

In contrast, analog may be worse in most cases in terms of distortions and noise, yet it works consistently in all practically encountered cases, which may be important for recording and reproduction of certain genres of music.

Increasing the sampling rate effectively rescales the problem: certain signal fragments and components which couldn’t be perceptually transparently captured at a lower sampling rate are now captured well enough at the increased sampling rate.

At the limit, sampling at increasing rate becomes perceptually equivalent to analog recording, sans the distortions and noise. At which point does it happen? It depends greatly on the characteristics of music, and on critical listening abilities of the person who tries to enjoy that music.

Correspondingly, the highest frequency that we can hope to encode with similar fidelity as DSD will be 44,100 / 2 / 64 = 344.5 Hz. Say goodbye to the "micro expression of transients and micro transients"!

I am going to highlight this last paragraph. This is 100% false. That is not how DSD works. The single bit in DSD is not equivalent to a single bit change in PCM. No direct comparisons can be made. Hence you conclusion cannot be made and can be assumed false.

Let me clarify. I wrote "encoded" meaning that we could use the remaining still available stream of one-bit values to encode in the same way that DSD does. Of course bits are used differently by PCM and DSD - pulse vs delta etc.

That was to illustrate the point that the amount of information per second remaining available, in the case if we’d decided to use 15 bits for encoding of dynamic range, is indeed equivalent to a very low-fi format.

There are two flaws in your statement of equivalence 11 bits and 0.03% distortion detection. More like 3 flaws. That distortion limit is at full scale.

To understand what I meant, look at the physical bits of the quietest in this context PCM-encoded signal. All the upper bits, which I called "used for encoding dynamic range", will be zero.

It is not that these specific bits of PCM stream would be always used for encoding dynamic range. What counts is the number of bits that we have to keep unused while encoding the quietest segment of music.

Secondly, please take into account that human hearing system is capable of adjusting its sensitivity, and symphony composers tend to use this factor fully.

The symphonies typically have quiet segments, when a neighboring spectator shuffling her purse may be pretty distracting, and they also have short bursts of apotheosis, with SPL falling just short of hearing system pain threshold.

In the context of a quiet segment, the perceived distortion level threshold is scaled down. That’s why I do indeed consider it as if it was a full-scale signal.

There are other factors of course: e.g. the equivalent loudness curve shifts.Yet if we only consider the most stable part of the curve, at mid-frequencies, the rule-of-thumb calculations generally work, plus-minus a bit.

Assume your stereo is set for 100db peaks, which is fairly loud and you have low distortion playback. There is a particular distortion level evident at that volume. In your analysis, you are claiming to be able to hear distortion at the bit level, on sounds that are only 70db. Are you claiming to be able to hear 0.03% distortion on a 70db peak signal.

That would depend on nature of the music fragment, right? And on my hearing ability. In general, I didn’t claim anything of the sort. Only that, as an order of magnitude estimation, an amp with 0.3% THD is usually considered low quality, an amp with 0.003% THD very high quality. The middle on logarithmic scale: 0.03%, was considered in enough accounts I found credible as a threshold of quality.

Further, CD is dithered. Dither improves the dynamic range where our hearing is most sensitive for added noise where it is not. That extends the dynamic range to where we are most sensitive to 110db. Your argument fails with that information.

Dithering is helpful in most practical cases. Yet, if you look at the mathematical derivations of the common dithering schemes, you’ll see that the characteristic duration of signal stability is a factor in calculations.

Similarly to the examples I gave earlier in this reply. If a signal is composed of slowly changing sinusoids, dithering helps a lot.

It a signal consists mostly of harmonic components quickly changing their amplitudes, non-harmonic transients, and frequently appearing/disappearing components, dithering is not as effective.

>>> So, for faithful reproduction of a symphony we would need

>>> 90 / 6 = 15 bits for encoding the dynamic range, and 14 bits for

>>> encoding the shape of the signal. 15 + 14 = 29 bits.

This is obviously not at all accurate. You are stacking flaws in your understanding of how digital works to come to incorrect conclusions.

I believe at that point I provided enough explanations. Your reactions are quite typical of engineers who consider the classic DSP based on Fourier Analysis the only true paradigm.

From my perspective, it is only absolutely true for abstract mathematical constructs.

It is nothing but useful approximation of real world. One ought to be very careful with the corner cases, where the abstractions stray too far away from the phenomena they are supposed to model.

The digital bit depth only needs to be large enough to encompass the full dynamic range.

As I highlighted, the approach you are advocating doesn’t address the need of having some bits left available for encoding the shape of signal faithfully enough to be perceived as distortions-free.

The theory I use explains well enough why the so-called Loudness Wars can be considered a rational, professionally responsible, reaction to deficiencies of the most widely used at the time audio recording format - CD.

This theory explains why some listeners still prefer listening to LP for some genres of music, despite the fact that, according to the classic theory, CD is vastly superior. Once again, this is a rational and responsible reaction.

The theory explains with good enough for me personally precision why most professional sound mixing and mastering studios didn’t advance beyond the 24/192 PCM format.

It also explains why some modern symphony recording engineers moved to 24/384 and DSD256 formats. And other otherwise unexplainable for me phenomena.

By shifting noise, we don’t even need that many bits for the dynamic range. DSD has 1 bit depth. The noise is shifted to provide large dynamic range. CD has 16 bits. The noise is shifted to increase the dynamic range.

DSD is a delta format. Formally, general DSD has unlimited bit depth, and thus dynamic range. It is only constrained in specific versions of the format to correspond to a set bit depth at an PCM-equivalent sampling rate.

The noise considerations started to amuse me lately. Practical examples were a trio of class-D power amplifies, highly regarded by ASR. I bought them over the years, evaluated, and quickly got rid of, due to intolerable for me distortions.

Yet SINAD of these amplifiers was excellent. Which made me look closely at SINAD measurement procedures. Long story short, SINAD is predicated on taking Fourier transform over a very long window, of a signal comprising of a set of sinusoids with equal and unchanging amplitudes.

Where all three failed miserably for me was reproduction of low-signal-level transients, something SINAD doesn’t capture all that well. Yet the theory I use explained their behavior rather precisely. It also predicted what power amplifiers would be more acceptable to me.

>>> However, in order for the quietest signal to be still

>>> distinguishable, it only needs to be 6 db, or 1 bit, above the noise

>>> floor. This leaves an equivalent of 11 bits for dynamic range,

>>> which is more than twice of the 5 bits of the usable CD dynamic

>>> range.

You are basing this conclusion on a stack of fundamental flaws. It does not represent reality. More accurate is that we can hear below the noise floor.

It depends on the nature of noise and nature of signal, doesn’t it? For white noise and a short sinusoidal burst, I’d agree with you. I’m more interested in a typical music signal, with spectrum close to pink noise, masked by pink noise. In that case, having it 6 dB over the noise floor results in more reliable perception.

>>> Viewed from this perspective, LP has twice as wide usable dynamic

>>> range in comparison with CD. But higher noise and distortions.

This is also based on a stack of flawed assumptions. It is incorrect.

Not on assumptions. On theories. Fitting experimental facts. The theory I use is more sophisticated than the classic one, taking into account analog characteristics of human hearing system.

On its simplest level, instead of considering just dynamic range, it also considers the shape of what the dynamic range is applied to. Once this is done, preference for LP in certain situations ceases to be a mystery.

Cochlea is not a Fourier transforming machine. In some regards it is more crude, yet in others it is far more advanced. As an example, it starts noticeably reacting only after observing two cycles of a pure sinusoid, virtually irrespective of frequency.

For higher frequencies, at 44.1 KHz sampling rate, this may correspond to only a few samples. The shape of a quickly changing signal can’t be faithfully captured by such small number of samples.

Once we get into signals comprised of quickly appearing and disappearing components, the simple intuition good enough for the previous example no longer works, and math becomes much heavier, yet fundamentals remain: the higher the sampling rate (assuming equal quantization accuracy), the deeper the bit depth (assuming equal timing accuracy), the better it gets.

And yes, I’m aware of the oversampling nature of practical ADC and DAC. Of the fact that internally they are sampling/reconstructing signal at significantly higher rates, and then encode adjustments not only into the slower-sampled values within the signal time range, but also outside it.

Still, Information Theory is a bitch. If there isn’t enough bits to encode the changes in the signal that would be noticed by cochlea, some meaningful information would be lost. I did some experiments on fragments of music that I recorded and mixed myself. The distortions of 16/44 compared to 24/192, albeit subtle, mostly manifested themselves as uneven rhythm of smaller-volume transients.

 

@mahler123

With all this esoterica being discussed, how does one account for the phenomena of the fact that most lps these days use digital files, and vinylista think they sound great, as long as they don’t know the truth?

Thinking in terms of LP vs Digital dichotomy doesn't explain it. More nuance is needed. At the very least, the Digital formats needs to be split onto Below The Transparency Threshold and Over the Transparency Threshold.

The Transparency Threshold depends on nature of music and hearing abilities of listener. It is different between typical pop music and untrained listeners vs classical symphonies and professional musicians or sound engineers.

Since this is an audiophile forum, I prefer to talk about the High Transparency Threshold. Here I deliberately use terminology different from official, e.g. High Definition Audio, to keep it free from marketing attachments.

In my previous posts, I gave hints as to why I believe CD format is significantly below the High Transparency Threshold, whereas 192/24 PCM and DSD128 are slightly above it. Going into even more detailed technical discussions here does not appear fruitful.

So, as I understand it, if an LP is pressed from a digital format that is over the High Transparency Threshold, it ought to sound no different compared to one pressed from an analog studio master of the same recording.

 

We are getting somewhere.

@thespeakerdude

From where I am sitting you have not provided one explanation because every single explanation or example you have used is wrong, stacking misunderstanding on top of misunderstanding.

This is precisely how it should feel, from the point of view of someone remaining in an old paradigm. New paradigm overturns some of the old paradigm's assumptions and conclusions, which is obviously "wrong" in the context of the old paradigm.

Fourier analysis is not a paradigm, it is a mathematical translation from time to frequency, it just is.

Mathematically, Fourier Analysis is a theory based on integral transforms with harmonic kernels. "Integral" means that there is an integral involved, calculated from low boundary of integration to high boundary of integration.

Direct Fourier Transform takes bounds from time domain. Reverse Fourier transform takes bounds from frequency domain. Time domain and frequency domain can be, as classes of specific cases, continuous or discrete.

This theory is beautiful in its simplicity. For instance, formulas for direct and reverse transforms, in their traditional formulation, only differ in one sign in one place. The simplicity affords efficient implementation of the transforms in computer code.

The accuracy, as I previously wrote, is based on suitable bandwidth limitations, and appropriate windowing functions, much which occur naturally in audio, but are still supplemented by the appropriate analog filters, over sampling, and digital processing.

And here we move away from the theory and arrive to a paradigm. The "suitable bandwidth limitations" remove information contained in original analog air pressure variations over time at the point of recording.

Central belief of the paradigm states that removal of frequency components beyond 20Hz and 20 KHz is perceptually benign, for all types of music, and all listeners. Technically, this is the crux of our disagreements. I do not subscribe to this belief.

People are not just guessing at the implementation and not considering what the underlying waveforms can and do look like. Let me break just one section down to illustrate your logic flaws and misunderstandings. It carries through to the rest of what you have wrote:

You start with a flawed premise, proceed to a flawed understanding of digitization, and finish with an incorrect understanding of reconstruction.

I value your opinion. Couldn't asked for a better illustration of what I had to endure in my prior discussions at ASR.

However, my professors, from leading European universities, and their teaching assistants, had other opinions, giving me straight As on all courses related to Fourier Analysis and DSP.

The theory I'm using today contains the classic Fourier Analysis, and classic DSP based on it, as subsets. I absolutely do use them in domains of their applicability, when I believe they are going to provide accuracy sufficient for a task at hand.

Yet there is more, which came mostly from research conducted by others over past three decades. Unfortunately, too much of it is still widely dispersed in numerous peer-reviewed papers, rather than concentrated in a few engineering handbooks.

Flawed premise: 12 KHz sine wave do not suddenly appear, starting at 0. As I previously wrote, we are dealing with a bandwidth limited and defined system. You cannot go from 0, silence, directly into what looks exactly like a sine wave. That transition exceeds the 20KHz (or whatever we are using). Also, the digitizer, filters, etc. will have been running and settled to required accuracy by the time this tone burst arrives. Whatever you send it, will have been limited in frequency, by design, by the analog filters preceding the digitizer.

When one writes enough code processing real-life music recorded with high enough fidelity, one absolutely starts believing that such music components do exist: going from zero to almost pain threshold in a matter of microseconds, and then rapidly decaying.

One of the best examples of music genres rich in such components that I know of is Indonesian Gamelan. It is a curious genre: worshiped by its devotees in native land, and almost completely ignored by listeners outside the region.

Even the best gamelan CD recordings of famous Indonesian ensembles sound to me like incoherent early practices. Live, classical gamelan compositions, played with passion by experienced musicians, sound heavenly to me.

Flawed understanding of Digitization: As written above, the digitizer was already running when the tone burst arrives. Whether the sample clock is shifted globally the equivalent of 1/8 of a 12KHz tone, or not, will have no impact on the digitization of the information in the band limited analog signal.

This depends greatly on the nature of the band-limiting filter used. For analog filters this statement is generally true, with understanding that perfect brick wall filters don't exist, so there are still some smaller artifacts to be expected because of that. For digital filters applied to stream of oversampled values in some ADC devices, not so much. 

Flawed understanding of reconstruction: When I reconstruct the analog signal, using the captured data, whether I use the original clock, or the shifted one, the resulting waveform that results will be exactly the same. In relationship to the data file, all the analog information will be shifted by about 10 useconds. That will happen equally on all channels. The waveforms will look exactly the same either case. One set of data files will have an extra 10 useconds of silence at the front of them (or at the end).

See my comment above. Depends on the nature of bandwidth-limiting filter.

I am sure you believe this, but you used flawed logic, a flawed understanding of the waveform, and a flawed understanding of digitization, reconstruction, and the associated math.

I don't believe so. I used more sophisticated understanding of those. Knowing, both from learning the theory and from practical experience, that absolute predicted perfection isn't practically achievable, and that one needs to very carefully look at what artifacts are produced by this or that digitization method, and whether the artifacts may be heard under certain conditions by certain listeners.

I went back and looked looked at the research. In lab controlled situations, humans can detect, a very specific signal up to 25db below the noise floor, A-weighted. That is not listening to music, that is an experiment designed to give a human the best possible chance. For vinyl, that means in a controlled experiment, maybe you could hear a tone at -95db referencing 0db as max. With CD, the same would be true at -110db (or more) due to the 100% use of dithering.

Research on masking of signal by noise is kind of 101 of psychoacoustics. What we already established is that I'm more interested in how practically encountered noise is masking practically encountered quiet music passages. I do realize that masking thresholds for specially constructed signals and noise patterns may be different. 

To be sure we are on the same page. Class-D amplifiers are analog amplifiers. They are not digital. I will correct you. Perception of distortion. You are making an assumption of something that is there, without proof it is there.

Yes an no. Implemented with analog circuitry, yes. But, at some point inside a class-D amp analog signal is transformed into a sequence of discrete +V and -V segments, starting and ending at analog time boundaries.

So, it is kind of a hybrid. Analog in time domain throughout, discrete at an intermediate stage in amplitude domain. Not what most people would call classically digital, yet not quite purely analog either.

Which theory is it that you are using? I noted many flaws in your understanding of critical elements of digital audio, and assertions that are also incorrect. I have already falsified your theory.

I see it differently. The old paradigm is falsified by phenomena for which it gives invalid predictions. For instance, according to the old paradigm, LPs shall be long gone, the way of cassette tape recorders and VCR video tapes. Yet LPs persisted, and the classic paradigm produces no convincing explanation as to why.

New paradigm not only explains why LPs persisted for so long, but also specifically predicts what they'll be replaced with. To repeat once again, to the best of my understanding, eventually they'll be replaced by digital recordings with information density equal to, or higher than, those of PCM 192/24 and DSD128 formats.

Perhaps not important to this discussion, but 16/44.1 is a delivery format. From what my colleagues tell me, is has not been used as a digitization format in decades, and depending on your point of demarcation, it has not been used as a digitization format since the 1980’s, as all the hardware internally samples at a higher rate and bit depth.

Yet another phenomenon not explained by the old paradigm.

According to the old paradigm, 16/44.1 format shall be sufficient for capturing and delivering any type of music, yet in practice all those "pesky producers and sound engineers", for some mysterious reasons, want to use digital formats providing higher information density.

The new paradigm not only qualitatively explains this phenomenon, but also accurately describes the numerical parameters of the formats that were found sufficient by trial and error by many highly qualified practitioners.

The new paradigm also explains why gear providing even higher information densities, easily available these days (e.g. I own several nice 32/768 ADC/DACs), isn't as widely used at its highest settings in practical music production.

 

 

@thespeakerdude

If there is any interest, this is probably the best single article I have discovered that explains digital audio. It is not light reading nor heavy reading. Dan, who put it together obviously put a lot of time into it. It is almost 20 years old so comments about processing power are no longer relevant, but everything else is. I have come across many articles on digital audio written by less technical people. They get the basics right, but they often make mistakes and they never go into the depth that this article provides. You may need to read it 2 or 3 times to understand well enough, but if you do, it will dispel a lot of misconceptions about digital audio.


https://lavryengineering.com/pdfs/lavry-sampling-theory.pdf

If you have any questions about the article I will try to answer them.

Thank you for referring this document. Oldie but goodie. Excellent for illustrating the older paradigm vs newer paradigm.

Let’s look at the graph there marked with "Let us begin by examining a band limited square wave". That’s what I meant by saying that quickly changing signals start looking ragged when band limited. The document goes into a detailed explanation of why this is happening. For a briefer explanation, one can peruse a Wikipedia article about Gibbs Phenomenon.

Note that what we see on a square wave is an extreme example. The underlying mechanism of the Gibbs Phenomenon is in action on any harmonic signal with changing magnitude - just to a lesser degree, depending on ratio between characteristic time of the harmonic components magnitude change and sampling interval.

The concentrated difference between the older and the newer paradigm is this:

Subscribers to the old paradigm believe that the wiggles we see on charts like that don’t ever affect perception of sound quality, as long as the signal to be band-limited is "music", and the upper boundary is set at 22 KHz.

The new paradigm tells us that it depends. That certain wiggles may affect perceived sound quality of certain music signals band-limited under the conditions above, for certain listeners.

By the way, 0.1% THD corresponds to a width of one pixel on a typical laptop display, if a graph like that is enlarged to fill the whole screen.

Basically, we can hear a difference that we can barely see on a graph.

If one sees any visual difference of a band-limited music signal compared to the original one, this should arise strong suspicion that such difference may be heard.

@johnread57

@Fair can you summarize on this issue?

Will do, referring to questions in the original post.

It’s obvious right? They sound different, and I’m sure they measure differently.

Yes and yes.

Well we know the dynamic range of cd’s is larger than vinyl.

This is debatable.

But do we have an agreed description or agreed measurements of the differences between vinyl and digital?

No. Because there are multiple - at least two - paradigms, defining certain important characteristics like dynamic range differently.

I know this is a hot topic so I am asking not for trouble but for well reasoned and detailed replies, if possible. And courtesy among us. Please.

Tried my best to abide.

I’ve always wondered why vinyl sounds more open, airy and transparent in the mid range. And of cd’s and most digital sounds quieter and yet lifeless than compared with vinyl. YMMV of course, I am looking for the reasons, and appreciation of one another’s experience.

In the first paradigm, CD is superior in sound quality to LP. Digitizing CDs and delivering their content via online streaming should have killed off LPs.

In the second paradigm theoretical framework, LP occupies a middle ground between CD and such perceptually transparent digital formats as PCM 192/24 and DSD128.

Correspondingly, the second paradigm maintains that CDs, physical or digitized, are not capable of superseding LPs. But perceptually transparent digital formats likely will.

 

@johnread57
@thespeakerdude

I did not present an opinion. I presented verifiable, researched, well understood, mathematical facts. Facts not disputed by those with the deepest understanding of the underlying math, and those able to adapt the math to practical implementation.

 

Below is an opinion. It misinterprets personal opinion, narrow market popularity, and different to "something". That something is only described in easily falsified claims, falsified with math, not an appeal to narrow market popularity.

Perhaps @Fair, can enlighten with at least 2 or 3 of these research papers he claims are hard to find? A new paradigm with 3 decades of research that legitimately calls into question all current signal processing and hearing knowledge should have many available sources to reference.

>>> I see it differently. The old paradigm is falsified by phenomena

>>> for which it gives invalid predictions.

>>> For instance, according to the old paradigm,

>>> LPs shall be long gone, the way of cassette tape recorders

>>> and VCR video tapes. Yet LPs persisted,

>>> and the classic paradigm produces no convincing explanation

>>> as to why.

Very well. This 2016 review  A Meta-Analysis of High Resolution Audio Perceptual Evaluation contains references to

"18 published experiments for which sufficient data could be obtained ... providing a meta-analysis involving over 400 participants in over 12,500 trials"

Conclusion is:

"Results showed a small but statistically significant ability of test subjects to discriminate high resolution content, and this effect increased dramatically when test subjects received extensive training."

 

Pair of charts below illustrates my statement about the growing LPs popularity and vanishing CDs purchases. In a wider context: digital streaming appears to be decimating CD sales, yet LPs have not been affected by that (or maybe even helped?).

CD sales in the US

LP sales in the US

 

@thespeakerdude

Perhaps @fair , can enlighten with at least 2 or 3 of these research papers he claims are hard to find? A new paradigm with 3 decades of research that legitimately calls into question all current signal processing and hearing knowledge should have many available sources to reference.

Not what I meant. I meant that advances in signal processing and understanding of how mammalian hearing system works were significant over the past three decades. Yet not all of the advances are reflected in engineering practice yet. 

CD format, as an example, was developed prior to that, and its designers couldn't take advantage of these advancements. Sadly, this format and its derivatives remain leading by volume for lossless online streaming.

Certain common handbooks on DSP and auditory science haven't been updated yet either. Leave alone the mass of practitioners who still work in the paradigm expressed in these handbooks.

I'll give you a couple of examples. First, a gradual understanding, over about quarter of a century, of the role of so-called Octopus Cells in the functioning of the hearing system: https://www.google.com/search?q=octopus+cells+hearing.

Second, influence of new micro-surgery and robotic devices and techniques, which became widely available in the past two decades, on the hearing system research and medical practice: https://www.google.com/search?q=cochlea+microsurgery+hearing+research.

Still I wait for this. A meta analysis of purely digital sources, some too old to be relevant due to hardware limitations and others with experimental flaws, does not support your hypothesis let alone suggest there is any new paradigm.

That's why I prefer to believe in results of meta-analysis. Some of the smaller-scale experiments support the hypothesis, some others don't. All are imprecise in one way or another. Yet with a larger array of data, statistical inference starts working with sufficient, quantifiable precision.

From where I stand, this meta-analysis shows that the old paradigm, supporting the notion that 16/44.1 can encode perceptually transparently any music for any human listener, is experimentally disproven, and thus the paradigm itself is falsified by the evidence.

I do appreciate the repartee as it demonstrates the vinyl argument.

You are welcome. Winning a gratitude, however small, of a tough intellectual opponent, is one of the best rewards one can hope for in a discussion.

This is just like the tube discussion.  Even though there are significant, identifiable differences between typical tube amplifiers and SS amplifiers with good design practices, differences that are highly audible, every discussion devolves into a debate between those who point out those differences and those who believe in some unseen, unmeasurable property that "must" exist.

I lost interest in this particular discussion quite long ago, after I was able to replicate the "tube sound" inside a common DAW, using easily available software plugins. I occasionally listen to tube amps owned by others, including very expensive ones, but keep coming away underwhelmed.

In a way, the proponents of the tube and other highly distorting amps are right that there are some properties of such amps that evoke hard to measure phenomena inherent in the human hearing system, yet these phenomena can be activated by other means too.

For instance, the Missing Fundamental effect. It makes an amp sound more warmly-bassy to some listeners, which can be pleasant on some source material. Yet, this effect is routinely used in Bluetooth boomboxes, via quite simple DSP algorithms. The downside is of course masking of midrange, so it is beneficial only for some music genres.

 

@cleeds

No, Fourier Analysis is not a theory. It's a theorem; it can be shown to be perfect with math. Unless you can show a fault in the actual math, all of your hand waving and word salad is for naught.

Fourier Analysis, as a theory, is a subset of Harmonic Analysis theory. It has dozens of theorems. If you are interested, pick one of these books - Harmonic Analysis on Amazon - and see for yourself.

A handful of the theorems is routinely used by DSP practitioners. For instance, Parseval's Theorem can be used for quick estimation of THD contributed by a device, via comparing pre- and post- waveforms.

Again, I'm an analog guy. But if we seek better digital, it can't be done without understanding how digital works. I'm all for hi-res and everything it takes to get the best out of it. But let's not pretend that somehow the underlying premise of digital audio is somehow broken. It isn't. Its problems are elsewhere.

One of the problems, as I see it, is simplification, excessive popularization, and at times even vulgarization of science, which became widespread in the Western world during the several past decades.

I guess I have to go that route as well, given the circumstances. So here it goes. A mathematical theorem is like a part of a legal contract: its words have precise meaning, often unexpectedly different from their everyday meaning; and it has small print.

As it relates to the Sampling Theorem, the phrase "contains frequencies" means something quite different from what everyday common sense would make one assume. And the theorem itself is just a paragraph in a long contract, with lots of small print.

Let's say you have a health insurance contract, and it covers your teenage son too. He rides electric bicycle. God forbid, he gets in an accident on the bicycle, hits his head, and requires expensive urgent care and rehabilitation.

Naturally, you assume that the insurance will cover it all, and you feel safe in belief that you'll be only out for deductibles and copays.

Suddenly, your insurance company sends you official letter saying that this event isn't going to be covered by them, because during the accident, according to a police report, your son was violating The Law.

Perhaps he was riding without a helmet. Perhaps he was riding on a walkway in a town with an ordnance prohibiting that. And there is a small print in the actual insurance contract stipulating its provisions null and void if injuries were sustained in the process of violating The Law.

As it relates to the Sampling Theorem, the signal has to fulfill very precise obligations, before the theorem can guarantee its accurate capture and reconstruction.

The meaning of the words describing these obligations is precisely defined elsewhere in the theory, in other definitions, lemmas, and theorems. Nothing wrong with the theory per se, at all.

Practical music doesn't fulfill these obligations, and thus the Sampling Theorem only works approximately. How well it works can be calculated too, using other parts of the theory, yet this is far more involved, and the answer is signal-dependent.

The vulgarization, in the case of CD format marketing, was in omission of the facts I described in the previous paragraph. Was it done on purpose or through honest mistake? I don't know. 

 

@thespeakerdude

>>> Practical music doesn’t fulfill these obligations,

>>> and thus the Sampling Theorem only works approximately.

>>> How well it works can be calculated too, using other parts

>>> of the theory, yet this is far more involved, and the answer is

>>> signal-dependent.

As you are the self declared expert @fair , complete with straight A’s in relevant courses, perhaps you could go through the basic math for us taking into account practical analog filters, typical over sampling and decimation filters, etc. As you know so much about this subject that should be very easy for you. You can tell us exactly how big the errors are of course.

To be frank, I'm bored at this point. Not getting anything of value back from you.

Clearly, you are not reading not only what I'm referencing, but also the parts of my posts that you are quoting.

I'm telling you that the errors are signal dependent, you are quoting this, and then are asking me to tell you exactly how big the errors are? 

You know, serious people run long simulations to answer this question for specific digitization schemes and sets of representative signals.

May return to answering your questions when I see your commitment to learning.

I've seen accounts of experimental results seemingly proving that CD format is perceptually transparent. Also, accounts of experiments seemingly proving the opposite.

Not satisfied with what I've seen in these papers, I conducted my own experiments over the past decade, where I controlled the whole recording and reproduction chain. The results, for me personally, were more definitive.

I have to disclose at that point that I'm educated, certified, and worked in the areas of Physics and Neurophysiology, including stints in several national labs, in Europe and USA.

So, my standard of inquiry is rather high. When the outcome of an experiment depends on accurate detection of one photon vs two arriving within a given microsecond, it has to be.

Experiments in psychoacoustics are hard, mostly because music signals and the final "instrument" - human hearing system - are very variable. Thus, I put more faith in meta-studies rather than in individual experiments, including my own.

Meta-studies, for instance like the one I already referred to in this thread:

 

Proponents of this or that point of view like focusing on one particular study, or a handful of them, proving their point. I don't believe that's the way to go.

For instance, the oft-cited Boston Audio Society study (BAS study), which seemingly proved the perceptual transparency of the 16/44.1 digital loop, didn't conduct one of the mandatory steps any experimental science professional would do - calibration.

If an experiment is to elucidate the importance of distortions, one absolutely has to take into consideration the nature and levels of distortions inherent in the gear involved in the experiment. What do we have in the case of the BAS study?

https://www.bostonaudiosociety.org/explanation.htm

"The Principal System

The playback equipment in this system consisted of an Adcom GTP-450 preamp and a Carver M1.5t power amplifier."

What is Carver M1.5t power amplifier?

 

SINAD reveals that the amp's distortions are high and rising with frequency. "... distortion-free range = 14 to 11 bits". "Distortions rises @ 66 watts". "Max power = 273 watts @ 44 dB SINAD".

What the designers of the study should have done could include at least, back in mid-2000s: taking the Principal System into a certified anechoic chamber, and measuring its performance, including its distortion profile, using certified calibrated instruments.

 

@cleeds

A math theorem is a principle that can be proven with math. The Fourier Transform is a theorem. It has been proven by both math and in actual implementation - it's the basis of both digital and analog audio. Your alternative paradigms aren't consistent with Fourier. You're probably having fun with your fanciful imaginings but that's all they are. If you want to make actual progress you'll have to accept the math. Anything else is futile and more than a bit silly.

Let's make it more interesting. Shall we?

You likely know who Bob Stuart is:

 

Despite his lifelong achievements and indisputable expert status, audiophile community met MQA format he co-created with significant controversy, e.g. 

 

 

I propose reading a peer-reviewed paper he co-authored:

 

 

And then telling us what Tidal did wrong to further stir up the MQA controversy. And perhaps what Tidal and others can do in the future to improve the MQA situation. You may think it can't be improved, and this would be interesting to hear too. 

I hope the ensuing discussion would be very relevant to topic at hand, because Tidal and other "Ultra HD" music streaming services hasn't yet made a dent in LP sales, but I believe they may do so in the near future.

 

Some references I found informative.

Level I

Popular level. Summarizes pros and cons, remaining in old paradigm, e.g. in regard to dynamic range estimation.

 

Level II

Somewhat technical level. Parts 1-5 provide good overview of LP technology. Parts 6-9 describe distortions inherent in it.  

 

 

Level III

Very technical level. This explains why, despite obviously inherently high level of distortions, LPs sound just fine to many people. Executive summary: this is because LP distortions are mostly of "right" kind, which human hearing system do not register with the same intensity as "wrong" distortions.

http://www.gedlee.com/Papers/Distortion_AES_I.pdf

http://www.gedlee.com/Papers/Distortion_AES_II.pdf