Saturday, September 9, 2023

IFI's Global Master Timing (GMT)





 

After giving up previously, I finally found some details on ifi's GMT clock besides the vague blurb included in the list of features on ifi's Zen DAC V2 page. The relevant page  was apparently part of an internal ifi presentation on various technologies used in ifi products, by their R&D division, for a business-oriented division.

The GMT clock was apparently developed for AMR's high-end gear, and has become ifi's universal clock, the best grades of which apparently have jitter on the order of 0.3 pS, and a very clean spectrum from 9 kHz to 15 kHz (which is all that's shown in the relevant plot, above, so perhaps it's the critical range). For information on how such low jitter can be obtained, see Analog-to-Digital Converter Clock Optimization: A Test Engineering Perspective.

Whatever the Zen DAC V2's jitter-rating is, its high end is as clean as any analog gear I've heard, which includes moving-coil cartridges and 45 RPM direct-to-disc LPs. This is partly due to the fact that it uses the TI/Burr-Brown DSD1793 Advanced Segment DAC, which has low jitter-sensitivity, low noise in general, and 24-bit resolution (see my Zen DAC V2 review for details). Naturally, the quality of the high end depends on the source material, and the cleanest CD I've heard is United We Swing by the Wynton Marsalis Septet.  

Reading the aforementioned presentation-page was the first time I'd run across the term "femto clocks," although I've done quite a bit of searching on the subject of jitter. It's another indication of how poorly the audio media has kept us informed of significant advancements in digital audio playback. But then considering that it's clearly downplaying the significance of audio polarity, it's apparently deliberately keeping us in the dark so that we'll flail about trying to obtain musical satisfaction from digital audio, and waste a lot of money on playback gear and recordings in the process. There are probably other, even darker motives, but what matters is that playing good CDs by means of well-designed digital playback gear which uses TI/Burr-Brown Advanced Segment DAC-chips, with a decent system otherwise, and with the correct audio polarity (although some CDs aren't very sensitive to polarity), is the least expensive approach to being able to listen to a particular piece of music when you're in the mood for it, and to obtain musical satisfaction from it.

Without further ado, here's the text from the web-page I found about GMT:


Femto Clocks – Picky about Phase Noise

Background

All Femto clocks are good, very good in fact. They exhibit jitter levels lower than most clock crystals which leads to better sonics. An oft quoted benchmark for accuracy is Femtoseconds (Fs) / parts per million [Fs for jitter, ppm for frequency error].

Back in 2008, before the word “Femto Clock” became all the rage, AMR developed a special type of clock in DP-777 [a $5K digital processor] as part of the “Global Master Timing” (GMT) and “Jitter less” technologies (why special? See below, because not all Femto clocks are the same).

We called it the GMT Clock platform (which is comprised of specialised hardware+software) as it is not just buying a “clock in a can” and job done.

Having worked with all sorts of clocks, including discrete, Rubidium, Superclocks and not the least Femto clocks over the years, we know them quite well.

All Femto Clocks exhibit excellent low phase-noise (measured jitter within the clock). However, as their origins lay in being part of SONET, the popular SONET targeted “Femto-clock” is less desirable as its best phase-noise performance is concentrated in the > 12KHz region (read: at the very top and way above the audible band, so benefits audio less).

As an example, this link highlights the use of Femtoclock technology in the telecommunications sector where they are spec'd for.

This [referring to the relevant jitter-spectrum plot at top of page] is an Optical Comm system (aka SONET - which is a Subset). We added the blue line to highlight the -70dBr region so that when referenced to the AP2 chart in the next section of the micro iDSD, it is more of an “apple to apple” comparison.

The spike at 50KHz is the "signal" As you can see, for quite a few KHz around this region, phase noise is low, this is what matters in this application.

However, the area around the green arrow is the most crucial human audible range of 20Hz > 20kHz where phase noise performance is less impressive in the region of -100dB to -70dB.

Explanation

Therefore, the key for AMR was to design a new system, the “GMT” clock platform which not only exhibits the lowest phase-noise in the crucial audible band, but offers precision (< 0.004ppm tolerance) adjustability with literally millions of possible frequencies  (as per the DP-777 "GMT" Technical Paper).

The GMT Clock system designed into the micro iDSD measures <280 Fs, comparable to many Femto-Clocks (because it was designed to give low jitter).

From the [GMT jitter-spectrum plot above], you can see that jitter in 9kHz > 15 kHz is very good, the micro iDSD noise floor goes all the way down to -150dB which is virtually across the board with no spikes.

How this benefits the user

Consistent, across the board negligible jitter means timing is supreme, with just the right amount of attack/decay and of course, tonal accuracy. We are really pleased with the very low jitter performance of the micro iDSD in the most crucial audible range – in fact, we would not mind if customers pit it against significantly more expensive DACs.

We hope you found this interesting as it sheds some light on the particular attention we have paid to parts performance and custom design in the micro iDSD (actually, we took it from the DP-777!).

Addendum: What about Rubidium Clocks?
 

[Above] is a chart of several types of rubidium clocks. What they all exhibit is many sharp spikes in phase noise. Even though they measure well, some down to -150dB, when they spike, noise levels jump up to -70dB to -90dB.

This is far from ideal which is why we have not used such clocks, neither in iFi nor AMR products. It all boils down to paying close attention to the specific clock/s used and its performance in the audible range.
 

Monday, August 21, 2023

Zen DAC V2 & TI/Burr-Brown Segment-DAC-chips




Keywords: "Burr-Brown, Advanced Segment, sound quality"

After the Zen DAC V2 is powered up for 24 hours [1] by means of a 5V 2A AC-to-DC adapter plugged into the Zen's External Power input, it sounds unbelievably good, with the advantages of analog and digital, and without the disadvantages of either. The source material seems to be the limiting factor, and while playing the United We Swing CD by the Wynton Marsalis Septet, I heard high-end detail that would probably impress any audiophile.

The Zen also has incredible low-level detail, due to the unique design of the TI/Burr-Brown Advanced Segment DAC-chip (ASD) used in the Zen (details below). As I wrote this, I was listening to the Monk Quartet with John Coltrane at Carnegie CD [2], and for the first time heard Monk quietly humming as he played.

After hearing the Zen, I can't stand my previous DAC, a Topping D10s, because I realized that it has no low-level or high-end detail, and produces lifeless "toy" audio with a noisy/harsh, artificially bright high end. This might be due to the Sabre DAC's relatively high jitter sensitivity (DAC output noise vs. clock jitter), combined with a relatively high-jitter clock. Regular delta-sigma DACs apparently require clocks with very low jitter to sound good, and such clocks aren't cheap.

After hearing CDs through the Zen, I realized that they have the POTENTIAL to sound better than LPs.  However, some copyright-holders want to retain control over the best versions of their recordings, so they reserve them for LPs since CDs can be copied exactly and easily. But most CDs are good enough for me, and some are fantastic.   

High-res sounds no better than CDs, according to High-Resolution Audio: Does it Sound Better? https://goldmund.com/does-high-resolution-audio-sound-better/​ by Goldmund Acoustic Laboratory. The higher sampling rates might have helped some ancient DACs with lousy digital interpolation filters, by reducing the need for digital interpolation, but only about 16 of the 24 bits are used for consumer releases. High-res formats were originally intended just for recording and processing.

TI/Burr-Brown Advanced Segment DAC-chips

What is probably the biggest breakthrough in digital audio playback history occurred in January 2001 with the introduction of TI/Burr-Brown model PCM1738 Advanced Segment DAC-chip (ASD). The Zen DAC uses the DSD1793 ASD, which is one of the most recent ASD-models. The ASD architecture (described below) is a more precise implementation of the Current Segment DAC (CSD) architecture, which is used in such models as the PCM1690, an 8-channel DAC-chip which according to Archimago's Musings' test of the Onkyo TX-NR1009 as a DAC, is an impressive performer.

But TI has never given this breakthrough the publicity which it deserves (such as small ads on audio-review websites with an endorsement from an audio expert and a link to a website with details), so that even now few people are aware of the existence of ASDs or CSDs, and even fewer recognize their significance. Their sound quality is typically attributed to their direct conversion of DSD (as if this would help PCM), by using names such as True Native or Bit Perfect. I didn't know about them until I got a Zen in 2023, wanted to know what makes it sounds so good, and investigated.     

The attached block diagram and waveform-illustration show the basic idea behind the ASD architecture, which is more properly known as the "Advanced segment DAC interpolated by sigma-delta," for reasons explained below. ("Sigma delta" is interchangeable with "delta sigma.")

The current-segment DAC (CS-DAC) in the block diagram (not to be confused with the current-segment architecture mentioned previously) is the actual DAC which converts digital inputs into analog outputs. It consists of 66 nominally equal current sources which are differentially switched between two outputs (positive and negative). The ICOB-decoder converts Bits 1-6 into a 62-bit  "thermometer-code" [3], and the delta-sigma modulator provides 4 such bits, but at 64x the sampling rate of the other 62 bits. When all are bits are off, the output (taken across the differential outputs) is -33 (-2.5 mA), when 33 are on, output is 0, and when all 66 are on, the output is +33 (+2.5 mA). So, it has a total of 67 levels, but since the 4 inputs from the sigma-delta modulator are switched at 64 times the rate of the others, they can represent any value between -2 and +2, with 24 binary-bit resolution, when filtered. This allows the total output to represent any value between +/-2.5 mA with 24-bit resolution.

Not all ASD-models use precisely the same configuration shown in the block diagram. For example, the PCM510xA family uses just a 32-bit sigma-delta modulator to drive the CS-DAC, probably because gear-manufacturers want to win the spec-war. However, the waveform-illustration probably still applies since it represents the fundamental concept behind every variation of the ASD architecture.

For purposes of the waveform-illustration, the ICOB-decoder (see the block diagram) is presumed to have an output of only 7 levels (0 and +/-3), when in reality it has 63 (0 and +/-31), i.e. 6 binary bits of resolution or (2exp6 - 1) levels.
 
The DWA block is a complex scrambler to randomize the connections between the CS-DAC's data sources (the ICOB-decoder and the sigma-delta modulator) and the DAC's inputs, in order to average-out mismatches between the CS-DAC's segments to obtain 24-bit linearity.

The characteristics of the on-chip digital interpolation filters (DIFs) [4] are also important, but the DIFs in the latest ASD-models sound great, although I'm not certain how they were in the first model, the PCM1738, introduced in 2001. The 1738 was used in some high-end gear, but it might have been combined with external digital filters.


Technical papers on ASDs

The attached waveform-illustration of the basic concept behind the advanced segment DAC is from a 2001 paper entitled "A 126dB D-Range Current-Mode Advanced Segment DAC"  by Norio Terada and Shige Nakao of TI-Japan. A 2000 paper, entitled "A 117dB D-Range Current-mode Multi-bit Audio DAC for PCM and DSD Audio Playback"  (apparently by the entire ASD development team) also contains useful information.  


A potential explanation for TI's cryptic ASD/CSD sales pitch

Every ASD/CSD-model's data sheet includes the statement "[model #] uses ... TI’s advanced segment DAC architecture to achieve excellent dynamic performance and improved tolerance to clock jitter." I'm not entirely certain what they mean by "dynamic performance," but dynamic range specs are listed under that heading in ASD data sheets. It might also have something to do with the ASD's 24-bit linearity and its clean transitions from one sample to the next, with minimal switching- and jitter-related noise in the output, as a result of the CS-DAC's number of levels, its differential nature, and the design of its differential switches.

The 2001 paper includes a jitter-sensitivity analysis which indicates that the jitter sensitivity of a multi-level DAC is approximately proportional to the full-scale range and inversely proportional to the number of levels, which in the case of the ASD's current-segment DAC is 67 levels (0 and 33 levels above and below). The paper doesn't mention the sigma-delta modulator's jitter-sensitivity, but it has five levels, and its amplitude-range at the CS-DAC's output is only about 6% of the total range (4 levels of 66 total), so it probably doesn't make a significant contribution to clock-jitter-induced noise at the ASD's output. I suppose that its small amplitude-range also gives it tighter control over low-level signals compared to a full-range DS-DAC. This, combined with its low clock-jitter-induced noise, would explain the incredibly clean low-level detail I've heard from the Zen.

The 2001 paper also includes a photo of an impressively clean -120dB, 5 mVp-p 1kHz sine-wave, produced by an ASD and displayed on an oscilloscope. The clock-jitter isn't specified, although the paper mentions that 200 pS was typical and 100 pS was about the minimum in audio gear at the time. Commonly-used modern crystal oscillators apparently have about 10 pS. RME's Steady Clock is rated at 0.1 pS.


ASD-endorsements by high-end audio-gear designers

From Bel Canto's chief engineer John Stronczer in the Stereo Times review of the Bel Canto DAC 3.7 :  
 
"At Bel Canto Design, WE STICK WITH THE CLASSIC PCM1792 DAC BECAUSE IT REPRESENTS THE BEST DAC PERFORMANCE ACHIEVABLE TODAY. This holds true for every aspect of the DAC - noise, distortion, dynamic performance and even the way the analog output electronics interfaces with the PCM1792 DAC is superior to other options."

Thorsten Loesch, the head of R&D at AMR/ifi when the Zen DACs were being developed, endorsed ASDs here (see the Reader-view page): 

"R2R DACs also have problems with low signal levels [due to linearity errors, which severely distort low-level signals], which is what Delta Sigma DACs avoid. Conversely Delta Sigma (including DSD) technology performs relatively poorly at high signal levels in comparison to multibit solutions.

That’s why we prefer hybrids; multibit architecture for high signal levels, (the so-called upper bits or the MSBs), and high speed (DSD256 or higher) Delta Sigma topology for lower signal levels. The Burr-Brown DSD1793 we’ve been using so much is one such hybrid and one of the best options for us."

According to "MLGrado" on the regular (non-Reader View) version of the aforementioned web-page:

Essentially the BB DSD1793 behaves like an R2R DAC.
[...]
The Burr Brown 'segment DAC' ... was thought by its makers to be A BREAKTHROUGH ADVANCEMENT OVER PURE R2R DACS OR PURE DS DACS. IT AIMS TO GIVE YOU THE BEST OF BOTH WORLDS, WORKING AROUND INHERENT WEAKNESSES. [emphasis added]

So one may indeed consider going to R2R a step backward. In any case, you would not have iFi's TRUE DSD conversion. R2R must convert DSD to PCM.

[end of excerpts]

ASDs are used in everything from Sirius XM receivers, CD players, DVD players, TVs, etc., to extreme high-end DACs. My cheap Vizio TV's "analog" sound quality puzzled me for a long time, but now I realize that it probably uses ASDs. Audio Research's latest DAC, the approximately $8K DAC9, uses the PCM1792A ASD. Bel Canto also uses them in their latest gear. In a Steve Hoffman forum entitled "Enamored by the Burr-Brown PCM1796 DAC chip," someone claimed that a Pioneer DVD player's audio section with ASDs blew the doors off of an expensive discrete R2R DAC. A Zen DAC Signature V2 review on Amazon claims that the Signature version sounds better than a $1500 DAC which uses R2R chips. A lot of companies with a reputation for audiophile sound on a budget, such as Music Hall, Musical Fidelity, Denon, and Onkyo use ASDs in some of their gear. Yamaha's WXAD-10 streaming-DAC uses PCM5121 ASDs and reportedly sounds "superb" and "engaging." Pro-ject uses ASDs in their recently-introduced CD players. The $130 Soundavo HP-DAC1 DAC-amp also uses ASDs and got a rave review on Amazon by a 70-ish vinyl junkie who wasn't impressed by digital previously. So, it's not very difficult to find DACs, streamers, receivers, etc. which incorporate ASDs and sound great. But almost nowhere, except in TI data-sheets for individual ASD-models, are they identified as ASDs, which is their most significant aspect as far as sound quality is concerned.

Regular sigma-delta ADC/DAC chips are used in studio-grade ADCs and DACs, so the standard sigma-delta technology can obviously provide reference-grade sound quality. (There apparently are no ADCs which use ASDs, which might mean that the ASD architecture is incompatible with the requirements of ADCs.) However, studio-grade ADCs and DACs typically cost thousands of dollars, partly because they have clocks with extremely low jitter. The $1300 RME ADI-2, which is popular among recording engineers, has less than 0.1 pS of jitter and sounds great with either AKM or Sabre DAC-chips.  

Zen's Output Stage

The Zen's output stage, as far as I can tell from various vague descriptions, consists of an OV2637A amplifier module, and associated circuitry to configure the module as an I-V converter and anti-aliasing filter in the Zen DAC. The module includes four single-ended amplifiers, each consisting of a JFET-input op-amp followed by a buffer such as a TI BUF634A (which can drive headphones), and each of which is used as one side of a differential pair, forming two differential pairs. It takes the form of a tiny circuit board with pins protruding from one end, and chips attached directly to it. Some of these chips are surface-mount devices, and others are bonded in place and connected to circuit-board traces via wire-bonds. The wire bonds and associated chips have to be protected, such as by epoxy encapsulation or with a cap bonded to the circuit board.

To provide sufficient supply voltages for the op-amps and buffers, the Zen has a voltage-doubler (a flip-flop, diodes, and caps), and the resulting dirty 10V supply is regulated down to about 9V, which is capacitively bypassed to provide high instantaneous current capacity. With a 9V supply, the maximum single-ended output would be about 8.5Vp-p, and the maximum differential output would be about 17Vp-p, which is a hefty signal from USB power or a 5V wall-wart. The Zen has single-ended and differential outputs for headphones and amps, so it's basically pro-grade.

Zen shows off good CDs and exposes flaws in bad ones

Esperanza Spalding's Radio Music Society CD didn't particularly thrill me when I listened to it through my delta-sigma-based DACs. But when I played it through the Zen DAC, I was amazed by the CD's sound quality, and mesmerized by the intricate music and Spalding's enchanting voice.

Another great CD is Alive in America by Steely Dan, whose main recording engineer Roger Nichols used it as an opportunity to prove how good he could make a CD sound, since it's a live CD and he didn't have to reserve the best version for LPs. So, he recorded and mixed it in analog, digitized the mixer's output with an Apogee AD-1000 (20 bits, 44.1 kHz), and used an Apogee UV-22 to convert the 20-bit data to 16-bit data that sounds like 20-bit data. Sign In Stranger sounds particularly good.

However, the Zen's excellent low-level detail is a two-edged sword, because it makes any flaw stand out. For example, the high end of my Kamakiriad CD (Reprise 9 45230-2) vaguely bothered me through the D10s, but through the Zen is obviously deliberately degraded. The LP probably has a great high end, and perhaps a more recent CD-release has a better high end.

Miscellaneous audio information

There is no need for precision CDs or expensive transports, due to the error-correction system built into the CD record/playback system. Precision CDs might contain better recordings than the ones used for regular CDs, because the superior sound quality can be attributed to the physical precision. But if the copyright holder reserves the best versions for LPs, no digital release will sound as good as the LP, although MP3's might be made from the best version since they use lossy compression.

Separate CD-transports cannot possibly reduce the CORRECTED error rate compared to that obtained from a cheap transport in a CD player, which is a few ppB for CDs in decent shape. If you doubt that CD read-errors can actually be corrected, explain how software can be stored on CDs (simply re-reading the erroneous data won't work). Using separate transports just adds compromises or cost to the task of synchronizing the transport's clock and the DAC-chip's clock. Ideally, the clock should be next to the DAC-chip's convert-command pin, to minimize jitter in the DAC's convert-command, and the clock's signal would be sent to the transport, which can tolerate higher levels of jitter. Any buffer-memories between the transport and the DAC-chip could also tolerate a higher level of jitter, as long as the DAC's data-inputs can tolerate the jitter in the buffer's output.

CD treatments don't reduce the corrected error-rate - they reduce laser-servo hunting, and thus reduce laser-servo-induced noise on the transport's power supply. The only CD players to benefit from CD-treatments were cheap early players with poor isolation between the transport power supply and the analog power supply, which allowed some of the laser-servo noise to get into the output.

Software is stored on CDs & DVDs, and it can't have any errors, so it's obviously possible to truly correct CD read-errors. Software has an extra layer of correction compared to audio, but it's rarely needed.

Some CDs, although very few, have inverted audio polarity, probably deliberately, in an attempt to prevent us from being able to rip the best version. So, if a CD just doesn't sound "right," try flipping the polarity. Flipping polarity doesn't make much difference if a recording is incoherent or has inconsistent polarities within it, or if the playback gear is incoherent. Examples of CDs with inverted polarity are Night by John Abercrombie, the CD layer of Michael Brecker's Pilgrimage SACD, and Bartok String Quartets 1-6 by the Alban Berg Quartet. Audio polarity can be inverted by flipping the polarity of each speaker connection, assuming that the speakers are passive, or by ripping the CD and using an audio-editing program to invert each file.

Some high-end gear, such as Audio Research preamps and Benchmark DACs, have polarity/invert switches, as does all sorts of pro-grade gear. Some DAC-chips, including some ASDs, have invert-functions. The fact that so little consumer-grade digital playback gear has polarity-control capability, although it would be cheap and easy to add it, supports my suspicion that inverted polarity is being used as a means of degrading the sound quality of some CDs.

To get the best sound from your system, periodically disconnect and reconnect all connections, including internal connections, AC connections and breakers (turn off everything that uses a significant amount of power before cycling breakers). Also use canned air to blow dust out of ventilated components, because dust acts as a parasitic circuit, especially in high humidity. In severe cases, it might help to scrub with alcohol and something like a toothbrush around components to eliminate all parasitic paths.

The optimal turntable configuration is belt-drive with a pivoting arm with a fixed head-shell. Direct drives have a feedback loop and their speed oscillates around the ideal speed, and tangential tonearms are a scam. Use a record cleaner such as a Spin-Clean, even if you have to scrimp on the TT or exceed your budget, because a really clean record sounds much better, and you can't undo record wear.

Notes

[1] Experts recommend leaving digital playback devices powered up all the time (unless they use a lot of power or have tubes which can't be shut down independently of the rest of the circuitry), and letting them warm up for at least 24 hours before judging them. For example, see Audioquest's technical paper entitled Evaluation of Digital Devices.

Any functional 5V 2A wall-wart with a 5.5mm OD/2.1mm ID barrel connector output can be used as a power source for the Zen. (See Prodigit's page entitled "How to test Noise from the Output of Power Supply" ) Some claim that higher-quality supplies improve the Zen's sound quality further, which I suppose might be possible.

When power is applied to the Zen's External Power input, the LED next to the input glows, and the USB connection isn't used for power. The LEDs on front indicate whether the USB input is active.

[2] The recording on the Monk Quartet at Carnegie CD was made on 11/29/57, using tube gear. ("Tube sound" strikes me as very coherent and alive. According to the VTL site, tubes sound better than other devices because they have a more linear gain, and need less feedback.) The tape was poorly labeled, put in storage, and forgotten, until it was discovered 48 years later in 2005, and digitized and digitally processed (see Recently Discovered Jazz Jewel Restored by Hip-Hop Legend). The resulting sound quality is amazing considering the age of the tape when it was digitized, and the performance was excellent.

[3] In a thermometer-code, each single-level increase is achieved by simply activating another bit, so that as the signal-level increases, more bits turn on, and none off, and as the signal level decreases, more bits turn off, and none on. Likewise, in a liquid thermometer, the liquid rises and falls with temperature, with no gaps.

[4] DIFs basically calculate samples to be inserted between the samples provided by a CD, for example, thus reducing the size of the steps in the "staircase" waveform at the output of the DAC, so that a simple, clean, and cheap high-frequency analog filter which doesn't affect the audio range can completely filter/smooth out the staircase.


Sunday, February 20, 2022

Were Steely Dan's actual analog masters digitized in the early 80's with a 3M deck?

Rev 2/28/22 (see Notes)

Steely Dan has a reputation for being audiophiles, and Roger Nichols, their main engineer, was a proponent of digitizing analog recordings to preserve them. So, I find it hard to believe that they didn't digitize their analog masters (either the multi-tracks, the original stereo masters, the  safety copies which Roger Nichols made when each album was new [1], or all of these), with a good ADC until 1998.

In fact, an article entitled Roger Nichols: Digital-To-Digital Transfers from the May 2006 issue of Sound on Sound indicates that Nichols used a 3M digital recorder to copy analog multi-tracks in 1982, and that he was transferring at least one of them to Pro Tools in 2006, which just happens to have been the 30th anniversary of Aja, when the legendary Cisco LP was released:

"I am currently transferring digital multitrack tapes from early (1982) projects into Pro Tools for surround mixing. I remembered some of the problems with early recordings that related to early A-D and D-A converter designs. During these transfers I wanted to correct any of the early shortfalls, if possible. Even though they were 16-bit recordings, transfers to 24-bit would help preserve the accuracy of the original recordings.

"These early 3M digital 32-track machines did not have digital outputs, so the transfers were to be made via analogue cables into new 24-bit converters [something  more advanced than the Apogee AD-8000 8-channel 24-bit ADC which he obtained in 1997/8, I presume]. There was no such thing as a 16-bit converter when the 3M machine was designed, so they used a unique combination of a 12-bit converter with an additional four bits of an 8-bit converter for gain ranging. [I gather that this means that 4 bits of the 8-bit DAC were used for controlling the reference voltage on a 12-bit multiplying DAC.] This required a very expensive HP spectrum analyser to set the tracking of all the converter elements.
[...]
Converter Tracking

"Let's compound the issue a little. A-D converters have the same problem. The error for each bit during the recording is added to the error for the bits during playback. In early digital machines they hand-matched A-D and D-A converters to match closely to get the best sound on each track. If you had to replace a converter, you were in big trouble unless you replaced both with a matched pair. SINCE A-D AND D-A CONVERTERS BASICALLY WORK THE SAME WAY, SOME MACHINES USED THE SAME CONVERTER FOR RECORDING AND PLAYBACK TO AVOID TRACKING PROBLEMS. [In other words, the linearity error of the recording was canceled out by playing it back through the same DACs which were part of the ADCs which were used in the A-to-D conversion-process. There would still be some quantization error of less than 1/2 LSB, but this would be way down in the noise, and dither might not have been required. The catch is that this required the same 3M deck which was used for making the recording to be used for playing it back, perhaps decades later, which would have been risky because converters can fail. So perhaps two or three recordings were made on separate decks running in parallel to ensure that at least one of the decks would survive for a few decades. It would have been an expensive approach, but it would have provided extra security because someone couldn't steal a tape and play it on just any deck without losing sound quality. It was if the ultimate sound quality was encrypted, and that the key to obtaining it was to play it on the same deck that was used for recording it.]

"There were linearity problems with the 3M machines, but you could set the D-A tracking to match the A-D tracking so that the throughput of each track was linear unto itself. This meant that what you recorded on a track was what you played back on that track. This would be good enough for the transfers I needed to make. I DID HAVE A DIGITAL INTERFACE BOARD THAT I BUILT FOR THE 3M [2], but if I'd used that, the digital transfers would have had the non-linearity of the A-D converter without the correction of the D-A converter. So, analogue transfers it was to be. [Well, actually, he wanted to make a 24-bit copy, and playing it back on the deck which was used for making the recording (i.e. each channel was played back through the same DAC which was used in that channel's ADC), and filtering it with the 3M deck's (analog) anti-aliasing filters [3] was the best way to precisely re-create the original waveform so that it could be digitized with a 24-bit ADC, and the resulting digital recording could be encrypted for security.]

So, as far as I'm concerned, the evidence is sufficient to conclude that Nichols made 3M digital copies of Dan's multi-track and original stereo analog masters back in the early 80's, and that they were converted to 24-bit copies which are being used for LPs. Any evidence or arguments against this conclusion can probably be explained away. For example, the fact that they made a surround-sound version of Gaucho, supposedly from the analog multi-track, and then never got around to the other albums, might have just been intended as a cover story for the existence of a 3M copies, and perhaps they concluded that it wasn't worth the effort to make surround-sound versions of the rest (not even their latest albums were released as surround-sound versions, as far as I know). If they want to retain control over those recordings, it's their right, and I'm satisfied with the 1999 Aja CD-release played on my Topping D10s (an awesome-sounding cheap DAC) and the knowledge that if I were to shell out $500 or so for a decent belt-drive turntable/preamp with a decent pivoting tonearm [4], about $100 for a good record-cleaner, and $50 for the LP, I could hear Aja in its original splendor, or even better, because the original recordings will last forever, and LP-recording and playback continue to improve.

Notes

Revisions

2/24/22 - Revamped after finding the 2006 Sound on Sound article by Roger Nichols using 3M decks to copy analog multi-track masters, which eliminated the need for a lot of the circumstantial evidence and speculation in the previous revision.

2/25/22 - Corrected a few conceptual errors, added some insights, and smoothed out some rough spots.

2/28/22 - Revised Note 4.


[1] Nichols stored the safety masters (15 ips analog copies of the original stereo analog masters) in his personal archive, and never played them, except for Aja, once, to make the 1982 digital master with a non-oversampling 16/44.1 Sony deck with the notorious dry/smeared-sounding input filters. (Some tracks from this version of Aja can be heard on the Very Best Of Steely Dan CD, a compilation made from the 1981/82 digital recordings. The Greatest Hits CD was apparently made with Apogee AD-1000 20-bit ADCs from a good analog master, and sounds good, although the analog master was created in 1978 and it was apparently digitized in 1993 when the AD-1000 was introduced, so it's not as clean as the LP-copy I had in the mid-80's.) Nichols could have used the 44.1 kHz oversampling deck which JVC introduced in 1982, and which made great recordings, but Dan probably didn't want to let such good digital copies out into the wild. The 1985 digital masters were apparently made with a Sony 1610 with Apogee input filters from the same analog tapes, which by then had seriously deteriorated so that the high end is little but a dry sheen, except for Katy Lied (dBx-encoded) and Gaucho. The Gaucho digital master has a slightly high speed, so that the slow cuts sound rushed, and the Aja CD has a peaky high-end boost which sounds dirty and artificial. The MoFi Aja and Gaucho CDs (which I haven't heard) were made from the same digital masters, and according to Nichols, in an article in Metal Leg 18, Gaucho has the same speed error as the commercial release. However, MoFi might have used different EQ on Aja. MoFi got their own oversampling ADC in 1988 from Theta Digital. It used BB ADCs in parallel to sample the input at presumably 176 kHz, and Motorola DSP chips to implement the digital filter that reduced the sampling rate to 44.1 kHz.

[2] Nichols obviously broke out the digital signals so that he could duplicate digital recordings without converting them to analog. 3M probably sold accessories for this purpose, but Nichols, who was famous for creating recording-related gear, might have found a better approach, or at least a less expensive one. If the duplicates were played back on the original recorder, their errors would be canceled out.

[3] It was easier to design anti-aliasing filters with a linear phase characteristic for converters with a 50 kHz sampling rate than for those with a 44.1 kHz sampling rate. But in about September of 1985, Apogee introduced aftermarket filters with linear phase response for non-oversampling 44.1 kHz decks. Oversampling 20-bit studio-grade converters were introduced by Apogee and others in 1993, and 24-bit units were introduced in 1997.

[4] Linear-tracking tonearms are expensive gimmicks designed to relieve rich people of their excess money, which I realized after designing a cheap one and wondering why the design hadn't been used previously. Even if a linear-tracking design manages to reduce tracking error, it won't sound significantly better, and it will have higher lateral inertia, which causes problems when playing records with off-center holes and/or warps. (There are tangential pivoting tonearms which reduce tracking error by rotating the headshell, but there's no concensus among audiophiles that they sound any better, and their prices are astronomical.) I also prefer pivoting arms because they allow a phono preamp to be placed near the pivot to minimize the length of the tonearm wiring. With modern op-amps, excellent phono preamps can be built into turntables. Well-Tempered Lab turntables, which are based on original thinking, use pivoting arms, although long ones in some cases. Unfortunately, they are considerably more expensive than what I would want to spend on a TT.

Sunday, September 19, 2021

How Telarc put their 50 kHz recordings onto the CD-layer of SACDs

Rev 9/20/21

While looking into whether Telarc's early 50 kHz digital recordings have been digitally converted to CD-grade digital by using an asynchronous sample-rate converter (which became available in about 2006), I ran across a 2004 review of Telarc hybrid SACD-60634 (Saint-SaĆ«ns, Symphony No. 3 “Organ”, Eugene Ormandy, Philadelphia Orchestra) which was made by converting the 50 kHz master to DSD using a dCS 972 digital-format converter, then to analog, and then to CD-grade PCM using a custom Telarc ADC. The article indicates that they had tried using a sample-rate converter to convert directly from 50 kHz to 44.1 kHz, but that the results didn't sound very good due to problems with sample-rate converters at that time. The Wikipedia article on Soundstream provides detailed information on the digital recorder which Telarc used for making their 50 kHz recordings, which could be released as 50 kHz FLACs to minimize the number of sample-rate conversions which it would undergo in the process of being converted to analog.

I've also learned that someone developed a way to convert 48 kHz recordings to 44.1 kHz very early in the digital era, and that Decca released a lot of its 48 kHz recordings on a CD-collection known as The Decca Sound, which according to The Decca Sound by S. Andrea Sundaram, doesn't sound so great, although it's not clear exactly why.

So, there are probably other examples of early digital recordings, with sampling rates other than 44.1 kHz, which were transferred to CDs early in the CD-era. Such recordings were released as LPs, so there had to be decent DACs to convert them to analog, which could have been digitized with something like a JVC VP-900, an oversampling ADC with a 16/44.1 output, apparently introduced in 1982. So, it would have had 16-bit linearity and a linear phase characteristic, so that its recordings would have good low-level detail, good high-end detail, and good imaging, but it was expensive. In September of 1985, Apogee introduced its linear-phase aftermarket input filters for the typical early digital recorder, and recording engineers adopted them in droves as quickly as possible, so that most digital recorders soon had a clean high end and good imaging. But before then, a lot of digital recordings had poor detail and imaging.

However, mass-market CD players in general were lousy until at least 2010, due to the sound quality of low-cost audio-DAC chips. According to Benchmark's app note entitled A Look Inside the New ES9028PRO Converter Chip and the New DAC3 [November 14, 2016]:

"It has been a little over 7 years since ESS Technology introduced the revolutionary ES9018 audio D/A converter chip. This converter delivered a major improvement in audio conversion and, for 7 years, it has held its position as the highest performing audio D/A converter chip. But a new D/A chip has now claimed this top position. Curiously the successor did not come from a competing company; it came from ESS. On October 19, 2016, ESS Technology announced the all-new ES9028PRO 32-bit audio D/A converter. In our opinion, ESS is now two steps ahead of the competition!"

So, I gather that the 9018 Sabre DAC introduced in 2010 was the first really good audio-DAC chip, and it would have been too expensive at that time to put in mass-market players. But now there are many good inexpensive audio DAC-chips (although Benchmark is still partial to Sabre DACs), and Sabre DACs are appearing in low-cost players and DACs. I have a $100 2017 Nobsound Bluetooth 4.2 Lossless Player with a 9018 Sabre DAC, and it's amazing, although it's crude compared to Benchmark's DACs.

But due to piracy fears, the best versions of some albums are reserved for high-res streaming and LPs. According to various audio experts, including high-res expert Mark Waldrep, PhD (a.k.a. Dr. AIX), whose website is RealHD-audio.com, high-res recordings and LPs sound better than CDs because they're mixed and mastered better, and not because of the recording format. So, CDs supposedly could sound as good as high-res or LPs, if they were mixed and mastered as well, and the low-level detail could be kept out of the dither-region, where it is mixed with noise which is intended to mask the severe low-level distortion of 16-bit digital. There was a period in CD-history known as the "loudness wars," when CDs were recorded at the highest possible level because they would sell better, perhaps because high recording levels kept the low-level details out of the dither-region. Unfortunately, this approach required excessive compression and might have led to clipping.

Friday, September 17, 2021

LXmini speaker system

As much as I'd like an affordable version of the B&W Nautilus speakers, which cost between $60K and $90K, someone probably would have started producing such a speaker by now if they were ever going to. They aren't patented, since when B&W tried to patent the principle, they found that someone had patented it about fifty years earlier (see Musical 'mollusc' is fifty years late https://www.newscientist.com/article/mg14920123-400-musical-mollusc-is-fifty-years-late/). It seems to me that the tapered coiled transmission line for the woofer could be made inexpensively by molding it out of special plastic in the form of left and right halves, lining them with special foam, and gluing them together.

Another attractive but probably also quite expensive line of speakers is made by PMC of the UK. They've refined the traditional transmission line concept, of which their Advanced Transmission Line web-page
(https://pmc-speakers.com/technology/atl) has one of the most succinct descriptions I've seen .

But there is an inexpensive speaker known as the LXmini (https://www.linkwitzlab.com/LXmini/Introduction.htm), which is based on transmission-line principles, and which according to many reviews, provides amazing sound quality for its price. This isn't surprising, considering that it was designed by the late great Siegfried Linkwitz, an electronics genius with a passion for designing the ideal loudspeaker. If you peruse his website, you'll find abundant evidence of his genius and passion for speaker-design. He had designed speakers even before the ones shown on his site, so he had quite a bit of experience, and the LXmini was his latest.

These are the best LXmini reviews I've found:

A) Stereophile article on an audio show where Linkwitz speakers were demonstrated (https://www.stereophile.com/content/rmaf-2014-reichert-sunday)

B) LXmini review from home theater reviewer (https://www.hometheatershack.com/threads/linkwitz-lab-lxmini-kit-speaker-performance-review.137434/)

C) LXC [LXmini knock-off] page (https://sites.google.com/site/cdenneler/home/lxc)

For more bass, there's the LXmini+2 system, which includes two dipole woofers, two more power amps, and a miniDSP 4x10 crossover instead of a 4x4HD. An advantage of dipole woofers is that they don't pressurize the room, at least as much, and therefore don't excite room resonances as much as typical subwoofers.

The 4x4HD crossover used with the LXmini has two digital inputs and an analog input, which is internally converted to digital. It also has a volume control which is controlled with an optional remote. Because I prefer to minimize conversions between the analog and digital realms, I had to find a replacement for my analog-output lossless player, and after considering many alternatives, decided to use one of my PCs as a music-server. This would allow me to use my collection of music files on 4GB USB flash drives, which I like because they streamline the process of de-cluttering, reorganizing, and defragmenting, and if one of them dies, I don't lose much, considering that each one costs just a few bucks.

For analog purists, there's also an LXmini analog electronic crossover, designed by Nelson Pass, who is famous for his high end audio electronics designs. It's available as a kit for about $200, which is a steal, considering its performance.

The PC has a Toslink output which I'd run to the Toslink input on the miniDSP crossover. To listen through headphones, I'd connect an FX-Audio DAC-amp to the PC. My TV would go to the miniDSP's USB input, and my old analog receiver's preamp outputs would go to the miniDSP's analog inputs, for listening to FM. But whenever I listen to FM, I'm reminded that it's become a wasteland, now that all the good music has gone to Sirius XM.

Monday, September 13, 2021

Top expert concludes that CD-grade digital is as good as we need

Mark Waldrep, PhD, a.k.a. Dr. AIX, is one of the top experts on high-res audio, if not THE top expert, as a look at his website, Real-HD Audio, should convince anyone. After decades of believing that high-res consumer recordings sound better than CDs, he has concluded that the reason that they have a reputation for superior sound quality is due to a variety of reasons, but not due to their higher resolution (see The Truth About High-Resolution Audio: Facts, Fiction and Findings). This is also the conclusion of other experts, such as Sean Olive, PhD, former president of the AES and now a director of Harman International's research division, and Goldmund Labs, where cost is no object and they would love an excuse to sell even more expensive digital gear than they already do. Dr. Waldrep conducted a survey, which might be ongoing, by releasing some of the high-res recordings which he made, and knows to be truly high-res, in both high-res and CD-grade, and having participants identify which is which by just listening to them, and fill out an on-line questionnaire. The aforementioned article contains a link to this survey.

However, the fact is that some high-res recordings, such as those available via lossless streaming, do sound better than the corresponding CD-grade recordings, although not necessarily because of the resolution. I surmise that these better-sounding recordings are streamed in high-res so that the music industry can attribute the higher quality to the higher resolution (to avoid having to explain why they're not released on CD), while preventing them from being copied digitally. The LP-version might sound better yet, because the record companies can put their best recordings on LPs without having to worry much about anyone being able to make an exact copy, because so few people can afford laser turntables, and those who do have laser TTs can be watched for indications of piracy.

I've devised a piracy-proof pay-per-play system (described in a previous post) which would allow us to download and store encrypted music files which could be decrypted only by the player with the corresponding decryption key, to avoid having to download lots of data every time the music is played (which might not be possible under all circumstances in which you would want to hear the music), while generating long-term income for the industry. The player would erase the key if any tampering were detected or if the internal batteries were allowed to drain excessively. If the key were erased, it could be replaced by a newly-generated decryption key at the factory under tight security (to avoid the risk of having to store the decryption keys anywhere besides the players), but all of the music files would have to be re-downloaded, which would act as a deterrent to tampering or letting the batteries discharge excessively. But if CD-grade files are sufficient, at least the downloads wouldn't be huge. Some people object to pay-per-play, but they could continue to use the existing inferior systems. I see it as a cost-effective alternative, because I'd be able to obtain the best recordings in digital form and avoid the expenses associated with the existing systems.

Tuesday, August 3, 2021

A proposal for a piracy-proof pay-per-play digital audio system

 Anyone with a TV knows that high-res recordings sound much better than CDs. My cheap old system can't reproduce very high frequencies very well (although once in a while I'm very satisfied with it), and yet I can clearly hear the superiority of TV-audio over CDs. We typically assume that this is due to the higher sampling rate and resolution of high-res digital, but according to various articles by audio authorities, including High-Resolution Audio: Does it Sound Better?​ by Goldmund Acoustic Laboratory (posted to Goldmund.com), so-called high-res recordings for consumers use only about 16 bits of the format's 24 bits, and sampling rates over 44.1 kHz are useless at best.

An article entitled Is high-resolution audio really as good as it sounds?, by Ian Paul of TechHive, cites experts who contend that high-def audio sounds better because "sound engineers often put more care and attention into higher-resolution recordings than they do to mass market CD releases." This implies that CDs could sound as good as high-res recordings if the engineers would do a better job of the mastering, but I suspect that the mediocre mastering is deliberate, because the music industry doesn't want to put their best recordings on CD, at least when the corresponding album is new and popular, because it's easy to make perfect copies of CDs. (There are error-correction systems built into the CD record/playback system. Software is stored on CDs, and it can't have any errors. CD-treatments didn't reduce the corrected error-rate - they improved the sound quality of early cheap CD players with poor isolation between the transport and analog power supplies, by reducing laser-servo hunting and the resulting noise on the analog power supply.) So, at first they give us versions made from substandard analog masters, or a poor mastering job, such as by using too much compression or bandwidth-limiting, or by shoving the detail down into the dither. [1]  Later, they sell superior versions based on superior source recordings and/or superior mastering, and in some cases inferior versions which supposedly sound better because they're made with high-resolution converters or from the original master, which is used only because it has deteriorated. There are quite a few exceptions to this strategy, such as Jean-Luc Ponty's CDs [2], but I'd rather that it didn't exist at all.

Other experts contend that high-res formats do provide audible benefits. Dan Lavry of Lavry Engineering, which manufactures high-end DACs and pro-grade ADCs, recommends a minimum sampling rate of 60 kHz (36% higher than 44.1 kHz), and apparently a minimum of 18 bits, which practically speaking requires the use of 24-bit formats with 88.2 kHz or 96 kHz sampling-rates. Anything higher than 96 kHz, he claims, only adds problems. Based on my experience listening to CD-grade digital, I tend to agree with Lavry, because the high end on even the best CDs doesn't seem sufficiently "airy" or "open." So it seems to me that we need a new standard consumer format, with a bit-depth of 20 bits and a sampling rate slightly higher than 60 kHz, to which the existing high-res formats could be converted cleanly and fairly easily. Such a standard would minimize file size while providing the highest audio quality.   

Good recordings and masterings are typically placed on LPs from the start, because LPs can't be copied exactly and easily, like CDs.  Some people spend hundreds of thousands of dollars for a Goldmund Reference turntable, in an attempt to extract all of the detail from their LPs. I've heard a $10K Linn Sondek LP12, which for me was a revelation in palpability, and Linn has recently introduced a new design for the LP12's platter-bearing. So, LPs can sound much better than most of us will ever know.

A record-cleaning machine is absolutely essential for anyone who's serious about getting the maximum detail from LPs, as well as minimizing wear. A "dustbuster" just doesn't cut it. When you buy your first turntable, get a cheaper turntable if necessary to get a record-cleaning machine. You can upgrade your turntable later, but you can't undo record wear.  

Tangential (linear-tracking) tonearms seem like a good idea, because they eliminate tracking error and skating force. According to Myles B. Astor, PhD, Senior Editor, Positive-Feedback.com, on AudioNirvana.org (https://www.audionirvana.org/forum/the-audio-vault/analog-playback/tonearms/49834-linear-tracking-tonearms),  

"High-end audio seems to be filled with those products that offer the promise of exceptional performance but are held back by some flaw or another. Those products that you know in your heart can be fixed if given a little TLC. Then you find out after spending ages with that component that no matter what you do that issue will never be resolved.

"One such product is linear tracking (air and mechanical). I'm sure that many of us here at one time or another (or still do) in their audio journey mounted a linear tracking arm on the turntable. Yes, in theory linear tracking arms offer the ultimate in LP playback. But in reality, linear tracking arms are held back many issues not the least of which include:

Movement and ability to maintain tangency to the groove [some tangential arms use sensors to detect deviations from tangency, and servo motors to move the arm along the track, which I always thought was a bad design]
Solidity and low octave reproduction [some tangential arms are too light and flimsy]
Susceptibility to warp wow with short arms [some tangential designs have very short arms]
High horizontal mass [a problem when playing records that aren't perfectly round, requiring the entire arm to move back and forth to track the record]
Air flow issues [with air-bearing arms]
Machining issues eg. the best air bearings might have a 1/10,000 of an inch tolerance and finding a material that isn't susceptible to temperature fluctuation is challenging
Freedom of movement [excessive friction]
Inability to maintain azimuth across the record [airplane azimuth: wing up/down - I don't know why this is an issue]
Setup can be frustrating

"Yet, some of the best sound I've ever heard is from air bearing arms such as the now sadly discontinued Air Tangent tonearm. The arm's resolution, sense of spaciousness and soundstage were something to behold. But at the same time, the arm wasn't the equal of a pivoted arm when it came to reproduction of the lowest octaves. The of course, there's the issue of convenience. After a while, who wants to hassle with air compressors, filters, tubing running everywhere in the room. Especially in smaller apartment quarters as opposed to someone with a house who could put the compressor in another room, garage or basement.

"But honestly, I do miss listening to those arms and sometimes I ponder putting a linear tracking in addition to a pivoted arm on my turntable. Maybe one day."

That day might have arrived. Bergmann Audio of Denmark has devised an elegant ball-bearing design (see photo). The design concept isn't entirely new, but the implementation is simpler and more precise than its predecessors, which can be seen on the aforementioned Positive Feedback page. I suppose that the bearings are permanently lubricated with something like a Teflon film on the bearing races. I gather that the bearings aren't sealed because doing so would add friction. The bearings and track could just be blasted with canned air periodically to get rid of dust. It could probably be implemented for a few hundred dollars in carbon fiber (the bearings and carriage would have to be metal, however). It's not yet on the market as of this writing, so perhaps they're still tweaking the design, or perhaps they're afraid that it would destroy the market for the ridiculously expensive air-bearing designs. A lot of people would probably get a turntable with the Bergmann ball-bearing tangential arm just because it's so ideal. 




 

But I'm not interested in the ultimate sound quality for myself - I just want to be able to enjoy music, and I find that one of the biggest factors in being able to enjoy a particular recording, assuming that the record/playback chain is fairly transparent, is being in the mood to listen to the recording. (My speakers leave much to be desired, but I can put up with them until someone comes out with a $1K pair knock-offs of the $60K B&W Nautilus, including the amps and analog electronic crossovers. The cabinets could be molded, the drivers don't have to be the ultimate, and the crossovers and amps could be combined into modules to be placed next to the speakers.)

So, I propose a system which would allow us to obtain good digital recordings, of whatever sampling rate and bit-depth we prefer (assuming that someone is willing to provide them, and as long as there are suitable players/DACs), without the possibility of piracy, and which would allow consumers to be charged a little each time they play their copy, so that the album doesn't have to be sold over and over again as better and better digital versions to keep making money. So, everyone would win, and it would be cost-effective for consumers, despite having to pay a little each time they play something. The source-file would be downloaded in encrypted form and stored in encrypted form on devices owned by the consumer. The files would be encrypted for the intended player, and only the decryption key stored on the player could decrypt them. Nobody would ever have access to the decryption key, which would be stored in EEPROM which would be erased within seconds if any of the following occurred:

A) The player's case were opened (which would require cutting or breaking the case, or special tools which would be illegal for consumers to obtain)
B) Any motion, such as a drill bit, were detected by motion sensors within the case
C) If the temperature got too low or too high, perhaps indicating a plan to kill the internal batteries, which would be highly reliable rechargeable batteries with a reasonable temperature range.
D) If the internal battery voltage got too low, perhaps indicating a plan to let the batteries die in order to access the decryption keys. The player would provide as much warning as possible to connect the AC adapter, depending on how quickly the batteries were being drained, before erasing the EEPROM. Replacement EEPROMs could be installed only under tightly controlled conditions at the factory, where it could be ascertained that nobody had had access to the decryption key on the EEPROM.

The player would also try to send an alert (including GPS data in this situation only, due to the potential for a crime in progress, possibly at some otherwise-unknown location) to the server immediately, or at least as soon as it could establish a link to the server.

There would obviously have to be special provisions for preventing the EEPROM from being erased during assembly or repair, when the case would be open. Perhaps the activation-code (which the server would normally provide, and which would be changed after every activation) could be entered, followed by a code to disable the motion sensors. Naturally, after the player is reassembled, the sensors would be re-enabled. This would be performed only at the factory under tightly controlled circumstances, and be done in a manner which would prevent anyone from having access to the decryption key or the activation code. (If the existing activation code and a new activation code aren't entered by a certain time and date, due to the owner's failure to pay the bill, the player would be deactivated, but obviously the decryption key wouldn't be erased, and the activation code would still be valid. When the owner pays the bill, the current and next activation codes would be sent the next time the player connects to the server. (There could be a button on the player to cause it to connect to the server, instead of waiting for it to do so automatically.) Obviously deactivation doesn't deactivate the player's ability to connect to the server - just the ability to play music. In fact, if the player fails to connect to the server every few hours, a music-industry security official would try to contact the owner to determine what's going on, and the police would pay a visit if the owner doesn't respond. Prospective buyers would be informed of these conditions to ensure that they are willing to accept them, and to let them know that there is no way to beat the system.

Huge rewards could be offered during the development phase for cracking the security, so that the best hackers would do their best to crack it.

The player would have analog outputs only, to ensure that nobody could build a device that would provide access to the unencrypted data. So, even the basic player would have an excellent DAC-section, which isn't difficult with modern DAC-chips. There could also be high-end models.

The player would keep track of what's played on it (the player would contain an FeRAM or MRAM - nonvolatile memories with better data-retention than flash memory at high temperatures, although with other problems such as relatively tiny capacity and relatively high write-current requirements). This information would be uploaded to the server at the end of the billing-cycle, and if the bill weren't paid in a reasonable amount of time, the player would be deactivated during one of its more frequent security-related connections to the server. The activation-code would be changed for each billing-cycle, and would be unique for every player.

When it's time to contact the server, which would be controlled by an internal clock (which would be synchronized to internet time whenever the player is connected to the server), the player would establish a secure link to the server by means of an IoT module. Although very little data would be sent over this link, it would have to be very secure because it would contain the authorization codes (current and next) and usage-data. The details would have to be worked out by a circuit designer who is familiar with each aspect of the overall design.

The manufacturing process would be automated and monitored by security personnel to ensure that nobody would ever have direct access to the decryption keys.

To convince people to switch to the new system, a representative selection of its music would be available in decrypted form for free.

Although quantum computers might some day be able to easily crack some types of asymmetric encryption (i.e. with separate encrypt and decrypt keys), there are already some types which quantum computers can't crack, and experts are working on others.

The server would serve files from all record labels, and be protected by a bullet-proof firewall, or a series of firewalls, which would limit inputs to the server to a specific format, and validate every input to ensure that it conforms to a valid selection and an activated player, and that the registered owner is making the request, such as via two-layer authentication and perhaps an iris scan, to absolutely prevent hacking.

There would also be a limit on how much someone could be charged per billing period, so that people can feel free to play as much music as they like, without worrying about receiving a massive bill. However, the amount would be sufficient to pay a fair amount to the music industry for any amount of music that someone would likely play in a billing period.

It would be convenient and inexpensive for the vast majority of people to download these files, such as free access to wi-fi hotspots for this purpose. Eliminating the need to download the music every time you want to listen to it is one of the main advantages of this system. High-res streaming requires a heavy-duty internet connection whenever listening, and universal high-res streaming would put a heavy load on the internet. If and when high-capacity NRAM becomes available, it will become feasible to download vast amounts of data to a wi-fi-connected phone at extreme rates, and store it indefinitely even at 200F. The files could be copied or transferred to a couple of large drives (primary and backup), which would serve as a personal music archive.

Some might object to this system by noting that the analog output could be digitized and stored as a regular FLAC-file, and played with something that wouldn't levy per-play charges, but the inconvenience, loss of quality, and lack of significant savings would prevent most people from committing this form of theft. Besides, this can already be done with LPs and high-res streaming.

Music could still be released on CDs for those who don't like the idea of paying a little each time they listen to a high-quality digital recording of their favorite music, and would rather pay more for a lower-quality recording on CD to avoid paying a small amount each time they listen to it.

Notes


[1] Dither is noise with a special frequency balance, with an amplitude of about 1 LSB p-p, added after mixing, etc. to mask the grittiness of 16-bit digital recordings, and to give the impression that there's 20 bits of dynamic range, so that for example fade-outs don't have audible cutoffs. With dither, bit 16 becomes a sort of DSD-bit, which allows it to represent details smaller than it normally would, but limited to low-frequency details, with high-frequency details converted to noise, which I believe explains the flakiness of the high-end on many CDs.

[2] The original CD-release of Ponty's album Open Mind (which includes outstanding solos by Chick Corea on synth, and George Benson on electric guitar) was apparently digitized with a Sony PCM-1610 16-bit, 44.1 KHz digital processor with stock input filters, so that the CD has a dry high end and mediocre imaging. However, the 1990 release was apparently digitized with a PCM-1610 with Apogee's superior input filters, which produced recordings with a detailed, liquid high end and good imaging, and the transfer to digital was otherwise done very well. The analog master was also excellent, so the CD is very clean and juicy. The music is great for unwinding. The 1990 release, which is sold by Amazon and perhaps others, is apparently distinguished from the original release only by differences of a few seconds in the listed duration-times of the cuts (the catalog number and packaging are the same for both releases). The 1990 release's listed duration-times are as follows: 1-8:05; 2-6:05; 3-4:57; 4-7:17; 5-5:14; 6-7:40.