From: Patent Tactics, George Brock-Nannestad Hello, I am just latching on to Don's mail - I absolutely agree, but I am not in agreement with a lot of what was written while I was away for a few days. First of all, I have to admit that it is unlikely that any pre-amp should tolerate a spark at the input - Tom Fine's complaint. In such a case, a circuit consisting of properly biased germanium diodes or transistors would reduce the max input to + or - 200 mV. The effect of feedback loop saturation is a very real one, and those amps with the best stationary signal performance are frequently the worst to display this on very fast input. Headroom and short feedback loops are what we need. I would have liked the discussion on analog vs. digital to have been a bit more informed, but on the other hand one cannot request any hardware and software user to have studied and understood a book like John Watkinson's 'The Art of Digital Audio' in order to permit them to write on the ARSCLIST. For this reason we must be thankful for the many who have taken time to elucidate a point; in this patchwork manner the general knowledge will increase. I am going to make a few notes that some will find boring, and I hope some will find faulty, because that will provoke them to express themselves as clearly as possible to correct me. The general problem is that sometimes something that is as clear as water to some is very murky to others, and the discussion does not take place in the same context - we all carry our own around with us. Now, most of what we are dealing with is bandwidth. Bandwidth and time resolution are linked in the world of resistors, capacitors, and inductors, but not so where our hearing is concerned. For solid proof of this go to Milind Kunchur's webpage - they have been mentioned before on this list. The so-called Nyquist criterion states that if you want to be able to reconstruct a signal from a sampled representation of it, you have to decide the highest frequency you want represented and then you have to sample at a rate that is at least twice the highest frequency. When you then reconstruct your signal by providing voltages at the sample rate according to the table of values that represents the audio signal, you will have your full bandwidth and dynamic range back, provided the resolution or bit-depth has been sufficient. It has been claimed that this stepwise presented waveform makes the signal unlistenable, but that is not the case, because you invariably smooth it by filtering, so nothing above your defined highest frequency gets out in the analog domain again. You will note that you are in full control: define the maximum frequency and define the resolution. If they are not sufficient to your purpose, go higher. In order to have a sampled representation of our analog sound signal (which is itself an electrical representation of a pressure variation around atmospheric pressure) we have to measure it at regular intervals defined by the sample rate and hold the value for as long it takes to obtain the tabular value - the value that goes into the digital word stream. Now, we have a problem with pure sampling of a waveform: if it has a frequency that is more than half of the sampling rate, that too will be sampled, but in this case under-sampled, which means that the result appears to be at a completely different and inharmonic and jarring frequency, an alias. Once we have adopted a sampling frequency we must simply ensure that no signal above half the sampling rate is available for sampling. This is done by filtering, so-called anti-alias filtering. With 44.1 kHz sampling rate, no signal above 22.05 kHz is permissible. On the other hand, we do want our 20 kHz bandwidth - this is what a young, pre-earbud ear can mostly hear. So, our filter has to go from full transmission at 20kHz to zero transmission at 22.05 kHz *). No problem, our telephone engineers have done this kind of exercise for 90 years. However, such sharp cut-off filters come with some frightful time delay distortion (phase to some), very audible down to 2 kHz. That was the situation for about 10 years in CD audio, until someone came up with the idea to correct the time response in the digital domain by means of a digital filter. The currency to pay for this is total delay time, but for something recorded a year ago, some microseconds do not matter. In other words, many of our problems come from the filter. If we increase the sampling rate we can make use of our more frequent samples in two ways: we can do a gentler filtering that does not have the delay effect at low frequencies, or we can increase our highest frequency. If we do the latter, we shall be able to increase the time resolution of our digital representation. With 20 kHz the maximum slope of the waveform is only one quarter of that of an 80 kHz bandwidth - reachable by a 192 kHz sampling frequency and a gentler anti-aliasing filter. Dependent on where you are on the scale of knowing electronics at graduate level, you will miss a discussion of one-bit sampling and also of dithering. I am not out to write a book. John Watkinson did. There is one thing that I would love to see in the market: an interactive program where you could take a number of input waveforms and vary the parameters bit depth, bandwidth, sampling rate, anti-alias filter type and parameters, all to experiment on, to see what things look like when you are in control. On the mock-ups based on this principle I have seen, one of the most impressive is "oops, where has it gone", when a frequency or a delay between two signals suddenly disappear. Or when two frequency response peaks close by suddenly merge into one. Understanding WHY they have gone seems to me to be very important. A couple of other items came up while we were at 78rpm reproduction: - the reason why we need an elevated bandwidth for recordings on rough surfaces is because that is where the noise signal is. If we are to deal properly with the noise it has to be represented correctly. And 80 kHz is a marked improvement and will also permit us to look into low-level artefacts at supersonic frequencies. I would refer you to my paper "What are the Sources of the Noises We Remove", Proceedings of AES 20th Int. Conf. 'Archiving, Restoration, and New Methods of Recording', Eds. Z. Vajda, H. Pichler, Budapest 5-7 October 2001, pp. 175-182. Unless someone else makes the claim, I may well have been the first to state that in restoration work using preservation copying, the bandwidth is not determined by the intended signal but by the noise. This was at the IASA Conference in Bogensee near Berlin in 1994. - one might consider that the pickup would be encountering a formidable vertical wall when it met a square wave. However, the recording is usually a velocity recording, that is the square waveshape is differentiated (1st derivative), which means it turns into a triangular wave. The problem is that with a high bandwidth, the corners of this triangular wave (where the square shifts from constant positive to constant negative and vice versa) are very sharp, and even that may be difficult to trace. George Alexandrovich made a 7" test record with a "square wave" - there are virtually no wiggles at the corners of the triangles. It is for testing pickups, but I have not dared to use it on anything but the ELP. - for the same reason a tick, modelled by a steep rise followed by steep fall becomes two ticks, one positive and one negative. - even if you remove these two peaks and interpolate or draw a waveform connecting the ends, the low frequency excitation of the whole cartridge-tonearm system (stylus, cantilever, bearing, cartridge mass, tonearm mass and resonances) still remains as a low-level thud - a tail. CEDAR started back in 1988 with a program developed by Peter Rayner while still at the British Library to remove not only the ticks, but also the tail. Well, for those of you who have not yet dropped off - best wishes, George *) Actually it is not zero, but determined by the noise floor in the desired digital representation or bit depth. Every bit added gives 6dB more headroom. ------------------------------------------ > On 11/04/2012, Tom Fine wrote: > > > > As for the idea that 192kHz is somehow inferior to 96k, that sounds > > like audiophoolery or outdated ideas based on older hardware. I would > > say it's unecessary overkill to sample at that high a rate, for the > > reasons you stated, but the playback results shouldn't sound inferior > > to 96k unless something is being done wrong by the equipment or the > > user. > > > The main advantage of high sampling rates is that it is easier for > software to distinguish clicks (impulses) from music. > > The waveform of a muted trumpet (as played by Buck Clayton, for > instance) is very similar to a rapid sequence of clicks, so software > operating at 44.1 KHz cannot automatically tell the difference. > > But the trumpet signal will fade out at the high frequency limit of the > recording equipment, while clicks will have higher frequency components. > > Regards > -- > Don Cox > [log in to unmask]