Re: Advice to an audiophobe ??

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



William Case wrote:
On Wed, 2008-12-24 at 08:30 +1030, Tim wrote:
On Tue, 2008-12-23 at 10:55 -0500, William Case wrote:
...
1. Gives me a whole range of adjustments for different channels.  (I
assume channels means different sources  e.g. Master, Headphone, PCM etc.).
Just to clarify, a source device generates an audio signal( line in, PCM (pulse code modulation = wav uncompressed audio), that might then get processed (eg volume control, master, headphone), and then sent to a destination (often an output device like a speaker connector or headphone jack).

Because a typical soundcard has an internal hardware mixer, it can usually mix together various inputs (sources) like CD input, mic input and recorded audio signals, and produce a single output signal (mostly in stereo=2 channels). When mixing together externally received signals, no main CPU processing power is used, unless you are trying to record to hard disk etc.

2. gives me two choices and 3., 4., 5. gives me only Master. Which should I choose and why?
capture means recording - usually from a line in (eg from a vcr or mp3 player etc), or from a microphone. So exclude those from your choice.

If I should be using HDA NVida (Alsa-mixer), why do I have PulseAudio
options?
Consider pulse audio to be a real-time digital mixer and volume control, where the audio calculations are performed inside your main CPU. In the default setup, once pulseaudio has done it's processing, it passes the result to the alsa driver which outputs the audio data to the soundcard. The soundcard turns the digital audio data into analog audio signals for use with amplifier, speakers, or headphones.

Pulseaudio also has enhanced capabilities like remembering that when you playback with xmms that you like to output via your amplifiers and speakers, but when you are viewing a flash video, to playback into your headphones instead, at a different level. Another capability let's you choose the destination playback device while the material is actually being played. A third capability let's the output go to an audio device on another machine. Obviously, this is a bit trickier to set up.

...
These individual mixer input controls should normally be left off if you
never use them, as they can each introduce noise (hiss, beeps and
burbles, etc.) to the system.
I will turn them off except for Master and Front.  I will experiment
with PC Speaker.  Of course these are only available to me if I use the
default alsa mixer setting.
It's not a one or other setting, both parts will still be involved; pulseaudio will process, mix, and attenuate sound signals, whereas alsa will drive the physical hardware. The setting you are seeing lets you decide whether to control the physical driver volume levels or the software generated pulseaudio volume controls. If you mute or turn the alsa master way down, it wont matter how high you turn the pulseaudio mixer, since the alsamixer comes after the pulseaudio one in the audio chain. (also true for the reverse).

If you play back a loud audio file, and turn both the pulseaudio source and master up full. Then change to the alsa setting. You can then use the also setting to set up an absolute maximum level that you would want to hear, by adjusting the master. Then you could go back to the pulseaudio setting to adjust the playback to a comfortable setting, and from then on only use the pulseaudio setting.

...
      * How is sound related to video ?

Sound is the sound, video is the picture...  The question is too vague
to be answerable.
In digital format, sound and vision are both represented with digital 1's and 0's. With all video and audio file types, there is a packing together of the audio and video information into the one file. The multiplexed file provides information about when to playback each frame of video in relation to the audio in the file. For example, an mpeg2 (dvd) file might have two frames of video, then 2 of audio, then 1 of video, two audio in an order to achieve a consistent throughput of audio and video data.

      * Why are there so many files associated with producing sound?
In digital audio, the most basic file type is waveform (.wav), where each momentary value of audio is stored, on a 1 for 1 basis. Experiments and calculations can show us that for something we store as quality musical recording we need to sample that momentary value at 44kHz (times per second) or higher so as not to disrupt our digital recording with audio aliases. Since we also seem to enjoy the spatial enhancement produced by stereo or more channels, the file needs to store both left and right information. Finally, we found that if we only store the digital value using a small no of bits per sample, when played back we hear a harsh, chunky sound, rather than the CD like quality of using 16 (or more) bits per sample. The catch with all that is it takes up a lot space.

To solve space issues (less a problem now that storage space costs a lot less), compression schemes were developed. Most take advantage of reducing the number of channels eg to mono, reducing the sampling rate, or the number quantizing levels (bits/sample); but this is done in context of the type of audio being compressed - eg human voices are typically of lower frequency, and can sampled at a slower rate, and with less levels.

The biggest jump in compression was with psychoacoustic modelling, where it was found that in a complex sound, a listener does not notice that certain frequency (pitch) sounds become inaudible (or masked) by other sounds.

The reason there is so many formats, is because developers were essentially competing to produce more highly compressed audio files, without noticeable change in quality, when using a certain type of audio, over a certain communication medium. Eg: when the fastest home internet connections were slow modems, compression made it possible to transmit voice signals over your internet connection. If you tried to transmit music of higher quality that voice, you would have large audio distortions that made it difficult to hear the original material.

You might like to play with the audio editor program audacity (perhaps from rpmfusion if you want to be able to import and save in certain compressed formats (mp3)). It shows you a graphical representation of the audio file, and eg lets you choose a zoom, start and stop position, and just play back small parts of a file, so that you can work out what the sound "looks" like to a computer.

Hope that helps a bit more ;-)
DaveT.

--
fedora-list mailing list
fedora-list@xxxxxxxxxx
To unsubscribe: https://www.redhat.com/mailman/listinfo/fedora-list
Guidelines: http://fedoraproject.org/wiki/Communicate/MailingListGuidelines

[Index of Archives]     [Current Fedora Users]     [Fedora Desktop]     [Fedora SELinux]     [Yosemite News]     [Yosemite Photos]     [KDE Users]     [Fedora Tools]     [Fedora Docs]

  Powered by Linux