Re: Advice to an audiophobe ??

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



William Case wrote:
On Wed, 2008-12-24 at 08:30 +1030, Tim wrote:
On Tue, 2008-12-23 at 10:55 -0500, William Case wrote:
...
1. Gives me a whole range of adjustments for different channels.  (I
assume channels means different sources  e.g. Master, Headphone, PCM etc.).
Just to clarify, a source device generates an audio signal( line in, PCM (pulse code modulation = wav uncompressed audio), that might then get processed (eg volume control, master, headphone), and then sent to a destination (often an output device like a speaker connector or headphone jack).
Because a typical soundcard has an internal hardware mixer, it can 
usually mix together various inputs (sources) like CD input, mic input 
and recorded audio signals, and produce a single output signal (mostly 
in stereo=2 channels). When mixing together externally received signals, 
no main CPU processing power is used, unless you are trying to record to 
hard disk etc.
2. gives me two choices and 3., 4., 5. gives me only Master. Which should I choose and why?
capture means recording - usually from a line in (eg from a vcr or mp3 player etc), or from a microphone. So exclude those from your choice.
If I should be using HDA NVida (Alsa-mixer), why do I have PulseAudio
options?
Consider pulse audio to be a real-time digital mixer and volume control, where the audio calculations are performed inside your main CPU. In the default setup, once pulseaudio has done it's processing, it passes the result to the alsa driver which outputs the audio data to the soundcard. The soundcard turns the digital audio data into analog audio signals for use with amplifier, speakers, or headphones.
Pulseaudio also has enhanced capabilities like remembering that when you 
playback with xmms that you like to output via your amplifiers and 
speakers, but when you are viewing a flash video, to playback into your 
headphones instead, at a different level. Another capability let's you 
choose the destination playback device while the material is actually 
being played. A third capability let's the output go to an audio device 
on another machine. Obviously, this is a bit trickier to set up.
...
These individual mixer input controls should normally be left off if you
never use them, as they can each introduce noise (hiss, beeps and
burbles, etc.) to the system.
I will turn them off except for Master and Front.  I will experiment
with PC Speaker.  Of course these are only available to me if I use the
default alsa mixer setting.
It's not a one or other setting, both parts will still be involved; pulseaudio will process, mix, and attenuate sound signals, whereas alsa will drive the physical hardware. The setting you are seeing lets you decide whether to control the physical driver volume levels or the software generated pulseaudio volume controls. If you mute or turn the alsa master way down, it wont matter how high you turn the pulseaudio mixer, since the alsamixer comes after the pulseaudio one in the audio chain. (also true for the reverse).
If you play back a loud audio file, and turn both the pulseaudio source 
 and master up full. Then change to the alsa setting. You can then use 
the also setting to set up an absolute maximum level that you would want 
to hear, by adjusting the master. Then you could go back to the 
pulseaudio setting to adjust the playback to a comfortable setting, and 
from then on only use the pulseaudio setting.
...
      * How is sound related to video ?
Sound is the sound, video is the picture...  The question is too vague
to be answerable.
In digital format, sound and vision are both represented with digital 1's and 0's. With all video and audio file types, there is a packing together of the audio and video information into the one file. The multiplexed file provides information about when to playback each frame of video in relation to the audio in the file. For example, an mpeg2 (dvd) file might have two frames of video, then 2 of audio, then 1 of video, two audio in an order to achieve a consistent throughput of audio and video data.
      * Why are there so many files associated with producing sound?
In digital audio, the most basic file type is waveform (.wav), where each momentary value of audio is stored, on a 1 for 1 basis. Experiments and calculations can show us that for something we store as quality musical recording we need to sample that momentary value at 44kHz (times per second) or higher so as not to disrupt our digital recording with audio aliases. Since we also seem to enjoy the spatial enhancement produced by stereo or more channels, the file needs to store both left and right information. Finally, we found that if we only store the digital value using a small no of bits per sample, when played back we hear a harsh, chunky sound, rather than the CD like quality of using 16 (or more) bits per sample. The catch with all that is it takes up a lot space.
To solve space issues (less a problem now that storage space costs a lot 
less), compression schemes were developed. Most take advantage of 
reducing the number of channels eg to mono, reducing the sampling rate, 
or the number quantizing levels (bits/sample); but this is done in 
context of the type of audio being compressed - eg human voices are 
typically of lower frequency, and can sampled at a slower rate, and with 
less levels.
The biggest jump in compression was with psychoacoustic modelling, where 
it was found that in a complex sound, a listener does not notice that 
certain frequency (pitch) sounds become inaudible (or masked) by other 
sounds.
The reason there is so many formats, is because developers were 
essentially competing to produce more highly compressed audio files, 
without noticeable change in quality, when using a certain type of 
audio, over a certain communication medium. Eg: when the fastest home 
internet connections were slow modems, compression made it possible to 
transmit voice signals over your internet connection. If you tried to 
transmit music of higher quality that voice, you would have large audio 
distortions that made it difficult to hear the original material.
You might like to play with the audio editor program audacity (perhaps 
from rpmfusion if you want to be able to import and save in certain 
compressed formats (mp3)). It shows you a graphical representation of 
the audio file, and eg lets you choose a zoom, start and stop position, 
and just play back small parts of a file, so that you can work out what 
the sound "looks" like to a computer.
Hope that helps a bit more ;-)
DaveT.

--
fedora-list mailing list
fedora-list@xxxxxxxxxx
To unsubscribe: https://www.redhat.com/mailman/listinfo/fedora-list
Guidelines: http://fedoraproject.org/wiki/Communicate/MailingListGuidelines

[Index of Archives]     [Current Fedora Users]     [Fedora Desktop]     [Fedora SELinux]     [Yosemite News]     [Yosemite Photos]     [KDE Users]     [Fedora Tools]     [Fedora Docs]

  Powered by Linux