The Auphonic Audio Inspector is shown on the status page of a finished production and displays details about what our algorithms are changing in audio files. It is possible to zoom and scroll within audio waveforms and the Audio Inspector might be used to manually check production result and input files.
If you just want to try the Auphonic Audio Inspector yourself, take a look at this Multitrack Audio Inspector Example.
- Seek in Audio Files
- Click or tap inside the waveform to seek in files. The red playhead will show the current audio position.
- Zoom In/Out
- Use the zoom buttons ([+] and [-]), the mouse wheel or zoom gestures on touch devices to zoom in/out the audio waveform.
- Scroll Waveforms
- If zoomed in, use the scrollbar or drag the audio waveform directly (with your mouse or on touch devices).
- Show Legend
- Click the [?] button to show or hide the Legend, which describes details about the visualizations of the audio waveform.
- Show Input Track(s)
- Use the Show Input link to display input track(s) of a production: now you can see and listen to input and output files for a detailed comparison. Please click directly on the waveform to switch/unmute a track - muted tracks are grayed out slightly:
Please click on the fullscreen button (bottom right) to switch to fullscreen mode.
Now the audio tracks use all available screen space to see all waveform details:
In fullscreen mode, it’s also possible to control playback and zooming with keyboard shortcuts:
Press [Space] to start/pause playback, use [+] to zoom in and [-] to zoom out.
Singletrack Algorithms Inspector¶
First, we discuss the analysis data of our Singletrack Post Production Algorithms.
The audio levels of output and input files, measured according to the ITU-R BS.1770 specification, are displayed directly as the audio waveform. Click on Show Input to see the input and output file. Only one file is played at a time, click directly on the Input or Output track to unmute a file for playback:
- Waveform Segments: Music and Speech (gold, blue)
- Music/Speech segments are displayed directly in the audio waveform: Music segments are plotted in gold/yellow, speech segments in blue (or light/dark blue).
- Waveform Segments: Leveler High/No Amplification (dark, light blue)
- Speech segments can be displayed in normal, dark or light blue:
Dark blue means that the input signal was very quiet and contains speech,
therefore the Adaptive Leveler has to use a high amplification value in this segment.
In light blue regions, the input signal was very quiet as well, but our classifiers decided that the signal should not be amplified (breathing, noise, background sounds, etc.).
- Background Segments: Leveler Fade Up/Down (yellow, orange)
- If the volume of an input file changes in a fast way, the Adaptive Leveler volume
curve will increase/decrease very fast as well (= fade) and should be placed in speech pauses.
Otherwise, if fades are too slow or during active speech, one will hear
pumping speech artifacts.
Exact fade regions are plotted as yellow (fade up, volume increase) and orange (fade down, volume decrease) background segments in the audio inspector.
- Horizontal Lines: Noise and Hum Reduction Profiles (red)
- Our Noise and Hiss Reduction and Hum Reduction algorithms segment the audio file in regions
with different background noise characteristics, which are displayed as red
horizontal lines in the audio inspector
(top lines for noise reduction, bottom lines for hum reduction).
Then a noise print is extracted in each region and a classifier decides if and how much noise reduction is necessary - this is plotted as a value in dB below the top red line.
The hum base frequency (50Hz or 60Hz) and the strength of all its partials is also classified in each region, the value in Hz above the bottom red line indicates the base frequency and whether hum reduction is necessary or not (no red line).
Multitrack Algorithms Inspector¶
If our Multitrack Post Production Algorithms are used, additional analysis data is shown in the audio inspector.
The audio levels of the output and all input tracks are measured according to the ITU-R BS.1770 specification and are displayed directly as the audio waveform. Click on Show Input to see all the input files with track labels and the output file. Only one file is played at a time, click directly into the track to unmute a file for playback:
- Input Tracks: Waveform Segments, Background Segments and Horizontal Lines
Input tracks are displayed below the output file including their track names. The same data as in our Singletrack Algorithms Inspector is calculated and plotted separately in each input track:
- Output Waveform Segments: Multiple Speakers and Music
- Each speaker is plotted in a separate, blue-like color - in the example above
we have 3 speakers (normal, light and dark blue) and you can see directly in the waveform
when and which speaker is active.
Audio from music input tracks are always plotted in gold/yellow in the output waveform, please try to not mix music and speech parts in music tracks (see also Multitrack Best Practice)!
Ducking, Background and Foreground Segments¶
Music tracks can be set to Ducking, Foreground, Background or Auto - for more details please see Automatic Ducking, Foreground and Background Tracks.
- Ducking Segments (light, dark orange)
- In Ducking, the level of a music track is reduced if one of the speakers is active,
which is plotted as a dark orange background segment in the output track.
Foreground music parts, where no speaker is active and the music track volume is not reduced, are displayed as light orange background segments in the output track.
- Background Music Segments (dark orange background)
- Here the whole music track is set to Background and won’t be amplified when speakers are inactive.
Background music parts are plotted as dark organge background segments in the output track.
- Foreground Music Segments (light orange background)
- Here the whole music track is set to Foreground and its level won’t be reduced when speakers are active.
Foreground music parts are plotted as light organge background segments in the output track.
You can try the ducking/background/foreground audio inspector yourself: Fore/Background/Ducking Audio Examples.
Audio Search, Chapters Marks and Videos¶
- Audio Search and Transcriptions
- If our Automatic Speech Recognition Integration is used, a time-aligned transcription text will be shown above the waveform.
You can use the search field to search and seek directly in the audio file.
See our Speech Recognition Audio Examples to try it yourself.
- Chapters Marks
- Chapter Mark start times are displayed in the audio waveform as black vertical lines.
The current chapter title is written above the waveform - see “This is Chapter 2” in the screenshot above.
Supported Audio Formats¶
- WAV, MP3, AAC/M4A and Opus
- These formats are supported in all major browsers: Firefox, Chrome, Safari, Edge, iOS Safari and Chrome for Android.
- FLAC is supported in Firefox, Chrome, Edge and Chrome for Android - see
FLAC audio format.
- Ogg Vorbis
- Only supported by Firefox, Chrome and Chrome for Android - for details please see Ogg Vorbis audio format.
We suggest to use a recent Firefox or Chrome browser for best performance.
Please see our blog post Audio File Formats and Bitrates for Podcasts for more details about audio formats.
Mobile Audio Inspector¶
Multiple responsive layouts were created to optimize the screen space usage on Android and iOS devices, so that the audio inspector is fully usable on mobile devices as well: tap into the waveform to set the playhead location, scroll horizontally to scroll waveforms, scroll vertically to scroll between tracks, use zoom gestures to zoom in/out, etc.
Unfortunately the fullscreen mode is not available on iOS devices (thanks to Apple), but it works on Android and is a really great way to inspect everything using all the available screen space: