FadeDetection

standard mode | Duration/silence category

Inputs

  • rms (vector_real) - rms values array

Outputs

  • fadeIn (matrix_real) - 2D-array containing start/stop timestamps corresponding to fade-ins [s] (ordered chronologically)

  • fadeOut (matrix_real) - 2D-array containing start/stop timestamps corresponding to fade-outs [s] (ordered chronologically)

Parameters

  • cutoffHigh (real ∈ (0, 1], default = 0.85) :

    fraction of the average RMS to define the maximum threshold

  • cutoffLow (real ∈ [0, 1), default = 0.2) :

    fraction of the average RMS to define the minimum threshold

  • frameRate (real ∈ (0, ∞), default = 4) :

    the rate of frames used in calculation of the RMS [frames/s]

  • minLength (real ∈ (0, ∞), default = 3) :

    the minimum length to consider a fade-in/out [s]

Description

This algorithm detects fade-in and fade-outs time positions in an audio signal given a sequence of RMS values. It outputs two arrays containing the start/stop points of fade-ins and fade-outs. The main hypothesis for the detection is that an increase or decrease of the RMS over time in an audio file corresponds to a fade-in or fade-out, repectively. Minimum and maximum mean-RMS-thresholds are used to define where fade-in and fade-outs occur.

An exception is thrown if the input “rms” is empty.

References:

[1] Fade (audio engineering) - Wikipedia, the free encyclopedia, http://en.wikipedia.org/wiki/Fade-in

Source code

See also

FadeDetection (streaming) RMS (standard) RMS (streaming)