U
    -e]                     @   s  d Z ddlZddlZddlmZ ddlmZ ddlmZ ddlm	Z	 ddl
mZ dd	lmZmZmZmZmZmZmZ dd
lmZ ddlmZ ddddddddgZejeeejejf dddZejeejdddZejeejdddZejeeejdddZddddejeeee e!eejdddZ"d d!ejeeeef  e!ejd"d#dZ#d$d%d&ej$ej$fejeeeeeef eejd'd(d)Z%d&ej$d$d%ej$d*ejeeeef eeeeejejf d+d,dZ&d&ej$d$d%ej$d*ejeeeef eeeejd+d-dZ'ed.d.d.d/ejeee ed ejd0d1d2Z(ed.d.d3ejeee ed  eejejf d0d4d2Z(ed.d.d3ejeee e!eejeejejf f d0d5d2Z(d6ddd/ejeee e!eejeejejf f d0d7d2Z(ed.d.d.d/ejeee ed ejd0d8d9Z)ed.d.d3ejeee ed  eejejf d0d:d9Z)d6ddd/ejeee e!eejeejejf f d0d;d9Z)dS )<a  
Effects
=======

Harmonic-percussive source separation
-------------------------------------
.. autosummary::
    :toctree: generated/

    hpss
    harmonic
    percussive

Time and frequency
------------------
.. autosummary::
    :toctree: generated/

    time_stretch
    pitch_shift

Miscellaneous
-------------
.. autosummary::
    :toctree: generated/

    remix
    trim
    split
    preemphasis
    deemphasis
    N   )core)	decompose)feature)util)ParameterError)AnyCallableIterableOptionalTupleUnionoverload)Literal)	ArrayLikehpssharmonic
percussivetime_stretchpitch_shiftremixtrimsplit)ykwargsreturnc                 K   sT   t | }tj|f|\}}t j|| j| jd d}t j|| j| jd d}||fS )a  Decompose an audio time series into harmonic and percussive components.

    This function automates the STFT->HPSS->ISTFT pipeline, and ensures that
    the output waveforms have equal length to the input waveform ``y``.

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        audio time series. Multi-channel is supported.
    **kwargs : additional keyword arguments.
        See `librosa.decompose.hpss` for details.

    Returns
    -------
    y_harmonic : np.ndarray [shape=(..., n)]
        audio time series of the harmonic elements
    y_percussive : np.ndarray [shape=(..., n)]
        audio time series of the percussive elements

    See Also
    --------
    harmonic : Extract only the harmonic component
    percussive : Extract only the percussive component
    librosa.decompose.hpss : HPSS on spectrograms

    Examples
    --------
    >>> # Extract harmonic and percussive components
    >>> y, sr = librosa.load(librosa.ex('choice'))
    >>> y_harmonic, y_percussive = librosa.effects.hpss(y)

    >>> # Get a more isolated percussive component by widening its margin
    >>> y_harmonic, y_percussive = librosa.effects.hpss(y, margin=(1.0,5.0))
    dtypelengthr   stftr   r   istftr   shape)r   r   r!   	stft_harm	stft_percy_harmy_perc r(   P/var/www/html/Darija-Ai-Train/env/lib/python3.8/site-packages/librosa/effects.pyr   <   s
    $
c                 K   s8   t | }tj|f|d }t j|| j| jd d}|S )aj  Extract harmonic elements from an audio time-series.

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        audio time series. Multi-channel is supported.
    **kwargs : additional keyword arguments.
        See `librosa.decompose.hpss` for details.

    Returns
    -------
    y_harmonic : np.ndarray [shape=(..., n)]
        audio time series of just the harmonic portion

    See Also
    --------
    hpss : Separate harmonic and percussive components
    percussive : Extract only the percussive component
    librosa.decompose.hpss : HPSS for spectrograms

    Examples
    --------
    >>> # Extract harmonic component
    >>> y, sr = librosa.load(librosa.ex('choice'))
    >>> y_harmonic = librosa.effects.harmonic(y)

    >>> # Use a margin > 1.0 for greater harmonic separation
    >>> y_harmonic = librosa.effects.harmonic(y, margin=3.0)
    r   r   r   r    )r   r   r!   r$   r&   r(   r(   r)   r   l   s    
c                 K   s8   t | }tj|f|d }t j|| j| jd d}|S )ax  Extract percussive elements from an audio time-series.

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        audio time series. Multi-channel is supported.
    **kwargs : additional keyword arguments.
        See `librosa.decompose.hpss` for details.

    Returns
    -------
    y_percussive : np.ndarray [shape=(..., n)]
        audio time series of just the percussive portion

    See Also
    --------
    hpss : Separate harmonic and percussive components
    harmonic : Extract only the harmonic component
    librosa.decompose.hpss : HPSS for spectrograms

    Examples
    --------
    >>> # Extract percussive component
    >>> y, sr = librosa.load(librosa.ex('choice'))
    >>> y_percussive = librosa.effects.percussive(y)

    >>> # Use a margin > 1.0 for greater percussive separation
    >>> y_percussive = librosa.effects.percussive(y, margin=3.0)
    r   r   r   r    )r   r   r!   r%   r'   r(   r(   r)   r      s    
)r   rater   r   c                K   st   |dkrt dtj| f|}tj|||dd|ddd}tt| jd | }tj|f| j	|d|}|S )	a  Time-stretch an audio series by a fixed rate.

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        audio time series. Multi-channel is supported.
    rate : float > 0 [scalar]
        Stretch factor.  If ``rate > 1``, then the signal is sped up.
        If ``rate < 1``, then the signal is slowed down.
    **kwargs : additional keyword arguments.
        See `librosa.decompose.stft` for details.

    Returns
    -------
    y_stretch : np.ndarray [shape=(..., round(n/rate))]
        audio time series stretched by the specified rate

    See Also
    --------
    pitch_shift :
        pitch shifting
    librosa.phase_vocoder :
        spectrogram phase vocoder
    pyrubberband.pyrb.time_stretch :
        high-quality time stretching using RubberBand

    Examples
    --------
    Compress to be twice as fast

    >>> y, sr = librosa.load(librosa.ex('choice'))
    >>> y_fast = librosa.effects.time_stretch(y, rate=2.0)

    Or half the original speed

    >>> y_slow = librosa.effects.time_stretch(y, rate=0.5)
    r   zrate must be a positive number
hop_lengthNn_fft)r*   r+   r,   r   r   )
r   r   r!   Zphase_vocodergetintroundr#   r"   r   )r   r*   r   r!   Zstft_stretchZlen_stretchZ	y_stretchr(   r(   r)   r      s    &

   Zsoxr_hqF)bins_per_octaveres_typescale)r   srn_stepsr1   r2   r3   r   r   c          	      K   sl   t |std| ddt| |  }tjt| fd|i|t|| |||d}t j|| jd dS )a  Shift the pitch of a waveform by ``n_steps`` steps.

    A step is equal to a semitone if ``bins_per_octave`` is set to 12.

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        audio time series. Multi-channel is supported.

    sr : number > 0 [scalar]
        audio sampling rate of ``y``

    n_steps : float [scalar]
        how many (fractional) steps to shift ``y``

    bins_per_octave : int > 0 [scalar]
        how many steps per octave

    res_type : string
        Resample type. By default, 'soxr_hq' is used.

        See `librosa.resample` for more information.

    scale : bool
        Scale the resampled signal so that ``y`` and ``y_hat`` have approximately
        equal total energy.

    **kwargs : additional keyword arguments.
        See `librosa.decompose.stft` for details.

    Returns
    -------
    y_shift : np.ndarray [shape=(..., n)]
        The pitch-shifted audio time-series

    See Also
    --------
    time_stretch :
        time stretching
    librosa.phase_vocoder :
        spectrogram phase vocoder
    pyrubberband.pyrb.pitch_shift :
        high-quality pitch shifting using RubberBand

    Examples
    --------
    Shift up by a major third (four steps if ``bins_per_octave`` is 12)

    >>> y, sr = librosa.load(librosa.ex('choice'))
    >>> y_third = librosa.effects.pitch_shift(y, sr=sr, n_steps=4)

    Shift down by a tritone (six steps if ``bins_per_octave`` is 12)

    >>> y_tritone = librosa.effects.pitch_shift(y, sr=sr, n_steps=-6)

    Shift up by 3 quarter-tones

    >>> y_three_qt = librosa.effects.pitch_shift(y, sr=sr, n_steps=3,
    ...                                          bins_per_octave=24)
    zbins_per_octave=z must be a positive integer.g       @r*   )Zorig_srZ	target_srr2   r3   r   )size)	r   Zis_positive_intr   floatr   Zresampler   Z
fix_lengthr#   )	r   r4   r5   r1   r2   r3   r   r*   Zy_shiftr(   r(   r)   r      s    F


	T)align_zeros)r   	intervalsr8   r   c                C   s   g }|r8t | }tt |d }t|t|g}|D ]6}|rT|t|| }|| d|d |d f  q<tj	|ddS )a^  Remix an audio signal by re-ordering time intervals.

    Parameters
    ----------
    y : np.ndarray [shape=(..., t)]
        Audio time series. Multi-channel is supported.
    intervals : iterable of tuples (start, end)
        An iterable (list-like or generator) where the ``i``th item
        ``intervals[i]`` indicates the start and end (in samples)
        of a slice of ``y``.
    align_zeros : boolean
        If ``True``, interval boundaries are mapped to the closest
        zero-crossing in ``y``.  If ``y`` is stereo, zero-crossings
        are computed after converting to mono.

    Returns
    -------
    y_remix : np.ndarray [shape=(..., d)]
        ``y`` remixed in the order specified by ``intervals``

    Examples
    --------
    Load in the example track and reverse the beats

    >>> y, sr = librosa.load(librosa.ex('choice'))

    Compute beats

    >>> _, beat_frames = librosa.beat.beat_track(y=y, sr=sr,
    ...                                          hop_length=512)

    Convert from frames to sample indices

    >>> beat_samples = librosa.frames_to_samples(beat_frames)

    Generate intervals from consecutive events

    >>> intervals = librosa.util.frame(beat_samples, frame_length=2,
    ...                                hop_length=1).T

    Reverse the beat intervals

    >>> y_out = librosa.effects.remix(y, intervals[::-1])
    r   .r   r   Zaxis)
r   Zto_mononpnonzeroZzero_crossingsappendlenr   Zmatch_eventsconcatenate)r   r9   r8   y_outZy_monozerosintervalr(   r(   r)   r   W  s    /
 i   i   <   )r   frame_lengthr+   top_dbref	aggregater   c                 C   sv   t j| ||d}tj|ddddf |dd}|jdkrlt||t|jd }tj|t	t|jd d}|| kS )a?  Frame-wise non-silent indicator for audio input.

    This is a helper function for `trim` and `split`.

    Parameters
    ----------
    y : np.ndarray
        Audio signal, mono or stereo

    frame_length : int > 0
        The number of samples per frame

    hop_length : int > 0
        The number of samples between frames

    top_db : number > 0
        The threshold (in decibels) below reference to consider as
        silence

    ref : callable or float
        The reference amplitude

    aggregate : callable [default: np.max]
        Function to aggregate dB measurements across channels (if y.ndim > 1)

        Note: for multiple leading axes, this is performed using ``np.apply_over_axes``.

    Returns
    -------
    non_silent : np.ndarray, shape=(m,), dtype=bool
        Indicator of non-silent frames
    )r   rD   r+   .r   N)rF   rE   r   r:   )
r   Zrmsr   Zamplitude_to_dbndimr;   Zapply_over_axesrangeZsqueezetuple)r   rD   r+   rE   rF   rG   Zmsedbr(   r(   r)   _signal_to_frame_nonsilent  s    )
rL   )rE   rF   rD   r+   rG   )r   rE   rF   rD   r+   rG   r   c                C   s   t | |||||d}t|}|jdkrfttj|d |d}t| jd ttj|d d |d}	nd\}}	t	dg| j
 }
t	||	|
d< | t|
 t||	gfS )aC  Trim leading and trailing silence from an audio signal.

    Parameters
    ----------
    y : np.ndarray, shape=(..., n)
        Audio signal. Multi-channel is supported.
    top_db : number > 0
        The threshold (in decibels) below reference to consider as
        silence
    ref : number or callable
        The reference amplitude.  By default, it uses `np.max` and compares
        to the peak amplitude in the signal.
    frame_length : int > 0
        The number of samples per analysis frame
    hop_length : int > 0
        The number of samples between analysis frames
    aggregate : callable [default: np.max]
        Function to aggregate across channels (if y.ndim > 1)

    Returns
    -------
    y_trimmed : np.ndarray, shape=(..., m)
        The trimmed signal
    index : np.ndarray, shape=(2,)
        the interval of ``y`` corresponding to the non-silent region:
        ``y_trimmed = y[index[0]:index[1]]`` (for mono) or
        ``y_trimmed = y[:, index[0]:index[1]]`` (for stereo).

    Examples
    --------
    >>> # Load some audio
    >>> y, sr = librosa.load(librosa.ex('choice'))
    >>> # Trim the beginning and ending silence
    >>> yt, index = librosa.effects.trim(y)
    >>> # Print the durations
    >>> print(librosa.get_duration(y), librosa.get_duration(yt))
    25.025986394557822 25.007891156462584
    rD   r+   rF   rE   rG   r   r+   r   r   )r   r   N)rL   r;   flatnonzeror6   r.   r   frames_to_samplesminr#   slicerH   rJ   asarray)r   rE   rF   rD   r+   rG   
non_silentr<   startendZ
full_indexr(   r(   r)   r     s&    /	

c                C   s   t | |||||d}tt|t}|d g}|d rP|dtdg |d rn|tt	|g t
jt||d}t|| jd }|d}|S )an  Split an audio signal into non-silent intervals.

    Parameters
    ----------
    y : np.ndarray, shape=(..., n)
        An audio signal. Multi-channel is supported.
    top_db : number > 0
        The threshold (in decibels) below reference to consider as
        silence
    ref : number or callable
        The reference amplitude.  By default, it uses `np.max` and compares
        to the peak amplitude in the signal.
    frame_length : int > 0
        The number of samples per analysis frame
    hop_length : int > 0
        The number of samples between analysis frames
    aggregate : callable [default: np.max]
        Function to aggregate across channels (if y.ndim > 1)

    Returns
    -------
    intervals : np.ndarray, shape=(m, 2)
        ``intervals[i] == (start_i, end_i)`` are the start and end time
        (in samples) of non-silent interval ``i``.
    rM   r   r   r   rN   )r      )rL   r;   rO   diffastyper.   insertarrayr=   r>   r   rP   r?   minimumr#   Zreshape)r   rE   rF   rD   r+   rG   rT   edgesr(   r(   r)   r     s$    "

.)coefzi	return_zf)r   r^   r_   r`   r   c                C   s   d S Nr(   r   r^   r_   r`   r(   r(   r)   preemphasisb  s    rc   )r^   r_   c                C   s   d S ra   r(   rb   r(   r(   r)   rc   m  s    c                C   s   d S ra   r(   rb   r(   r(   r)   rc   x  s    g
ףp=
?c                C   s   t jd| g| jd}t jdg| jd}|dkrTd| dddf  | dddf  }t |}tjj||| t j|| jdd\}}|r||fS |S )	a	  Pre-emphasize an audio signal with a first-order differencing filter:

        y[n] -> y[n] - coef * y[n-1]

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        Audio signal. Multi-channel is supported.

    coef : positive number
        Pre-emphasis coefficient.  Typical values of ``coef`` are between 0 and 1.

        At the limit ``coef=0``, the signal is unchanged.

        At ``coef=1``, the result is the first-order difference of the signal.

        The default (0.97) matches the pre-emphasis filter used in the HTK
        implementation of MFCCs [#]_.

        .. [#] https://htk.eng.cam.ac.uk/

    zi : number
        Initial filter state.  When making successive calls to non-overlapping
        frames, this can be set to the ``zf`` returned from the previous call.
        (See example below.)

        By default ``zi`` is initialized as ``2*y[0] - y[1]``.

    return_zf : boolean
        If ``True``, return the final filter state.
        If ``False``, only return the pre-emphasized signal.

    Returns
    -------
    y_out : np.ndarray
        pre-emphasized signal
    zf : number
        if ``return_zf=True``, the final filter state is also returned

    Examples
    --------
    Apply a standard pre-emphasis filter

    >>> import matplotlib.pyplot as plt
    >>> y, sr = librosa.load(librosa.ex('trumpet'))
    >>> y_filt = librosa.effects.preemphasis(y)
    >>> # and plot the results for comparison
    >>> S_orig = librosa.amplitude_to_db(np.abs(librosa.stft(y)), ref=np.max, top_db=None)
    >>> S_preemph = librosa.amplitude_to_db(np.abs(librosa.stft(y_filt)), ref=np.max, top_db=None)
    >>> fig, ax = plt.subplots(nrows=2, sharex=True, sharey=True)
    >>> librosa.display.specshow(S_orig, y_axis='log', x_axis='time', ax=ax[0])
    >>> ax[0].set(title='Original signal')
    >>> ax[0].label_outer()
    >>> img = librosa.display.specshow(S_preemph, y_axis='log', x_axis='time', ax=ax[1])
    >>> ax[1].set(title='Pre-emphasized signal')
    >>> fig.colorbar(img, ax=ax, format="%+2.f dB")

    Apply pre-emphasis in pieces for block streaming.  Note that the second block
    initializes ``zi`` with the final state ``zf`` returned by the first call.

    >>> y_filt_1, zf = librosa.effects.preemphasis(y[:1000], return_zf=True)
    >>> y_filt_2, zf = librosa.effects.preemphasis(y[1000:], zi=zf, return_zf=True)
    >>> np.allclose(y_filt, np.concatenate([y_filt_1, y_filt_2]))
    True

    See Also
    --------
    deemphasis
          ?r   NrW   .r   r   r_   )r;   rS   r   
atleast_1dscipysignallfilter)r   r^   r_   r`   bar@   Zz_fr(   r(   r)   rc     s    L$
$c                C   s   d S ra   r(   rb   r(   r(   r)   
deemphasis  s    rm   c                C   s   d S ra   r(   rb   r(   r(   r)   rm     s    c                C   s   t jd| g| jd}t jdg| jd}|dkrt jt| jdd dg | jd}tjj||| |d\}}|d| | dd	df  | dddf  d
|  |t 	| jd   8 }n*t 
|}tjj||| || jd\}}|r||fS |S dS )a  De-emphasize an audio signal with the inverse operation of preemphasis():

    If y = preemphasis(x, coef=coef, zi=zi), the deemphasis is:

    >>> x[i] = y[i] + coef * x[i-1]
    >>> x = deemphasis(y, coef=coef, zi=zi)

    Parameters
    ----------
    y : np.ndarray [shape=(..., n)]
        Audio signal. Multi-channel is supported.

    coef : positive number
        Pre-emphasis coefficient.  Typical values of ``coef`` are between 0 and 1.

        At the limit ``coef=0``, the signal is unchanged.

        At ``coef=1``, the result is the first-order difference of the signal.

        The default (0.97) matches the pre-emphasis filter used in the HTK
        implementation of MFCCs [#]_.

        .. [#] https://htk.eng.cam.ac.uk/

    zi : number
        Initial filter state. If inverting a previous preemphasis(), the same value should be used.

        By default ``zi`` is initialized as
        ``((2 - coef) * y[0] - y[1]) / (3 - coef)``. This
        value corresponds to the transformation of the default initialization of ``zi`` in ``preemphasis()``,
        ``2*x[0] - x[1]``.

    return_zf : boolean
        If ``True``, return the final filter state.
        If ``False``, only return the pre-emphasized signal.

    Returns
    -------
    y_out : np.ndarray
        de-emphasized signal
    zf : number
        if ``return_zf=True``, the final filter state is also returned

    Examples
    --------
    Apply a standard pre-emphasis filter and invert it with de-emphasis

    >>> y, sr = librosa.load(librosa.ex('trumpet'))
    >>> y_filt = librosa.effects.preemphasis(y)
    >>> y_deemph = librosa.effects.deemphasis(y_filt)
    >>> np.allclose(y, y_deemph)
    True

    See Also
    --------
    preemphasis
    rd   re   Nr   r   rf   rW   .r      )r;   r[   r   rA   listr#   rh   ri   rj   Zarangerg   rY   )r   r^   r_   r`   rk   rl   r@   zfr(   r(   r)   rm     s"    @$&
 )*__doc__numpyr;   Zscipy.signalrh    r   r   r   r   Zutil.exceptionsr   typingr   r	   r
   r   r   r   r   Ztyping_extensionsr   Znumpy.typingr   __all__Zndarrayr   r   r   r7   r   r.   strboolr   r   maxrL   r   r   rc   rm   r(   r(   r(   r)   <module>   s:  !$ 0**B[  B
;
P
F

`
