一种用于截幅音频修复中的自适应一致迭代硬阈值算法
详细信息    查看全文 | 推荐本文 |
  • 英文篇名:An Adaptive Consistent Iterative Hard Thresholding Alogorith for Audio Declipping
  • 作者:邹霞 ; 吴彭龙 ; 孙蒙 ; 张星昱
  • 英文作者:ZOU Xia;WU Penglong;SUN Meng;ZHANG Xingyu;The Army Engineering University of PLA;
  • 关键词:音频信号处理 ; 截幅失真 ; 自适应门限 ; 一致迭代硬阈值
  • 英文关键词:Audio signal processing;;Clipping distortion;;Adaptive threshold;;Consistent Iterative Hard Thresholding(CIHT)
  • 中文刊名:DZYX
  • 英文刊名:Journal of Electronics & Information Technology
  • 机构:陆军工程大学指挥控制工程学院;
  • 出版日期:2018-12-18 10:15
  • 出版单位:电子与信息学报
  • 年:2019
  • 期:v.41
  • 基金:国家自然科学基金(61402519);; 江苏省优秀青年基金(BK20180080)~~
  • 语种:中文;
  • 页:DZYX201904023
  • 页数:7
  • CN:04
  • ISSN:11-4494/TN
  • 分类号:168-174
摘要
一致迭代硬阈值(CIHT)算法在处理音频截幅失真中具有较好的性能。但是,在截幅程度较大时音频截幅修复的性能会下降。因此,该文提出一种基于自适应门限的改进算法。该算法自动估计音频信号截幅程度,根据估计的截幅程度信息,自适应调整算法中的截幅程度因子。与近年来提出的CIHT算法和一致字典学习算法(CDL)相比,该文所提算法能更好地重建音频信号,特别在音频信号截幅失真严重的情况。该算法的运算复杂度与CIHT相近,与CDL相比,拥有更快的运行速度,有利于实时实现。
        Audio clipping distortion can be solved by the Consistent Iterative Hard Thresholding(CIHT)algorithm, but the performance of restoration will decrease when the clipping degree is large, so, an algorithm based on adaptive threshold is proposed. The method estimates automatically the clipping degree, and the factor of the clipping degree is adjusted in the algorithm according to the degree of clipping. Compared with the CIHT algorithm and the Consistent Dictionary Learning(CDL) algorithm, the performance of restoration by the proposed algorithm is much better than the other two, especially in the case of severe clipping distortion. Compared with CDL, the computational complexity of the proposed algorithm is low like CIHT,compared with CDL, it has faster processing speed, which is beneficial to the practicality of the algorithm.
引文
[1]JANSSEN A,VELDHUIS R,and VRIES L.Adaptive interpolation of discrete-time signals that can be modeled as autoregressive processes[J].IEEE Transactions on Acoustics,Speech,and Signal Processing,1986,34(2):317 -330.doi:10.1109/TASSP.1986.1164824.
    [2]ABEL J S and ABEL J S.Restoring a clipped signal[C].IEEE International Conference on Acoustics,Speech,and Signal Processing,Toronto,Canada,1991,3:1745-1748.doi:10.1109/ICASSP.1991.150655.
    [3]SIMON J G,PATRICK J,and WILLIAM N W.Statistical model-based approaches to audio restoration and analysis[J].Journal of New Music Research,2001,30(4):323-338.doi:10.1076/jnmr.30.4.323.7489.
    [4]ADLER A,EMIYA V,and JAFARI M G.Audio Inpainting[J].IEEE Transactions on Audio,Speech,and Language Processing,2012,20(3):922-932.doi:10.1109/TASL.2011.2168211.
    [5]ADLER A,EMIYA V,and JAFARI M G.A constrained matching pursuit approach to audio declipping[C].IEEEInternational Conference on Acoustics,Speech,and Signal Processing,Prague,Czech Republic,2011:329-332.doi:10.1109/ICASSP.2011.5946407.
    [6]DEFRAENE B,MANSOUR N,and HERTOGH S D.Declipping of audio signals using perceptual compressed sensing[J].IEEE Transactions on Audio,Speech,and Language Processing,2013,21(12):2627-2637.doi:10.1109/TASL.2013.2281570.
    [7]FOUCART S and NEEDHAM T.Sparse recovery from saturated measurements[J].Information and Inference:AJournal of the IMA,2017,6(2):196-212.doi:10.1093/imaiai/iaw020.
    [8]OZEROV A,BILEN C,and PEREZ P.Multichannel audio declipping[C].IEEE International Conference on Acoustics,Speech,and Signal Processing,Shanghai,China,2016:659-663.
    [9]KAI S,KOWALSKI M,and DORFLER M.Audio declipping with social sparsity[C].IEEE International Conference on Acoustics,Speech,and Signal Processing,Florence,Italy,2014:1577-1581.doi:10.1109/ICASSP.2014.6853863.
    [10]KITIC S,JACQUES L,and MADHU N.Consistent iterative hard thresholding for signal declipping[C].IEEEInternational Conference on Acoustics,Speech,and Signal Processing,Vancouver,Canada,2013:5939-5943.doi:10.1109/ICASSP.2013.6638804.
    [11]RENCKER L,BACH F,WANG Wenwu,et al.Consistent dictionary learning for signal declipping[C].International Conference on Latent Variable Analysis and Signal Separation,Guildford,UK,2018:446-455.doi:10.1007/978-3-319-93764-9_41.
    [12]LECUE G and FOUCART S.An IHT algorithm for sparse recovery from sub-exponential measurements[J].IEEESignal Processing Letters,2017,24(3):1280-1283.doi:10.1109/LSP.2017.2721500.
    [13]HINES A,SKOGLUND J,and KOKARAM A.Robustness of speech quality metrics to background noise and network degradations:Comparing ViSQOL,PESQ and POLQA[C].IEEE International Conference on Acoustics,Speech,and Signal Processing,Vancouver,Canada,2013:3697-3701.doi:10.1109/ICASSP.2013.6638348.
    [14]何孝月.基于EPESQ的VoIP语音质量评估的研究与实现[D].[硕士论文],中南大学,2008.doi:10.7666/d.y1326186.HE Xiaoyue.Speech Quality Evaluation of VoIP Based on EPESQ[D].[Master dissertation],Central South University,2008.doi:10.7666/d.y1326186.