<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>U-Net on 语音/音频论文速递</title>
    <link>https://nanless.github.io/audio-paper-digest-blog/tags/u-net/</link>
    <description>Recent content in U-Net on 语音/音频论文速递</description>
    <generator>Hugo</generator>
    <language>zh-cn</language>
    <lastBuildDate>Wed, 29 Apr 2026 00:00:00 +0000</lastBuildDate>
    <atom:link href="https://nanless.github.io/audio-paper-digest-blog/tags/u-net/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Bridging the Front-End and Back-End for Robust ASR via Cross-Attention-Based U-Net</title>
      <link>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-bridging-the-front-end-and-back-end-for-robust/</link>
      <pubDate>Wed, 29 Apr 2026 00:00:00 +0000</pubDate>
      <guid>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-bridging-the-front-end-and-back-end-for-robust/</guid>
      <description>语音识别 | 7.0/10</description>
    </item>
    <item>
      <title>Diff-vs: Efficient Audio-Aware Diffusion U-Net for Vocals Separation</title>
      <link>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-diff-vs-efficient-audio-aware-diffusion-u-net-for/</link>
      <pubDate>Wed, 29 Apr 2026 00:00:00 +0000</pubDate>
      <guid>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-diff-vs-efficient-audio-aware-diffusion-u-net-for/</guid>
      <description>语音分离 | 7.5/10</description>
    </item>
    <item>
      <title>FUN-SSL: Full-Band Layer Followed by U-Net With Narrow-Band Layers for Multiple Moving Sound Source Localization</title>
      <link>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-fun-ssl-full-band-layer-followed-by-u-net-with/</link>
      <pubDate>Wed, 29 Apr 2026 00:00:00 +0000</pubDate>
      <guid>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-fun-ssl-full-band-layer-followed-by-u-net-with/</guid>
      <description>声源定位 | 8.0/10</description>
    </item>
    <item>
      <title>TVP-UNet: Threshold Variance Penalty U-Net for Voice Activity Detection in Dysarthric Speech</title>
      <link>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-tvp-unet-threshold-variance-penalty-u-net-for/</link>
      <pubDate>Wed, 29 Apr 2026 00:00:00 +0000</pubDate>
      <guid>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-29-tvp-unet-threshold-variance-penalty-u-net-for/</guid>
      <description>语音活动检测 | 7.0/10</description>
    </item>
    <item>
      <title>ArtifactNet: Detecting AI-Generated Music via Forensic Residual Physics</title>
      <link>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-21-artifactnet-detecting-ai-generated-music-via/</link>
      <pubDate>Tue, 21 Apr 2026 00:00:00 +0000</pubDate>
      <guid>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-21-artifactnet-detecting-ai-generated-music-via/</guid>
      <description>这篇论文旨在解决AI生成音乐检测中普遍存在的泛化能力差的问题。当前主流方法（如CLAM、SpecTTTra）通过学习AI音乐的声音特征，在面对未见过的生成器时性能急剧下降。作者提出了一个核心假设：当前主流AI音乐生成器（如Suno, Udio）都依赖神经音频编解码器（如EnCodec）的残差矢量量化</description>
    </item>
    <item>
      <title>语音/音频论文速递 2026-04-21</title>
      <link>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-21/</link>
      <pubDate>Tue, 21 Apr 2026 00:00:00 +0000</pubDate>
      <guid>https://nanless.github.io/audio-paper-digest-blog/posts/2026-04-21/</guid>
      <description>共分析 34 篇语音/AI 论文</description>
    </item>
  </channel>
</rss>
