<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[PhonoByte: Short Summary]]></title><description><![CDATA[Short Summary on latest Speech and Audio AI]]></description><link>https://sankar1535.substack.com/s/short-summary</link><image><url>https://substackcdn.com/image/fetch/$s_!4m5b!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73c1ee4-049d-4623-9003-7d1cf146709e_408x408.png</url><title>PhonoByte: Short Summary</title><link>https://sankar1535.substack.com/s/short-summary</link></image><generator>Substack</generator><lastBuildDate>Mon, 04 May 2026 10:10:56 GMT</lastBuildDate><atom:link href="https://sankar1535.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Sankar Mukherjee]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[sankar1535@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[sankar1535@substack.com]]></itunes:email><itunes:name><![CDATA[Sankar Mukherjee]]></itunes:name></itunes:owner><itunes:author><![CDATA[Sankar Mukherjee]]></itunes:author><googleplay:owner><![CDATA[sankar1535@substack.com]]></googleplay:owner><googleplay:email><![CDATA[sankar1535@substack.com]]></googleplay:email><googleplay:author><![CDATA[Sankar Mukherjee]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[BitNet Distillation: The 3-Stage Pipeline That Delivers Full-Precision LLM Performance at 1.58-bit Ultra-Low Quantization]]></title><description><![CDATA[BitNet Distillation Paper: [Link] Github: [Link]]]></description><link>https://sankar1535.substack.com/p/bitnet-distillation-the-3-stage-pipeline</link><guid isPermaLink="false">https://sankar1535.substack.com/p/bitnet-distillation-the-3-stage-pipeline</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Fri, 17 Oct 2025 13:05:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-zqU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>BitNet Distillation Paper: [<a href="https://arxiv.org/pdf/2510.13998">Link</a>]                          Github: [<a href="https://github.com/microsoft/BitNet">Link</a>]</p><div><hr></div><h3>The Problem</h3><p>The core challenge addressed is the <strong>difficulty of deploying Large Language Models (LLMs) in downstream applications</strong>, especially on resource-constrained devices like smartphones, due to their rapidly escalating size, which leads to prohibitive memory consumption and computational overhead.</p><p>Specifically related to ultra-low precision quantization:</p><ul><li><p>Existing extreme low-bit LLMs, such as the 1.58-bit (ternary values ${-1, 0, 1}$) BitNet, require pretraining from scratch on large-scale corpora to achieve competitive accuracy, which incurs substantial computational and energy overhead.</p></li><li><p>Directly fine-tuning existing full-precision LLMs into 1.58-bit precision (referred to as BitNet-SFT) for specific downstream tasks is often unstable, results in significant performance degradation, and exhibits <strong>poor scalability</strong>. The performance gap relative to the full-precision baseline tends to widen as the model size increases (e.g., from 0.6B to 4B, the gap relative to FP16 baseline grows from 13.9 to 15.3 in one experiment).</p></li><li><p>Key challenges identified when attempting to fine-tune pre-trained full-precision LLMs into 1.58-bit BitNet include <strong>performance degradation</strong>, <strong>poor scalability</strong>, and <strong>training instability</strong>.</p></li></ul><h3>The Solution Suggested</h3><p>The solution proposed is <strong>BitNet Distillation (BitDistill)</strong>, which is described as a tailored distillation framework and a scaling-friendly Quantization-Aware Training (QAT) framework.</p><ul><li><p>BitDistill is a <strong>lightweight pipeline</strong> designed to fine-tune off-the-shelf full-precision LLMs (such as Qwen) into <strong>1.58-bit precision</strong> (ternary weights ${-1, 0, 1}$) for specific downstream tasks.</p></li><li><p>The goal of BitDistill is to achieve strong task-specific performance with minimal computational cost and to bridge the gap between extreme 1.58-bit quantization and practical deployment.</p></li><li><p>BitDistill allows 1.58-bit quantized LLMs to achieve downstream performance comparable to their full-precision counterparts.</p></li></ul><h3>How the Solution Was Achieved</h3><p>BitDistill is a three-stage training pipeline designed to mitigate the identified challenges of instability and poor scalability:</p><h4>Stage 1: Modeling Refinement</h4><p>This stage addresses optimization instability caused by excessively large activation variance in low-bit quantized models.</p><ul><li><p>The architecture of the LLM is modified by integrating the <strong>SubLN module</strong>.</p></li><li><p>SubLN layers are inserted at specific positions inside each transformer block, specifically right before the output projection of the Multi-Head Self-Attention (MHSA) module and before the output projection of the Feed-Forward Network (FFN).</p></li><li><p>This design ensures that hidden representations entering quantized projection layers are variance-stabilized, preventing activation scale explosion and improving training stability and task performance.</p></li></ul><h4>Stage 2: Continued Pre-Training (Crucial Warm-up Step)</h4><p>This stage mitigates the poor scalability issue that arises when directly fine-tuning full-precision weights to 1.58-bit representations using limited downstream training tokens.</p><ul><li><p>A small amount of pretraining corpus (e.g., only 10B tokens sampled from the FALCON corpus) is used to fine-tune the modeling-modified LLMs from Stage 1. This cost is virtually negligible compared to pre-training 1.58-bit BitNet from scratch (approximately 4T tokens).</p></li><li><p>This warm-up step enables the BitNet models to rapidly adapt to a feature space better suited for 1.58-bit optimization, making the weight distribution more similar to BitNet trained from scratch (concentrating weights near transition boundaries), thus improving downstream performance and preventing convergence to suboptimal local minima.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-zqU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-zqU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 424w, https://substackcdn.com/image/fetch/$s_!-zqU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 848w, https://substackcdn.com/image/fetch/$s_!-zqU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 1272w, https://substackcdn.com/image/fetch/$s_!-zqU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-zqU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png" width="1146" height="628" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:628,&quot;width&quot;:1146,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:222314,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/176410792?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-zqU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 424w, https://substackcdn.com/image/fetch/$s_!-zqU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 848w, https://substackcdn.com/image/fetch/$s_!-zqU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 1272w, https://substackcdn.com/image/fetch/$s_!-zqU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92e958d8-5059-41fe-bb62-2e9f47a2813d_1146x628.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Stage 3: Distillation-based Fine-tuning</h4><p>This stage recovers the accuracy of the full-precision teacher model by incorporating two types of knowledge distillation during the downstream task fine-tuning phase.</p><ul><li><p><strong>Logits Distillation (LD):</strong> This technique minimizes the Kullback&#8211;Leibler divergence between the output probability distributions (logits) of the full-precision teacher (FP16) and the 1.58-bit student.</p></li><li><p><strong>Multi-Head Attention Distillation (AD):</strong> Based on the MiniLM series, this encourages the 1.58-bit student to capture the fine-grained structural dependencies embedded in the FP16 teacher&#8217;s attention patterns. This distillation is typically performed only at a single layer rather than across all layers to provide greater optimization flexibility.</p></li></ul><p>The total loss for Stage 3 combines the standard cross-entropy loss ($L_{CE}$), the logits distillation loss ($L_{LD}$), and the attention distillation loss ($L_{AD}$).</p><h3>The Results of These Solutions</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sEyh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sEyh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 424w, https://substackcdn.com/image/fetch/$s_!sEyh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 848w, https://substackcdn.com/image/fetch/$s_!sEyh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 1272w, https://substackcdn.com/image/fetch/$s_!sEyh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sEyh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png" width="1091" height="730" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:730,&quot;width&quot;:1091,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:201747,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/176410792?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!sEyh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 424w, https://substackcdn.com/image/fetch/$s_!sEyh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 848w, https://substackcdn.com/image/fetch/$s_!sEyh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 1272w, https://substackcdn.com/image/fetch/$s_!sEyh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff452ceb0-a85a-4b42-9669-02086cbf4c3a_1091x730.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>BitDistill achieved significant efficiency gains while maintaining high performance:</p><ul><li><p><strong>Performance:</strong> BitDistill achieved <strong>performance comparable to full-precision counterparts</strong> across model size and benchmarks (classification and summarization), with only marginal differences observed in most cases.</p><ul><li><p>For example, on the 4B model classification task (MNLI), FP16-SFT achieved 91.48 accuracy, while BitDistill achieved 91.40.</p></li><li><p>BitDistill models consistently achieved performance close to full-precision fine-tuning across alternative backbones tested (Gemma and Qwen2.5).</p></li><li><p>The framework can effectively leverage a higher-quality FP16 teacher (e.g., distilling a 0.6B student with a 4B teacher) to achieve greater downstream task gains, sometimes even surpassing FP16 models of the same size.</p></li></ul></li><li><p><strong>Memory Efficiency:</strong> BitDistill enabled up to <strong>10&#215; memory savings</strong>. For instance, a 0.6B Qwen3 model reduced memory usage from 1.20 GB (FP16-SFT) to 0.11 GB (BitDistill).</p></li><li><p><strong>Inference Speed:</strong> BitDistill enabled <strong>2.65&#215; faster inference on CPUs</strong> compared to 1.58-bit BitNet, and achieved up to a <strong>2&#215; inference speedup</strong> compared to FP16 baselines. For example, inference speed increased from 427 tokens/s (FP16-SFT) to 1,135 tokens/s (BitDistill).</p></li><li><p><strong>Scalability:</strong> Unlike direct fine-tuning (BitNet-SFT), BitDistill <strong>preserves scalability</strong>, allowing performance to remain comparable to full-precision counterparts across all model sizes.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!olUg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!olUg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 424w, https://substackcdn.com/image/fetch/$s_!olUg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 848w, https://substackcdn.com/image/fetch/$s_!olUg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 1272w, https://substackcdn.com/image/fetch/$s_!olUg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!olUg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png" width="1136" height="354" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:354,&quot;width&quot;:1136,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:134934,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/176410792?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!olUg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 424w, https://substackcdn.com/image/fetch/$s_!olUg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 848w, https://substackcdn.com/image/fetch/$s_!olUg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 1272w, https://substackcdn.com/image/fetch/$s_!olUg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e517e0c-2f17-4988-854a-077adf37a90c_1136x354.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div></li><li><p><strong>Ablation Studies:</strong> All three stages (Modeling Refinement, Continued Pre-Training, and Distillation-based Fine-tuning) were found to be complementary, as excluding any one stage led to a non-trivial drop in downstream performance. The combination of Logits Distillation and Multi-Head Attention Distillation provided the most consistent performance.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I1Fn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I1Fn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 424w, https://substackcdn.com/image/fetch/$s_!I1Fn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 848w, https://substackcdn.com/image/fetch/$s_!I1Fn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 1272w, https://substackcdn.com/image/fetch/$s_!I1Fn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I1Fn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png" width="472" height="262.04137931034484" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:322,&quot;width&quot;:580,&quot;resizeWidth&quot;:472,&quot;bytes&quot;:82840,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/176410792?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a546bd-8cd9-4228-a066-0fbe390d82e8_1129x322.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I1Fn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 424w, https://substackcdn.com/image/fetch/$s_!I1Fn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 848w, https://substackcdn.com/image/fetch/$s_!I1Fn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 1272w, https://substackcdn.com/image/fetch/$s_!I1Fn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd17201ea-c815-4977-b49c-27bd9c5889c5_580x322.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div></li><li><p><strong>Compatibility with Quantization Techniques:</strong> Further investigation demonstrated that BitDistill is <strong>complementary</strong> to and compatible with various existing post-training and weight-quantization methods (such as Block-Quant, GPTQ, and AWQ). Models consistently benefited from the proposed framework regardless of the underlying quantization method, often matching the full-precision baseline.</p></li></ul><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[MOSS-Speech]]></title><description><![CDATA[Hidden State Divergence: The Imperative for Architectural Split Preserving Textual Integrity in Speech LLMs]]></description><link>https://sankar1535.substack.com/p/moss-speech</link><guid isPermaLink="false">https://sankar1535.substack.com/p/moss-speech</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Tue, 07 Oct 2025 07:43:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ae5U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hLLH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hLLH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 424w, https://substackcdn.com/image/fetch/$s_!hLLH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 848w, https://substackcdn.com/image/fetch/$s_!hLLH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 1272w, https://substackcdn.com/image/fetch/$s_!hLLH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hLLH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png" width="971" height="425" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:425,&quot;width&quot;:971,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:120358,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/175505313?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hLLH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 424w, https://substackcdn.com/image/fetch/$s_!hLLH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 848w, https://substackcdn.com/image/fetch/$s_!hLLH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 1272w, https://substackcdn.com/image/fetch/$s_!hLLH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd30a376-2bca-4654-90f8-22f8e43062c5_971x425.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#129504; Spoken dialogue systems have advanced toward end-to-end designs, yet most still rely on intermediate text guidance, creating bottlenecks in latency and expressivity.</p><p>&#9889;Preliminary studies reveal that hidden-state alignment between speech and text deteriorates and diverges in deeper layers of a joint Transformer backbone. Why should a single shared architecture handle modality-specific generation in its final blocks? &#129300; This divergence risks degrading the pretrained text LLM&#8217;s reasoning and knowledge when new speech capabilities are added.</p><p>&#128640; The core innovation is MOSS-Speech &#8212; overcoming this by implementing a modality-based layer-splitting design with a frozen pre-training strategy &#10052;&#65039;.</p><p>The model routes the shared hidden state &#8212; after leveraging joint fusion in early layers &#8212; into separate modality-specific branches (text &#128221; and speech &#128266;) starting at the 32nd of 36 Transformer blocks.</p><p>This design solves cross-modal misalignment in deep layers by reserving final layers for modality-specific generation, enabling deep fusion early while preventing divergence near output &#128260;.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ae5U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ae5U!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 424w, https://substackcdn.com/image/fetch/$s_!ae5U!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 848w, https://substackcdn.com/image/fetch/$s_!ae5U!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 1272w, https://substackcdn.com/image/fetch/$s_!ae5U!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ae5U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png" width="829" height="698" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:698,&quot;width&quot;:829,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ae5U!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 424w, https://substackcdn.com/image/fetch/$s_!ae5U!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 848w, https://substackcdn.com/image/fetch/$s_!ae5U!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 1272w, https://substackcdn.com/image/fetch/$s_!ae5U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F852e178a-c4af-4a97-bf5e-18f7ae35448a_829x698.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Critically, this approach, with frozen text parameters, mitigates reasoning and knowledge degradation typically seen when extending LLMs to new modalities &#129513;.</p><p>As a full streaming architecture, the encoder&#8217;s WER of 10.80% is slightly higher than 9.17% for block-causal GLM-4-Voice. The decoder, though achieving better intelligibility and quality &#127775;, shows a marginal trade-off in speaker similarity on English and Chinese benchmarks versus CosyVoice 2.</p><p>Model initialization used the Qwen-3-8B backbone, trained across two stages using ~4M hours of speech data &#9201;&#65039; &#8212; 690k hours English and 952k hours Chinese interleaved speech&#8211;text &#8212; followed by two epochs of supervised fine-tuning on 1.5M synthetic multimodal QA pairs &#127911;&#128218;.</p><p>The pre-trained model preserved textual ability and improved speech modeling, gaining 69.53 on Chinese spoken StoryCloze vs 54.39 for GLM-4-Voice &#128200;.</p><p>In supervised fine-tuning, it achieved SOTA spoken QA (77.33 on LLaMA-QA) and high perceived speech quality (4.37 UTMOS) &#127942;.</p><p>This architecture enables reasoning transfer to speech while paving the way for speech-native models with seamless, expressive human&#8211;AI interaction &#129309;&#128172;.</p><p>&#128279; More:<a href="http://arxiv.org/abs/2510.00499"> arxiv.org/abs/2510.00499</a></p>]]></content:encoded></item><item><title><![CDATA[Revolutionizing Text-to-Speech: How Differentiable Reward Optimization is Changing the Game]]></title><description><![CDATA[This paper presents a Differentiable Reward Optimization method to enhance the performance of neural codec language model-based text-to-speech systems by directly predicting rewards from neural codec]]></description><link>https://sankar1535.substack.com/p/revolutionizing-text-to-speech-how</link><guid isPermaLink="false">https://sankar1535.substack.com/p/revolutionizing-text-to-speech-how</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Fri, 11 Jul 2025 03:12:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!H-NT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h6><strong><a href="https://arxiv.org/abs/2507.05911">Paper</a></strong></h6><p>The world of text-to-speech (TTS) synthesis has been transformed by large language models, but training these systems to produce high-quality, controllable speech has remained a significant challenge. Traditional reinforcement learning from human feedback (RLHF) approaches for TTS are computationally expensive and complex. However, groundbreaking research from Alibaba's Tongyi Lab introduces a novel solution: <strong>Differentiable Reward Optimization (DiffRO)</strong>.</p><h2>The Problem with Traditional TTS Training</h2><p>Current neural codec language model-based TTS systems face several hurdles when implementing RLHF:</p><p><strong>Computational Overhead</strong>: Unlike text-based language models, TTS systems require additional backend flow matching and vocoder models to convert discrete tokens into audio. This creates a massive computational burden when generating training data.</p><p><strong>Limited Sample Diversity</strong>: Generated TTS samples often exhibit high similarity, making it difficult to distinguish between positive and negative examples for reward model training.</p><p><strong>Complex Evaluation</strong>: TTS quality depends on multiple factors&#8212;pronunciation accuracy, naturalness, speaker similarity, and emotional expression&#8212;making simple binary classification insufficient.</p><h2>Introducing DiffRO: A Paradigm Shift</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H-NT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H-NT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 424w, https://substackcdn.com/image/fetch/$s_!H-NT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 848w, https://substackcdn.com/image/fetch/$s_!H-NT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 1272w, https://substackcdn.com/image/fetch/$s_!H-NT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H-NT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png" width="1456" height="647" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:647,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H-NT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 424w, https://substackcdn.com/image/fetch/$s_!H-NT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 848w, https://substackcdn.com/image/fetch/$s_!H-NT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 1272w, https://substackcdn.com/image/fetch/$s_!H-NT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e09fac-993e-425f-8fb9-745aa8fcf031_1600x711.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>DiffRO addresses these challenges through three key innovations:</p><h3>1. Token-Level Reward Prediction</h3><p>Instead of synthesizing audio to evaluate quality, DiffRO predicts rewards directly from neural codec tokens. This approach leverages the fact that codec tokens should contain all necessary information from the input text. By using an ASR-style approach to predict the original text from tokens, the system can assess pronunciation accuracy without expensive audio synthesis.</p><h3>2. Differentiable Training Process</h3><p>Traditional RLHF requires complex reinforcement learning loops. DiffRO uses the Gumbel-Softmax technique to make the reward function differentiable, enabling direct optimization through standard backpropagation. This eliminates the need for Proximal Policy Optimization (PPO) or Direct Preference Optimization (DPO) strategies.</p><h3>3. Multi-Task Reward (MTR) Model</h3><p>Rather than focusing solely on pronunciation, DiffRO employs a multi-task reward model that simultaneously handles:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dXCs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dXCs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 424w, https://substackcdn.com/image/fetch/$s_!dXCs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 848w, https://substackcdn.com/image/fetch/$s_!dXCs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 1272w, https://substackcdn.com/image/fetch/$s_!dXCs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dXCs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png" width="434" height="370.94017094017096" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:936,&quot;resizeWidth&quot;:434,&quot;bytes&quot;:145429,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/168045481?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dXCs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 424w, https://substackcdn.com/image/fetch/$s_!dXCs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 848w, https://substackcdn.com/image/fetch/$s_!dXCs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 1272w, https://substackcdn.com/image/fetch/$s_!dXCs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b63179a-9f3f-4c3e-81a5-3a7e5f24a7e6_936x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ul><li><p>Automatic Speech Recognition (ASR) for pronunciation accuracy</p></li><li><p>Speech Emotion Recognition (SER) for emotional expression</p></li><li><p>Speech Quality Assessment (SQA) for overall audio quality</p></li><li><p>Age and Gender prediction for speaker characteristics</p></li></ul><h2>Impressive Results Across Multiple Dimensions</h2><p>The researchers tested DiffRO on the challenging seed-tts-eval benchmark, demonstrating significant improvements across multiple languages and evaluation criteria:</p><p><strong>Pronunciation Accuracy</strong>: The system achieved state-of-the-art word error rates across Chinese, English, and challenging text cases, with particularly impressive results on hard-to-pronounce content.</p><p><strong>Cross-Lingual Performance</strong>: Remarkably, DiffRO improved performance on Japanese and Korean without language-specific training data, suggesting the method helps models learn universal pronunciation principles.</p><p><strong>Emotional Control</strong>: The multi-task approach enabled sophisticated emotional expression control, with the system achieving high accuracy across Happy, Sad, and Angry categories. Most impressively, the model learned to generate natural audio events like laughter and breathing sounds without explicit training.</p><p><strong>Competitive Performance</strong>: DiffRO outperformed existing models including F5-TTS and GPT-SoVITS on emotion benchmarks, demonstrating its practical superiority.</p><h2>Beyond the Technical Achievement</h2><p>What makes DiffRO particularly exciting is its potential to democratize high-quality TTS development. By reducing computational requirements and simplifying the training process, this approach makes advanced speech synthesis more accessible to researchers and developers with limited resources.</p><p>The method's ability to control multiple speech attributes simultaneously&#8212;from pronunciation accuracy to emotional expression&#8212;opens new possibilities for personalized AI assistants, content creation tools, and accessibility applications.</p><h2>Looking Forward</h2><p>While DiffRO represents a significant leap forward, the researchers acknowledge areas for future development. The current approach primarily optimizes the language model component, while speaker-related attributes are more heavily influenced by the flow matching and vocoder stages. Future work aims to extend DiffRO to these components for even more comprehensive control.</p><p>The success of DiffRO suggests we're entering a new era of efficient, controllable speech synthesis. As the demand for natural, emotionally-aware AI interactions continues to grow, methods like DiffRO will be crucial for building the next generation of voice-enabled applications.</p><div><hr></div><p><em>This research represents a fundamental shift in how we approach TTS optimization, moving from expensive, complex training procedures to efficient, direct optimization methods. The implications extend far beyond technical improvements&#8212;they point toward a future where high-quality, controllable speech synthesis becomes a standard capability rather than a specialized achievement.</em></p>]]></content:encoded></item><item><title><![CDATA[Anthropic’s Modular Code Protocol (MCP)]]></title><description><![CDATA[&#129521; LLMs are powerful&#8212;but brittle.]]></description><link>https://sankar1535.substack.com/p/anthropics-modular-code-protocol</link><guid isPermaLink="false">https://sankar1535.substack.com/p/anthropics-modular-code-protocol</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Sun, 01 Jun 2025 07:10:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!EQKe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&#129521; <strong>LLMs are powerful&#8212;but brittle. Anthropic&#8217;s Modular Code Protocol (MCP) introduces a clean way to build AI agents that are safe, modular, and production-ready.</strong></p><p>Most agent systems today tangle prompts, tool logic, and data access. MCP offers a better blueprint by <strong>separating model reasoning, application execution, and user intent</strong> into three distinct, composable layers.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EQKe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EQKe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 424w, https://substackcdn.com/image/fetch/$s_!EQKe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 848w, https://substackcdn.com/image/fetch/$s_!EQKe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 1272w, https://substackcdn.com/image/fetch/$s_!EQKe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EQKe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png" width="1456" height="780" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:780,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EQKe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 424w, https://substackcdn.com/image/fetch/$s_!EQKe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 848w, https://substackcdn.com/image/fetch/$s_!EQKe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 1272w, https://substackcdn.com/image/fetch/$s_!EQKe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44797dd9-5de2-4d67-b41f-cc1d60f8d399_1600x857.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>&#128295; <strong>Tools</strong> (<em>Model-controlled</em>)<br> Functions the model can <em>invoke</em>, not generate &#8212; e.g.<br> &#8226; create_invoice()<br> &#8226; reply_to_customer()<br> &#8226; flag_transaction()<br> Tools give the model agency through well-defined actions.</p><p>&#128194; <strong>Resources</strong> (<em>Application-controlled</em>)<br> Scoped data made available to the model, such as:<br> &#8226; CRM records<br> &#8226; Order history<br> &#8226; Account ledgers or transaction logs<br> Resources are visible to the model but immutable &#8212; keeping data access safe and auditable.</p><p>&#129504; <strong>Prompts</strong> (<em>User-controlled</em>)<br> Reusable, structured templates that describe what the model should <em>do</em>, like:<br> &#8226; &#8220;Summarize a customer complaint&#8221;<br> &#8226; &#8220;Extract payment info into JSON&#8221;<br> &#8226; &#8220;Suggest an action plan for overdue invoices&#8221;</p><div><hr></div><p>&#9881;&#65039; <strong>The MCP design is client-server:</strong></p><ul><li><p>The <strong>MCP Server</strong> exposes tools, resources, and prompt templates.</p></li><li><p>The <strong>MCP Client</strong> assembles these at runtime, calling the model as needed.</p></li></ul><p>&#128161; <em>Customer Support Example:</em> Summarize the last 5 support tickets (prompt), read a customer's complaint history (resource), and respond via Zendesk (tool).<br> &#128161; <em>Finance Agent Example:</em> Analyze quarterly reports (resource), extract KPIs (prompt), and generate an invoice summary for review (tool).</p><div><hr></div><p>&#9989; <strong>Why MCP Matters for Real-World Agents<br></strong> &#8226; <strong>Modularity</strong> &#8211; Swap in tools or data views without rewriting the whole system<br> &#8226; <strong>Governance</strong> &#8211; Models never get raw access to databases or APIs<br> &#8226; <strong>Safety</strong> &#8211; Logic lives in code, not fragile prompt hacks<br> &#8226; <strong>Observability</strong> &#8211; Each step (data &#8594; decision &#8594; action) is transparent</p><div><hr></div><p>This model is language-agnostic and future-proof. LLMs become reasoning engines, while execution and intent stay in human hands.</p><p>MCP is ideal for teams building <strong>internal copilots</strong>, <strong>compliance-aware finance agents</strong>, or <strong>customer support assistants</strong> that need traceability, control, and adaptability.</p><p>&#128204; If you&#8217;re building with OpenAI, Claude, or local LLMs, this is a pattern worth adopting early.</p><p>More details: </p><div id="youtube2-kQmXtrmQ5Zg" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;kQmXtrmQ5Zg&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/kQmXtrmQ5Zg?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div>]]></content:encoded></item><item><title><![CDATA[A Glimpse into the First Duplex S2S Model Without Speech Pre-training]]></title><description><![CDATA[Interacting with AI through speech can feel unnatural due to delays and rigid turn-taking.]]></description><link>https://sankar1535.substack.com/p/a-glimpse-into-the-first-duplex-s2s</link><guid isPermaLink="false">https://sankar1535.substack.com/p/a-glimpse-into-the-first-duplex-s2s</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Wed, 28 May 2025 12:06:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WPja!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Interacting with AI through speech can feel unnatural due to delays and rigid turn-taking.</p><p>&#127897;&#65039; <strong>NVIDIA researchers</strong> have taken a significant step toward more fluid, real-time spoken dialogue with a novel <strong>duplex speech-to-speech (S2S) model</strong> designed for seamless interaction &#8212; including <strong>robust user barge-in</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WPja!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WPja!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 424w, https://substackcdn.com/image/fetch/$s_!WPja!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 848w, https://substackcdn.com/image/fetch/$s_!WPja!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 1272w, https://substackcdn.com/image/fetch/$s_!WPja!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WPja!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png" width="1456" height="595" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:595,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WPja!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 424w, https://substackcdn.com/image/fetch/$s_!WPja!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 848w, https://substackcdn.com/image/fetch/$s_!WPja!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 1272w, https://substackcdn.com/image/fetch/$s_!WPja!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F58568586-6d6f-48d8-97a8-ed35abde506c_1600x654.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128269; <strong>What&#8217;s new:<br></strong> Introducing a <strong>novel duplex S2S architecture</strong> that directly models <strong>simultaneous user and agent streams</strong> using <strong>channel fusion</strong>.<br> &#10024; It's the <strong>first duplex S2S model</strong> demonstrated <em>without requiring speech pretraining</em>, simplifying the development process from any Large Language Model (LLM).<br> &#128230; To support reproducibility, this is also the <strong>first openly available duplex S2S model</strong> with full <strong>training and inference code</strong>.</p><p></p><p>&#128736;&#65039; <strong>How it works:</strong></p><ul><li><p>Simultaneously takes continuous <strong>user and agent speech/text streams</strong>.</p></li><li><p>Uses <strong>channel fusion</strong> to model concurrent inputs.</p></li><li><p>User speech is processed by a <strong>pre-trained streaming encoder</strong>.</p></li><li><p>Agent output is generated using a <strong>codec</strong> (default: personalised <strong>NanoCodec</strong>).</p></li><li><p>Trained via <strong>multi-channel next token prediction</strong>, handling text and speech in parallel with turn-level alignments.</p></li><li><p>Leverages a <strong>causal speech encoder</strong> and <strong>text LLM</strong> for streaming capabilities.<br><br></p></li></ul><p>&#128200; <strong>Results:</strong></p><ul><li><p>Outperforms previous duplex models like <em>Moshi</em>.</p></li><li><p>&#128483;&#65039; <strong>Barge-in success rate:</strong> 94.5% vs. 55.1%</p></li><li><p>&#9201;&#65039; <strong>Barge-in latency:</strong> 0.69s vs. 0.81s</p></li><li><p>&#128266; <strong>Higher speech quality:</strong> greater than <strong>4.0 UTMOS</strong></p></li><li><p>&#129504; <strong>Better reasoning scores</strong> across evaluation datasets</p></li><li><p>&#9889; <strong>Lower bitrate:</strong> 0.6 kbps &#8212; <em>half</em> of some prior work</p></li><li><p>&#128161; <strong>Codec personalization</strong> boosts performance, enabling it to beat higher-bitrate codecs on reconstruction metrics.<br><br></p></li></ul><p>&#127757; <strong>Why it matters:<br></strong>This research tackles key limitations of traditional turn-based systems, enabling <strong>natural, responsive, and efficient</strong> human-computer dialogue.</p><p>By removing the need for extensive speech pre-training, it <strong>democratises real-time S2S development</strong>, making it easier to adapt a variety of LLMs for spoken conversations.</p><p>With <strong>open-source availability</strong>, <strong>lower resource requirements</strong>, and <strong>state-of-the-art conversational performance</strong>, this marks a major milestone in the journey to next-gen, highly interactive AI agents.</p><p>&#128207; The paper also introduces <strong>systematic metrics</strong> for evaluating conversational behaviour&#8217;s like <strong>turn-taking</strong> and <strong>barge-in</strong>, pushing the field forward.</p><p>More Details: <a href="https://arxiv.org/abs/2505.15670">https://arxiv.org/abs/2505.15670</a></p>]]></content:encoded></item><item><title><![CDATA[Architecting Real-time Speech Interaction: Step-Audio's Approach to Seamless Tool Calling]]></title><description><![CDATA[Building truly intelligent real-time speech interaction systems presents significant technical hurdles.]]></description><link>https://sankar1535.substack.com/p/architecting-real-time-speech-interaction</link><guid isPermaLink="false">https://sankar1535.substack.com/p/architecting-real-time-speech-interaction</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Wed, 21 May 2025 14:20:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Eu9U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Building truly intelligent real-time speech interaction systems presents significant technical hurdles. A key challenge arises when the system needs to access external information or services &#8211; performing "tool calls" or "API integrations". Traditionally, integrating these steps into a cascading ASR-LLM-TTS pipeline could introduce latency, requiring the system to wait for API responses before generating the final spoken output.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Eu9U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Eu9U!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 424w, https://substackcdn.com/image/fetch/$s_!Eu9U!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 848w, https://substackcdn.com/image/fetch/$s_!Eu9U!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 1272w, https://substackcdn.com/image/fetch/$s_!Eu9U!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Eu9U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png" width="741" height="420" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:420,&quot;width&quot;:741,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Eu9U!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 424w, https://substackcdn.com/image/fetch/$s_!Eu9U!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 848w, https://substackcdn.com/image/fetch/$s_!Eu9U!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 1272w, https://substackcdn.com/image/fetch/$s_!Eu9U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9e1176e-e3ff-40ea-96ab-ffc7acc6914a_741x420.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Step-Audio, the first production-ready open-source framework for intelligent speech interaction, addresses this with an innovative <strong>asynchronous tool invocation</strong> mechanism. Rather than a simple sequential process, the architecture specifically <strong>decouples the text-based tool processing from the audio generation pipelines</strong>.</p><h3>Why is this decoupling significant?</h3><p>Real-time text responses and their corresponding audio streams have a substantial bitrate disparity. By separating the processes, Step-Audio allows for the <strong>parallel execution of external service queries</strong> (like knowledge retrieval) and speech synthesis.</p><h3>The result?</h3><p>This design <strong>eliminates waiting time for audio rendering when a tool call is required</strong>, significantly <strong>enhancing interaction fluidity</strong>. The system manages complex tasks by augmenting its cognitive architecture with this crucial tool calling ability.</p><p>This architectural choice is a technical leap forward, enabling smoother, more natural conversational dynamics even when the system needs to look up information or interact with external systems.</p><p>More Details: <a href="https://arxiv.org/abs/2502.11946">https://arxiv.org/abs/2502.11946</a></p><p></p>]]></content:encoded></item><item><title><![CDATA[VITA-Audio: Real-Time Speech Generation, Redefined]]></title><description><![CDATA[Introducing VITA-Audio, a groundbreaking end-to-end speech model that dramatically reduces latency in real-time speech applications.]]></description><link>https://sankar1535.substack.com/p/vita-audio-real-time-speech-generation</link><guid isPermaLink="false">https://sankar1535.substack.com/p/vita-audio-real-time-speech-generation</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Wed, 14 May 2025 13:31:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ll8y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Introducing <strong>VITA-Audio</strong>, a groundbreaking end-to-end speech model that dramatically reduces latency in real-time speech applications.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ll8y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ll8y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 424w, https://substackcdn.com/image/fetch/$s_!Ll8y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 848w, https://substackcdn.com/image/fetch/$s_!Ll8y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 1272w, https://substackcdn.com/image/fetch/$s_!Ll8y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ll8y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png" width="1396" height="1292" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1292,&quot;width&quot;:1396,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ll8y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 424w, https://substackcdn.com/image/fetch/$s_!Ll8y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 848w, https://substackcdn.com/image/fetch/$s_!Ll8y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 1272w, https://substackcdn.com/image/fetch/$s_!Ll8y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F61e4beed-eecc-4466-8f03-fdf47e9b81d0_1396x1292.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#127775; <strong>Why it matters:<br></strong>Traditional systems suffer from high first-token delay, limiting truly natural voice interactions. VITA-Audio is the <strong>first multi-modal LLM</strong> capable of generating audio output <strong>during the initial forward pass</strong> &#8212; achieving <strong>zero audio token and generation delay</strong>.</p><p>&#128736;&#65039; <strong>How it works:</strong></p><ul><li><p>The core innovation is the Multiple<strong> Cross-modal Token Prediction (MCTP)</strong> module.</p></li><li><p>The MCTP modules were designed based on attention map visualizations and masking experiments, which revealed that LLM hidden states contain enough localized context for audio token generation.</p></li><li><p>These findings showed that the speech model focuses mainly on nearby text hidden states rather than the full semantic context, enabling the use of lightweight modules for audio prediction.</p></li><li><p>Inspired by the isomorphic Multi-Token Prediction (MTP) framework used in models like DeepSeek V3,<strong> MCTP</strong> modules generate audio tokens efficiently without extra LLM passes.</p></li><li><p>End-to-end architecture with audio encoder, LLM backbone, and audio decoder.</p></li><li><p>Interleaved text-audio token modeling retains language quality.</p></li><li><p>Progressive four-stage training ensures accurate speech synthesis.</p><p></p></li></ul><p>&#9889; <strong>Results:</strong></p><ul><li><p>Up to <strong>5x faster inference</strong> (Turbo mode).</p></li><li><p>First audio chunk latency slashed from <strong>236ms to 53ms.</strong></p></li><li><p><strong>State-of-the-art results</strong> on speech-to-speech tasks and top-tier TTS quality.</p></li><li><p>Fully open-source and reproducible.</p></li></ul><p>&#128266; <strong>Why it&#8217;s a game changer:<br></strong> VITA-Audio unlocks real-time speech-to-speech interaction for assistants, agents, and conversational AI &#8212; setting a <strong>new standard</strong> for latency-sensitive applications.</p><p>&#128194; More details: https://github.com/VITA-MLLM/VITA-Audio</p>]]></content:encoded></item><item><title><![CDATA[How Voila Speaks with Nuance: The Power of Structured Interleaved Alignment ]]></title><description><![CDATA[Voila is a new family of voice-language foundation models.]]></description><link>https://sankar1535.substack.com/p/how-voila-speaks-with-nuance-the</link><guid isPermaLink="false">https://sankar1535.substack.com/p/how-voila-speaks-with-nuance-the</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Mon, 12 May 2025 13:30:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9O7v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Voila is a new family of voice-language foundation models. At the heart of Voila&#8217;s ability to generate <em>expressive and synchronized speech</em> lies a key innovation: <strong>Structured Interleaved Alignment</strong> &#8212; a smarter way to connect <em>what&#8217;s said</em> with <em>how it sounds</em>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9O7v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9O7v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 424w, https://substackcdn.com/image/fetch/$s_!9O7v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 848w, https://substackcdn.com/image/fetch/$s_!9O7v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 1272w, https://substackcdn.com/image/fetch/$s_!9O7v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9O7v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png" width="1292" height="1284" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1284,&quot;width&quot;:1292,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9O7v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 424w, https://substackcdn.com/image/fetch/$s_!9O7v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 848w, https://substackcdn.com/image/fetch/$s_!9O7v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 1272w, https://substackcdn.com/image/fetch/$s_!9O7v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e01e2f7-60e0-4c9e-8b03-b076cf697317_1292x1284.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To enable a large language model (LLM) to understand and generate voice, Voila first transforms raw audio into <strong>discrete semantic and acoustic tokens</strong> using its neural audio codec, the Voila-Tokenizer. These tokens become part of the LLM's vocabulary &#8212; making audio a <em>first-class citizen</em> in language understanding and generation.</p><p>But here&#8217;s the twist &#128071;</p><p>Rather than batching all the text followed by all the audio (as many models do), Voila interleaves them in a <strong>structured, fine-grained sequence</strong>:<br> &#10145;&#65039; "&lt;Hello&gt;" + &lt;audio tokens&gt;<br> &#10145;&#65039; "&lt;I&gt;" + &lt;audio tokens&gt;<br> &#10145;&#65039; "&lt;am&gt;" + &lt;audio tokens&gt;<br> &#10145;&#65039; "&lt;Voila&gt;" + &lt;audio tokens&gt;</p><p>This <strong>tight pairing</strong> of each semantic unit with its corresponding acoustic representation creates a <strong>clear and explicit alignment</strong> &#8212; crucial for generating nuanced speech that feels natural and well-paced. This design differs from prior approaches, such as Spirit-LM and Unified Spoken Dialog Model (USDM), which also adopt interleaved text-audio formats but do so with looser coupling.</p><p>&#10024; Why does this matter?</p><ul><li><p>It strengthens the LLM&#8217;s understanding of how meaning and sound relate.</p></li><li><p>It simplifies training by reducing the ambiguity of loose alignments.</p></li><li><p>It empowers Voila to handle TTS, voice-based instructions, and multimodal reasoning with remarkable clarity.</p></li></ul><p>Once interleaved, the tokens are embedded &#8212; with text tokens repeated to match multi-layer audio tokens &#8212; averaged, and passed through the LLM. The output is then decoded by an audio transformer and finally rendered into high-quality speech by the Voila-Tokenizer.</p><p>This alignment strategy isn&#8217;t just a technical detail &#8212; it&#8217;s a foundational innovation that unlocks more human-like interaction between language and voice.</p><p>&#128266; Welcome to a new era of audio-native AI. </p><p>More details: https://github.com/maitrix-org/Voila</p>]]></content:encoded></item><item><title><![CDATA[INTP (Intelligibility Preference Speech Dataset): Making Zero-Shot TTS More Robust and Intelligible]]></title><description><![CDATA[Despite impressive progress, zero-shot Text-to-Speech (TTS) models still struggle with challenging linguistic scenarios &#8212; think tongue twisters, repeated words, code-switching, and cross-lingual synthesis.]]></description><link>https://sankar1535.substack.com/p/intp-intelligibility-preference-speech</link><guid isPermaLink="false">https://sankar1535.substack.com/p/intp-intelligibility-preference-speech</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Sun, 11 May 2025 16:12:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!rvh1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Despite impressive progress, zero-shot Text-to-Speech (TTS) models still struggle with <em>challenging linguistic scenarios</em> &#8212; think tongue twisters, repeated words, code-switching, and cross-lingual synthesis. A recent study directly tackles these weaknesses through <strong>preference alignment</strong> and a new dataset: <strong>INTP (Intelligibility Preference Speech Dataset)</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rvh1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rvh1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 424w, https://substackcdn.com/image/fetch/$s_!rvh1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 848w, https://substackcdn.com/image/fetch/$s_!rvh1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 1272w, https://substackcdn.com/image/fetch/$s_!rvh1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rvh1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png" width="1204" height="838" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:838,&quot;width&quot;:1204,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rvh1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 424w, https://substackcdn.com/image/fetch/$s_!rvh1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 848w, https://substackcdn.com/image/fetch/$s_!rvh1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 1272w, https://substackcdn.com/image/fetch/$s_!rvh1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa8234362-0cb8-4c8b-8715-9936a4b3d1f7_1204x838.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128269; <strong>What&#8217;s New:</strong></p><ul><li><p>Introduced <strong>INTP</strong>, a curated set of ~250K preference pairs targeting tough intelligibility cases.</p></li><li><p>Adapted <strong>Direct Preference Optimization (DPO)</strong> to align multiple TTS architectures: AR, Flow-Matching, and Masked Generative.</p></li><li><p>Preference signals were constructed using WER-based intra/inter-model comparisons and LLM-generated perturbations.<br></p></li></ul><p>&#128200; <strong>Why It Matters:</strong></p><ul><li><p>Across diverse zero-shot TTS models (e.g., ARS, F5-TTS, MaskGCT, CosyVoice 2, Ints), observed <strong>significant WER reductions</strong> and improved naturalness (N-CMOS), without sacrificing speaker similarity.</p></li><li><p>INTP shows <strong>weak-to-strong generalization</strong>&#8212;enhancing even strong models it wasn&#8217;t trained on.<br></p></li></ul><p>&#127744; Its also demonstrated a <em>flywheel effect</em>&#8212;a scalable loop of data and model improvement through iterative preference alignment.</p><p>&#128226; The authors will release the INTP dataset, DPO-based alignment code, and improved model checkpoints in Amphion toolkit to support further research.</p><p>More details: https://intalign.github.io/</p>]]></content:encoded></item><item><title><![CDATA[Ready to hear your LLM speak? Breakthrough in efficient Text-to-Speech with PEFT!]]></title><description><![CDATA[Large Language Models have redefined text-based AI, but giving them a voice without costly retraining has been a hurdle.]]></description><link>https://sankar1535.substack.com/p/ready-to-hear-your-llm-speak-breakthrough</link><guid isPermaLink="false">https://sankar1535.substack.com/p/ready-to-hear-your-llm-speak-breakthrough</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Mon, 05 May 2025 13:30:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!yHOh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Large Language Models have redefined text-based AI, but giving them a voice without costly retraining has been a hurdle. Traditional methods often require building speech models from scratch or computationally heavy full fine-tuning, sometimes impacting the LLM's original text prowess.</p><p>&#127908; <strong>Introducing TTS-Llama</strong> &#8212; a breakthrough approach that brings voice to LLMs using <strong>Parameter-Efficient Fine-Tuning (PEFT)</strong> techniques like LoRA. No need to retrain the whole model or sacrifice text performance.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yHOh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yHOh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 424w, https://substackcdn.com/image/fetch/$s_!yHOh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 848w, https://substackcdn.com/image/fetch/$s_!yHOh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 1272w, https://substackcdn.com/image/fetch/$s_!yHOh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yHOh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png" width="1218" height="1132" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1132,&quot;width&quot;:1218,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yHOh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 424w, https://substackcdn.com/image/fetch/$s_!yHOh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 848w, https://substackcdn.com/image/fetch/$s_!yHOh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 1272w, https://substackcdn.com/image/fetch/$s_!yHOh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F849ce86e-cfcf-4767-87d0-66dc9a35f621_1218x1132.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>&#128161; What&#8217;s New:</h2><p>With LoRA, we fine-tune just a tiny fraction of the LLM&#8217;s parameters &#8212; specifically the input embeddings, prediction head, and injected adapters &#8212; keeping the original model frozen. The result?<br> &#9989; State-of-the-art TTS with <strong>minimal compute<br></strong> &#9989; No compromise on the LLM&#8217;s original capabilities</p><h2>&#128295; How It Works:</h2><ol><li><p><strong>Text In &#8594;</strong> A fine-tuned Llama 3-8B generates high-level semantic tokens that capture meaning <em>and</em> prosody.</p></li><li><p><strong>Semantic Tokens &#8594; Acoustic Features:</strong> An acoustic language model (MusicGen) decodes these into fine-grained audio representations.</p></li><li><p><strong>Audio Out:</strong> A neural vocoder synthesizes the final waveform.</p></li></ol><p>This modular design is powered by a <strong>semantic tokenizer</strong> (extracting 4,096 discrete tokens) and an <strong>acoustic tokenizer</strong> (via convolutional autoencoder with residual quantization).</p><h2>&#128202; Training:</h2><ul><li><p>&#128483; 50K hours of LibriHeavy speech + 60K synthetic sentences for text normalization.</p></li><li><p>&#129518; 1.4B trainable parameters in TTS-Llama.</p></li><li><p>&#128421; Trained on 8&#215; H100 GPUs, batch size 256, 20K steps.</p></li></ul><h2>&#128293; Results:</h2><p>TTS-Llama achieved <strong>state-of-the-art zero-shot TTS</strong>:</p><ul><li><p><strong>Human-likeness MOS:</strong> 3.07</p></li><li><p><strong>Audio quality MOS:</strong> 3.47<br> (Beating VoiceCraft's 2.85 and 3.17)</p></li><li><p>It also excelled in TTS performance on text normalization tasks.</p></li></ul><h2>&#127757; Why It Matters:</h2><p>This work proves you don&#8217;t need to build speech models from scratch. By applying PEFT to existing LLMs, we can inject <em>new modalities</em> like speech efficiently and scalable. It's a massive step toward truly <strong>multimodal AI</strong>.</p><p>&#128196; <strong>More here</strong>: https://arxiv.org/abs/2410.20336</p>]]></content:encoded></item><item><title><![CDATA[Rethinking Efficiency in State-of-the-Art Neural Codecs: Variable Bitrate Residual Vector Quantization (VRVQ)]]></title><description><![CDATA[&#128269; What&#8217;s New?]]></description><link>https://sankar1535.substack.com/p/rethinking-efficiency-in-state-of</link><guid isPermaLink="false">https://sankar1535.substack.com/p/rethinking-efficiency-in-state-of</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Fri, 02 May 2025 13:31:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4m5b!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73c1ee4-049d-4623-9003-7d1cf146709e_408x408.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1><strong>&#128269; What&#8217;s New?</strong></h1><p>The cutting edge of neural audio compression has largely settled on <strong>Residual Vector Quantization (RVQ)</strong>, showcasing impressive performance at low bitrates.</p><p>But in a world of diverse audio content, from complex music to simple silence, can a codec truly be optimal when it rigidly uses the same number of codebooks for <em>every</em> frame?</p><p>Introducing <strong>Variable Bitrate Residual Vector Quantization (VRVQ)</strong> &#8212; a novel approach that brings <em>true variable bitrate coding</em> to RVQ-VAE-based audio codecs for the first time.</p><p>Unlike conventional codecs that use a <em>fixed</em> number of codebooks per frame, <strong>VRVQ dynamically adjusts</strong> codebook usage based on signal complexity. This makes audio coding more efficient, especially for simple content like silence or low-energy frames.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h-zK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h-zK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 424w, https://substackcdn.com/image/fetch/$s_!h-zK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 848w, https://substackcdn.com/image/fetch/$s_!h-zK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 1272w, https://substackcdn.com/image/fetch/$s_!h-zK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h-zK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png" width="691" height="290" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:290,&quot;width&quot;:691,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h-zK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 424w, https://substackcdn.com/image/fetch/$s_!h-zK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 848w, https://substackcdn.com/image/fetch/$s_!h-zK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 1272w, https://substackcdn.com/image/fetch/$s_!h-zK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd3a7e2c-43e6-4e51-8dc1-f9abfbe04a05_691x290.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1>&#129504; <strong>How It Works</strong></h1><ul><li><p>VRVQ approach introduces an importance map derived from an intermediate feature map of the encoder.</p></li><li><p>This importance map is used to generate a time-varying binary mask that controls which codebooks are used for each frame.</p></li><li><p>To enable effective training of this importance map alongside the codec, a novel gradient estimation method for the non-differentiable masking operation, utilizing a smooth surrogate function was developed.</p></li><li><p>The entire framework is jointly trained using a rate-distortion loss optimization.</p><p></p></li></ul><h1><strong>&#128202; Results</strong></h1><ul><li><p><strong>VRVQ significantly outperforms</strong> a state-of-the-art DAC baseline in SI-SDR under many conditions.</p></li><li><p>Scalability: Increasing codebooks from 8 &#8594; 16 improved performance further and matched constant bitrate baselines.<br></p></li></ul><h4>&#9888;&#65039; <strong>Limitation</strong>: </h4><p>At very low bitrates, the mask overhead slightly reduces variable bitrate gains &#8212; but the tradeoff is minimal for practical use cases.</p><h1>&#128204; <strong>Why It Matters</strong></h1><p>Fixed-bitrate codecs waste capacity on trivial audio. VRVQ solves this &#8212; offering smarter, adaptive compression for neural audio systems. It's a step toward <strong>more flexible, efficient, and high-quality audio codecs</strong> grounded in modern machine learning.</p><p></p><p>&#127911; <strong>Explore the project + audio samples</strong>:<br> &#128279;<a href="https://yoongi43.github.io/VBRRVQ.github.io/"> https://yoongi43.github.io/VBRRVQ.github.io/</a></p>]]></content:encoded></item><item><title><![CDATA[Muyan-TTS: Open-Source LLM-Based TTS Optimized for Podcasting—Built on a $50K Budget ]]></title><description><![CDATA[The Text-to-Speech (TTS) landscape is undergoing a transformation&#8212;powered by Large Language Models (LLMs) that significantly improve semantic understanding and speech naturalness.]]></description><link>https://sankar1535.substack.com/p/muyan-tts-open-source-llm-based-tts</link><guid isPermaLink="false">https://sankar1535.substack.com/p/muyan-tts-open-source-llm-based-tts</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Wed, 30 Apr 2025 13:30:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!u1Lo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The Text-to-Speech (TTS) landscape is undergoing a transformation&#8212;powered by Large Language Models (LLMs) that significantly improve semantic understanding and speech naturalness. But despite these advancements, most LLM-based TTS systems remain closed-source, resource-intensive, or ill-suited for real-world audio applications like podcasting.</p><p>Introducing <strong>Muyan-TTS</strong>, a fully open-source, trainable TTS system <strong>purpose-built for podcast scenarios</strong>&#8212;achieving <strong>high naturalness, zero-shot synthesis, and fast inference</strong> within a <strong>$50,000 total training budget</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!u1Lo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!u1Lo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 424w, https://substackcdn.com/image/fetch/$s_!u1Lo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 848w, https://substackcdn.com/image/fetch/$s_!u1Lo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 1272w, https://substackcdn.com/image/fetch/$s_!u1Lo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!u1Lo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png" width="1438" height="1008" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1008,&quot;width&quot;:1438,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!u1Lo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 424w, https://substackcdn.com/image/fetch/$s_!u1Lo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 848w, https://substackcdn.com/image/fetch/$s_!u1Lo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 1272w, https://substackcdn.com/image/fetch/$s_!u1Lo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8940eda0-9c0c-46e3-96be-3cdef827ff83_1438x1008.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#128269; <strong>Key Innovations:</strong></p><ul><li><p>&#9989; <strong>LLM-Based Synthesis</strong>: Uses a pre-trained <strong>LLaMA-3.2-3B</strong> as a replacement for traditional AR models&#8212;bridging the gap between LLM expressiveness and TTS fidelity.</p></li><li><p>&#9989; <strong>Parallel Audio-Text Tokenization</strong>: Text is tokenized with the LLM&#8217;s tokenizer, while audio is quantized using GPT-SoVITS audio tokens, enabling alignment of speech and text representations.</p></li><li><p>&#9989; <strong>VITS-Based Decoder</strong>: Fine-tuned for podcast audio to mitigate hallucinations and enhance pronunciation via a structured G2P module.</p></li></ul><p>&#9881;&#65039; <strong>System Overview:</strong></p><ol><li><p><strong>&#128229; Data Pipeline:</strong> 150K+ hours of multilingual raw audio collected and filtered down to 100K+ high-quality hours using Whisper, FunASR, and MOS/NISQA thresholds. Audio cleaning used Music Source Separation, DeEcho, DeReverb, and more. (~$30K in GPU compute)</p></li><li><p><strong>&#129504; LLM Pre-training:</strong> 15 epochs on 80 A100 GPUs&#8212;training LLaMA-3.2-3B on audio-text pairs via 1024 learned audio tokens. (~$19.2K)</p></li><li><p><strong>&#127919; SFT (Speaker Adaptation):</strong> Efficient post-training using a few minutes of target speaker data (~15 mins/GPU node per hour of speech).</p></li><li><p><strong>&#128266; Decoder Training:</strong> Fine-tuned SoVITS on a curated 10K-hour MOS&gt;4.5 podcast set. (~$1.34K)<br></p></li></ol><p>&#128202; <strong>Evaluation Highlights:</strong></p><ul><li><p><strong>&#128200; WER:</strong> 3.44% (LibriSpeech), 4.09% (SEED) &#8212; outperforming Spark-TTS, FireRedTTS, and GPT-SoVITS.</p></li><li><p><strong>&#128483;&#65039; MOS:</strong> 4.58 (LibriSpeech), 4.32 (SEED) &#8212; strong perceptual quality, especially after SFT (4.97 MOS).</p></li><li><p><strong>&#128101; Speaker Similarity:</strong> Improved from 0.37 &#8594; 0.46 (SIM) post-SFT.</p></li><li><p><strong>&#9889; Inference Speed:</strong> Fastest among all tested (Step-Audio, CosyVoice2, Spark-TTS, FireRedTTS, and GPT-SoVITS v3) models (real-time factor = 0.33).</p><p></p></li></ul><p>&#127919; <strong>Why It Matters:</strong></p><ul><li><p>First open-source LLM-TTS stack optimized end-to-end for <strong>long-form podcast content</strong>.</p></li><li><p>Fully reproducible training pipeline: data collection, LLM training, decoding, and inference&#8212;all open.</p></li><li><p>Achieves <strong>production-grade synthesis</strong> and <strong>real-time capabilities</strong> on <strong>modest compute</strong>&#8212;busting the myth that LLM-based TTS is out of reach for small labs and indie developers.</p></li></ul><p>&#128736;&#65039; Explore the code, pre-trained models, and docs:<a href="https://github.com/MYZY-AI/Muyan-TTS"> https://github.com/MYZY-AI/Muyan-TTS</a></p><p></p>]]></content:encoded></item><item><title><![CDATA[A New Era in Audio AI: Introducing Kimi-Audio]]></title><description><![CDATA[The shift in audio AI is underway &#8212; from narrow, task-specific systems to unified, foundation-level architectures.]]></description><link>https://sankar1535.substack.com/p/a-new-era-in-audio-ai-introducing</link><guid isPermaLink="false">https://sankar1535.substack.com/p/a-new-era-in-audio-ai-introducing</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Tue, 29 Apr 2025 11:23:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!A3Cr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The shift in audio AI is underway &#8212; from narrow, task-specific systems to unified, foundation-level architectures. Just like NLP had its LLM moment, audio is entering its own revolution.</p><p>Kimi-Audio is a major leap: an <strong>open-source, real-time, instruction-following foundation model</strong> for audio <strong>understanding, generation, and conversation</strong> &#8212; all within one architecture.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!A3Cr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!A3Cr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 424w, https://substackcdn.com/image/fetch/$s_!A3Cr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 848w, https://substackcdn.com/image/fetch/$s_!A3Cr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 1272w, https://substackcdn.com/image/fetch/$s_!A3Cr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!A3Cr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png" width="1456" height="1099" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1099,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!A3Cr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 424w, https://substackcdn.com/image/fetch/$s_!A3Cr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 848w, https://substackcdn.com/image/fetch/$s_!A3Cr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 1272w, https://substackcdn.com/image/fetch/$s_!A3Cr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9c7c370-5686-4037-8199-e22be36aaccc_1600x1208.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>&#129504; What makes it different?</strong></p><p>Kimi-Audio uses a hybrid input representation to bridge text and audio:<br> &#8226; Discrete semantic tokens (12.5Hz) from an ASR-based tokenizer<br> &#8226; Continuous acoustic vectors from Whisper, also downsampled to 12.5Hz</p><p>These are fed into a shared transformer initialized from a large language model, with:<br> &#8226; Dual heads for generating both text and discrete audio tokens<br> &#8226; A streaming detokenizer for real-time synthesis using flow matching + BigVGAN<br> &#8226; A novel look-ahead mechanism that smooths audio at chunk boundaries</p><p><strong>The result</strong>: low-latency, high-quality, multimodal generation from a single, unified system.</p><p><strong>&#128640; Why it matters</strong></p><p>Kimi-Audio changes how we build audio systems:<br> &#8226; Unified architecture replaces separate ASR, TTS, and sound models<br> &#8226; Real-time, instruction-following capabilities open new application domains<br> &#8226; Designed to support diverse data: speech, sound, music, and text<br> &#8226; Open-source release promotes transparency, reproducibility, and extensibility</p><p>This model is part of a broader vision for <strong>general-purpose audio intelligence</strong> &#8212; where models can hear, understand, and speak in natural, context-aware ways.</p><p><strong>&#128736;&#65039; Built for the community</strong></p><p>Kimi-Audio was pre-trained on 13M+ hours of audio and fine-tuned on 300K hours of labeled tasks. It comes with:<br> &#8226; Full codebase<br> &#8226; Released model checkpoints<br> &#8226; Evaluation toolkit</p><p>&#128279; Explore:<a href="https://github.com/MoonshotAI/Kimi-Audio"> https://github.com/MoonshotAI/Kimi-Audio</a></p>]]></content:encoded></item><item><title><![CDATA[Synergistic Speech Synthesis: Bridging Modalities for Enhanced LLM-driven Text-to-Speech]]></title><description><![CDATA[The landscape of text-to-speech synthesis has been significantly transformed by the integration of large language models through discrete tokenization, enabling remarkable control over various vocal attributes via speech prompt conditioning.]]></description><link>https://sankar1535.substack.com/p/synergistic-speech-synthesis-bridging</link><guid isPermaLink="false">https://sankar1535.substack.com/p/synergistic-speech-synthesis-bridging</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Mon, 21 Apr 2025 11:47:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QsaL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The landscape of text-to-speech synthesis has been significantly transformed by the integration of large language models through discrete tokenization, enabling remarkable control over various vocal attributes via speech prompt conditioning. Yet, does the necessity for discrete speech tokenization inherently compromise the nuanced acoustic fidelity and real-world applicability of these advanced architectures?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QsaL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QsaL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 424w, https://substackcdn.com/image/fetch/$s_!QsaL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 848w, https://substackcdn.com/image/fetch/$s_!QsaL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 1272w, https://substackcdn.com/image/fetch/$s_!QsaL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QsaL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png" width="1456" height="996" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:996,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QsaL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 424w, https://substackcdn.com/image/fetch/$s_!QsaL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 848w, https://substackcdn.com/image/fetch/$s_!QsaL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 1272w, https://substackcdn.com/image/fetch/$s_!QsaL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95397805-0fa7-4d42-88de-761b38b4ddfa_1600x1094.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This work introduces GOAT-TTS, a novel dual-branch architecture designed to optimise LLM-based text-to-speech generation. The framework uniquely incorporates a modality-alignment branch that captures continuous acoustic embeddings, fostering a bidirectional understanding between paralinguistic features and semantic text representations without strict transcript alignment. Complementing this, the speech-generation branch employs modular fine-tuning on the LLM's top layers, preserving its native linguistic comprehension while optimising for speech token prediction, further enhanced by a multi-token prediction mechanism for real-time streaming.</p><p>This dual-branch approach directly tackles the irreversible loss of acoustic details associated with speech prompt quantization and circumvents the limitations imposed by the requirement for precisely aligned speech-text pairs. By using continuous acoustic embeddings, GOAT-TTS maintains rich paralinguistic information, and the transcript-free modality alignment significantly broadens its deployment flexibility. Furthermore, the selective fine-tuning strategy mitigates the catastrophic forgetting of the LLM's original text comprehension capabilities during the adaptation for speech token generation.</p><p>The training involved a two-stage process, starting with modality-alignment training using speech-text continuation pairs derived from ASR corpora and TTS-generated speech, followed by speech-generation training on approximately 150k hours of real-world conversational data, employing a layer-wise parameter freezing strategy and multi-token prediction.</p><p>Experimental evaluations on standard datasets demonstrate that this approach achieves performance comparable to state-of-the-art TTS models in non-streaming settings and significantly outperforms them in streaming scenarios, while also proving effective in generating high-quality dialect-specific speech data capable of enhancing dialect ASR model performance.</p><p>This innovation suggests a promising trajectory towards more robust and versatile LLM-driven speech synthesis systems capable of high-fidelity output and seamless integration into real-world applications, particularly those involving diverse linguistic and acoustic environments.</p><p>More Details: <a href="https://arxiv.org/html/2504.12339v1">Paper</a></p>]]></content:encoded></item><item><title><![CDATA[Tokens with Meaning: Redefining Audio Language Modeling through Semantic Compression]]></title><description><![CDATA[When bitrates shrink but understanding deepens, is compression still a compromise?]]></description><link>https://sankar1535.substack.com/p/tokens-with-meaning-redefining-audio</link><guid isPermaLink="false">https://sankar1535.substack.com/p/tokens-with-meaning-redefining-audio</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Fri, 18 Apr 2025 07:39:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ma0q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The recent surge in audio language models highlights the crucial role of transforming raw audio into discrete tokens, enabling the application of large language model architectures to the auditory world. However, existing audio tokenisers often process individual audio segments independently, potentially overlooking broader contextual cues across time.</p><p>Could we achieve richer semantic representations and more efficient compression by explicitly modelling the inter-frame relationships in audio, rather than treating each segment in isolation? This question arises from the limitations of frame-based encoding prevalent in prior methods like Encodec.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ma0q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ma0q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 424w, https://substackcdn.com/image/fetch/$s_!ma0q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 848w, https://substackcdn.com/image/fetch/$s_!ma0q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 1272w, https://substackcdn.com/image/fetch/$s_!ma0q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ma0q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png" width="1456" height="605" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:605,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ma0q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 424w, https://substackcdn.com/image/fetch/$s_!ma0q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 848w, https://substackcdn.com/image/fetch/$s_!ma0q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 1272w, https://substackcdn.com/image/fetch/$s_!ma0q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2abe031d-a9ff-4aec-99b8-7b56ad96c67c_1600x665.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The core innovation lies in a novel <strong>query-based compression strategy</strong>, where a set of learnable query tokens interacts with audio frames via transformer layers to capture holistic, cross-frame context. This approach contrasts with frame-by-frame encoding, allowing the model to glean more semantic information and represent audio with fewer tokens.</p><p>This innovation directly tackles the problem of information redundancy and the lack of explicit context modelling in existing low-bitrate audio codecs. By using query tokens, the model can focus on extracting the most salient information across the entire audio sequence, leading to more efficient compression and semantically richer tokens.</p><p>One potential drawback is that the introduction of more complex mechanisms like the query-based attention might lead to increased computational demands during training and inference compared to simpler frame-based methods, although the paper argues for lower bitrate and comparable efficiency.</p><p>The ALMTokenizer was trained on approximately 4,500 hours of diverse audio data, encompassing speech (LibriTTS, MLS), sound (AudioSet), and music (Million Song Dataset). The model employs transformer networks with up to 24 self-attention layers and is trained for 200k steps using AdamW with a learning rate of 1e-4, across 4 NVIDIA A100-80G GPUs.</p><p>ALMTokenizer achieves competitive reconstruction at a lower bitrate than SOTA tokenizers like Encodec and MimiCodec, while outperforming them in downstream tasks like TTS, ASR, and audio captioning. MUSHRA tests also confirm strong subjective quality across speech, music, and sound.</p><p>This work points to a promising future for audio language models, showing that context-aware tokenization enables better compression with richer semantics&#8212;paving the way for more efficient and powerful multimodal AI systems.</p><p>More details: <a href="https://arxiv.org/html/2504.10344v1">paper</a></p>]]></content:encoded></item><item><title><![CDATA[Real-Time Audio's Neural Frontier: Bridging Quality and Latency]]></title><description><![CDATA[Recent strides in neural audio codecs have shown remarkable promise in discretizing audio signals with minimal bit usage while maintaining high fidelity, playing a crucial role in areas like real-time communication and speech language models.]]></description><link>https://sankar1535.substack.com/p/real-time-audios-neural-frontier</link><guid isPermaLink="false">https://sankar1535.substack.com/p/real-time-audios-neural-frontier</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Wed, 16 Apr 2025 05:41:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!eY0E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Recent strides in neural audio codecs have shown remarkable promise in discretizing audio signals with minimal bit usage while maintaining high fidelity, playing a crucial role in areas like real-time communication and speech language models. Yet, the prevalence of non-causal structures in these advanced codecs often leads to excessively high latency, hindering their applicability in scenarios demanding immediate interaction.</p><p>Can we truly achieve top-tier audio quality with neural codecs without sacrificing the crucial element of real-time responsiveness demanded by interactive applications?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eY0E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eY0E!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 424w, https://substackcdn.com/image/fetch/$s_!eY0E!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 848w, https://substackcdn.com/image/fetch/$s_!eY0E!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 1272w, https://substackcdn.com/image/fetch/$s_!eY0E!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eY0E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png" width="1456" height="397" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:397,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eY0E!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 424w, https://substackcdn.com/image/fetch/$s_!eY0E!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 848w, https://substackcdn.com/image/fetch/$s_!eY0E!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 1272w, https://substackcdn.com/image/fetch/$s_!eY0E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff48bd6db-8093-445f-85e1-e1d4ee075e14_1600x436.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The core innovation lies in the introduction of <strong>StreamCodec</strong>, a fully causal and symmetric encoder-decoder architecture operating in the MDCT domain, coupled with a novel <strong>residual scalar-vector quantizer (RSVQ)</strong>. This RSVQ sequentially employs scalar quantizer for coarse audio contours and improved vector quantizer to refine acoustic details in a residual manner.</p><p>This design directly tackles the trade-off between coding quality and latency inherent in streamable neural audio codecs by ensuring fully causal operation for low latency and using RSVQ to compensate for any quality loss arising from this structural constraint. While the paper does not explicitly mention specific disadvantages of this innovation, the complexity of the RSVQ might introduce some computational overhead, although the results suggest high efficiency.</p><p>Experiments were conducted on the 16 kHz LibriTTS (approximately 263 hours) and 48 kHz VCTK (approximately 43 hours) datasets. StreamCodec achieved a ViSQOL score of 4.30 at 1.5 kbps on LibriTTS and demonstrated a near 20 times real-time generation speed on a CPU with a lightweight 7M parameter model.</p><p>The development of StreamCodec signifies a significant step towards achieving high-fidelity, low-latency neural audio compression, paving the way for enhanced real-time communication applications and potentially influencing the design of future streamable audio processing systems.</p><p>More details: <a href="https://lnkd.in/dFz2zACN">https://lnkd.in/dFz2zACN</a></p>]]></content:encoded></item><item><title><![CDATA[Decoupled Training: Pioneering Offline Quantization for Neural Audio Compression]]></title><description><![CDATA[Neural audio codecs, particularly those employing neural networks to compress waveforms into discrete tokens, have become vital for recent advancements in audio generative models, with many state-of-the-art methods relying on the end-to-end training of autoencoders with a quantization bottleneck.]]></description><link>https://sankar1535.substack.com/p/decoupled-training-pioneering-offline</link><guid isPermaLink="false">https://sankar1535.substack.com/p/decoupled-training-pioneering-offline</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Fri, 11 Apr 2025 08:01:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gh3X!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Neural audio codecs, particularly those employing neural networks to compress waveforms into discrete tokens, have become vital for recent advancements in audio generative models, with many state-of-the-art methods relying on the end-to-end training of autoencoders with a quantization bottleneck.</p><p>But if the joint training of autoencoders and quantizers has been the dominant paradigm, does this fundamentally limit the exploration of more powerful and flexible quantization techniques developed independently?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gh3X!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gh3X!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 424w, https://substackcdn.com/image/fetch/$s_!gh3X!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 848w, https://substackcdn.com/image/fetch/$s_!gh3X!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 1272w, https://substackcdn.com/image/fetch/$s_!gh3X!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gh3X!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png" width="494" height="451.78752886836025" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:792,&quot;width&quot;:866,&quot;resizeWidth&quot;:494,&quot;bytes&quot;:166053,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/161009187?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!gh3X!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 424w, https://substackcdn.com/image/fetch/$s_!gh3X!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 848w, https://substackcdn.com/image/fetch/$s_!gh3X!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 1272w, https://substackcdn.com/image/fetch/$s_!gh3X!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b9f65c0-ef9e-40ae-a399-874eddeb6ad9_866x792.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The core innovation lies in a <strong>three-stage strategy termed QINCODEC</strong>, which involves pre-training a continuous compression autoencoder, followed by an offline quantization of the latent representations using a state-of-the-art trainable quantizer like QINCO2, and an optional final finetuning of the decoder.</p><p>This decoupling of training addresses the limitations of end-to-end approaches by allowing the use of any off-the-shelf quantizer, especially those with complex training procedures unsuitable for online updates, and simplifies the training process significantly.</p><p>A potential trade-off of this approach is that the offline quantization is inherently limited by the quality of the pre-trained latent representations and the bitrate, which can impact performance, particularly at lower bitrates.</p><p>The compression models were trained on 1-second, 44.1kHz audio clips from the WavCaps dataset, which contains around 400k general sounds, using 8 A100 GPUs for 1 million steps with an effective batch size of 240.</p><p>QINCODEC achieved <strong>competitive results at 8 kbps and outperformed state-of-the-art methods like DAC and ENCODEC at 16 kbps</strong> in various objective and subjective metrics, demonstrating the viability of non-end-to-end training.</p><p>This work suggests a future where audio codec design can be more modular and adaptable, amortising the cost of autoencoder pre-training and enabling the integration of advanced offline quantization techniques to potentially improve compression efficiency and quality for audio generative models.<br><br>More details: <a href="https://arxiv.org/abs/2503.19597">paper</a></p>]]></content:encoded></item><item><title><![CDATA[The Simplicity of Sound: A Leaner Approach to Audio Compression]]></title><description><![CDATA[Recent advancements in neural audio codecs have largely focused on achieving superior audio quality through increasingly intricate architectures that often employ multiple quantizers for encoding.]]></description><link>https://sankar1535.substack.com/p/the-simplicity-of-sound-a-leaner</link><guid isPermaLink="false">https://sankar1535.substack.com/p/the-simplicity-of-sound-a-leaner</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Thu, 10 Apr 2025 09:23:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!zdsB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zdsB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zdsB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 424w, https://substackcdn.com/image/fetch/$s_!zdsB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 848w, https://substackcdn.com/image/fetch/$s_!zdsB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 1272w, https://substackcdn.com/image/fetch/$s_!zdsB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zdsB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png" width="1456" height="736" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:736,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:245202,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://sankar1535.substack.com/i/161004729?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zdsB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 424w, https://substackcdn.com/image/fetch/$s_!zdsB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 848w, https://substackcdn.com/image/fetch/$s_!zdsB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 1272w, https://substackcdn.com/image/fetch/$s_!zdsB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248149a0-cd17-44e0-b947-565dc1e80726_1464x740.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Recent advancements in neural audio codecs have largely focused on achieving superior audio quality through increasingly intricate architectures that often employ multiple quantizers for encoding. But in the relentless pursuit of fidelity, have we overlooked the practical limitations imposed by the sheer computational cost and complexity of these multi-stage quantization processes?</p><p>This research introduces <strong>SQCodec, a novel neural audio codec that pioneers a return to elegance by demonstrating that high-fidelity audio compression can be achieved with a significantly more lightweight architecture relying on a single quantizer</strong>. Its core innovation lies in a carefully engineered design featuring streamlined convolutional networks, local Transformer modules, and the introduction of TConv, a mechanism specifically designed to capture both short- and long-term acoustic variations efficiently.</p><p>By adopting a single quantizer and a lightweight structure, this innovation directly tackles the inherent challenges associated with complex multi-quantizer systems, such as their considerable computational and memory demands which hinder real-world scalability. Furthermore, it simplifies the tokenisation process, avoiding the hierarchical token streams that complicate integration with downstream generative models and require specialised aggregation operations.</p><p>While the move to a single quantizer inherently presents a risk of reduced codebook capacity, potentially impacting reconstruction accuracy, this work addresses it through the implementation of Finite Scalar Quantization (FSQ), which supports large codebooks without the typical collapse issues. It's also worth noting that at very low bitrates, the Mel Spectrogram Distance (MEL) of this codec may be slightly higher compared to those explicitly optimising for it, as the training prioritises STFT distance.</p><p>The model was trained on a diverse set of audio data, including speech from LibriSpeech and Common Voice, music from MTG-Jamendo, and general sounds from FSD50K, showcasing its ability to learn across different acoustic domains. This training was efficiently conducted on a single NVIDIA RTX 4090 GPU, highlighting the model's reduced computational demands.</p><p>Crucially, experimental results demonstrate that <strong>SQCodec achieves audio quality comparable to state-of-the-art multi-quantizer codecs like DAC, while requiring an order of magnitude fewer parameters and significantly less computational overhead</strong>. At ultra-low bitrates, it often surpasses the performance of other single-quantizer models like WavTokenizer across various objective metrics.</p><p>This work paves the way for a future where high-fidelity neural audio compression is more accessible and integrable, particularly in resource-constrained scenarios and for applications like multimodal large language models and real-time communication. The success of this lightweight single-quantizer design opens exciting new avenues for optimising efficient audio processing and hardware-aware implementations.</p><p>More details: <a href="https://arxiv.org/abs/2504.04949">Paper</a></p>]]></content:encoded></item><item><title><![CDATA[Advancing Text-to-Speech: Acoustic Models and Vocoders in Focus]]></title><description><![CDATA[For the technically curious: let&#8217;s dissect the nuts and bolts of TTS pipelines&#8212;a domain where innovation is turning text into hyper-realistic speech at scale.]]></description><link>https://sankar1535.substack.com/p/advancing-text-to-speech-acoustic</link><guid isPermaLink="false">https://sankar1535.substack.com/p/advancing-text-to-speech-acoustic</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Thu, 10 Apr 2025 07:05:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kDwB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kDwB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kDwB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 424w, https://substackcdn.com/image/fetch/$s_!kDwB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 848w, https://substackcdn.com/image/fetch/$s_!kDwB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 1272w, https://substackcdn.com/image/fetch/$s_!kDwB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kDwB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png" width="1456" height="519" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:519,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kDwB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 424w, https://substackcdn.com/image/fetch/$s_!kDwB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 848w, https://substackcdn.com/image/fetch/$s_!kDwB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 1272w, https://substackcdn.com/image/fetch/$s_!kDwB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6bd4fcab-abe1-47be-9fbd-62eb962c8f8d_1600x570.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For the technically curious: let&#8217;s dissect the <strong>nuts and bolts of TTS pipelines</strong>&#8212;a domain where innovation is turning text into hyper-realistic speech at scale.</p><blockquote><p><strong>&#128273; Acoustic Models</strong></p></blockquote><p>At the heart of TTS, acoustic models predict features (e.g., mel-spectrograms) that capture the nuances of human speech:</p><ul><li><p><strong>RNN-based Models</strong>: Early stars like Tacotron 2 combined sequential processing and attention mechanisms to align text with speech, but autoregressive decoding limits their speed.</p></li><li><p><strong>CNN-based Models</strong>: Enter Deep Voice and ParaNet, with parallel processing and faster inference by capturing dependencies across entire sequences simultaneously.</p></li><li><p><strong>Transformer-based Models</strong>: TransformerTTS and FastSpeech revolutionized TTS by leveraging self-attention for global context modeling, enabling long-term prosody and rhythm. FastSpeech 2 even integrated pitch, energy, and duration predictors for enhanced expressiveness.</p></li><li><p><strong>LLM-based Approaches</strong>: Cutting-edge models like PromptTTS and InstructTTS bridge natural language and speech synthesis, offering fine-grained control over tone, emotion, and prosody using text-based prompts.</p></li></ul><blockquote><p><strong>&#128266; Speech Vocoders</strong></p></blockquote><p>Turning acoustic features into waveforms that sound natural is no small feat. Here&#8217;s what&#8217;s shaping this critical step:</p><ul><li><p><strong>GAN-based Vocoders</strong>: HiFi-GAN and Parallel WaveGAN leverage adversarial training to generate natural speech efficiently while maintaining frequency and time-domain coherence.</p></li><li><p><strong>Diffusion-based Vocoders</strong>: WaveGrad and DiffWave are redefining quality by iteratively refining waveforms from noise, though they challenge real-time applications with computational intensity.</p></li><li><p><strong>RNN/CNN Vocoders</strong>: Models like WaveRNN and Parallel WaveNet balance speed and quality, introducing multi-band strategies and non-autoregressive techniques for faster synthesis.</p></li></ul><blockquote><p><strong>&#128202; Representation Matters</strong></p></blockquote><ul><li><p><strong>Continuous Features</strong> (e.g., mel-spectrograms): Offer rich detail for expressiveness and prosody control but demand more compute.</p></li><li><p><strong>Discrete Tokens</strong> (e.g., quantized units): Compact and efficient, perfect for LLM-based and zero-shot TTS, but sometimes lose nuanced speech details.</p></li></ul><p>Whether you're optimizing <strong>FastSpeech for real-time synthesis</strong>, leveraging <strong>HiFi-GAN for natural voices</strong>, or exploring <strong>end-to-end systems like VITS</strong>, the choices in acoustic modeling and vocoders determine your system&#8217;s balance between <strong>quality, efficiency, and flexibility</strong>.</p><p>&#128161; Want to learn more about this emerging space? Explore the details here:<br> &#128073;<a href="https://arxiv.org/abs/2412.06602"> </a><strong><a href="https://arxiv.org/abs/2412.06602">https://arxiv.org/abs/2412.06602</a></strong></p><p>&#128161; <strong>What&#8217;s next?</strong> Diffusion-based real-time vocoders? Transformer-vocoder hybrids? Share your ideas or challenges&#8212;let&#8217;s push TTS further!</p>]]></content:encoded></item><item><title><![CDATA[Revolutionizing Audio Generation: Introducing X-Codec]]></title><description><![CDATA[The landscape of audio generation has been evolving rapidly, thanks to advancements in Large Language Models (LLMs).]]></description><link>https://sankar1535.substack.com/p/revolutionizing-audio-generation</link><guid isPermaLink="false">https://sankar1535.substack.com/p/revolutionizing-audio-generation</guid><dc:creator><![CDATA[Sankar Mukherjee]]></dc:creator><pubDate>Thu, 10 Apr 2025 06:51:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!jYFW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jYFW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jYFW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 424w, https://substackcdn.com/image/fetch/$s_!jYFW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 848w, https://substackcdn.com/image/fetch/$s_!jYFW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 1272w, https://substackcdn.com/image/fetch/$s_!jYFW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jYFW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png" width="960" height="535" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:535,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jYFW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 424w, https://substackcdn.com/image/fetch/$s_!jYFW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 848w, https://substackcdn.com/image/fetch/$s_!jYFW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 1272w, https://substackcdn.com/image/fetch/$s_!jYFW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72660d51-5ea1-4e91-b21c-11b1e19eb49f_960x535.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The landscape of audio generation has been evolving rapidly, thanks to advancements in Large Language Models (LLMs). However, there's been a persistent challenge: <strong>existing audio codecs, like EnCodec and HuBERT, fall short in maintaining semantic and acoustic integrity when used for audio LLM tasks</strong>.</p><p>Let's dive into the key pain points:<br> &#128313; <strong>Acoustic Codecs</strong> (e.g., EnCodec): While great for compressing audio, these struggle with "low-level fluctuations" in audio data, forcing LLMs to rely on massive datasets for training and causing slow convergence.<br> &#128313; <strong>Semantic Codecs</strong> (e.g., HuBERT): These excel at high-level semantic representation but ignore critical acoustic details, leading to two-stage frameworks for generation (semantic first, acoustic later).<br> &#128313; <strong>Hybrid Attempts</strong>: Models like SpeechTokenizer try to combine semantic and acoustic features but rely on complex two-stage AR+NAR frameworks.</p><div><hr></div><p><strong>&#127897;&#65039; Introducing X-Codec: Redefining Audio Encoding for LLMs &#128640;</strong></p><p>Enter <strong>X-Codec</strong>, a groundbreaking approach to audio encoding that integrates <strong>semantic information directly into the codec</strong>. Here's how it&#8217;s transforming audio generation:</p><p>&#128313; <strong>"X-Shaped" Architecture</strong>:<br> X-Codec combines <strong>semantic and acoustic features</strong> in a unified Residual Vector Quantization (RVQ) structure. This ensures tokens capture both semantic richness and acoustic fidelity, enabling accurate and streamlined audio generation.</p><p>&#128313; <strong>Semantic Reconstruction Loss</strong>:<br> By introducing a loss function after the RVQ stage, X-Codec guarantees that semantic integrity is preserved throughout the generation process.</p><p>&#128313; <strong>Improved Performance Across Tasks</strong>:<br> X-Codec drastically reduces WER in text-to-speech tasks while excelling in music continuation and text-to-sound applications. Its uniform tokenization ensures compatibility with existing LLMs without requiring structural changes.</p><p>&#128313; <strong>Enhanced Phonetic Discriminability</strong>:<br> Better phonetic precision compared to traditional acoustic codecs, especially with increased quantization layers.</p><p>&#128313; <strong>Versatility &amp; Compatibility</strong>:<br> Whether it's speech, music, or sound generation, X-Codec&#8217;s semantic-aware tokens deliver superior results without disrupting existing workflows.</p><p>&#128196; Read the full paper: <a href="https://arxiv.org/html/2408.17175v3">Paper</a></p><p>X-Codec is paving the way for more accurate, efficient, and versatile audio generation. What potential applications do you see for this innovation in your domain?</p>]]></content:encoded></item></channel></rss>