<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Tongyi Lab</title>
    <description>The latest articles on DEV Community by Tongyi Lab (@alibaba_tongyi_lab_25ad9f).</description>
    <link>https://dev.to/alibaba_tongyi_lab_25ad9f</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3600585%2F9fc831ed-e6e1-4fc6-9ca0-2087a3adb0d8.png</url>
      <title>DEV Community: Tongyi Lab</title>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/alibaba_tongyi_lab_25ad9f"/>
    <language>en</language>
    <item>
      <title>Jan 23, 2026 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 23 Jan 2026 05:56:09 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/jan-23-2026-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-ogf</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/jan-23-2026-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-ogf</guid>
      <description>&lt;p&gt;Hello, creators and builders,&lt;br&gt;
With the full Qwen3-TTS family now open-sourced this week, anyone can now design, clone, or fine-tune voices with studio-grade quality, in 10 languages, using just 0.6B parameters.&lt;br&gt;
Let’s dive in.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h3&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-TTS Full Model Family Open-Sourced&lt;/strong&gt;&lt;br&gt;
Qwen3-TTS is officially live. We’ve open-sourced the full family—VoiceDesign, CustomVoice, and Base—bringing high quality to the open community.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;5 models (0.6B &amp;amp; 1.7B)&lt;/li&gt;
&lt;li&gt;Free-form voice design &amp;amp; cloning&lt;/li&gt;
&lt;li&gt;Support for 10 languages&lt;/li&gt;
&lt;li&gt;SOTA 12Hz tokenizer for high compression&lt;/li&gt;
&lt;li&gt;Full fine-tuning support&lt;/li&gt;
&lt;li&gt;&lt;p&gt;SOTA performance&lt;br&gt;
Everything is out now—weights, code, and paper. Enjoy. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/QwenLM/Qwen3-TTS" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/collections/Qwen/qwen3-tts" rel="noopener noreferrer"&gt;Hugging Face&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelscope.cn/collections/Qwen/Qwen3-TTS" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://qwen.ai/blog?id=qwen3tts-0115" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/QwenLM/Qwen3-TTS/blob/main/assets/Qwen3_TTS.pdf" rel="noopener noreferrer"&gt;Paper&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/spaces/Qwen/Qwen3-TTS" rel="noopener noreferrer"&gt;Hugging Face Demo&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelscope.cn/studios/Qwen/Qwen3-TTS" rel="noopener noreferrer"&gt;ModelScope Demo&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.alibabacloud.com/help/en/model-studio/qwen-tts-voice-design" rel="noopener noreferrer"&gt;API&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Wan 2.6 Reference-to-Video Available in ComfyUI&lt;/strong&gt;&lt;br&gt;
Great news for visual creators: Wan 2.6 Reference-to-Video is now officially integrated into ComfyUI. Now you can turn 1–2 reference clips into short videos with natural motion, camera movement, and visual style.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Learn motion, camera angles, and style from reference videos&lt;/li&gt;
&lt;li&gt;Combine up to two clips for blended motion and visual style&lt;/li&gt;
&lt;li&gt;Generate 5–10 second videos at 720p or 1080p&lt;/li&gt;
&lt;li&gt;Maintain character consistency and natural movement&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;🔗 &lt;a href="https://blog.comfy.org/p/wan26-reference-to-video" rel="noopener noreferrer"&gt;Check out the blog&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Refine Lighting Effects: Qwen-Edit-2511_LightingRemap_Alpha0.2 from zooeyy&lt;/strong&gt;&lt;br&gt;
This LoRA by zooeyy is specifically designed for "color-block-guided relighting." The LoRA intelligently reconstructs lighting and renders atmosphere on subjects or scenes based on the position and hue of color blocks in the input image, while automatically removing the blocks themselves to produce natural, high-fidelity illumination.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/zooeyy/Qwen-Edit-2511_LightingRemap_Alpha0.2?spm=a2ty_o01.29997173.0.0.38145171sOwdmV&amp;amp;file=Qwen-Edit-2511_LightingRemap_Alpha0.2" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo8hnxl39ku0m4vps5jsp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo8hnxl39ku0m4vps5jsp.png" alt=" " width="800" height="423"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Object Remover: Qwen-Image-Edit-2511-Object-Remover from prithivMLmods&lt;/strong&gt;&lt;br&gt;
Qwen-Image-Edit-2511-Object-Remover by prithivMLmods is an adapter (LoRA) developed for Qwen’s Qwen-Image-Edit-2511 image-to-image model, specifically designed for precise object removal from images. The model removes specified objects while preserving the background and remaining elements, maintaining realism and original visual details. &lt;br&gt;
👉 &lt;a href="https://huggingface.co/prithivMLmods/Qwen-Image-Edit-2511-Object-Remover?spm=a2ty_o01.29997173.0.0.38145171sOwdmV" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft78hazxuh5i3qgw1q91r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft78hazxuh5i3qgw1q91r.png" alt=" " width="800" height="500"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Image Style Customization Workflow: AmazingZImageWorkflow from martin-rizzo&lt;/strong&gt;&lt;br&gt;
This workflow for Z-Image-Turbo by martin-rizzo expands the ComfyUI base workflow with additional features, particularly focused on high-quality image styles and user-friendly functionality, while also integrating an image refiner and a simple upscaler. Its standout feature is a dedicated Style Selector with 18 customizable presets, allowing you to switch artistic directions instantly. &lt;br&gt;
👉 &lt;a href="https://github.com/martin-rizzo/AmazingZImageWorkflow" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frdwcfywtki2erp8giy90.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frdwcfywtki2erp8giy90.png" alt=" " width="800" height="933"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Z-Image Nodes Collection: ComfyUI-ZImagePowerNodes from martin-rizzo&lt;/strong&gt;&lt;br&gt;
This toolkit by martin-rizzo solves common pain points by offering one-click style switching and high-speed "Turbo" configurations (optimizing for as few as 7 steps). &lt;br&gt;
It effectively brings precision and ease-of-use to the Z-Image ecosystem.&lt;br&gt;
👉&lt;a href="https://github.com/martin-rizzo/ComfyUI-ZImagePowerNodes" rel="noopener noreferrer"&gt; Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq2v71b0tvou5b5qvaa3i.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq2v71b0tvou5b5qvaa3i.png" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New model releases &amp;amp; upgrades&lt;/li&gt;
&lt;li&gt;AI research breakthroughs&lt;/li&gt;
&lt;li&gt;Open-source tools you can use today&lt;/li&gt;
&lt;li&gt;Community highlights that inspire&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>aigc</category>
      <category>opensource</category>
      <category>qwen</category>
    </item>
    <item>
      <title>Jan 16, 2026 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 16 Jan 2026 08:04:59 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/jan-16-2026-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-46d9</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/jan-16-2026-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-46d9</guid>
      <description>&lt;p&gt;Hello, creators and builders,&lt;br&gt;
While this week didn’t bring new model releases or research papers from our lab, it was anything but quiet.&lt;br&gt;
In fact, it was a brilliant showcase of community ingenuity — developers building, refining, and reimagining what’s possible with AIGC ecosystem.&lt;br&gt;
Let’s celebrate what you built this week.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Powerful ControlNet Union: &lt;a href="https://huggingface.co/alibaba-pai/Qwen-Image-2512-Fun-Controlnet-Union" rel="noopener noreferrer"&gt;Qwen-Image-2512-Fun-Controlnet-Union&lt;/a&gt; from &lt;a href="https://huggingface.co/alibaba-pai" rel="noopener noreferrer"&gt;alibaba-pai&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
A breakthrough in control: this new ControlNet Union from alibaba-pai integrates Canny, HED, Depth, Pose, MLSD, Scribble, and Inpainting into a single model — all built on 5 layer blocks of Qwen-Image-2512. No more juggling models. Just one pipeline, infinite control.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/alibaba-pai/Qwen-Image-2512-Fun-Controlnet-Union" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftmf5x2mtuw2kairpu6j2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftmf5x2mtuw2kairpu6j2.png" alt=" " width="800" height="569"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Unblur-Upscale: &lt;a href="https://huggingface.co/prithivMLmods/Qwen-Image-Edit-2511-Unblur-Upscale" rel="noopener noreferrer"&gt;Qwen-Image-Edit-2511-Unblur-Upscale&lt;/a&gt; from &lt;a href="https://huggingface.co/prithivMLmods" rel="noopener noreferrer"&gt;prithivMLmods&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
This new adapter for Qwen-Image-Edit-2511 is designed to unblur and upscale images to high resolution. The model enhances image clarity by reducing blur, restoring fine details, and improving overall sharpness while preserving natural textures and realistic colors.&lt;br&gt;
A masterclass in efficiency and visual fidelity.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/prithivMLmods/Qwen-Image-Edit-2511-Unblur-Upscale" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fog6shvkvttvo1ejspljo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fog6shvkvttvo1ejspljo.png" alt=" " width="800" height="496"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Smater Control: &lt;a href="https://huggingface.co/alibaba-pai/Z-Image-Turbo-Fun-Controlnet-Union-2.1" rel="noopener noreferrer"&gt;Z-Image-Turbo-Fun-Controlnet-Union-2.1-2601&lt;/a&gt; from &lt;a href="https://huggingface.co/alibaba-pai" rel="noopener noreferrer"&gt;Alibaba-PAI&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
Alibaba-pai just updated Z-Image-Turbo-Fun-Controlnet-Union-2.1-2601. The team has fixed the inference speed bug and significantly improved robustness by restructuring the dataset with multi-resolution control images (512~1536).&lt;br&gt;
Highlight: A new 8-step distilled version is now available. It solves the blurriness issue found in previous tests and finally restores the model's true acceleration capabilities.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/alibaba-pai/Z-Image-Turbo-Fun-Controlnet-Union-2.1" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdpfx0u2x93cvcw9n2x1r.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdpfx0u2x93cvcw9n2x1r.png" alt=" " width="800" height="768"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Image Rotation Restoration: &lt;a href="https://huggingface.co/dx8152/Qwen-Image-Edit-2511-Gaussian-Splash" rel="noopener noreferrer"&gt;Qwen-Image-Edit-2511-Gaussian-Splash&lt;/a&gt; from &lt;a href="https://huggingface.co/dx8152" rel="noopener noreferrer"&gt;dx8152&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
Generate a PLY point cloud --&amp;gt; Adjust angle in editor --&amp;gt;Refine with Qwen-Image-Edit-2511 Gaussian Splash LoRA.&lt;br&gt;
It accurately restores complex perspective shifts. &lt;br&gt;
As shown in the demo, it handles 3D rotation and can even restore high-def details from blurry close-ups. Within a 45° range, the consistency is unmatched.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/dx8152/Qwen-Image-Edit-2511-Gaussian-Splash" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Anime Sketch Extractor: &lt;a href="https://huggingface.co/yeq6x/QwenImageEdit_LoRA" rel="noopener noreferrer"&gt;QwenImageEdit_LoRA&lt;/a&gt; from &lt;a href="https://huggingface.co/yeq6x" rel="noopener noreferrer"&gt;yeq6x&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
Interested in seeing the clean line art behind your favorite anime characters? This new LoRA for Qwen-Image-Edit-2509 effectively extracts sketches from existing images. &lt;br&gt;
It’s a handy tool for anyone wanting to study character structure.&lt;/p&gt;

&lt;p&gt;A must-have for animators and illustrators.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/yeq6x/QwenImageEdit_LoRA" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc5r212xkuntswhfa9r9z.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc5r212xkuntswhfa9r9z.png" alt=" " width="800" height="266"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Sampling Optimized for Z-Image: &lt;a href="https://github.com/LAOGOU-666/ComfyUI-LG_SamplingUtils" rel="noopener noreferrer"&gt;ComfyUI-LG_SamplingUtils&lt;/a&gt; from &lt;a href="https://github.com/LAOGOU-666" rel="noopener noreferrer"&gt;LAOGOU-666&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
ComfyUI-LG_SamplingUtils is a comprehensive toolset designed for ComfyUI by LAOGOU-666, providing a series of practical sampling nodes that make operations more intuitive and convenient. This extension focuses on advanced sampling techniques, particularly optimized for Flow Matching models like ZImage and Lumina2.&lt;br&gt;
👉 &lt;a href="https://github.com/LAOGOU-666/ComfyUI-LG_SamplingUtils" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv0n6a0ecr9oxiwym20v1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv0n6a0ecr9oxiwym20v1.png" alt=" " width="800" height="695"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  🔥 Events &amp;amp; Challenges
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;New Challenge: #starringwithwan Is LIVE!&lt;/strong&gt;&lt;br&gt;
The #starringwithwan Challenge is officially open! Use the “Starring” feature in Wan App to create a video where you share the screen with our lead AI character Rowan and Ewan.🏆 Top 20 creators win 1-month Premium Membership (redeemable code)&lt;br&gt;
🗓️ Deadline: January 28, 2026&lt;br&gt;
👀 How to Enter&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Download the Wan App: &lt;a href="https://wan.video/#wan-app?cref=mkt&amp;amp;cinfo=twitter" rel="noopener noreferrer"&gt;https://wan.video/#wan-app?cref=mkt&amp;amp;cinfo=twitter&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Use the “Starring” feature to create a video with our lead character&lt;/li&gt;
&lt;li&gt;Post your video on X with hashtags #starringwithwan &amp;amp; #wanapp&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;🔗 &lt;a href="https://x.com/Alibaba_Wan/status/2011633882156728426" rel="noopener noreferrer"&gt;Learn More&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New model releases &amp;amp; upgrades&lt;/li&gt;
&lt;li&gt;AI research breakthroughs&lt;/li&gt;
&lt;li&gt;Open-source tools you can use today&lt;/li&gt;
&lt;li&gt;Community highlights that inspire&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>aigc</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Jan 9, 2026 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 09 Jan 2026 10:13:08 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/jan-9-2026-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-5cn3</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/jan-9-2026-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-5cn3</guid>
      <description>&lt;p&gt;🎄 Happy New Year!&lt;br&gt;
We hope you enjoyed a restful holiday season filled with joy, creativity, and maybe even a few AI experiments by the fireplace. As we step into 2026, we’re more inspired than ever by what this community has built — and what we’ll create together in the year ahead.&lt;br&gt;
To kick off the new year, we’re thrilled to give you our first gift of 2026: *&lt;em&gt;Wan App is now live on iOS &amp;amp; Android! *&lt;/em&gt;🎁&lt;br&gt;
Please note: Wan App is rolling out gradually and may not yet be available in all countries or regions. We’re working hard to bring it to you as quickly as possible. Scan the QR code below and give it a try!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4eg5n3far6s3j20bs7cx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4eg5n3far6s3j20bs7cx.png" alt=" " width="608" height="406"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This week also brings groundbreaking releases, let’s dive in.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Qwen-Image-2512: Finer Details, Greater Realism&lt;/strong&gt;&lt;br&gt;
We are thrilled to announce the Qwen-Image-2512 open-source release! This December update pushes the boundaries of our text-to-image foundational model, moving from "AI-generated" looks to true photorealism.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Enhanced Human Realism : We’ve eliminated the artificial "AI look" by capturing intricate facial details—like wrinkles and pores—and ensuring better adherence to body postures.&lt;/li&gt;
&lt;li&gt;Finer Natural Detail : Experience notably more detailed rendering of landscapes, misty waterfalls, and animal fur with distinct, individual strands.&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Advanced Text Rendering : Achieve professional-grade layout for complex infographics and PPT slides with unprecedented textual accuracy.&lt;br&gt;
Try it now: &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/Qwen/Qwen-Image-2512" rel="noopener noreferrer"&gt;Qwen Chat&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/Qwen/Qwen-Image-2512" rel="noopener noreferrer"&gt;Hugging Face&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelscope.ai/models/Qwen/Qwen-Image-2512" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/QwenLM/Qwen-Image" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://qwen.ai/blog?id=qwen-image-2512" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/spaces/Qwen/Qwen-Image-2512" rel="noopener noreferrer"&gt;Hugging Face Demo&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelscope.cn/aigc/imageGeneration" rel="noopener noreferrer"&gt;ModelScope Demo&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelstudio.console.alibabacloud.com/?tab=doc#/doc/?type=model&amp;amp;url=2840914_2&amp;amp;modelId=group-qwen-image-max" rel="noopener noreferrer"&gt;API&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen Code v0.6.0: Smarter, More Connected&lt;/strong&gt;&lt;br&gt;
Your AI coding assistant just got better:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Experimental Skills: Introduced experimental Skills feature for extended capabilities&lt;/li&gt;
&lt;li&gt;VS Code Enhancements: Improved extension description with download links and clickable bash toolcall outputs&lt;/li&gt;
&lt;li&gt;Commands Support: Added /compress and /summary commands for non-interactive &amp;amp; ACP usage&lt;/li&gt;
&lt;li&gt;Multi-Provider Support: Added Gemini and Anthropic providers with normalized authentication configuration&lt;/li&gt;
&lt;li&gt;Enhancements &amp;amp; Stability: Improved testing reliability with fixed flaky integration tests, enhanced Windows compatibility through CLI path resolution, updated OAuth client for Figma MCP server, streamlined SDK release workflows, and clearer README documentation for faster onboarding.
🔗 &lt;a href="https://github.com/QwenLM/qwen-code/releases/tag/v0.6.0" rel="noopener noreferrer"&gt;Check out the full changelog&lt;/a&gt;
👉 Get started in Terminal
&lt;/li&gt;
&lt;/ul&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;npm install -g @qwen-code/qwen-code@latest
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;MAI-UI: The Foundation GUI Agent Family&lt;/strong&gt;&lt;br&gt;
We’re releasing MAI-UI—a family of foundation GUI agents. It natively integrates MCP tool use, agent user interaction, device–cloud collaboration, and online RL, establishing state-of-the-art results in general GUI grounding and mobile GUI navigation, surpassing Gemini-2.5-Pro, Seed1.8, and UI-Tars-2 on AndroidWorld.&lt;br&gt;
To meet real-world deployment constrains, MAI-UI includes a full-spectrum of sizes, including 2B, 8B, 32B and 235B-A22B variants. We open-sourced two models: MAI-UI-2B and MAI-UI-8B.&lt;br&gt;
Technical Highlight:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;MCP tool use: MAI-UI natively support MCP tool use, compressing long, fragile UI operation sequences into a few API calls.&lt;/li&gt;
&lt;li&gt;Agent user interaction: MAI-UI proactively ask clarifying questions when user instructions are ambiguous or incomplete.&lt;/li&gt;
&lt;li&gt;Device-cloud collaboration: MAI-UI can dynamically select on-device or cloud execution based on task execution state and data sensitivity.&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Online RL: Significant experimental gains from scaling parallel environments from 32 to 512 (+5.2 points) and increasing environment step budget from 15 to 50 (+4.3 points).&lt;br&gt;
Get started:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/Tongyi-MAI/MAI-UI" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://tongyi-mai.github.io/MAI-UI/" rel="noopener noreferrer"&gt;Project page&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/Tongyi-MAI/MobileWorld" rel="noopener noreferrer"&gt;MobileWorld benchmark&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://tongyi-mai.github.io/MobileWorld/" rel="noopener noreferrer"&gt;MobileWorld homepage&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-VL-Embedding &amp;amp; Qwen3-VL-Reranker: Advanced Multimodal Retrieval &amp;amp; Cross-Modal Understanding&lt;/strong&gt;&lt;br&gt;
Meet Qwen3-VL-Embedding and Qwen3-VL-Reranker:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Built upon the robust Qwen3-VL foundation model&lt;/li&gt;
&lt;li&gt;Processes text, images, screenshots, videos, and mixed modality inputs&lt;/li&gt;
&lt;li&gt;Supports 30+ languages&lt;/li&gt;
&lt;li&gt;Achieves state-of-the-art performance on multimodal retrieval benchmarks&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Two-stage retrieval architecture:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Embedding Model – generates semantically rich vector representations in a unified embedding space&lt;/li&gt;
&lt;li&gt;Reranker Model – computes fine-grained relevance scores for enhanced retrieval accuracy&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Developer-friendly capabilities:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Configurable embedding dimensions&lt;/li&gt;
&lt;li&gt;Task-specific instruction customization&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Embedding quantization support for efficient and cost-effective downstream deployment&lt;br&gt;
Now Available at:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Hugging Face：&lt;a href="https://huggingface.co/collections/Qwen/qwen3-vl-embedding" rel="noopener noreferrer"&gt;Qwen3-VL-Embedding&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Hugging Face：&lt;a href="https://huggingface.co/collections/Qwen/qwen3-vl-reranker" rel="noopener noreferrer"&gt;Qwen3-VL-Reranker&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/QwenLM/Qwen3-VL-Embedding" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://qwen.ai/blog?id=qwen3-vl-embedding" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/QwenLM/Qwen3-VL-Embedding/blob/main/assets/qwen3vlembedding_technical_report.pdf" rel="noopener noreferrer"&gt;Tech Report&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧠 Research Breakthroughs
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;MobileWorld: A Next-Gen Benchmark for Real-World Mobile Agents&lt;/strong&gt;&lt;br&gt;
Meet MobileWorld — a revolutionary benchmark from the MAI Team at Tongyi Lab that transcends the limitations of traditional ones by realistically simulating users’ complex real-world demands:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Substantially increased task difficulty: Featuring long-horizon, cross-app workflows, tasks require an average of 27.8 steps (nearly double that of AndroidWorld), with 62.2% of tasks necessitating coordination across multiple apps, ensuring strong alignment with real-life usage scenarios.&lt;/li&gt;
&lt;li&gt;Novel task paradigms: Introducing agent-user interaction tasks and MCP-argumented tasks, which challenge agents’ abilities to interpret ambiguous instructions and make tool-calling decisions.&lt;/li&gt;
&lt;li&gt;A robust and reproducible evaluation environment: Built on a self-hosted app ecosystem, Docker containers, and AVD snapshots, this infrastructure guarantees consistent, fair, and replicable experimental conditions.
Evaluation results reveal a stark reality: even the current state-of-the-art (SOTA) models achieve only a 51.7% success rate, with end-to-end models peaking at just 20.9%. On the agent-user interaction and MCP-argumented tasks, mainstream agents’ success rate drops nearly to zero, highlighting a significant gap between agent's capabilities and real-world deployment readiness.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The codebase is now open-source:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://github.com/Tongyi-MAI/MobileWorld" rel="noopener noreferrer"&gt;GitHub &lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://tongyi-mai.github.io/MobileWorld/" rel="noopener noreferrer"&gt;Website&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2512.19432" rel="noopener noreferrer"&gt;arXiv&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Higging Face Wrapped 2025: 2 Papers from Qwen Upvoted As Top10&lt;/strong&gt;&lt;br&gt;
We’re honored that both the Qwen3 Technical Reportand Group Sequence Policy Optimization (GSPO)were featured in Hugging Face’s Wrapped 2025 Top 10 most upvoted papers.&lt;br&gt;
Thank you to the entire Qwen team — and to you, our community — for your upvotes.&lt;/p&gt;




&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;See Qwen3-VL “Think” Before It Speaks: comfyui-prompt-generator from d3cker&lt;/strong&gt;&lt;br&gt;
We are stoked to recommend the "comfyui-prompt-generator" by d3cker. This custom node is a total powerhouse, especially when using Qwen3-VL-8B-Thinking—it actually displays its "thinking process" before spitting out the perfect prompt.&lt;br&gt;
&lt;a href="https://www.reddit.com/r/comfyui/comments/1pxaq1f/custom_node_image_based_prompt_generator/" rel="noopener noreferrer"&gt;👉 Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1i7ij10expqnbdmzzbk9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1i7ij10expqnbdmzzbk9.png" alt=" " width="800" height="310"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AnyPose LoRA: AnyPose from lilylilith&lt;/strong&gt;&lt;br&gt;
Made in mind with the new Qwen Image Edit 2511 lightning LoRA for fast inference, with just a single reference image as a pose guide, you can pilot any image to follow that pose with this LoRA.&lt;br&gt;
&lt;a href="https://huggingface.co/lilylilith/AnyPose" rel="noopener noreferrer"&gt;👉 Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F12flrbfwhdzyxiac3uas.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F12flrbfwhdzyxiac3uas.png" alt=" " width="800" height="680"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Upscale2K LoRA: Qwen-Image-Edit-2511-Upscale2K from valiantcat&lt;/strong&gt;&lt;br&gt;
This is a model for High-definition magnification of the picture, trained on Qwen/Qwen-Image-Edit-2511, and it is mainly used for losslessly enlarging images to approximately 2K size, injecting a serious dose of clarity and texture into every frame.&lt;br&gt;
&lt;a href="https://huggingface.co/valiantcat/Qwen-Image-Edit-2511-Upscale2K" rel="noopener noreferrer"&gt;👉 Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3duzcg46gf2cjqa9i3fl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3duzcg46gf2cjqa9i3fl.png" alt=" " width="800" height="656"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Speed Meets Aesthetics: Qwen-Image-2512 Turbo V2.0 from Wuli-art&lt;/strong&gt;&lt;br&gt;
Wuli Team has released V2.0 of their Qwen-Image-2512 Turbo LoRA. &lt;br&gt;
Optimized for 4-8 steps, it offers a perfect balance of insane speed and high-aesthetic output. A vital resource for efficient local deployment and high-fidelity generation.&lt;br&gt;
&lt;a href="https://huggingface.co/Wuli-art/Qwen-Image-2512-Turbo-LoRA" rel="noopener noreferrer"&gt;👉 Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1phxdvfq6swgirm82xx1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1phxdvfq6swgirm82xx1.png" alt=" " width="800" height="402"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New model releases &amp;amp; upgrades&lt;/li&gt;
&lt;li&gt;AI research breakthroughs&lt;/li&gt;
&lt;li&gt;Open-source tools you can use today&lt;/li&gt;
&lt;li&gt;Community highlights that inspire&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>android</category>
      <category>ios</category>
      <category>news</category>
    </item>
    <item>
      <title>Dec 26, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 26 Dec 2025 07:30:55 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-26-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-910</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-26-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-910</guid>
      <description>&lt;p&gt;🎄 Merry Christmas and Happy New Year!&lt;br&gt;
As 2025 comes to a close, we want to extend our deepest gratitude to each of you for your creativity and support this year. Your experiments, feedback, and brilliant creations have been the heartbeat of our open ecosystem.&lt;br&gt;
As a final gift of the year, we’re excited to share the newest models and tools born in this last week of 2025.&lt;br&gt;
Let’s take a look at what’s just landed.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing Qwen-Image-Layered: native image decomposition, fully open-sourced&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Why it stands out&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Photoshop-grade layering: Physically isolated RGBA layers with true native editability&lt;/li&gt;
&lt;li&gt;Prompt-controlled structure: Explicitly specify 3–10 layers — from coarse layouts to fine-grained details&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Infinite decomposition: Keep drilling down: layers within layers, to any depth of detail&lt;/p&gt;

&lt;p&gt;🔗 Get started:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/Qwen/Qwen-Image-Layered" rel="noopener noreferrer"&gt;Hugging Face&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelscope.cn/models/Qwen/Qwen-Image-Layered" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://github.com/QwenLM/Qwen-Image-Layered" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qwen.ai/blog?id=qwen-image-layered" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2512.15603" rel="noopener noreferrer"&gt;Technical Report (arXiv)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/spaces/Qwen/Qwen-Image-Layered" rel="noopener noreferrer"&gt;Live Demo (HF)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelscope.cn/studios/Qwen/Qwen-Image-Layered" rel="noopener noreferrer"&gt;Live Demo (ModelScope)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;New Open-Source End-to-End Voice Model: Fun-Audio-Chat&lt;/strong&gt;&lt;br&gt;
We’re open-sourcing Fun-Audio-Chat — an end-to-end voice model that’s more than just a chatbot.&lt;br&gt;
It’s your AI voice partner:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Empathetic: Understands emotion, tone, and intent&lt;/li&gt;
&lt;li&gt;Action-oriented: Follows voice commands to complete tasks&lt;/li&gt;
&lt;li&gt;End-to-end S2S architecture: lower latency, higher efficiency.&lt;/li&gt;
&lt;li&gt;Dual-resolution design: ~50% lower GPU cost&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Leader in multiple benchmarks (OpenAudioBench, MMAU, etc.).&lt;br&gt;
Open, efficient, and deeply useful.&lt;br&gt;
🔗 Try it:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/FunAudioLLM/Fun-Audio-Chat" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/FunAudioLLM/Fun-Audio-Chat-8B" rel="noopener noreferrer"&gt;Hugging Face&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelscope.cn/models/FunAudioLLM/Fun-Audio-Chat-8B" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://funaudiollm.github.io/funaudiochat" rel="noopener noreferrer"&gt;Demo&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;New Qwen3-TTS Lineup: VoiceDesign &amp;amp; VoiceClone&lt;/strong&gt;&lt;br&gt;
Create, control, and clone voices—faster and more expressive than ever.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;VoiceDesign-VD-Flash&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fully controllable speech via free-form text instructions — tone, rhythm, emotion, persona&lt;/li&gt;
&lt;li&gt;No preset voices. Design your own unique vocal identity&lt;/li&gt;
&lt;li&gt;Outperforms GPT-4o-mini-tts &amp;amp; Gemini-2.5-pro on role-play benchmarks&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;VoiceClone-VC-Flash&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Clone any voice from just 3 seconds of audio&lt;/li&gt;
&lt;li&gt;Generate speech in 10 languages (CN / EN / JP / ES + more)&lt;/li&gt;
&lt;li&gt;15% lower WER vs. ElevenLabs &amp;amp; GPT-4o-Audio in multilingual tests&lt;/li&gt;
&lt;li&gt;Context-aware cadence for more natural delivery&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;🔗 Try it now&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="http://chat.qwen.ai" rel="noopener noreferrer"&gt;Qwen Chat&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qwen.ai/blog?id=qwen3-tts-vc-voicedesign" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen-Image-Edit-2511:  Stronger Consistency &amp;amp; Real-World Image Editing&lt;/strong&gt;&lt;br&gt;
What’s new in 2511:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Stronger multi-person consistency for group photos and complex scenes&lt;/li&gt;
&lt;li&gt;Built-in popular community LoRAs — no extra tuning required&lt;/li&gt;
&lt;li&gt;Enhanced industrial &amp;amp; product design generation&lt;/li&gt;
&lt;li&gt;Reduced image drift with dramatically improved character &amp;amp; identity consistency&lt;/li&gt;
&lt;li&gt;Improved geometric reasoning, including construction lines and structural edits&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;From identity-preserving portrait edits to high-fidelity multi-person fusion and practical engineering &amp;amp; design workflows, 2511 pushes image editing to the next level.&lt;br&gt;
🔗 Try it now&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://chat.qwen.ai/?inputFeature=image_edit" rel="noopener noreferrer"&gt;Qwen Chat (Image Edit)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/Qwen/Qwen-Image-Edit-2511" rel="noopener noreferrer"&gt;Hugging Face&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelscope.cn/models/Qwen/Qwen-Image-Edit-2511" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Z-Image Turbo: #1 Open-Weight Text-to-Image Model in the Artificial Analysis Image Arena&lt;/strong&gt;&lt;br&gt;
According to Artificial Analysis, Z-Image Turbo now ranks #1 among all open-weight image models in the Artificial Analysis Image Arena.&lt;br&gt;
Why it leads:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Only $5/1k images on Alibaba Cloud &lt;/li&gt;
&lt;li&gt;Runs on consumer with just 16GB of memory&lt;/li&gt;
&lt;li&gt;Apache 2.0 open source license
A 6B powerhouse that proves: high quality doesn’t require high cost.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frc16kbm0zcih6p85ly0n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frc16kbm0zcih6p85ly0n.png" alt=" " width="800" height="507"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Portrait Photography: BEYOND REALITY Z IMAGE 1.0 from Nurburgring&lt;/strong&gt;&lt;br&gt;
This model, fine-tuned from Z-Image-Turbo, optimizes skin textures and environmental details while maintaining analog film aesthetics. It is available in both BF16 and FP8 versions, the latter being compatible with 8GB VRAM hardware.&lt;br&gt;
👉 &lt;a href="https://modelscope.cn/models/Nurburgring/BEYOND_REALITY_Z_IMAGE/summary" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;br&gt;
● New model releases &amp;amp; upgrades&lt;br&gt;
● AI research breakthroughs&lt;br&gt;
● Open-source tools you can use today&lt;br&gt;
● Community highlights that inspire&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>news</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Dec 19, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 19 Dec 2025 07:47:35 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-19-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-51o3</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-19-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-51o3</guid>
      <description>&lt;p&gt;Hello, creators and builders,&lt;br&gt;
This week was a harvest of breakthroughs in voice and video AI.&lt;br&gt;
From Wan2.6 — our cinematic multimodal generation model that brings characters to life with consistent appearance, voice, and cinematic storytelling — to Fun-ASR and Fun-CosyVoice 3, our speech models now available with open-source versions, the future of expressive AI has never felt closer.&lt;/p&gt;

&lt;p&gt;Let’s dive in.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing Wan2.6: The Cinematic Multimodal Generation Model&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Starring: Cast characters from reference videos into new scenes. Support human or human-like figures, enabling complex multi-person and human-object interactions with appearance and voice consistency.&lt;/li&gt;
&lt;li&gt;Intelligent Multi-shot Narrative: Turn simple prompts into auto-storyboarded, multi-shot videos. Maintain visual consistency and upgrade storytelling from single shots to rich narratives.&lt;/li&gt;
&lt;li&gt;Native A/V Sync: Generate multi-speaker dialogue with natural lip-sync and studio-quality audio. It doesn’t just look real - it sounds real.&lt;/li&gt;
&lt;li&gt;Cinematic Quality: 15s 1080p HD generation with comprehensive upgrades to instruction adherence, motion physics, and aesthetic control.&lt;/li&gt;
&lt;li&gt;Advanced Image Synthesis and Editing: Deliver cinematic photorealism with precise control over lens and lighting. Support multi-image referencing for commercial-grade consistency and faithful aesthetic transfer.&lt;/li&gt;
&lt;li&gt;Storytelling with Structure: Generate interleaved texts and images powered by real-world knowledge and reasoning capabilities, enabling hierarchical and structured visual narratives.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;🔗 &lt;a href="https://wan.video/" rel="noopener noreferrer"&gt;Try Wan 2.6 yourself&lt;/a&gt; (150 Free Credits Everyday!)&lt;/p&gt;

&lt;p&gt;🔗 &lt;a href="https://modelstudio.alibabacloud.com/#/modeListType/wan" rel="noopener noreferrer"&gt;API&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fun-ASR Upgrade: Noise-robust, Multilingual, Customizabe ASR&lt;/strong&gt;&lt;br&gt;
We’re thrilled to unveil the newest evolution of Fun-ASR, our enterprise-grade end-to-end Automatic Speech Recognition model — now more noise-robust, more multilingual, and more customizable than ever. We’re also releasing the lightweight Fun-ASR-Nano (0.8B) model as open source.&lt;/p&gt;

&lt;p&gt;Major Upgrades in Fun-ASR&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Achieves 93% accuracy in real-world noisy environments such as conferences, metro stations, and in-car speech.&lt;/li&gt;
&lt;li&gt;Breakthrough in lyric recognition: accurately transcribes vocals even with strong background music or rap-style delivery.&lt;/li&gt;
&lt;li&gt;Supports 31 languages, with enhanced performance for East Asian &amp;amp; Southeast Asian languages including Japanese and Vietnamese.&lt;/li&gt;
&lt;li&gt;Covers 7 major Chinese dialect groups and 26 regional accents with high precision.&lt;/li&gt;
&lt;li&gt;The RAG-based solution boosts enterprise-grade customization by raising the hotword limit from 1,000 to 10,000 without compromising accuracy.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Fun-ASR-Nano (0.8B) Released as Open Source&lt;/strong&gt;&lt;br&gt;
Lightweight yet highly noise-resistant ASR model optimized for: Compute-sensitive scenarios, Edge devices, and Low-latency real-time recognition&lt;/p&gt;

&lt;p&gt;🔗 Now available on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://github.com/FunAudioLLM/Fun-ASR" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://funaudiollm.github.io/funasr/" rel="noopener noreferrer"&gt;github.io&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelscope.cn/models/FunAudioLLM/fun-asr-nano" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/FunAudioLLM/Fun-ASR-Nano-2512" rel="noopener noreferrer"&gt;Hugging Face&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/spaces/FunAudioLLM/Fun-ASR-Nano" rel="noopener noreferrer"&gt;Demo&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Fun-CosyVoice 3: The Next-Generation Text-to-Speech Model&lt;/strong&gt;&lt;br&gt;
Fun-CosyVoice 3, our next-generation text-to-speech model — now faster, more expressive, and officially open-sourced.&lt;/p&gt;

&lt;p&gt;What’s New in Fun-CosyVoice 3:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;50% lower first-token latency with full bidirectional streaming TTS, enabling true real-time “type-to-speech” experiences.&lt;/li&gt;
&lt;li&gt;Significant improvement in Chinese–English code-switching, with WER (Word Error Rate) reduced by 56.4%.&lt;/li&gt;
&lt;li&gt;Enhanced zero-shot voice cloning: replicate a voice using only 3 seconds of audio, now with improved consistency and emotion control.&lt;/li&gt;
&lt;li&gt;Support for 30+ timbres, 9 languages, 18 Chinese dialect accents, and 9 emotion styles, with cross-lingual voice cloning capability.&lt;/li&gt;
&lt;li&gt;Achieves significant improvements across multiple standard benchmarks, with a 26% relative reduction in character error rate (CER) on challenging scenarios (test-hard), and certain metrics approaching those of human-recorded speech.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Fun-CosyVoice 3 (0.5B) Now Open Source&lt;/strong&gt;&lt;br&gt;
We’re releasing a lightweight yet powerful 0.5B-parameter version with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Zero-shot voice cloning&lt;/li&gt;
&lt;li&gt;Local deployment support&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Outperforms popular open-source TTS models across evaluated metrics.&lt;br&gt;
🔗 Explore &amp;amp; Download:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://modelscope.cn/models/FunAudioLLM/Fun-CosyVoice3-0.5B" rel="noopener noreferrer"&gt;Modelscope&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://github.com/FunAudioLLM/CosyVoice" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://funaudiollm.github.io/cosyvoice3/" rel="noopener noreferrer"&gt;github.io&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/FunAudioLLM/Fun-CosyVoice3-0.5B-2512" rel="noopener noreferrer"&gt;Huggingface&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen Code v0.5.0: Smarter AI coding assistant&lt;/strong&gt;&lt;br&gt;
What’s new:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;VSCode Integration: Bundled CLI into VSCode release package with improved cross-platform compatibility&lt;/li&gt;
&lt;li&gt;Native TypeScript SDK: Seamlessly integrate with Node/TS&lt;/li&gt;
&lt;li&gt;Smart Session Management: Auto-saves and continue conversations&lt;/li&gt;
&lt;li&gt;Support for OpenAI-compatible reasoning models, including DeepSeek V3.2, Kimi-K2, and more&lt;/li&gt;
&lt;li&gt;Control custom tools via SDK-hosted servers&lt;/li&gt;
&lt;li&gt;Russian Language Support: Added internationalization with Russian language option&lt;/li&gt;
&lt;li&gt;Enhanced User Experience: Terminal bell setting for audio notifications and session resume command display&lt;/li&gt;
&lt;li&gt;Testing &amp;amp; Stability: Better Ubuntu shell support, faster SDK timeouts, and rock-solid test stability&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉Get started in Terminal: &lt;br&gt;
npm install -g @qwen-code/qwen-code&lt;br&gt;
🔗 &lt;a href="https://github.com/QwenLM/qwen-code/releases/tag/v0.5.0" rel="noopener noreferrer"&gt;Check out the full changelog&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Children’s Storytelling: COOLKIDS LoRA from Clumsy_Trainer&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This Z-Image-Turbo LoRA captures the whimsy, warmth, and visual charm of children’s illustration — perfect for picture books, educational content, or animated shorts.&lt;br&gt;
The generations feel like pages from a beloved storybook.&lt;br&gt;
👉 &lt;a href="https://t.co/88TCy31PWq" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbrtb4osn9uwqguomubfa.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbrtb4osn9uwqguomubfa.png" alt=" " width="622" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;** Portrait Polisher: AWPortrait-Z from Shakker-Labs**&lt;br&gt;
AWPortrait-Z is a native noise-reduction LoRA that polishes Z-Image's portrait capabilities. From "relit" lighting to authentic skin texture, it is a massive quality-of-life upgrade for character generation.&lt;br&gt;
👉&lt;a href="https://huggingface.co/Shakker-Labs/AWPortrait-Z" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8v5kyjmt51ks19oezrzv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8v5kyjmt51ks19oezrzv.png" alt=" " width="800" height="480"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Z-Image Workflow Masterpiece from luneva&lt;/strong&gt;&lt;br&gt;
This Z-Image workflow generates pixel-level realistic details for both foregrounds and backgrounds at incredible speeds.&lt;br&gt;
No brute force, no upscaling needed—just pure, high-density realism. A must-try for the community.&lt;br&gt;
👉&lt;a href="https://t.co/Yr0ymKU2Ko" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fezm89lp8z1nduh91twk5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fezm89lp8z1nduh91twk5.png" alt=" " width="500" height="800"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  🔥 Upcoming Events
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;WAN MUSE+ Season 3 “IN CHARACTER” Now Live&lt;/strong&gt;&lt;br&gt;
We’re thrilled to launch WAN MUSE+ Season 3: “IN CHARACTER” — a global creative challenge inviting you to explore identity, narrative, and AI expression.&lt;br&gt;
Prize Pool: Up to $14,000&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Best Narrative / Best Animated Short Award / Best Visual / Best PSA Award &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Nomination &amp;amp; Special Inspiration Awards&lt;br&gt;
How to Enter:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Post on TikTok / IG / X / YouTube with hashtags: #incharacter  #wanmuse  #wan&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;AIGC Platforms: SeaArt.Ai, WaveSpeedAI, Tensor.Art&lt;br&gt;
🔗 &lt;a href="https://wan.video/activity/muse-enlist" rel="noopener noreferrer"&gt;Full details&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New model releases &amp;amp; upgrades&lt;/li&gt;
&lt;li&gt;AI research breakthroughs&lt;/li&gt;
&lt;li&gt;Open-source tools you can use today&lt;/li&gt;
&lt;li&gt;Community highlights that inspire&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now → &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>llm</category>
      <category>genai</category>
    </item>
    <item>
      <title>Dec 12, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 12 Dec 2025 05:59:28 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-12-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-375c</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-12-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-375c</guid>
      <description>&lt;p&gt;Hello, builders and researchers,&lt;br&gt;
This week was nothing short of extraordinary for Qwen — a true harvest of research milestones, product breakthroughs, and community-powered innovation.&lt;br&gt;
From multilingual TTS that sounds human to RL methods that train smarter, we’re witnessing the full arc of what open, thoughtful AI can become.&lt;/p&gt;

&lt;p&gt;Let’s dive in.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-Omni-Flash (2025-12-01): Smarter, More Human&lt;/strong&gt;&lt;br&gt;
What's improved:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Enhanced multi-turn video/audio understanding - conversations flow naturally&lt;/li&gt;
&lt;li&gt;Customize your AI's personality through system prompts (think roleplay scenarios!)&lt;/li&gt;
&lt;li&gt;Smarter language handling + rock-solid support: 119 text languages | 19 speech&lt;/li&gt;
&lt;li&gt;Voices indistinguishable from humans&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;🔗 Try it now:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://chat.qwen.ai" rel="noopener noreferrer"&gt;Qwen Chat&lt;/a&gt;: click the VoiceChat and VideoChat button (bottom-right)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qwen.ai/blog?id=qwen3-omni-20251201" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/spaces/Qwen/Qwen3-Omni-Demo" rel="noopener noreferrer"&gt;Demo&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelscope.cn/studios/Qwen/Qwen3-Omni-Demo" rel="noopener noreferrer"&gt;Demo&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelstudio.console.alibabacloud.com/?tab=doc#/doc/?type=model&amp;amp;url=2840914_2&amp;amp;modelId=qwen3-omni-flash-realtime-2025-12-01" rel="noopener noreferrer"&gt;Realtime API&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelstudio.console.alibabacloud.com/?tab=doc#/doc/?type=model&amp;amp;url=2840914_2&amp;amp;modelId=qwen3-omni-flash-2025-12-01" rel="noopener noreferrer"&gt;Offline API&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-TTS (version 2025-11-27): Voices That Feel Real&lt;/strong&gt;&lt;br&gt;
We've leveled up on what matters most: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;More Personalities: Over 49 high-quality voices, from cute and playful to wise and stern. Find your perfect match! &lt;/li&gt;
&lt;li&gt;Global Reach: Now speaks 10 languages (zh, en, de, it, pt, es, ja, ko, fr, ru) &amp;amp; authentic dialects (Minnan, Wu, Cantonese, Sichuan, Beijing, Nanjing, Tianjin, Shaanxi) &lt;/li&gt;
&lt;li&gt;Insanely Natural: The rhythm and speed adapt just like a real person. It's uncanny. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;🔗 Try it now:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://chat.qwen.ai" rel="noopener noreferrer"&gt;Qwen Chat&lt;/a&gt;: click Response → Read aloud&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qwen.ai/blog?id=qwen3-tts-1128" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelstudio.console.alibabacloud.com/?tab=doc#/doc/?type=model&amp;amp;url=2840914_2&amp;amp;modelId=qwen3-tts-flash-realtime-2025-11-27" rel="noopener noreferrer"&gt;Realtime API&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelstudio.console.alibabacloud.com/?tab=doc#/doc/?type=model&amp;amp;url=2840914_2&amp;amp;modelId=qwen3-tts-flash-2025-11-27" rel="noopener noreferrer"&gt;Offline API&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://huggingface.co/spaces/Qwen/Qwen3-TTS-Demo" rel="noopener noreferrer"&gt;Demo&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://modelscope.cn/studios/Qwen/Qwen3-TTS-Demo" rel="noopener noreferrer"&gt;Demo&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen Code v0.2.2 → v0.3.0: Stream JSON + Global Ready&lt;/strong&gt;&lt;br&gt;
Two breakthrough features:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Stream JSON Support&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;• &lt;code&gt;--output-format stream-json&lt;/code&gt; for streaming output&lt;/p&gt;

&lt;p&gt;• &lt;code&gt;--input-format stream-json&lt;/code&gt; for structured input&lt;/p&gt;

&lt;p&gt;• 3-tier adapter architecture + complete session management&lt;/p&gt;

&lt;p&gt;• Endless possibilities for SDK integration, automation tools, CI/CD pipelines!&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Full Internationalization&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;• Built-in EN/CN interface + custom language pack extensions&lt;/p&gt;

&lt;p&gt;• &lt;code&gt;/language ui zh-CN&lt;/code&gt; - One-click UI switching&lt;/p&gt;

&lt;p&gt;• &lt;code&gt;/language output Chinese&lt;/code&gt; - Set AI output language&lt;/p&gt;

&lt;p&gt;• Global developers welcome to contribute your local language packs! &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Security &amp;amp; Stability Leap Forward&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;🔗 &lt;a href="https://github.com/QwenLM/qwen-code" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Qwen Learn Mode — Your Personal AI Learning Tutor&lt;/strong&gt;&lt;br&gt;
In Qwen Learn Mode, Qwen Chat turns information into understanding that actually sticks.Powered by our Qwen3-Max model and grounded in cognitive psychology, it designs a learning path tailored to the way you think.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Guides you through Socratic-style dialogue, instead of just giving you answers&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Adapts to your current level, like a tutor who always works in your optimal learning zone&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Builds mental scaffolds so you can handle complex logic without feeling overwhelmed&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;✨ &lt;a href="https://chat.qwen.ai/?inputFeature=learn" rel="noopener noreferrer"&gt;Try Learn Mode&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  🧠 Research Breakthroughs
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing SAPO: A Smoother Path to RL Training&lt;/strong&gt;&lt;br&gt;
We introduce Soft Adaptive Policy Optimization (SAPO) — a smooth, stable, and highly effective RL method for training large language models.&lt;br&gt;
SAPO replaces hard boundaries with a continuous, temperature‑controlled gate that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Smooth trust‑region behavior → no abrupt gradient drop&lt;/li&gt;
&lt;li&gt;Sequence-level coherence → align sequence‑level behavior&lt;/li&gt;
&lt;li&gt;Token-level adaptivity → preserves useful gradients &amp;amp; boosts sample efficiency&lt;/li&gt;
&lt;li&gt;Asymmetric temperatures → significantly improved stability, esp. in MoE models
What does this mean in practice?&lt;/li&gt;
&lt;li&gt;Longer stable RL runs&lt;/li&gt;
&lt;li&gt;Higher Pass@1&lt;/li&gt;
&lt;li&gt;Stronger performance on Qwen3‑VL across math, coding &amp;amp; multimodal tasks&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;📄 &lt;a href="https://arxiv.org/abs/2511.20347" rel="noopener noreferrer"&gt;Paper on arXiv&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;📚 &lt;a href="https://qwen.ai/blog?id=sapo" rel="noopener noreferrer"&gt;Technical Blog&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Model Milestone: Z-Image-Turbo Ranks #1&lt;/strong&gt;&lt;br&gt;
According to ArtificialAnlys, Z-Image-Turbo now ranks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;#1 Open Source Model&lt;/li&gt;
&lt;li&gt;Top 10 Overall — the only open model on the list
With high-fidelity outputs, $5/1k pricing, and full open source, this is generative AI that’s accessible, affordable, and community-driven.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpkns9ei0fndvl8sus6n7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpkns9ei0fndvl8sus6n7.png" alt=" " width="800" height="262"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-4B: The #1 Base Model for Fine-Tuning&lt;/strong&gt;&lt;br&gt;
A rigorous benchmark on small language models by distil labs shows:&lt;br&gt;
Qwen3-4B emerges as the #1 base model for fine-tuning, matching or exceeding a 120B teacher model on 7 out of 8 tasks.&lt;br&gt;
If you need maximum accuracy with efficient compute, Qwen3-4B is your starting point.&lt;br&gt;
If you are looking for maximum accuracy with efficient compute, Qwen3-4B is your top choice.&lt;/p&gt;

&lt;p&gt;📄 &lt;a href="https://www.distillabs.ai/blog/we-benchmarked-12-small-language-models-across-8-tasks-to-find-the-best-base-model-for-fine-tuning" rel="noopener noreferrer"&gt;Read the full report&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;XiYan-SQL: #1 on All Open BIRD-CRITIC Leaderboards&lt;/strong&gt;&lt;br&gt;
XiYan-SQL is an innovative natural language–to–SQL conversion framework designed to address the performance challenges large language models face in SQL generation tasks.&lt;br&gt;
XiYan-SQL just hit #1 across all open BIRD-CRITIC (SWE-SQL) leaderboards, the real-world SQL diagnostic benchmark from academia + Google Cloud, built from actual database errors and tricky queries.&lt;br&gt;
Why XiYan-SQL matters &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Not just text → SQL: it diagnoses and fixes failing queries.&lt;/li&gt;
&lt;li&gt;Handles complex ops (INSERT / UPDATE / DELETE) across messy, multi-dialect DBs.&lt;/li&gt;
&lt;li&gt;Remains robust even on unseen, out-of-distribution databases.
What this means in practice:&lt;/li&gt;
&lt;li&gt;More reliable SQL debugging in real, production-style environments&lt;/li&gt;
&lt;li&gt;Stronger robustness for messy and evolving data stacks&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Community Celebration: WanMuse+ “Heartbeat” Winners Announced&lt;/strong&gt;&lt;br&gt;
The winners of WanMuse+ Season 2: “Heartbeat” have been revealed.To every creator who showed AI what it means to feel a heartbeat — we see you, we honor you, and we’re inspired by you.&lt;/p&gt;

&lt;p&gt;🎉 Congratulations to all finalists and winners!&lt;/p&gt;

&lt;p&gt;🔗 &lt;a href="https://x.com/Alibaba_Wan/status/1998623091740782632" rel="noopener noreferrer"&gt;Learn More&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Light Migration LoRA: Qwen-Edit-2509-Light-Migration from dx8152&lt;/strong&gt;&lt;br&gt;
Say goodbye to unnatural lighting artifacts.This Light Migration LoRA from dx8152 for Qwen-Image-Edit-2509 solves the “secondary lighting” headache by seamlessly transferring lighting conditions across scenes — preserving realism without hallucination.&lt;/p&gt;

&lt;p&gt;👉 &lt;a href="https://huggingface.co/dx8152/Qwen-Edit-2509-Light-Migration?spm=a2ty_o01.29997173.0.0.38145171tHdK9w" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyofufecc9ap4xoncl4d5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyofufecc9ap4xoncl4d5.png" alt=" " width="800" height="286"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Upscale LoRA: Qwen-Image-Edit-2509-Upscale2K from starsfriday&lt;/strong&gt;&lt;br&gt;
No more pixelated outputs.This Upscale LoRA from starsfriday losslessly magnifies your generations to ~2K/4K resolution while preserving fine details — perfect for turning rough concepts into production-ready visuals.&lt;/p&gt;

&lt;p&gt;🔗 &lt;a href="https://huggingface.co/starsfriday/Qwen-Image-Edit-2509-Upscale2K" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fddozf4sudlg92tai8g7u.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fddozf4sudlg92tai8g7u.png" alt=" " width="800" height="631"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;br&gt;
● New model releases &amp;amp; upgrades&lt;br&gt;
● AI research breakthroughs&lt;br&gt;
● Open-source tools you can use today&lt;br&gt;
● Community highlights that inspire&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>opensource</category>
      <category>coding</category>
    </item>
    <item>
      <title>Dec 5, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 05 Dec 2025 05:42:25 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-5-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-3ja7</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-5-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-3ja7</guid>
      <description>&lt;p&gt;Hello, builders and visionaries,&lt;br&gt;
This week, local AI got a major upgrade — and your workflows just got sharper, faster, and more expressive.&lt;br&gt;
Let’s dive in.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h3&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-Next Lands on llama.cpp&lt;/strong&gt;&lt;br&gt;
Big news for local AI enthusiasts: llama.cpp (PR #16095) just added support for Qwen3-Next — Qwen’s new hybrid architecture!&lt;br&gt;
You can now run Qwen3-Next locally with efficient CPU/GPU inference.&lt;br&gt;
🔗 &lt;a href="https://github.com/ggml-org/llama.cpp/pull/16095" rel="noopener noreferrer"&gt;View the PR&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Z-Image Milestone: 1 Million ComfyUI Workflow Downloads in One Week&lt;/strong&gt;&lt;br&gt;
Our Z-Image ComfyUI workflow just crossed 1,000,000 downloads in 7 days — a historic moment for open generative AI.&lt;br&gt;
Thank you for making Z-Image not just fast and small — but beloved.&lt;br&gt;
🔗 &lt;a href="https://huggingface.co/Comfy-Org/z_image_turbo" rel="noopener noreferrer"&gt;Learn more&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Z-Image-Turbo ControlNet Union Now Live&lt;/strong&gt;&lt;br&gt;
Tired of managing separate ControlNet files?Alibaba-PAI just dropped a Z-Image-Turbo ControlNet Union that handles Canny, Depth, Pose, HED, and MLSD in a single model.&lt;br&gt;
Trained on 1 million images at 1328px resolution, so it actually respects high-res details.&lt;br&gt;
🔗 &lt;a href="https://huggingface.co/alibaba-pai/Z-Image-Turbo-Fun-Controlnet-Union?spm=a2ty_o01.29997173.0.0.381451712MwKsR" rel="noopener noreferrer"&gt;Download on Hugging Face&lt;/a&gt;&lt;/p&gt;




&lt;h3&gt;
  
  
  ✨ Community Spotlights
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Concept Sliders for Z-Image: sliders-for-windows from sdbds (Qing Long)&lt;/strong&gt;&lt;br&gt;
Want more texture without changing your subject? sdbds drops this LoRA for Z-Image using Concept Sliders to dig deeper into the model.&lt;br&gt;
It enhances fine details and lighting rendering while keeping your original content 100% intact. No hallucinations, just high-definition polish. &lt;br&gt;
👉 &lt;a href="https://github.com/sdbds/sliders-for-windows/tree/qinglong?spm=a2ty_o01.29997173.0.0.381451712MwKsR" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F098hju9ziqm2ab8jnhgo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F098hju9ziqm2ab8jnhgo.png" alt=" " width="764" height="764"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Children's Drawings: z_image_turbo_childrens_drawings from ostris&lt;/strong&gt;&lt;br&gt;
ostris released this "Children's Drawings" model as part of a full training tutorial with AI Toolkit. Come for the crayon scribbles, stay for the dev knowledge. Perfect for when you need that "my 5-year-old drew this" energy in your generations.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/ostris/z_image_turbo_childrens_drawings" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcy7xovb4gexabg29fv22.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcy7xovb4gexabg29fv22.png" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Technically Color Z: Technically-Color-Z-Image-Turbo from renderartist&lt;/strong&gt;&lt;br&gt;
Technically Color Z by renderartist is meticulously crafted to capture the unmistakable essence of classic film. This LoRA greatly enhances the depth and brilliance of hues, creating more realistic yet dreamlike textures, lush greens, brilliant blues, and sometimes even the distinctive glow seen in classic productions, making your outputs look truly like they've stepped right off a silver screen.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/renderartist/Technically-Color-Z-Image-Turbo" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbnkni59msmwo1xjnlwp4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbnkni59msmwo1xjnlwp4.png" alt=" " width="800" height="1081"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Pixel Art: elusarca-pixel-art-style-lora-zimage-turbo from reverentelusarca&lt;/strong&gt;&lt;br&gt;
If you’re generating retro assets, you need this in your workflow. The Pixel Art LoRA for Z-Image-Turbo by reverentelusarca cleans up the noise and forces a much crispier grid than the default output. &lt;br&gt;
A must-have for that 16-bit aesthetic. &lt;br&gt;
👉 &lt;a href="https://huggingface.co/reverentelusarca/elusarca-pixel-art-style-lora-zimage-turbo?spm=a2ty_o01.29997173.0.0.381451712MwKsR" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F87ercl9dvzpbsfyces9e.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F87ercl9dvzpbsfyces9e.png" alt=" " width="800" height="568"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Material Transfer: Qwen-Edit-2509-Material-transfer from oumoumad&lt;/strong&gt;&lt;br&gt;
oumoumad dropped a Material Transfer LoRA for Qwen-Image-Edit-2509. You can take a plain render (like a car seat or cabin), feed it a material board or shaderball, and it applies the texture instantly.&lt;br&gt;
CMF workflow just got a whole lot faster.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/oumoumad/Qwen-Edit-2509-Material-transfer" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe3wpbeo2fg33ku1zubta.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe3wpbeo2fg33ku1zubta.png" alt=" " width="800" height="375"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;br&gt;
● New model releases &amp;amp; upgrades&lt;br&gt;
● AI research breakthroughs&lt;br&gt;
● Open-source tools you can use today&lt;br&gt;
● Community highlights that inspire&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>devdiscuss</category>
      <category>programming</category>
    </item>
    <item>
      <title>Dec 5, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 05 Dec 2025 05:42:25 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-5-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-5eo1</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/dec-5-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-5eo1</guid>
      <description>&lt;p&gt;Hello, builders and visionaries,&lt;br&gt;
This week, local AI got a major upgrade — and your workflows just got sharper, faster, and more expressive.&lt;br&gt;
Let’s dive in.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-Next Lands on llama.cpp&lt;/strong&gt;&lt;br&gt;
Big news for local AI enthusiasts: llama.cpp (PR #16095) just added support for Qwen3-Next — Qwen’s new hybrid architecture!&lt;br&gt;
You can now run Qwen3-Next locally with efficient CPU/GPU inference.&lt;br&gt;
🔗 &lt;a href="https://github.com/ggml-org/llama.cpp/pull/16095" rel="noopener noreferrer"&gt;View the PR&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Z-Image Milestone: 1 Million ComfyUI Workflow Downloads in One Week&lt;/strong&gt;&lt;br&gt;
Our Z-Image ComfyUI workflow just crossed 1,000,000 downloads in 7 days — a historic moment for open generative AI.&lt;br&gt;
Thank you for making Z-Image not just fast and small — but beloved.&lt;br&gt;
🔗 &lt;a href="https://huggingface.co/Comfy-Org/z_image_turbo" rel="noopener noreferrer"&gt;Learn more&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Z-Image-Turbo ControlNet Union Now Live&lt;/strong&gt;&lt;br&gt;
Tired of managing separate ControlNet files?Alibaba-PAI just dropped a Z-Image-Turbo ControlNet Union that handles Canny, Depth, Pose, HED, and MLSD in a single model.&lt;br&gt;
Trained on 1 million images at 1328px resolution, so it actually respects high-res details.&lt;br&gt;
🔗 &lt;a href="https://huggingface.co/alibaba-pai/Z-Image-Turbo-Fun-Controlnet-Union?spm=a2ty_o01.29997173.0.0.381451712MwKsR" rel="noopener noreferrer"&gt;Download on Hugging Face&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  ✨ Community Spotlights
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Concept Sliders for Z-Image: sliders-for-windows from sdbds (Qing Long)&lt;/strong&gt;&lt;br&gt;
Want more texture without changing your subject? sdbds drops this LoRA for Z-Image using Concept Sliders to dig deeper into the model.&lt;br&gt;
It enhances fine details and lighting rendering while keeping your original content 100% intact. No hallucinations, just high-definition polish. &lt;br&gt;
👉 &lt;a href="https://github.com/sdbds/sliders-for-windows/tree/qinglong?spm=a2ty_o01.29997173.0.0.381451712MwKsR" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F098hju9ziqm2ab8jnhgo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F098hju9ziqm2ab8jnhgo.png" alt=" " width="764" height="764"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgsa3xyqqj9x9t4offfwj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgsa3xyqqj9x9t4offfwj.png" alt=" " width="764" height="764"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Children's Drawings: z_image_turbo_childrens_drawings from ostris&lt;/strong&gt;&lt;br&gt;
ostris released this "Children's Drawings" model as part of a full training tutorial with AI Toolkit. Come for the crayon scribbles, stay for the dev knowledge. Perfect for when you need that "my 5-year-old drew this" energy in your generations.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/ostris/z_image_turbo_childrens_drawings" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcy7xovb4gexabg29fv22.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcy7xovb4gexabg29fv22.png" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Technically Color Z: Technically-Color-Z-Image-Turbo from renderartist&lt;/strong&gt;&lt;br&gt;
Technically Color Z by renderartist is meticulously crafted to capture the unmistakable essence of classic film. This LoRA greatly enhances the depth and brilliance of hues, creating more realistic yet dreamlike textures, lush greens, brilliant blues, and sometimes even the distinctive glow seen in classic productions, making your outputs look truly like they've stepped right off a silver screen.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/renderartist/Technically-Color-Z-Image-Turbo" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Pixel Art: elusarca-pixel-art-style-lora-zimage-turbo from reverentelusarca&lt;/strong&gt;&lt;br&gt;
If you’re generating retro assets, you need this in your workflow. The Pixel Art LoRA for Z-Image-Turbo by reverentelusarca cleans up the noise and forces a much crispier grid than the default output. &lt;br&gt;
A must-have for that 16-bit aesthetic. &lt;br&gt;
👉 &lt;a href="https://huggingface.co/reverentelusarca/elusarca-pixel-art-style-lora-zimage-turbo?spm=a2ty_o01.29997173.0.0.381451712MwKsR" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Material Transfer: Qwen-Edit-2509-Material-transfer from oumoumad&lt;/strong&gt;&lt;br&gt;
oumoumad dropped a Material Transfer LoRA for Qwen-Image-Edit-2509. You can take a plain render (like a car seat or cabin), feed it a material board or shaderball, and it applies the texture instantly.&lt;br&gt;
CMF workflow just got a whole lot faster.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/oumoumad/Qwen-Edit-2509-Material-transfer" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/..." alt="Uploading image" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;📬 Want More? Stay Updated.&lt;br&gt;
Every week, we bring you:&lt;br&gt;
● New model releases &amp;amp; upgrades&lt;br&gt;
● AI research breakthroughs&lt;br&gt;
● Open-source tools you can use today&lt;br&gt;
● Community highlights that inspire&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>devdiscuss</category>
      <category>programming</category>
    </item>
    <item>
      <title>Nov28, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 28 Nov 2025 07:49:16 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/nov28-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-kbb</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/nov28-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-kbb</guid>
      <description>&lt;p&gt;Hello, community,&lt;/p&gt;

&lt;p&gt;This week, research and community converged in perfect harmony.&lt;br&gt;
On the global stage, our work on Gated Attention was honored with the NeurIPS 2025 Best Paper Award. And right here, in the open, we launched Z-Image: an open-source, 6-billion-parameter model that delivers top-tier image generation for everyone, everywhere.&lt;/p&gt;

&lt;p&gt;But as always, the real magic came from you.&lt;/p&gt;

&lt;p&gt;This week reminded us of a simple truth: Great AI isn’t built in isolation — it’s co-created.&lt;/p&gt;

&lt;p&gt;You read our papers.You fine-tune our models.You build tools we never imagined.And you push us to be better.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing Z-Image: A High Performance, Open, and Accessible Image Generation Model&lt;/strong&gt;&lt;br&gt;
We are pleased to introduce Z-Image, an efficient 6-billion-parameter foundation model for image generation.&lt;br&gt;
Through systematic optimization, it proves that top-tier performance is achievable without relying on enormous model sizes, delivering strong results in photorealistic generation and bilingual text rendering that are comparable to leading commercial models.&lt;br&gt;
We are publicly releasing two specialized models on Z-Image: Z-Image-Turbo for generation and Z-Image-Edit for editing. The model code, weights, and an online demo are now publicly available to encourage community exploration and use. With this release, we aim to promote the development of generative models that are accessible, low-cost, and high-performance.&lt;br&gt;
📄 &lt;a href="https://tongyi-mai.github.io/Z-Image-blog/" rel="noopener noreferrer"&gt;Blog&lt;/a&gt;&lt;br&gt;
📌 &lt;a href="https://github.com/Tongyi-MAI/Z-Image" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;&lt;br&gt;
📌 &lt;a href="https://modelscope.ai/models/Tongyi-MAI/Z-Image-Turbo/summary" rel="noopener noreferrer"&gt;ModelScope&lt;/a&gt;&lt;br&gt;
📌 &lt;a href="https://huggingface.co/Tongyi-MAI/Z-Image-Turbo" rel="noopener noreferrer"&gt;HuggingFace&lt;/a&gt;&lt;br&gt;
📌 &lt;a href="https://modelscope.cn/studios/Tongyi-MAI/Z-Image-Gallery" rel="noopener noreferrer"&gt;Z-Image gallery&lt;/a&gt;&lt;br&gt;
&lt;em&gt;P.S. Z-Image Turbo is already #1 on Hugging Face’s trending models and Spaces. Thank you, community — you’re moving faster than we are!&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwpmqrhrmgh4brkzcwuj8.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwpmqrhrmgh4brkzcwuj8.jpeg" alt=" " width="800" height="410"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa5dlpq45anc44vsujsfk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa5dlpq45anc44vsujsfk.png" alt=" " width="800" height="409"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📚 Research Breakthroughs
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;NeurIPS 2025 Best Paper Award&lt;/strong&gt;&lt;br&gt;
We are deeply honored to announce that our paper“&lt;strong&gt;Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free&lt;/strong&gt;” has been awarded the NeurIPS 2025 Best Paper Award.&lt;br&gt;
Reflections from the Selection Committee: &lt;em&gt;This paper represents a substantial amount of work that is possible only with access to industrial scale computing resources, and the authors’ sharing of the results of their work, which will advance the community’s understanding of attention in large language models, is highly commendable, especially in an environment where there has been a move away from open sharing of scientific results around LLMs.&lt;/em&gt;&lt;br&gt;
📖 &lt;a href="https://blog.neurips.cc/2025/11/26/announcing-the-neurips-2025-best-paper-awards/" rel="noopener noreferrer"&gt;Read the announcement&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-VL Technical Report Now on arXiv&lt;/strong&gt;&lt;br&gt;
The full story behind Qwen3-VL is now out on arXiv&lt;br&gt;
From pretraining to post-training, architecture to infra, data to evaluation, we’ve packed in the details for anyone building on vision-language models.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;3 models &amp;gt;1M downloads in just over a month&lt;/li&gt;
&lt;li&gt;Qwen3-VL-8B leads with 2M+ downloads&lt;/li&gt;
&lt;li&gt;Built on the shoulders of Qwen2.5-VL (2800+ citations in &amp;lt;10 months!)
Whether you’re fine-tuning, deploying, or researching VLMs — this is your playbook.
📄 &lt;a href="https://arxiv.org/pdf/2511.21631" rel="noopener noreferrer"&gt;Read the full paper on arXiv&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Turn Portraits Into Cartoons: Qwen-Image-Edit-2509-Caricature-LoRA from drbaph&lt;/strong&gt;&lt;br&gt;
This LoRA from drbaph transforms input images into sketched caricature art with exaggerated features. It's an image-to-image model that takes your photo as input and creates humorous, artistic caricature representations of people and animal subjects with emphasized facial features and characteristics.&lt;br&gt;
👉 &lt;a href="https://t.co/jppiz399fs?spm=a2ty_o01.29997173.0.0.38145171BQ38VC" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn1nq7oxupkwypbufk35u.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn1nq7oxupkwypbufk35u.gif" alt=" " width="508" height="764"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Light Restoration V2: Qwen-Image-Edit-2509-Light_restoration from dx8152&lt;/strong&gt;&lt;br&gt;
dx8152 is moving at lightning speed! The V2 update of their Light Restoration LoRA now lets you scrub lighting from any reference image to build better training pairs.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/dx8152/Qwen-Image-Edit-2509-Light_restoration" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhj5ij8myoan8ei37vid9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhj5ij8myoan8ei37vid9.png" alt=" " width="800" height="206"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Day/Night Shift: Qwen-Edit-Loras from lividtm&lt;/strong&gt;&lt;br&gt;
Need a clean Day/Night shift? lividtm has you covered. This LoRA for Qwen-Image-Edit-2509 handles 2K resolution while keeping scene details locked. Simple trigger words, high fidelity.&lt;br&gt;
 👉 &lt;a href="https://huggingface.co/lividtm/Qwen-Edit-Loras" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnn528pm5u4vj0ave60l2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnn528pm5u4vj0ave60l2.png" alt=" " width="800" height="216"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;Every week, we bring you:&lt;br&gt;
● New model releases &amp;amp; upgrades&lt;br&gt;
● AI research breakthroughs&lt;br&gt;
● Open-source tools you can use today&lt;br&gt;
● Community highlights that inspire&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release.&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>news</category>
      <category>machinelearning</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Nov21, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 21 Nov 2025 07:33:17 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/nov21-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-3kop</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/nov21-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-3kop</guid>
      <description>&lt;p&gt;Hello, creators, engineers, and visionaries,&lt;/p&gt;

&lt;p&gt;Before we dive in this week, we have a milestone to share — and it belongs to you.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;10 million users are now creating with Qwen Chat!&lt;/strong&gt; Not just asking questions, but writing code, designing images, uncovering insights, and bringing invisible visions to life.&lt;/p&gt;

&lt;p&gt;This week wasn’t just about releases. It was about awakening new possibilities.&lt;/p&gt;

&lt;p&gt;From an agent system that evolves itself, to video models climbing the global leaderboards — we’re witnessing AI innovation and creativity, powered by your ingenuity.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
Subscribe Now：&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing AgentEvolver: An Open-Source Self-Evolving Agent System&lt;/strong&gt;&lt;br&gt;
We’re thrilled to open-source AgentEvolver —an end-to-end, self-evolving training framework that unifies self-questioning, self-navigating, and self-attributing into a cohesive system. It empowers agents to autonomously improve their capabilities, aiming for efficient, cost-effective, and continuous capability evolution.&lt;br&gt;
AgentEvolver provides three Self-Evolving Mechanisms from Environment to Policy:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Automatic Task Generation (Self-Questioning) – Explore the environment and autonomously create diverse tasks, eliminating costly manual dataset construction.&lt;/li&gt;
&lt;li&gt;Experience-guided Exploration (Self-Navigating) – Summarize and reuse cross-task experience, guiding higher-quality rollouts and improving exploration efficiency.&lt;/li&gt;
&lt;li&gt;Attribution-based Credit Assignment (Self-Attributing) – Process long trajectories to uncover the causal contribution of intermediate steps, enabling fine-grained and efficient policy optimization.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Built on a service-oriented dataflow architecture, AgentEvolver seamlessly integrates environment sandboxes, LLMs, and experience management into modular services. &lt;br&gt;
AgentEvolver achieves superior results while using substantially fewer parameters than larger baseline models, according to the AppWorld and BFCL-v3 benchmarks.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://github.com/modelscope/AgentEvolver" rel="noopener noreferrer"&gt;GitHub Repository&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2511.10395" rel="noopener noreferrer"&gt;Technical Report&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Qwen Code v0.2.1 Released: Smarter, Faster, Cleaner&lt;/strong&gt;&lt;br&gt;
We shipped 8 versions (v0.1.0-&amp;gt;v0.2.1) in 17 days and here's the new leap: &lt;br&gt;
Free Web Search: Support for multiple providers. Qwen OAuth users get 2000 free searches per day!&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Smarter Code Editing: New fuzzy matching pipeline reduces errors and saves tokens—fewer retries needed.&lt;/li&gt;
&lt;li&gt;More Control: Fine-tune AI behavior with temperature, top_p, and max tokens settings.&lt;/li&gt;
&lt;li&gt;Better IDE Integration: Enhanced Zed IDE support with todo and task management tools.&lt;/li&gt;
&lt;li&gt;Cleaner Output: Tool responses now use plain text instead of complex JSON—easier for AI to understand.&lt;/li&gt;
&lt;li&gt;Improved Search: Better file filtering (respects &lt;code&gt;.gitignore&lt;/code&gt;), smarter search tools, and standardized naming.&lt;/li&gt;
&lt;li&gt;Faster Performance: Multi-stage normalization pipeline for zero-overhead matching, better Unicode handling, and optimized output limits.&lt;/li&gt;
&lt;li&gt;Bug Fixes: Fixed token limits for multiple models, improved cross-platform support (macOS &amp;amp; Windows), and better stability.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Try it now:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://github.com/QwenLM/qwen-code" rel="noopener noreferrer"&gt;https://github.com/QwenLM/qwen-code&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://github.com/QwenLM/qwen-code/releases" rel="noopener noreferrer"&gt;https://github.com/QwenLM/qwen-code/releases&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Model Milestone: Wan2.5-Preview landed in the Top 5 on LMArena leaderboards&lt;/strong&gt;&lt;br&gt;
This week, we've seen a new milestone of Wan 2.5-Preview with 2 models — i2v and t2i — landed in the Top 5 on the Image-to-Video and Text-to-Image LMArena leaderboards.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Wan2.5-i2v-preview → #3 on Image-to-Video Leaderboard&lt;/li&gt;
&lt;li&gt;Wan2.5-t2i-preview → #5 on Text-to-Image Leaderboard&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://wan.video/" rel="noopener noreferrer"&gt;Try it now&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wan Powers ElevenLabs’ New Image &amp;amp; Video Platform&lt;/strong&gt;&lt;br&gt;
We’re proud to see Wan among the leading models powering ElevenLabs’ new creative platform — ElevenLabs Image &amp;amp; Video (Beta).&lt;br&gt;
&lt;a href="https://elevenlabs.io/image-video" rel="noopener noreferrer"&gt;Try it on ElevenLabs&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;SGLang Diffusion Joins the Ecosystem — With Wan &amp;amp; Qwen Support!&lt;/strong&gt;&lt;br&gt;
SGLang Diffusion brings SGLang’s state-of-the-art performance to image &amp;amp; video generation. And yes — it now supports Wan, Qwen-Image, and Qwen-Image-Edit, and other major open-source video and image generation models.&lt;br&gt;
We love seeing this kind of ecosystem synergy — this is how AI grows.&lt;br&gt;
&lt;a href="https://t.co/lpETRCgjEx" rel="noopener noreferrer"&gt;SGLang Diffusion&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Multi-Angle Relighting LoRA: Qwen-Edit-2509-Multi-Angle-Lighting from dx8152&lt;/strong&gt;&lt;br&gt;
Introducing Qwen-Edit-2509-Multi-Angle-Lighting from dx8152, a LoRA that lets you paint with light.&lt;br&gt;
The idea is simple: use a control map + text prompt to change the lighting. It's still in the early stages (V1), but the potential here is huge. &lt;br&gt;
&lt;a href="https://huggingface.co/dx8152/Qwen-Edit-2509-Multi-Angle-Lighting" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Manga Coloring LoRA: PanelPainter V2&lt;/strong&gt;&lt;br&gt;
"PanelPainter V2" just dropped, and it's a total glow-up. It's not just a helper anymore; this LoRA is trained to handle the coloring on its own. It's not perfect (consistency is still tricky ), but it's a massive step in the right direction.&lt;br&gt;
&lt;a href="https://t.co/eMr9mH3yKL" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Nunchaku-quantized versions of Qwen-Image-Edit-2509: nunchaku-qwen-image-edit-2509 from nunchaku-tech&lt;/strong&gt;&lt;br&gt;
nunchaku-tech dropped quantized versions of the 2509 model, and the big news is the pre-fused Lightning models. We're talking 4-step and 8-step edits.&lt;br&gt;
This is a must-grab for anyone who wants high-speed, low-VRAM image editing.&lt;br&gt;
&lt;a href="https://huggingface.co/nunchaku-tech/nunchaku-qwen-image-edit-2509" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Realistic Photography LoRA: boreal-qwen-image from kudzueye&lt;/strong&gt;&lt;br&gt;
This LoRA from kudzueye is an experimental LoRA designed for realistic photography. &lt;br&gt;
There's a ComfyUI workflow included to get you started.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://huggingface.co/kudzueye/boreal-qwen-image" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Preserving Subjects While Editing Images: Qwen-Image-Edit-InSubject from peteromallet&lt;/strong&gt;&lt;br&gt;
This LoRA from peteromallet is fine-tune for QwenEdit and significantly improves its ability to preserve subjects while making edits to images. It works effectively with both single subjects and multiple subjects in the same image. &lt;br&gt;
&lt;a href="https://huggingface.co/peteromallet/Qwen-Image-Edit-InSubject" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Book Flatten and Crop LoRA: book_flatten_and_crop_qwen_image_edit_2509 from tarn59&lt;/strong&gt;&lt;br&gt;
Need to fix those split-page book scans?&lt;br&gt;
Tarn59 just solved that with a new LoRA for Qwen-Image-Edit-2509. It flattens the page, crops the image, and magically removes the middle crease. Works best if you play around with the aspect ratio to match your book.&lt;br&gt;
&lt;a href="https://huggingface.co/tarn59/book_flatten_and_crop_qwen_image_edit_2509" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;FLAT/LOG Style Images: QwenEdit2509-FlatLogColor from tlennon-ie&lt;/strong&gt;&lt;br&gt;
AI images usually come "pre-cooked" with too much contrast, which is a nightmare for color grading. &lt;/p&gt;

&lt;p&gt;tlennon-ie created a brilliant fix with Qwen-Image-Edit-2509. It converts generations into a flat, LOG-style profile—basically a digital negative that preserves shadow and highlight details.&lt;br&gt;
Perfect if you need to match AI assets with professional video footage.&lt;br&gt;
&lt;a href="https://huggingface.co/spaces/tlennon-ie/QwenEdit2509-FlatLogColor" rel="noopener noreferrer"&gt;Try it here&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  🔥 Upcoming Events
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Meet Qwen in Seoul (Dec10): AMD’s AI Developer Meetup&lt;/strong&gt;&lt;br&gt;
AMD’s AI Developer Meetup in Seoul (Dec 10) is filling FAST.As a key partner, we’re bringing you the future of generative AI — live, hands-on, and free.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Dec 10 | 📍 Seoul, Aloft Gangnam&lt;/li&gt;
&lt;li&gt;Free limited-edition swag for all attendees&lt;/li&gt;
&lt;li&gt;Register now — spots are limited: &lt;a href="https://luma.com/0yxjboie" rel="noopener noreferrer"&gt;https://luma.com/0yxjboie&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;What You’ll Experience:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Qwen-Image Technology Deep Dive&lt;/li&gt;
&lt;li&gt;Korean Enterprise AI &amp;amp; Cloud Case Studies&lt;/li&gt;
&lt;li&gt;🎨 Hands-On Workshop: Qwen-Image × LoRA&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;→ Fine-tune your own LoRA with Qwen-Image&lt;/p&gt;

&lt;p&gt;→ Train &amp;amp; infer using DiffSynth-Studio on AMD MI300x GPUs&lt;/p&gt;

&lt;p&gt;→ Build custom visual models — from zero to masterpiece&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wan Muse “Heartbeat”Creative Challenge — The Shortlist Is Here&lt;/strong&gt;&lt;br&gt;
The Professional Category Shortlist for Wan Muse Season 2: “Heartbeat” is now live.&lt;br&gt;
📌 Public Review Period: November 18–21, 2025&lt;br&gt;
👉 View All Shortlisted Works&lt;br&gt;
🔍 Found an issue? We take fairness seriously. Report violations (real-name required):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Not AI-generated by Wan&lt;/li&gt;
&lt;li&gt;Plagiarism or copyright breach&lt;/li&gt;
&lt;li&gt;Content policy violation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;📩 Email: &lt;a href="mailto:tongyiwanxiang@service.aliyun.com"&gt;tongyiwanxiang@service.aliyun.com&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;📬 Want More? Stay Updated.&lt;br&gt;
Every week, we bring you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New model releases &amp;amp; upgrades&lt;/li&gt;
&lt;li&gt;AI research breakthroughs&lt;/li&gt;
&lt;li&gt;Open-source tools you can use today&lt;/li&gt;
&lt;li&gt;Community highlights that inspire&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release: &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>programming</category>
      <category>llm</category>
    </item>
    <item>
      <title>Nov14, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 14 Nov 2025 08:45:49 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/nov14-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-ca1</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/nov14-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-ca1</guid>
      <description>&lt;p&gt;Hello, creators and innovators,&lt;/p&gt;

&lt;p&gt;This week, as we share the latest from our lab, the real magic happened beyond it - in the hands of developers, artists, and builders around the world.&lt;/p&gt;

&lt;p&gt;From stunning image edits with Qwen-Image, to expressive, style-shifting generations with Wan, the open-source community is turning our foundational models into something extraordinary.&lt;/p&gt;

&lt;p&gt;These aren't just LoRAs. They're personal expressions of what's possible when open models meet bold imagination.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;br&gt;
&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;Subscribe Now → &lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing Qwen DeepResearch 2511, a major upgrade to our DeepResearch model&lt;/strong&gt;&lt;br&gt;
Qwen DeepResearch 2511 is now live and ready to transform how you explore, analyze, and synthesize knowledge.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Dual Mode: Choose Normal for speed, or flip to Advanced and let the AI dive deep - spending extra time for a more thorough analysis.&lt;/li&gt;
&lt;li&gt;File Uploads Enabled: Now you can easily upload your documents or images for the AI to analyze!&lt;/li&gt;
&lt;li&gt;Boosted Search Power: Drastically improved search efficiency &amp;amp; depth. It reads more, understands deeper, and delivers better answers - in less time.&lt;/li&gt;
&lt;li&gt;Precise Report Control: Command the report format - word count, paragraphs, &amp;amp; content! Get comprehensive reports with enhanced citation reliability.&lt;/li&gt;
&lt;li&gt;&lt;p&gt;All-New UX: Our new decoupled architecture delivers a smoother, more responsive user experience.&lt;br&gt;
👉 Try it now:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://chat.qwen.ai/?inputFeature=deep_research" rel="noopener noreferrer"&gt;Qwen Chat&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://qwen.ai/download" rel="noopener noreferrer"&gt;QwenChat APP&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Light Restoration (built on Qwen-Image-Edit-2509): Qwen-Image-Edit-2509-Light_restorationfrom dx8152&lt;/strong&gt;&lt;br&gt;
This LoRA from &lt;a href="https://huggingface.co/dx8152" rel="noopener noreferrer"&gt;dx8152&lt;/a&gt; removes unwanted shadows and fixes exposure with astonishing naturalness - perfect for photographers and digital artists.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/dx8152/Qwen-Image-Edit-2509-Light_restoration" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;br&gt;
🎥 &lt;a href="https://x.com/dx8152/status/1986960553252991463/video/1?s=20" rel="noopener noreferrer"&gt;Video demo&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frj1vltz10ic8ypv9x61g.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frj1vltz10ic8ypv9x61g.png" alt=" " width="800" height="308"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Photo Upscale (built on Qwen-Image-Edit-2509): Qwen-Edit-2509-Upscale-LoRA from vafipas663&lt;/strong&gt;&lt;br&gt;
Need to rescue those blurry, low-res, or noisy old photos? This Upscale-LoRA from &lt;a href="https://huggingface.co/vafipas663" rel="noopener noreferrer"&gt;vafipas663&lt;/a&gt; does an amazing job at enhancing photography, fixing noise, and destroying JPEG artifacts.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/vafipas663/Qwen-Edit-2509-Upscale-LoRA" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh94r7q8nq7rpww0qqrc2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fh94r7q8nq7rpww0qqrc2.png" alt=" " width="695" height="641"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Skin Realism (built on Qwen-Image-Edit-2509): qwen-edit-skin from tlennon-ie&lt;/strong&gt;&lt;br&gt;
Want your portraits to have a natural skin texture? This LoRA from &lt;a href="https://huggingface.co/tlennon-ie" rel="noopener noreferrer"&gt;tlennon-ie&lt;/a&gt; specifically addresses the nuances of human skin, adding detail and realism that may not be present in the original generations.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/tlennon-ie/qwen-edit-skin" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0vzrk1v7n3pdf8ge07sb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0vzrk1v7n3pdf8ge07sb.png" alt=" " width="800" height="292"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Photo-to-Anime (built on Qwen-Image-Edit-2509): Qwen-Image-Edit-2509-Photo-to-Animefrom autoweeb&lt;/strong&gt;&lt;br&gt;
This LoRA from &lt;a href="https://huggingface.co/autoweeb" rel="noopener noreferrer"&gt;autoweeb&lt;/a&gt; turns any photo into a stunning anime image. Just try a simple prompt like "transform into anime" and watch the magic happen.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/autoweeb/Qwen-Image-Edit-2509-Photo-to-Anime" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8x0zqqj5tobmu244ea1j.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8x0zqqj5tobmu244ea1j.png" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Generate the Next Scene (built on Qwen-Image-Edit-2509): next-scene-qwen-image-lora-2509 from lovis93&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://huggingface.co/lovis93" rel="noopener noreferrer"&gt;lovis93&lt;/a&gt;'s "Next Scene" LoRA understands camera motion - dolly shots, push-ins, pans - and generates seamless transitions between frames. V2 is even better.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/lovis93/next-scene-qwen-image-lora-2509" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9zigfu4upsjt2j2n412g.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9zigfu4upsjt2j2n412g.gif" alt=" " width="500" height="281"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3D Chibi (built on Qwen-Image-Edit-2509): Qwen-Edit-3DChibi-LoRA from rsshekhawat&lt;/strong&gt;&lt;br&gt;
Ready to turn your entire photo library into an adorable 3D Chibi world? &lt;a href="https://huggingface.co/rsshekhawat" rel="noopener noreferrer"&gt;rsshekhawat&lt;/a&gt;'s LoRA creates high-quality, highly detailed 3D Chibi Style images.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/rsshekhawat/Qwen-Edit-3DChibi-LoRA" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Firfcesa1iwz7vt1bsze8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Firfcesa1iwz7vt1bsze8.png" alt=" " width="800" height="397"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Background White to Scene (built on Qwen-Image-Edit-2509): Qwen-Image-Edit-2509-White_to_Scene from dx8152&lt;/strong&gt;&lt;br&gt;
Tired of product shots stuck on white? This LoRA from &lt;a href="https://huggingface.co/dx8152" rel="noopener noreferrer"&gt;dx8152&lt;/a&gt; is an Image Fusion tool that lets you take any object and seamlessly place it into a brand new scene using trigger words like "change white background to scene".&lt;br&gt;
👉 &lt;a href="https://huggingface.co/dx8152/Qwen-Image-Edit-2509-White_to_Scene" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Farr7dq8xmu5ffpvrbz0n.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Farr7dq8xmu5ffpvrbz0n.gif" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;In-Scene Image Editing (built on Qwen-Image-Edit): qwen-image-edit-inscene-lora from flymy-ai&lt;/strong&gt;&lt;br&gt;
Want to change the action in your photo without breaking the scene? &lt;a href="https://huggingface.co/flymy-ai" rel="noopener noreferrer"&gt;flymy-ai&lt;/a&gt;'s InScene LoRA is specialized to understand complex in-scene commands, maintain coherence, and handle object positioning like a pro.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/flymy-ai/qwen-image-edit-inscene-lora" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo0am70hexboi2qltg1ga.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo0am70hexboi2qltg1ga.png" alt=" " width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Apply Texture (built on Qwen-Image-Edit-2509): apply_texture_qwen_image_edit_2509 from tarn59&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://huggingface.co/tarn59" rel="noopener noreferrer"&gt;tarn59&lt;/a&gt;'s wild "Apply Texture" LoRA lets you apply any texture to any object. Just use Apply ... texture to ... to trigger the image generation!&lt;br&gt;
👉 &lt;a href="https://huggingface.co/tarn59/apply_texture_qwen_image_edit_2509" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6u1h5vcgxrpuop6jtfja.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6u1h5vcgxrpuop6jtfja.png" alt=" " width="800" height="277"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Animal to Human Avatar (built on Qwen-Image-Edit-2509): Qwen-Edit-2509-Anishift-LoRA from hiru13do37&lt;/strong&gt;&lt;br&gt;
Ever wondered what your pet would look like as a cool human? &lt;a href="https://huggingface.co/hiru13do37" rel="noopener noreferrer"&gt;hiru13do37&lt;/a&gt;'s Anishift LoRA transforms animals into hyper-stylized human avatars - with personality.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/hiru13do37/Qwen-Edit-2509-Anishift-LoRA" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8kam79yu71sjv1d6phk3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8kam79yu71sjv1d6phk3.png" alt=" " width="800" height="850"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;De-Cartoon Everything (built on Qwen-Image-Edit-2509): QwenEdit-Anything2Real_Alpha from lrzjason&lt;/strong&gt;&lt;br&gt;
This LoRA from &lt;a href="https://huggingface.co/lrzjason" rel="noopener noreferrer"&gt;lrzjason&lt;/a&gt; is designed to convert illustrations, anime, cartoons, paintings, and other non-photorealistic images into convincing photographs while preserving the original composition and content.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/lrzjason/QwenEdit-Anything2Real_Alpha" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;br&gt;
&lt;a href="https://huggingface.co/lrzjason/QwenEdit-Anything2Real_Alpha" rel="noopener noreferrer"&gt;https://huggingface.co/lrzjason/QwenEdit-Anything2Real_Alpha&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fast Image Editing (built on Qwen-Image-Edit): eigen-banana-qwen-image-edit from eigen-ai-labs&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://huggingface.co/eigen-ai-labs" rel="noopener noreferrer"&gt;eigen-ai-labs&lt;/a&gt; dropped Eigen-Banana - optimized for fast, high-quality image editing with text prompts. This model enables efficient text-guided image transformations with reduced inference steps while maintaining excellent quality. (Non-commercial use only.)&lt;br&gt;
👉 &lt;a href="https://huggingface.co/eigen-ai-labs/eigen-banana-qwen-image-edit" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F59l86nuuggj83fnw7lt4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F59l86nuuggj83fnw7lt4.png" alt=" " width="800" height="265"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Push-In Camera (built on Wan2.1): Motion-Lora-Camera-Push-In-Wan-14B-720p-I2V from lovis93&lt;/strong&gt;&lt;br&gt;
This LoRA from &lt;a href="https://huggingface.co/lovis93" rel="noopener noreferrer"&gt;lovis93&lt;/a&gt; was trained on 100 clips to introduce realistic, high-quality push-in drone camera motion into your generations, enhancing your creations by delivering natural camera dynamics across various styles and scenes.&lt;br&gt;
👉 &lt;a href="https://huggingface.co/lovis93/Motion-Lora-Camera-Push-In-Wan-14B-720p-I2V" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;br&gt;
&lt;strong&gt;Slow-Motion Enhancement (built on Wan2.2): ComfyUI-PainterI2V from 绘画小子(Douyin creator)&lt;/strong&gt;&lt;br&gt;
ComfyUI-PainterI2V from 绘画小子(Douyin creator) specifically fixes the slow-motion issue in 4-step LoRAs (e.g., lightx2v) with reduced slow-motion drag, enhanced camera movement, optimized single-frame image-to-video workflows, and plug &amp;amp; play features.&lt;br&gt;
👉 &lt;a href="https://github.com/princepainter/ComfyUI-PainterI2V" rel="noopener noreferrer"&gt;Try it Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flphhg6lguw5pa3ryee1z.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flphhg6lguw5pa3ryee1z.gif" alt=" " width="304" height="384"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;📬 Want More? Stay Updated.&lt;br&gt;
Every week, we bring you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;New model releases &amp;amp; upgrades&lt;/li&gt;
&lt;li&gt;AI research breakthroughs&lt;/li&gt;
&lt;li&gt;Open-source tools you can use today&lt;/li&gt;
&lt;li&gt;Community highlights that inspire&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;👉 &lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;Subscribe to The Tongyi Weekly and never miss a release.&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>genai</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Nov7, 2025 | The Tongyi Weekly: Your weekly dose of cutting-edge AI from Tongyi Lab</title>
      <dc:creator>Tongyi Lab</dc:creator>
      <pubDate>Fri, 07 Nov 2025 08:27:11 +0000</pubDate>
      <link>https://dev.to/alibaba_tongyi_lab_25ad9f/nov7-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-4n4c</link>
      <guid>https://dev.to/alibaba_tongyi_lab_25ad9f/nov7-2025-the-tongyi-weekly-your-weekly-dose-of-cutting-edge-ai-from-tongyi-lab-4n4c</guid>
      <description>&lt;p&gt;Hello, community!&lt;/p&gt;

&lt;p&gt;We’re Tongyi Lab — the AI research institute under Alibaba Group, and the team behind Qwen, Wan, Tongyi Fun, and a growing ecosystem of models and frameworks loved by millions of developers worldwide.&lt;/p&gt;

&lt;p&gt;From this week forward, we will be sharing the latest updates and breakthroughs from Tongyi and bring you directly from our lab to your desk — weekly.&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;Subscribe Now&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;Welcome to this week's update. In the past week, we've seen exciting updates from our open-source projects like Qwen and AgentScope.&lt;/p&gt;

&lt;h2&gt;
  
  
  📣 Model Release &amp;amp; Updates
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Introducing Qwen3-Max-Thinking-Preview: An Early Preview of Qwen3-Max-Thinking&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;We're excited to announce that Qwen3-Max-Thinking-Preview is now available on Qwen Chat! This is an early preview of Qwen3-Max-Thinking.&lt;/p&gt;

&lt;p&gt;Even at this intermediate stage, this model demonstrates remarkable potential (100% score) on challenging reasoning benchmarks like AIME 2025 and HMMT when augmented with tool use and scaled test-time compute.&lt;/p&gt;

&lt;p&gt;Try it in Qwen Chat and Alibaba Cloud API:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://chat.qwen.ai/?thinking=true" rel="noopener noreferrer"&gt;Qwen Chat&lt;/a&gt;&lt;br&gt;
&lt;a href="https://modelstudio.console.alibabacloud.com/?tab=doc#/doc/?type=model&amp;amp;url=2840914_2&amp;amp;modelId=qwen3-max-preview" rel="noopener noreferrer"&gt;Alibaba Cloud API &lt;/a&gt;（enable_thinking=True）&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AgentScope Updates: New Agents, Enhanced Features, and More&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This week, we've upgraded AgentScope - our open-source framework for building agentic applications - with exciting new samples and features, making it easier than ever to build, deploy, and scale intelligent agent systems:&lt;/p&gt;

&lt;p&gt;𝐍𝐞𝐰 𝐀𝐠𝐞𝐧𝐭 𝐈𝐦𝐩𝐥𝐞𝐦𝐞𝐧𝐭𝐚𝐭𝐢𝐨𝐧𝐬: we open-sourced two new yet powerful agent applications built on AgentScope: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Alias-Agent: A versatile LLM-empowered agent application that flexibly handles diverse real-world tasks within a secure sandbox environment: &lt;a href="https://github.com/agentscope-ai/agentscope-samples/tree/main/alias" rel="noopener noreferrer"&gt;https://github.com/agentscope-ai/agentscope-samples/tree/main/alias&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data-Juicer Agent: An intelligent multi-agent system that enables natural language-driven data processing by seamlessly integrating AgentScope with Data-Juicer: &lt;a href="https://github.com/agentscope-ai/agentscope-samples/tree/main/data_juicer_agent" rel="noopener noreferrer"&gt;https://github.com/agentscope-ai/agentscope-samples/tree/main/data_juicer_agent&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;𝐂𝐨𝐫𝐞 𝐂𝐚𝐩𝐚𝐛𝐢𝐥𝐢𝐭𝐢𝐞𝐬 𝐄𝐱𝐩𝐚𝐧𝐬𝐢𝐨𝐧: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Agentic RL Support: Fine-tune workflows using Trinity-RFT with minimal code changes: &lt;a href="https://github.com/agentscope-ai/agentscope/tree/main/examples/training/react_agent" rel="noopener noreferrer"&gt;https://github.com/agentscope-ai/agentscope/tree/main/examples/training/react_agent&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Long-term Memory: Integrated ReMe toolkit for personal, task, and tool-level memory management: &lt;a href="https://github.com/agentscope-ai/agentscope/tree/main/examples/functionality/long_term_memory/reme" rel="noopener noreferrer"&gt;https://github.com/agentscope-ai/agentscope/tree/main/examples/functionality/long_term_memory/reme&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;𝐀𝐠𝐞𝐧𝐭𝐒𝐜𝐨𝐩𝐞-𝐒𝐚𝐦𝐩𝐥𝐞𝐬: &lt;/p&gt;

&lt;p&gt;We introduced a curated collection of ready-to-use agent implementations and full-stack applications built with AgentScope: &lt;a href="https://github.com/agentscope-ai/agentscope-samples" rel="noopener noreferrer"&gt;https://github.com/agentscope-ai/agentscope-samples&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;𝐑𝐮𝐧𝐭𝐢𝐦𝐞 𝐔𝐩𝐠𝐫𝐚𝐝𝐞𝐬:&lt;/p&gt;

&lt;p&gt;We've upgraded the AgentScope Runtime to make it easier to deploy and interact with agents: App-like Agent Deployment, Python SDK, and GUI &amp;amp; Desktop-enabled Sandboxes: &lt;a href="https://github.com/agentscope-ai/agentscope-runtime" rel="noopener noreferrer"&gt;https://github.com/agentscope-ai/agentscope-runtime&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  🧩 Ecosystem Highlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-VL Lands on llama.cpp&lt;/strong&gt;&lt;br&gt;
Qwen3-VL—our state-of-the-art vision-language model—is now available on llama.cpp! You can now run this powerful model entirely on your personal device, with native support for on CPU, CUDA, Metal, Vulkan, and other backends. &lt;/p&gt;

&lt;p&gt;We’ve also released GGUF weights for all variants—from 2B up to 235B.&lt;/p&gt;

&lt;p&gt;Download &amp;amp; explore:&lt;/p&gt;

&lt;p&gt;Hugging Face: &lt;a href="https://huggingface.co/collections/Qwen/qwen3-vl" rel="noopener noreferrer"&gt;https://huggingface.co/collections/Qwen/qwen3-vl&lt;/a&gt;&lt;br&gt;
ModelScope: &lt;a href="https://modelscope.cn/collections/Qwen3-VL-5c7a94c8cb144b" rel="noopener noreferrer"&gt;https://modelscope.cn/collections/Qwen3-VL-5c7a94c8cb144b&lt;/a&gt;&lt;br&gt;
PR: &lt;a href="https://github.com/ggml-org/llama.cpp/pull/16780" rel="noopener noreferrer"&gt;https://github.com/ggml-org/llama.cpp/pull/16780&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Qwen3-Max-Preview Entered the Top Tier of Arena Expert Leaderboard&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The Qwen3-Max-Preview continues to rank near the top of the new Arena Expert Leaderboard, showcasing its ability to handle challenging prompts from real users. &lt;/p&gt;

&lt;p&gt;Arena Expert is a new LMArena evaluation framework to identify the toughest, most expert-level prompts from real users, powering a new Expert leaderboard.&lt;/p&gt;

&lt;p&gt;Check out the Arena Expert Leaderboard: &lt;a href="https://lmarena.ai/leaderboard" rel="noopener noreferrer"&gt;https://lmarena.ai/leaderboard&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  ✨ Community Spotlights
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Qwen-Edit LoRA Model Hits Top 5 on Hugging Face - from Developer @dx8152&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Shoutout to developer @dx8152! The LoRA model Qwen-Edit-2509-Multiple-angles, built atop Qwen-Image-Edit-2509, surged to #5 on Hugging Face’s download chart—an inspiring example of what’s possible when foundational models empower creators.&lt;/p&gt;

&lt;p&gt;Download Link: &lt;a href="https://huggingface.co/dx8152/Qwen-Edit-2509-Multiple-angles" rel="noopener noreferrer"&gt;https://huggingface.co/dx8152/Qwen-Edit-2509-Multiple-angles&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Article content&lt;br&gt;
Demo: Qwen-Edit-2509-Multiple-angles&lt;/p&gt;

&lt;h2&gt;
  
  
  📬 Want More? Stay Updated.
&lt;/h2&gt;

&lt;p&gt;This is just one week of what’s coming.&lt;/p&gt;

&lt;p&gt;Every week, we bring you:&lt;/p&gt;

&lt;p&gt;New model releases &amp;amp; upgrades&lt;br&gt;
AI research breakthroughs&lt;br&gt;
Open-source tools you can use today&lt;br&gt;
Community highlights that inspire&lt;/p&gt;

&lt;p&gt;👉 Subscribe to The Tongyi Weekly and never miss a release:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.linkedin.com/build-relation/newsletter-follow?entityUrn=7392460924453945345" rel="noopener noreferrer"&gt;Subscribe Now&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thank you for being part of this journey.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Tongyi Lab is a research institution under Alibaba Group dedicated to artificial intelligence and foundation models, focusing on the research, development, and innovative applications of AI models across diverse domains. Its research spans large language models (LLMs), multimodal understanding and generation, visual AIGC, speech technologies, and more.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>llm</category>
      <category>programming</category>
    </item>
  </channel>
</rss>
