Baidu just released a reasoning model π₯ ERNIE-4.5-21B-A3B-Thinking
huggingface.co/baidu/ERNIE-...
β¨ Small MoE - Apache 2.0
β¨ 128K context length for deep reasoning
β¨ Efficient tool usage capabilities
@adinayakup.bsky.social
AI Research @Hugging Face π€ Contributing to the Chinese ML community.
Baidu just released a reasoning model π₯ ERNIE-4.5-21B-A3B-Thinking
huggingface.co/baidu/ERNIE-...
β¨ Small MoE - Apache 2.0
β¨ 128K context length for deep reasoning
β¨ Efficient tool usage capabilities
Baidu just released a thinking model π₯ ERNIE-4.5-21B-A3B-Thinking
huggingface.co/baidu/ERNIE-...
β¨ Small MoE - Apache 2.0
β¨ 128K context length for deep reasoning
β¨ Efficient tool usage capabilities
MiniCPM4.1π₯New edge-side LLM built for efficiency + reasoning from OpenBMB
huggingface.co/openbmb/Mini...
β¨ 8B - Apache 2.0
β¨ Hybrid reasoning model: deep reasoning +fast inference.
β¨5x faster on edge chips, 90% smaller (BitCPM)
β¨Trained on UltraClean + UltraChat v2 data
Inverse IFEval π₯New benchmark from Bytedance & MAP
huggingface.co/datasets/m-a...
huggingface.co/papers/2509....
Testing LLMs on their ability to override biases & follow adversarial instructions.
β¨ 8 challenge types
β¨ 1,012 CN/EN Qs across 23 domains
β¨ Human-in-the-loop + LLM-as-a-Judge
Added the Kwai team from Kuaishou in the heatmap
huggingface.co/spaces/zh-ai...
Klear-46B-A2.5π₯ a sparse MoE LLM developed by the Kwai-Klear Team at Kuaishou
huggingface.co/collections/...
β¨ 46B total / 2.5B active - Apache2.0
β¨ Dense-level performance at lower cost
β¨ Trained on 22T tokens with progressive curriculum
β¨ 64K context length
Latest update from Moonshot AI
Kimi K2 >>> Kimi K2-Instruct-0905π₯
huggingface.co/moonshotai/K...
β¨ 32B activated / 1T total parameters
β¨ Enhanced agentic coding intelligence
β¨ Better frontend coding experience
β¨ 256K context window for long horizon tasks
β¨ Supports 33 languages, including 5 ethnic minority languages in China π
β¨ Including a translation ensemble model: Chimera-7B
β¨ Full pipeline: pretrain > CPT > SFT > enhancement > ensemble refinement > SOTA performance at similar scale
Hunyuan-MT-7B π₯ open translation model released by Tencent
huggingface.co/collections/...
β¨ 560B total / ~27B active MoE β MIT license
β¨ 128k context length + advanced reasoning
β¨ ScMoE design: 100+ TPS inference
β¨ Stable large-scale training + strong agentic performance
From food delivery to frontier AI π Meituan, the leading lifestyle platform just dropped its first open SoTA LLM: LongCat-Flash π₯
huggingface.co/meituan-long...
β¨ Large-scale triplet dataset (content, style, stylized)
β¨ Disentangled learning: style alignment + content preservation
β¨ Style Reward Learning (SRL) for higher fidelity
β¨ USO-Bench: 1st benchmark for style & subject jointly
β¨ SOTA results on subject consistency & style similarity
USO π¨ Unified customization model released by Bytedance research
Demo
huggingface.co/spaces/byted...
Model
huggingface.co/bytedance-re...
Paper
huggingface.co/papers/2508....
β¨ Direct raw audio: text & speech ,no ASR+LLM+TTS pipeline
β¨ High-IQ reasoning: RL + CoT for paralinguistic cues
β¨ Multimodal RAG + tool calling
β¨ Emotion, timbre, dialect & style control
β¨ SOTA on ASR, paralinguistic, speech dialog
Step-Audio 2π₯ New end to end multimodal LLM for audio & speech, released by StepFun
huggingface.co/collections/...
>Applications: AI-as-a-service, test bases, new standards
>Open-source: support communities, encourage contributions (incl. university credits & recognition), foster new application approaches, and build globally impactful ecosystems π
>Talent, policy & safety frameworks: secure sustainable growth
β¨Highlights:
>Models: advance theory, efficient training/inference, evaluation system
>Data: high-quality datasets, IP/copyright reform, new incentives
>Compute: boost chips & clusters, improve national network, promote cloud standardization, and ensure inclusive, efficient, green, secure supply.
π¨π³ Chinaβs State Council just released its βAI+β Action Plan (2025)
huggingface.co/spaces/zh-ai...
β¨Goal: By 2035, AI will deeply empower all sectors, reshape productivity & society
β¨Focus on 6 pillars:
>Science & Tech
>Industry
>Consumption
>Public welfare
>Governance
>Global cooperation
β¨ SOTA vision language capability
β¨ 96Γ video token compression > high-FPS & long video reasoning
β¨ Switchable fast vs deep thinking modes
β¨ Strong OCR, document parsing, supports 30+ languages
MiniCPM-V 4.5 π New MLLM for image, multi-image & video understanding, running even on your phone, released by OpenBMB
huggingface.co/openbmb/Mini...
InternVL3.5 π₯ New family of multimodal model by Shanghai AI lab @opengvlab
huggingface.co/collections/...
β¨ 1B Β· 2B Β· 4B Β· 8B Β· 14B Β· 38B ο½ MoE β 20B-A4B Β· 30B-A3B Β· 241B-A28B πApache 2.0
β¨ +16% reasoning performance, 4.05Γ speedup vs InternVL3
Intern-S1-mini π₯ lightweight open multimodal reasoning model by Shanghai AI Lab.
huggingface.co/internlm/Int...
β¨ Efficient 8B LLM + 0.3B vision encoder
β¨ Apache 2.0
β¨ 5T multimodal pretraining, 50%+ in scientific domains
β¨ Dynamic tokenizer for molecules & protein sequences
Seed-OSS π₯ The latest open LLM from Bytedance Seed team
huggingface.co/collections/...
β¨ 36B - Base & Instruct
β¨ Apache 2.0
β¨ Native 512K long context
β¨ Strong reasoning & agentic intelligence
β¨ 2 Base versions: with & without synthetic data
β¨DeepSeek V3.1 just dropped on @hf.co
huggingface.co/collections/...
β¨ Apache 2.0
β¨ Semantic + Appearance Editing: rotate, restyle, add/remove
β¨ Precise Text Editing β edit CN/EN text, keep style
huggingface.co/spaces/Qwen/...
Before my vacation: Qwen releasing.
When I came back: Qwen still releasing
Respect!!π«‘
Qwen Image Edit π₯ the image editing version of Qwen-Image by Alibaba Qwen
huggingface.co/Qwen/Qwen-Im...
π₯ July highlights from Chinese AI community
huggingface.co/collections/...
Iβve been tracking things closely, but Julyβs open-source wave still managed to surprise me.
Canβt wait to see whatβs coming next! π
Qwen team did it again!π₯
They just released Qwen3-Coder-30B-A3B-Instruct on the hub
huggingface.co/Qwen/Qwen3-C...
β¨ Apache 2.0
β¨30B total / 3.3B active (128 experts, 8 top-k)
β¨ Native 256K context, extendable to 1M via Yarn
β¨ Built for Agentic Coding
β¨ 321B total / 32B active - Apache 2.0
β¨ MFA + AFD : cutting decoding cost by up to 70% vs. DeepSeek-V3
β¨ 4T image-text pretraining: strong visionβlanguage grounding
β¨ Modular, efficient, deployable: runs on just 8Γ48GB GPUs
Itβs here! After the WAIC announcement, StepFun has just dropped Step 3 π₯ their latest multimodal reasoning model on the hub.
Model: huggingface.co/stepfun-ai/s...
Paper: huggingface.co/papers/2507....