π₯ July highlights from Chinese AI community
huggingface.co/collections/...
Iβve been tracking things closely, but Julyβs open-source wave still managed to surprise me.
Canβt wait to see whatβs coming next! π
@adinayakup.bsky.social
AI Research @Hugging Face π€ Contributing to the Chinese ML community.
π₯ July highlights from Chinese AI community
huggingface.co/collections/...
Iβve been tracking things closely, but Julyβs open-source wave still managed to surprise me.
Canβt wait to see whatβs coming next! π
Qwen team did it again!π₯
They just released Qwen3-Coder-30B-A3B-Instruct on the hub
huggingface.co/Qwen/Qwen3-C...
β¨ Apache 2.0
β¨30B total / 3.3B active (128 experts, 8 top-k)
β¨ Native 256K context, extendable to 1M via Yarn
β¨ Built for Agentic Coding
β¨ 321B total / 32B active - Apache 2.0
β¨ MFA + AFD : cutting decoding cost by up to 70% vs. DeepSeek-V3
β¨ 4T image-text pretraining: strong visionβlanguage grounding
β¨ Modular, efficient, deployable: runs on just 8Γ48GB GPUs
Itβs here! After the WAIC announcement, StepFun has just dropped Step 3 π₯ their latest multimodal reasoning model on the hub.
Model: huggingface.co/stepfun-ai/s...
Paper: huggingface.co/papers/2507....
Qwen3-30B-A3B-Thinking-2507 π₯ latest step in scaling thinking capabilities from Alibaba_Qwen
huggingface.co/Qwen/Qwen3-3...
β¨ 30B total / 3B active - Apache 2.0
β¨ Native 256K context
β¨ SOTA coding, alignment, agentic reasoning
Skywork UniPic π₯a unified autoregressive multimodal model for image understanding, generation, & editing, by Skywork
huggingface.co/collections/...
β¨ 1.5 B - MIT License
β¨ Runs on RTX 4090
β¨ Truly unified architecture
Qwen just released Qwen3-30B-A3B-Instruct-2507 π₯ an upgrade to the non-thinking mode model
huggingface.co/Qwen/Qwen3-3...
β¨ 30B MoE / 3.3B active - Apache 2.0
β¨ Strong gains in reasoning, math, coding, & multilingual tasks
β¨ Native support for 256K long-context inputs
β¨ 5B/14B - Apache2.0
β¨ Cinematic-level aesthetics (lighting, tone, composition)
β¨ Massive training data (+83% videos)β smoother motion
β¨ Supports image-only video generation, even without a prompt.
Wan2.2 π₯A video diffusion model with MoE just released by Alibaba_Wan
huggingface.co/Wan-AI/Wan2....
huggingface.co/Wan-AI/Wan2....
GLM-4.5 π₯ The largest open models yet from Zhipu AI.
Built for intelligent agents with unified capabilities: reasoning, coding, tool use.
huggingface.co/collections/...
β¨ 355B total / 32B active - MIT license
β¨ Hybrid modes: Thinking mode for complex tasks/ Non-thinking mode for instant replies
Panshi η£η³ πͺ¨ Scientific Model by the Chinese Academy of Sciences
huggingface.co/ScienceOne-A...
huggingface.co/ScienceOne-A...
β¨ 8B/32B- Apache2.0
β¨ Trained on scientific data & laws across math, physics, chemistry, bio, etc.
β¨ Supports 300+ tools, 170M+ papers, autonomous scientific planning
TencentHunyuan released their first 3D world model: Hunyuan World 1.0 π₯
huggingface.co/tencent/Huny...
β¨From a single prompt to explorable 3D scenes in minutes
β¨Supports Immersive roaming / Semantic-level interactivity / Physics-ready simulation
Shanghai AI Lab released their latest multimodal reasoning model: Intern-S1 π₯
huggingface.co/internlm/Int...
β¨Apache 2.0
β¨235B MoE LLM (Qwen3)+ 6B Vision Encoder(InternViT)
β¨Pretrained on 5T tokens (50% scientific)
β¨Excels at chemistry, protein folding, math, physics
Coze is going open source π₯
Coze is a full-stack platform for building & managing AI Agents built by ByteDance
β¨ Coze Studio β Visual dev + plugin framework github.com/coze-dev/coz...
β¨ Coze Loop β Prompt testing, evals, monitoring github.com/coze-dev/coz...
More to come. Stay tuned!
Qwen3-235B-A22B-Thinking-2507 π§ new reasoning model by Qwen
huggingface.co/Qwen/Qwen3-2...
β¨ 235B total / 22B active (8 experts)
β¨ 256K context window
β¨ Agent-ready with tool use & <think> reasoning mode
Hope the team gets some well-deserved rest this weekend after all the massive releases π
Ming-lite-omni v1.5 π₯ upgrade version of Ming-lite-omni, by AntGroup
huggingface.co/inclusionAI/...
β¨ 20.3B / 3B active - MoE
β¨ SOTA video understanding via 3D MRoPE + curriculum learning
β¨ Real time speech synthesis + dialect support
β¨ Enhanced multimodal generation with ID & scene consistency
β¨ Highly Customizable: Supports custom terms, domain prompts, and translation memory for accurate, context-aware results.
β¨ Fast and affordable: $0.5 per million tokens.
Qwen is on fire this week π₯
Qwen3-MTπ Translation model supports 92 languages. Demo is available on @hf.co
huggingface.co/spaces/Qwen/...
Qwen3-Coder π» agentic code model by Alibaba_Qwen
huggingface.co/collections/...
β¨ 480B total, 35B activated MoE
β¨ Agentic Coding + Browser Use β Top code model performance
β¨ 256K context (up to 1M via Yarn) for repo-scale understanding
KAT-V1 π₯ a LLM that tackles overthinking by switching between reasoning and direct answers, by Kuaishou.
huggingface.co/Kwaipilot/KA...
β¨ 40B
β¨ Step-SRPO: smarter reasoning control via RL
β¨ MTP + Distillation: efficient training, lower cost
From paper to project page in one clickπ
AnyCoder π₯ turns research PDFs into structured, shareable project pages in seconds!
huggingface.co/spaces/akhal...
Powered by 8 SoTA open models on @hf.co
Seed-X π₯ a suite of multilingual translation models released by ByteDance.
huggingface.co/collections/...
β¨ instruction/reinforcement learning/reward model
β¨ Supports 28 languages, bidirectional translation
β¨ 7B with mistral architecture
β¨ Excels across science, law, finance, literature & more
β¨ Early-exit branches (3Bβ7B)
β¨ Shared weights, seamless switching
β¨ Fine-tuned for chat β complex reasoning
TeleAI δΈε½η΅δΏ‘ launched AI Flow βοΈ a new AI architecture where intelligence flows across cloud -edge - devices.
Theyβve also released one of the projectβs key pillars on the Hub: AI Flow Ruyi ε¦ζ π₯ a scalable Familial Model.
Model:
huggingface.co/TeleAI-AI-Fl...
Paper:
huggingface.co/papers/2506....
Kimi-K2 π₯πA trillion-parameter MoE model focused on long context, code, reasoning, and agentic behavior.
huggingface.co/collections/...
β¨ Base & Instruct
β¨ 1T total / 32B active - Modified MIT License
β¨ 128K context length
β¨ Muon optimizer for stable trillion-scale training
So excited to finally release our first robot today: Reachy Mini
A dream come true: cute and low priced, hackable yet easy to use, powered by open-source and the infinite community.
Read more and order now at huggingface.co/blog/reachy-...
POLARπ»ββοΈ New reward modeling by Shanghai AI Lab.
huggingface.co/collections/...
β¨ 7B/8B - Apache 2.0
β¨ Scalable policy discriminative pretraining
β¨ Easy RLHF with minimal preference data
Skywork-Reward-V2π₯ Reward models by Skywork AI
huggingface.co/collections/...
β¨ 0.6B - 8B
β¨ Trained on 26M human-LLM preference pairs
β¨ 0.6B > 27B in many tasks