博客 资源 关于 搜索 主题
AI Development

AI News Digest: KIMI K2.6 Incoming, Qwen 3.6 on M5 Max, Claude System Prompts Leaked

更新于 2026年4月19日

分类: AI Development
分享

AI news digest April 19 2026 — KIMI K2.6, Qwen 3.6 on M5 Max, Claude system prompts, ChatGPT diagnosis

Sunday morning, April 19, 2026. KIMI K2.6 is on the horizon, someone is running Qwen 3.6 on an M5 Max and calling it “as good as Claude,” Simon Willison tracked the evolution of Claude’s system prompts as a git timeline, and a 23-year-old woman used ChatGPT to self-diagnose a rare genetic disorder after years of being misdiagnosed. Here’s the digest.


KIMI K2.6 Is Coming

The KIMI K2.6 announcement hit r/LocalLLaMA with 449 upvotes and 85 comments. Moonshot AI’s next model is generating serious buzz in the open-weight community.

KIMI K2 was already competitive with frontier models on coding and reasoning tasks. K2.6 is expected to push that further. The timing puts it directly alongside Qwen 3.6 in the race for best open-weight model — and the competition is only accelerating.

For developers who’ve been watching the open-weight space heat up this week, this is another data point: the pace of releases is compressing. We’re getting meaningful new models every few days now, not every few months.

→ r/LocalLLaMA: KIMI K2.6 SOON !!


Qwen 3.6 on M5 Max: As Good as Claude

The Qwen 3.6 momentum isn’t slowing down. A post titled “I’m running Qwen3.6-35B-A3B with 8-bit quant and 64k context through OpenCode on my MBP M5 Max 128GB and it’s as good as Claude” hit 477 points with 233 comments.

That’s a bold claim — and the comment section largely agreed. The M5 Max with 128GB of unified memory is becoming the reference hardware for serious local inference. Running a 35B MoE model with 64k context at usable speeds, locally, with no API calls, no rate limits, and no identity verification. For developers who value privacy and control, this is the setup.

Meanwhile, Qwen3.6-35B-A3B-Uncensored-Wasserstein-GGUF also hit 133 points. Uncensored fine-tunes are appearing within days of the base model release now. The community pipeline from “model drops” to “production-ready GGUF with specific traits” has gotten remarkably fast.

For anyone running local models on macOS, the M5 Max is looking like the sweet spot for running frontier-competitive models without cloud dependencies.

→ r/LocalLLaMA: Qwen3.6-35B-A3B on M5 Max — “as good as Claude”


Simon Willison Diffs Claude’s System Prompts

Simon Willison published two pieces back-to-back that are worth reading together:

  1. “Changes in the system prompt between Claude Opus 4.6 and 4.7” — A detailed diff showing exactly what Anthropic changed in the system instructions between model versions.
  2. “Claude system prompts as a git timeline” — He built a tool to track system prompt changes over time using git, turning an opaque process into an auditable history.

This is the kind of transparency work that matters. System prompts shape model behavior in ways users don’t see, and Anthropic has been more open about publishing them than most labs. Willison’s git timeline approach means the community can now track changes systematically instead of relying on screenshots and hearsay.

The practical implication: if you’re building on Claude’s API and notice behavior changes between versions, the system prompt diff is the first place to look. It’s not always the model weights — sometimes it’s the instructions.

→ Simon Willison: Changes in the system prompt between Claude Opus 4.6 and 4.7

→ Simon Willison: Claude system prompts as a git timeline


ChatGPT Diagnosed a Rare Genetic Disorder

A 23-year-old woman used ChatGPT to self-diagnose a rare genetic disorder after years of being misdiagnosed by doctors. The post hit 340 points and 146 comments on r/ChatGPT.

Stories like this keep surfacing. AI as a first-pass diagnostic tool for rare conditions is becoming a recurring pattern — patients with unusual symptom clusters that don’t map neatly to common diagnoses are finding that LLMs can cross-reference symptoms against a broader literature base than any single specialist.

The usual caveats apply: LLMs hallucinate, they’re not doctors, and self-diagnosis carries real risks. But the signal in these stories is consistent. For rare conditions where the diagnostic journey averages 5-7 years and multiple specialists, having an AI pattern-match across your full symptom profile has genuine utility — as a starting point for the right conversation with the right doctor.

→ r/ChatGPT: Woman, 23, Self-Diagnosed Her Rare Genetic Disorder Using ChatGPT


Quick Hits

  • “Is it just me or is ChatGPT being a dick lately?” hit 485 points on r/ChatGPT with 194 comments. Model behavior drift continues to be a sore spot for users.

→ r/ChatGPT: Is it just me or is ChatGPT being a dick lately?

  • “That time Anthropic played 2.5 million ChatGPT users” cleared 866 points. The AI company rivalry memes are getting stronger.

→ r/ChatGPT: That time Anthropic played 2.5 million ChatGPT users

  • Someone built a tiny world model game that runs locally on iPad — 181 points on r/LocalLLaMA. The “AI on edge devices” trend continues to produce impressive demos.

→ r/LocalLLaMA: I made a tiny world model game that runs locally on iPad

  • Zero-Copy GPU Inference from WebAssembly on Apple Silicon hit Hacker News. If you’re building inference pipelines on Apple hardware, this is a significant performance optimization path.

→ Zero-Copy GPU Inference from WebAssembly on Apple Silicon

  • Reconstructing a Dead USB Protocol cleared 161 points on r/programming — a deep reverse-engineering writeup going from unknown chip to working implementation. Classic hacker content.

→ Reconstructing a Dead USB Protocol: From Unknown Chip to Working Implementation


Takeaways

  1. The open-weight race is a weekly sprint now. KIMI K2.6 incoming while Qwen 3.6 is still dominating the conversation. Models are shipping faster than the community can benchmark them.
  2. Apple Silicon is the local inference platform. M5 Max running 35B models with 64k context at Claude-competitive quality. The hardware moat for cloud inference is eroding.
  3. System prompt transparency matters. Willison’s git-based tracking of Claude’s system prompts sets a standard that other labs should follow. Users deserve to understand what’s shaping model behavior.
  4. AI diagnostics keep proving out. Another rare disease story. The pattern is too consistent to dismiss, even with the valid concerns about hallucination risk.
  5. ChatGPT user sentiment is souring. 485 points on “being a dick lately” plus ongoing behavior drift complaints suggest OpenAI has a user experience problem to address.

Yesterday’s digest covered Qwen 3.6 dominating LocalLLaMA, GPT Image Gen 2, and the danger of modern open source. The weekend news cycle hasn’t slowed down.

分类 AI Development
分享

相关文章

将最新的AI见解发送到您的收件箱

了解最新的趋势、教程和行业见解。加入信任我们新闻通讯的开发人员社区。

仅新账户。提交您的电子邮件即表示您同意我们的 隐私政策