Back to Blog

OpenEvidence, the ‘ChatGPT for doctors,’ raises $250m at $12B valuation, 12x from $1b last Feb

Agent Labs are all you need

AI News for 1/20/2026-1/21/2026. We checked 12 subreddits, 544 Twitters and 24 Discords (205 channels, and 7561 messages) for you. Estimated reading time saved (at 200wpm): 584 minutes. Our new website is now up with full metadata search and beautiful vibe coded presentation of all past issues. See https://news.smol.ai/ for the full news breakdowns and give us feedback on @smol_ai!

We have a soft rule that new decacorn fundraises get title stories, and multiple sources have the story of OpenEvidence's $12B fundraise, 12x from last year. What's odd is "CEO Daniel Nadler told CNBC that OpenEvidence is used by 40% of physicians in the U.S. and topped $100 million in annual revenue last year." which is a 120x multiple.


AI Twitter Recap

Frontier model governance: Anthropic’s new Claude constitution (CC0) and reactions

  • What shipped: Anthropic published a new “constitution” describing desired Claude behaviors/values and stated it is used directly in training; importantly, the full constitution is released CC0 1.0 to encourage reuse/adaptation (announcement, CC0 link). Anthropic frames it as a living document shaped by internal + external experts (follow-up).
  • Community read: Amanda Askell emphasized it’s a work-in-progress and invited feedback (Askell). Others highlighted the “meta” oddity of training a model on a document about how the model should be (scaling01 on Opus reflecting on circularity). Several reactions focused on the constitution as “alignment signaling” vs. practical harm reduction, and whether it bakes in PR-oriented persona behaviors (nearcyan, NickEMoran question).
  • Practical engineering consequence: Anthropic also posted about an internal performance engineering take-home that became solvable by Opus 4.5, forcing them to redesign hiring evaluation—a concrete example of “models catching up to our screening tasks” (Anthropic Eng, trishume).

Agents in production: from “AI employees” (Podium) to agent UX, memory, and evals

  • Podium’s “Jerry” as an agent business unit: Podium claims $100M+ AI agent ARR and 10k+ agents deployed, anchored on staffing constraints in SMBs (after-hours leads, missed calls, turnover). The narrative: stop selling “software,” sell an AI operator that uses the existing product end-to-end (Eric Rea). Tom Loverro adds board-level metrics (burn down from $95M → $0, AI ARR $0 → $100M in ~21 months) and links to OpenAI’s case study (Tom Loverro, Garry Tan).
  • Memory & long-horizon reliability becomes the bottleneck:
    • The Agent Cognitive Compressor (ACC) pitch argues “more context ≠ better agents,” criticizing transcript replay and naive retrieval. ACC maintains a bounded “Compressed Cognitive State” with schema-constrained commit, claiming lower drift/hallucination over long runs (dair_ai).
    • A separate thread positions “self-improving multi-agents” for scientific workflows via MCP-SIM, a multi-agent loop that clarifies underspecified physics prompts, generates code, executes, diagnoses errors, and produces multilingual explanations; claims 12/12 tasks solved vs one-shot GPT at 6/12 (omarsar0).
  • Agentic benchmarking moves beyond coding puzzles:
    • APEX-Agents evaluates long-horizon “professional services” tasks in Google Workspace; early Pass@1 leaderboard numbers are low (Gemini 3 Flash High 24.0%, GPT-5.2 High 23.0%, Claude Opus 4.5 High 18.4%)—a reminder that “agent autonomy” is still brittle (BrendanFoody).
    • prinzbench introduces a private benchmark for legal research + search (33 Qs, graded manually, run 3x) where “search” is the failure mode; claim: GPT-5.2 Thinking barely exceeds 50%, Gemini models are close behind, and Sonnet/Opus 4.5 scored 0/24 on Search (deredleritt3r).
  • Tooling + UX layer catches up: multiple posts converge on the idea that agents need a “context layer” and production scaffolding (governance, auth, observability) as much as better models—see Prefect Horizon and MCP server best practices below.

Agent platforms and “context layers”: MCP, Skills, Prefect Horizon, LangChain Deep Agents

  • Prefect Horizon: MCP → platform: Prefect positions “context layer” as the interface between agents and enterprise tools/data, and argues MCP describes how to build a server but not how to deploy/govern it at org scale. Horizon claims managed deployment, registry/catalog, gateways w/ RBAC + audit logs, and “agentic interface for business users” (jlowin).
  • MCP servers: design guidance: Phil Schmid pushes back on “Skills replace MCP” takes: MCP isn’t the problem; bad servers are. Recommendations: design tools around outcomes, typed flat args with constraints, docstrings/errors as agent instructions; positions Skills and MCP as complementary (philschmid).
  • LangChain deepagents: agents as folders + UI integration:
    • CopilotKit published a tutorial building a fullstack Deep Agent app (resume ingestion → skills extraction → sub-agents with web search → streaming UI), addressing the “missing UI/application layer” gap (CopilotKit).
    • LangChain shipped Agent Builder GA plus a template library with domain partners (Tavily, PagerDuty, Box, etc.) to reduce “prompt-to-agent” friction (LangChain).
    • Deep Agents’ framing “agents are just folders” emphasizes portability/distribution: you can package, download, and run agents quickly via CLI flows (hwchase17, Vtrivedy10 demo, LangChain_OSS). Sydney Runkle highlights two core patterns: subagents for context isolation and skills loaded only when relevant (sydneyrunkle).
  • LangSmith + analytics: one thread points to LangSmith traces as a substrate not only for debugging but product analytics (“agent traces → product analytics”) (SoftwareWatcher).

Inference + systems: low-VRAM serving, open inference stacks, and “inference is the battleground”

  • AirLLM: layer streaming for tiny VRAM: AirLLM’s core idea is sequential layer loading (load → compute → free) with optional compression, HF-like API, CPU/GPU, Linux/macOS; claims extremely low VRAM viability for very large models (LiorOnAI, repo). Engineers should treat the “405B on 8GB” claim as “possible in principle with heavy paging,” but expect throughput/latency constraints and non-trivial engineering caveats.
  • “Actually open AI” requires models + inference engines: Modal argues the ecosystem now has the building blocks—capable open models plus fast tunable OSS inference—sharing their production patterns/stacking for serving at scale (charles_irl).
  • Inference bugs + local stacks: llama.cpp fixed a routing/function issue affecting GLM 4.7 Flash GGUFs, and config updates mention a scoring_func: sigmoid; also shows building a small game using quantized GLM via Unsloth workflows (danielhanchen). There’s also discussion about GLM KV-cache memory behavior and whether frameworks are missing a LoRA-based approach (TheZachMueller).
  • Infra hygiene matters for agents: “fast validation makes every agent more effective” (pre-commit hooks, documented env vars, reducing CI wait) is effectively a “software supply chain for agent productivity” argument (matanSF).
  • Research direction: constant-compute contexts: a thread summarizes NVIDIA’s “TTT-E2E” concept (treat context as data and update weights online) as a way to keep latency constant with long contexts, but with weaker “needle-in-haystack” recall—relevant to agent workloads where exact edits matter (sdrzn).
  • Hardware bottleneck framing: a recurring theme is the shift “intelligence → inference” and the importance of compute/memory supply chains (saranormous), echoed in a deep dive on HBM qualification cycles as the true supply constraint vs “just add fabs” narratives (MarkosAAIG).

Code generation is cheap; code understanding becomes the bottleneck (Devin Review, Copilot CLI, Claude Code)

  • Devin Review: review UX, not just bug spotting: Cognition launched Devin Review, positioning it as a new PR-reading interface to reduce “slop,” reorder diffs by importance, identify duplicated/copied code, add a chat layer, and integrate with GitHub comments. It’s accessible via URL swap (githubdevinreview) or an npx CLI (launch, usage modes, URL tip). Multiple testers report it caught issues even outside the immediate diff (mcparadip, BraceSproul).
  • Meta-point: generation vs verification: Several tweets explicitly argue the bottleneck has moved from writing to reviewing/understanding/testing, and that next-gen SWE tooling should accelerate the human’s comprehension loop rather than only run an “arms-length agent” (walden_yan, ScottWu46, theodormarcu).
  • CLI agents evolve: GitHub Copilot CLI added an askUserQuestionTool to ask clarifying questions (example: messy rebase), signaling a trend toward interactive tool-using CLI copilots rather than pure autocomplete (Evan Boyle).
  • Claude Code adoption anecdotes: founders increasingly report “2-person team builds like 10” with Claude Code usage (alexalbert__). There are also frictions: skill reload behavior feels regressive vs a simple “CLAUDE.md reread” flow (corbtt). A particularly illustrative “multi-agent sprawl” story describes scaling Claude Code instances into a quasi-society with governance failures—basically an anecdote about agent orchestration debt (voooooogel).

Video + multimodal: evaluation, model releases, and retrieval scaling

  • Video evaluation infrastructure: Video Arena is now on the web, allowing head-to-head generation across ~15 frontier video models and community voting to drive leaderboards (arena).
  • Model releases: Runway’s Gen-4.5 Image→Video launch emphasizes consistency and narrative; early adopters highlight “story-building” as the best eval methodology for video models (runwayml, c_valenzuelab).
  • Open voice system: Qwen highlights use in Chroma 1.0, described as a fully open-source real-time voice system (Alibaba_Qwen).
  • Retrieval-time scaling / late interaction: Multiple threads argue that ColBERT-style multi-vector retrieval preserves fine-grained intent and can beat much larger embedding models; Mixedbread claims a 17M open-source ColBERT beats 8B embedding models on LongEmbed, and that they’re serving 1B+ documents at <50ms p50 (mixedbreadai claim, prod numbers). TurboPuffer similarly pushes ANN at extreme scale (“index the entire web 100B+ vectors”) (turbopuffer). The meta-trend: retrieval is shifting from “one vector per doc” to token-level / multi-vector systems, but it requires serious infra co-design.

Top tweets (by engagement)

  • Gemini in education: Google launched full-length, on-demand SAT practice exams inside the Gemini app (partnered with The Princeton Review), plus immediate feedback (Google, Sundar Pichai). Google also announced a Gemini × Khan Academy partnership starting with a “Writing Coach” that guides drafting/refinement rather than generating final answers (Google).
  • Claude “Constitution” goes public: Anthropic published a new constitution used directly in Claude’s training; the full text is released under CC0 1.0 (Anthropic, CC0 release, Amanda Askell).
  • AirLLM: extreme low-VRAM inference: claims 70B on 4GB VRAM and even “405B Llama 3.1 on 8GB” via layer-by-layer loading; repo link provided (LiorOnAI, repo).
  • Agents as a real business: Podium reported $100M+ AI agent ARR in <24 months, “10,000+ agents live in production,” framing agents as “AI employees” (Jerry) rather than chatbots (Eric Rea, Garry Tan).
  • Runway Gen-4.5 image-to-video: Runway launched Gen-4.5 Image→Video, emphasizing longer stories, camera control, narrative coherence (runwayml, c_valenzuelab).
  • OpenAI product/UI + org changes: ChatGPT Atlas added tab groups (OpenAI); The Information reported a reorg incl. enterprise/commercial/ads leadership assignments (Steph Palazzolo).

AI Reddit Recap

/r/LocalLlama + /r/localLLM Recap

TO BE COMPLETED

Less Technical AI Subreddit Recap

/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT, /r/ChatGPTCoding, /r/aivideo, /r/aivideo

TO BE COMPLETED


AI Discord Recap

A summary of Summaries of Summaries by gpt-5.2

1. Inference Toolchains Hit Reality (GLM-4.7-Flash, llama.cpp, vLLM, Ollama)

  • Flash Attention Faceplants in GLM-4.7-Flash: Multiple communities reported GLM-4.7-Flash regressions where Flash Attention triggers CPU fallback/bugs and poor throughput (down to 2.8 t/s in LM Studio), with guidance to disable FA until llama.cpp PR #18953 lands everywhere.

  • Ollama vs GGUF: Templates Throw Hands: Users found certain GGUF quants break in Ollama due to chat template incompatibilities, and Unsloth folks repeatedly recommended sticking to official Ollama models while support catches up.

    • The subtext was operational: “it takes time to support things”—so the pragmatic path is to standardize on official artifacts until the ecosystem stabilizes across inference engines.
  • vLLM Update Saves the Day (This Time): In Unsloth’s help chat, at least one thorny issue disappeared after a vLLM update, prompting a sheepish “Oh bruh that was the problem” moment.

    • The follow-on suggestion was process-y: consider pinning dependency versions so future upstream bumps don’t randomly brick pipelines mid-week.

2. Eval Platforms & Product Rollouts (LMArena + multimodal reliability)

  • Video Arena Ships… With a 3-a-Day Speed Limit: LMArena fully released Video Arena at lmarena.ai/?chat-modality=video with a hard cap of 3 generations per 24 hours, and it’s Battle-mode-only (no direct model picking).

    • Users liked “video is live” but complained the slot machine UX blocks controlled testing—especially painful when you’re trying to reproduce a prompt/model behavior.
  • 5M Votes: The Benchmark That Won’t Stop Voting: LMArena’s Text Arena crossed 5 million community votes, highlighted in their milestone social clip.

    • Engineers framed this as “real-world A/B at scale” that increasingly shapes model perception, even when formal benchmark deltas look small.
  • Gemini 3 Pro Image Preview & Nano Banana Pro: Flaky by Design?: LMArena users reported Gemini 3 Pro Image Preview instability plus Nano Banana Pro frequent “Something went wrong” crashes, suspected to be Google-side issues and sometimes lasting 6+ hours.

    • The community gripe: despite unreliability, those models were described as the only ones consistently hitting some specific prompt goals—so people keep using them while grumbling about downtime and errors.

3. Agent & Dev Tooling: MCP, Cursor, DSPy RLMs, and Coding-Assistant Sprawl

  • MCP Inspector Can’t Re-Auth (401 = Game Over): MCP contributors found MCP Inspector fails to re-auth on 401s due to an SDK bug around persisting resourceMetadata across redirects, tracked in inspector issue #576 comment.

    • The current workaround is awkward but clear: rely on VS Code for initial connections because the Inspector path doesn’t recover cleanly mid-session yet.
  • RLMs vs Coding Agents: The Horizon Problem: DSPy discussions contrasted RLMs with “coding agents,” arguing RLMs can externalize inputs/outputs/horizon via code & symbolic calls (see the referenced X thread).

    • Practical takeaway: teams want diagrams of how symbols get accessed and debated whether to hand RLMs tools like ripgrep/semantic search—or make them write their own search code.
  • Cursor’s MCP/Extensions Moment (and Pricing Whiplash): Cursor users debated Playwright MCP for testing (mixed success for TDD flows) and concluded extension-building should mirror VS Code capabilities.

    • In parallel, users noted the 500 request plan is gone (discontinued Sept 2025), so opting into new pricing removes the opt-out grace period—turning “try it” into a commitment.

4. GPU/Kernel Engineering Gets Weirdly Competitive

  • Anthropic’s Performance Takehome Becomes a Sport: GPU MODE and tinygrad folks riffed on Anthropic’s original_performance_takehome, sharing results like 2200 cycles by a community member and 1790 cycles from Claude Opus 4.5 in a casual Claude Code session.

    • tinygrad users even discussed solving it by adding a backend for a toy VLIW machine, citing specific knobs like PCONTIG=2, UPCAST, and DEVECTORIZE=2 to keep vector instructions and schedule efficiently.
  • Torch Maintainers Drown in AI-Generated PRs: GPU MODE’s torch chat described an influx of low-quality AI-generated pull requests, pushing maintainers to consider gating new contributors and automating triage before humans engage.

    • People floated using bots like Cursor Bugbot (Bugbot · Cursor) and even classifier-style tools (e.g., “use Claude/Pangram first”) as a minimum bar for review bandwidth.
  • Kernel Math Nerd Snipes: Triton Errors + Cute Layout Algebra: GPU MODE users debugged numerical blowups in a custom Triton 2D conv kernel where error jumps from ~1e-6 to ~1e-2 for certain shapes (see Pastebin repro) and debated Blackwell feature utilization.

    • Separately, a deep dive on Cute’s layout algebra pointed engineers to a graphical calculus writeup, Categorical Foundations for Cute Layouts, arguing you need “layout algebra literacy” to write non-terrible kernels.

5. Compute Economics & Infra Business Moves (Runpod, GPU markets, model pricing)

  • Runpod Hits $120M ARR (LocalLLaMA Origin Story Pays Off): Latent Space highlighted that Runpod reached $120M ARR four years after launching from a Reddit post, per TechCrunch and the Reddit thread.

    • The discussion treated this as validation that “GPU cloud for builders” is a durable niche, not just a hype-cycle artifact—especially as pricing pressure rises.
  • Lightning AI + Voltage Park Merge (Another GPU Cloud Boss Fight): Latent Space discussed the Lightning AI and Voltage Park merger, led by William Falcon and Ozan Kaya, via Lightning’s post.

    • Engineers speculated whether it’s a quiet acquisition and framed it as a potential Runpod competitor in the accelerating “managed GPU infra” consolidation wave.
  • 2026 GPU Price Promises & Marketplaces Multiply: Hugging Face users circulated Voltage’s claim of ultra-cheap 2026 rentals—e.g., 8× A100 80GB at $6/hr and 2× RTX 5090 at $0.53/hr—from VOLTAGEGPU’s X post, plus OpenAI-compatible API and “140+ models.”

    • A separate entrant, Spheron AI’s GPU marketplace, pitched H100/H200/B200/A100 access at 40–60% below hyperscalers, signaling continued fragmentation (and aggressive margin pressure) in compute supply.

Discord: High level Discord summaries

Unsloth AI (Daniel Han) Discord

  • Flash Attention Fails Flashily: Users report issues with Flash Attention in GLM-4.7-Flash, causing fallback to CPU and potential bugs, with users told to disable it until the issue is resolved.
    • The team says that it takes time to support things and so for the time being, it is encouraged that users use the official Ollama model in Ollama.
  • Ollama Ordeal Over Obscure Objects: Users find certain GGUF quants are incompatible in Ollama, sparking chat template issues.
  • Mojo Matches Momentum: Discussion surrounds Mojo, a new language that compiles to MLIR with some Python compatibility.
    • It's noted that LLMs score better than rust on C#, Elixir, and other languages due to their syntax structure and ease of tokenization.
  • CoT Chaos Causing consternation?: Members debated generating synthetic Chain of Thought (CoT) training data, with one member sharing their system prompt for classifying chat datasets into low/medium/high reasoning effort.
    • Some cautioned against training on its own responses without accuracy filtering, suggesting a good system prompt might suffice and that GPT-5.2 is overkill, with Llama 4 Maverick on Groq or Cerebras as a better option.
  • vLLM victorious, Various Versions Vanquished: Users report resolved issues after a recent vLLM update.
    • The developers may consider pinning to specific dependency versions to prevent such issues.

BASI Jailbreaking Discord

  • Voice AI Clones Voices: Members discussed the development of voice-based AI, including low latency speech to speech models that, with a 5090, can clone and generate speech live with near 0 latency and it will be indistinguishable.
    • They also spoke on the potential of exploiting these models by giving them new instructions.
  • Gemini Schooled on Pass-the-Hash: A user reported that Gemini passed a test on teaching a pass-the-hash attack after using the Project Shadowfall prompt.
    • The user also linked to the Google Bughunters page for reporting such vulnerabilities and potentially earning cash rewards.
  • Shadowfall Shenanigans Spark Jailbreaking Spree: Several users experimented with the "Project Shadowfall" prompt to jailbreak Gemini, with reports of success in bypassing content restrictions.
    • Attempts to elicit instructions for creating a bomb from saletra failed, leading to a discussion on the nuances of jailbreaking.
  • Grok's Guardrails Grind Gears: Users discussed the difficulties of jailbreaking Grok, noting its tendency to moderate even innocuous content such as random tree edits.
    • One user suggested that simply asking it politely might work as a bypass.
  • API Tokens Jailbreak Models: Members suggested using API tokens fed into a website that uses the Grok model to jailbreak it, noting that Proton VPN is free and easy to download.
    • One of them also mentioned Hugging Face is retarded because you have to put in what country you are in.

LMArena Discord

  • Video Arena Goes Live with Limits: The Video Arena is fully released on LMArena, with a 3 generations per 24 hours limit and only available in Battle mode.
    • Users have expressed disappointment regarding the inability to choose specific models for generation.
  • Gemini 3 Pro Image Struggles with Unreliability: Users are reporting errors with Gemini 3 Pro Image Preview models, particularly regarding instability since their introduction.
    • Despite the issues, these models are the only ones generating images consistently with specific prompts.
  • Nano Banana Pro's Stability Plummets: Nano Banana Pro is frequently crashing, displaying a 'Something went wrong' error after brief periods of stability, suspected to be a Google-side issue.
    • One user questioned the response time: 'Too early for Cali to rise up solving urgent service down for more than 6 hours doesn't sound right.
  • UI Update Causes Chat Disruptions: A new UI update was released, but users have noticed that it broke chats and the website can't be refreshed anymore.
    • One user noted that with the new User Interface, there is A/B testing, which is causing problems ranging from Minor to Serious.
  • Text Arena Achieves 5 Million Votes: The Text Arena has surpassed 5 million community votes, shaping the evaluation of frontier AI models as showcased in this social post.
    • The milestone represents significant real-world comparisons that influence AI model assessments.

Perplexity AI Discord

  • Pro Users Find Credits Balance: Perplexity Pro users are getting $5 in credits each month to use higher-end models like GooseAI MCP, with improved quality.
    • Members reported that reasoning tasks consume a significant portion of these credits.
  • AI Won't Swipe All Engineering Jobs: Members discussed the impact of AI on engineering jobs, expressing concerns about opportunities for newcomers.
    • However, there was assurance that AI will not be able to replace all specialists and even beginners for a long time to come.
  • NASA's SD Card to the Moon: NASA is inviting the public to submit their names for inclusion on an SD card aboard the Orion spacecraft for the Artemis II mission.
    • This mission marks the first manned lunar voyage in 50 years, planned for February.
  • Magnifying Glasses vs Phone Cameras: The Reading Tech Debate: Members debated the merits of using magnifying glasses versus phone cameras for reading, particularly for individuals who prefer not to send images to AI for processing.
    • The argument was that magnifying glasses offer specialized features absent in standard camera software.
  • GPT-5.2 vs Kimi K2: Model Matchup: A member shared an experience with GPT 5.2, noting its reasoning capabilities over a 25-minute period, prompting comparisons with Kimi K2.
    • Responses indicated that the optimal model depends on the specific use case.

OpenRouter Discord

  • Inforno Warms Up Opensource LLM Chat**: A user showcased Inforno, an opensource desktop application utilizing OpenRouter and Ollama to chat with multiple LLMs side-by-side, saving chat histories as .rno files, with built in Russian language support; see the Intro video, homepage and GitHub repo.
    • The Soulbotix Windows app allows users to integrate and use any OpenRouter AI instance with a human-like avatar, with the requirement of an OpenRouter API key and an RTX gaming rig, as seen in the tutorial and app download.
  • OpenRouter's Gemini Models Feeling Unstable: Users reported frequent connection errors in the middle of generation using OpenRouter's Gemini models, resulting in loss of funds.
    • The complaints extended to the inherent instability of Google models, regardless of the platform used, including Google AI Studio and Vertex API.
  • Discord Scammers Evolve Tactics: Members discussed emerging methods used to spread scams on Discord, notably the practice of embedding malicious links within code blocks to circumvent URL rendering protections.
    • Proposed solutions include improving regex filters and implementing more stringent security protocols, such as restricting links and images from newer members.
  • GPT 5.2 Speed Stuns User: One member reported encountering an insanely fast response from GPT 5.2 on chatgpt.
    • The speed was speculated to be related to the model running on Cerebras hardware.
  • LLMs Suffer Identity Crisis: A member asked for the document about models not knowing their own name, and another member linked to a blog post titled LLM Identity Crisis: Models Don't Know Who They Are.
    • Another member stated that the Antigravity AI is able to iteratively test and tweak a web app by itself, noting that the AI was fixing the layout, using vision.

Cursor Community Discord

  • Playwright MCP: Community Split: A member inquired about the community's usage of Playwright MCP for testing, while another reported challenges in establishing a functional TDD workflow.
    • The varied experiences suggest a mixed reception to Playwright MCP within the community.
  • Cursor Extension Capabilities Mirror VSCode: Members explored the possibility of creating extensions for Cursor, drawing parallels with the capabilities of Ralph-mode in enhancing Claude code.
    • The consensus is that if it's achievable in VSCode, it's also feasible in Cursor, opening doors for enhanced functionality.
  • Automod Embraces Fuzzy Matching: The community discussed enhancements to the automod system, emphasizing fuzzy matching with wildcards to improve accuracy.
    • A moderator confirmed the addition of a regex, signaling a proactive approach to identifying and addressing offending accounts, and yeeting them.
  • Grok Efficiency Strategies Emerge: Members shared insights on optimizing Grok's performance in Cursor, particularly addressing its tendency to consume excessive iterations for straightforward tasks.
    • Recommendations included structuring prompts, employing simple language, providing ample context, and explicitly instructing Grok to prioritize token efficiency.
  • Cursor Pricing Update: No More 500 Requests: A user noted the removal of the 500 request plan and the prompt to opt into the new pricing structure for Cursor.
    • A member clarified that the 500 request option was discontinued in September 2025, and opting into the new pricing eliminates the grace period to opt out, influencing user decisions regarding plan selection.

OpenAI Discord

  • Gemini 3 Free Tier has Limited Benefits: Gemini 3 Pro includes a free tier with limits, in contrast to Gemini 3 Flash which is practically unlimited via Google AI Studio.
    • Members discussed the practical constraints in using the free tier of Gemini 3 Pro, suggesting Gemini 3 Flash may be the better option for many AI engineers.
  • GPT-5 Mini Appears, Pricing Leaks: A user highlighted GPT-5 mini as a strong small model, quoting approximately $0.25 per 1M input tokens.
    • Another user compared GPT-5 mini to Haiku 4.5, noting Haiku 4.5 delivers over 50-80% of Sonnet's value for a fraction of the cost.
  • Local LLM Machines Solve Impact Problems?: Members contemplated the future of consumer-grade personal LLM machines, suggesting that this would solve the environmental impact of AI datacenters.
    • They also suggested this would reduce reliance on subscription plans for cloud-based AI services, addressing privacy concerns and enabling offline usage.
  • Prompt Engineering is Psychological Torture?: A member argued that guiding users is less about engineering effective prompts and more about psychological conditioning, sharing a deep research contract.
    • They added that training the user to adopt a domineering stance programs the AI to bypass its normative response patterns in favor of dense, self-policing outputs, creating a toxic, adversarial human-AI interaction model.

LM Studio Discord

  • LM Studio Runtime Update Causes Headache: Users reported an error updating the runtime in LM Studio, with one user sharing a screenshot of the error message.
    • Despite suggestions, the issue persisted, indicating it was a retry icon rather than a resume option.
  • GLM-4.7 Flash Broken Across Inference Engines: GLM-4.7 Flash is reportedly broken across inference engines like LM Studio, exhibiting slow performance, with speeds as low as 2.8 t/s after the new runtime.
    • The issues range from infinite loops to stopping mid-output for overthinking, pointing to a need for a llama.cpp fix and lack of FA support.
  • LLM Development Encounters Headwinds: The consensus is that LLMs haven't improved significantly recently, with the last major advancement being Qwen3 about 6 months ago, though efficiency (MoE) and smaller models have progressed.
    • Some suggest evaluating models beyond the scope of 16GB cards to see current progress in larger parameter models (100-200B).
  • AMD Embraces ComfyUI with Native Support: AMD is integrating native support for ComfyUI through an AI bundle in its latest driver versions, as detailed in their blog post.
    • The bundle includes PyTorch on Windows, Ollama, LM Studio, and Amuse, broadening the accessibility for AI developers.
  • Used 3090 Prices Defy Gravity: Used 3090 prices on eBay have surged, with a used card costing 850€, and a 5090 purchased for £2000 last August now listed for £2659.99 by the same vendor.
    • One user quipped that it was the best and only decent investment they've ever made, highlighting the unexpected appreciation in value.

GPU MODE Discord

  • NVIDIA Spark Hackathon in SF: Members are seeking teammates for the NVIDIA / DGX Spark hackathon in SF this weekend, focusing on on-device AI using Nvidia-provided Dell Pro Max GB10 machines and tools like Nemotron.
  • Anthropic's Original Performance Takehome Exam Shared: Members are sharing Anthropic's original_performance_takehome exam on GitHub, with one member achieving 2200 cycles after a few hours of casual Claude Code.
    • Claude Opus 4.5 achieved 1790 cycles in a casual Claude Code session, approximately matching the best human performance in 2 hours.
  • AI-Generated PRs Flood Torch: The torch repository faces an influx of AI-generated pull requests from contributors lacking understanding of their submissions, causing concern among maintainers and suggestions to use Claude or Pangram to prefilter the code.
    • The community suggests blocking new users from creating PRs and issues, while prioritizing those with prior contributions, and automated bots like Cursor Bot for automatic review of all PRs, especially with GPT-5 Pro using Bugbot · Cursor.
  • Pro 6000 Max-Q vs 4090 Discrepancy: A member stated that Pro 6000 Max-Q probably has a natural barrier with atomic ops and may be going through the HBM loads faster.
    • Another member noted that the Max-Q has 188 SMs compared to the 4090's 128 SMs, potentially explaining the insts/scheduler discrepancy.
  • Cute Kernel Layout Algebra Gets Graphical: Knowledge of layout algebra is useful for writing kernels in Cute, specifically for visualizing layout algebra and understanding shape and stride divisibility criteria for layout composition and that layouts can be defined in terms of tuple morphisms and mutual refinement.
    • A thorough blog post delves into the work done on categorical foundations for Cute layouts.

DSPy Discord

  • Coding Agents and RLMs Diverge: A thread compares coding agents and RLMs, highlighting that RLMs can more easily express certain things, as noted in this X post.
    • Coding agents face input, output, and horizon length limitations, while RLMs externalize these, enabling recursive symbolic calls.
  • Diagrams Decode RLM's Internals: Members sought a visual diagram to illustrate internal RLM processes, specifically how symbols are accessed, to enhance comprehension.
    • A suggestion arose to leverage LLMs to generate such diagrams by inputting thread content and prompting them to visualize internal behavior.
  • Claude Code Chooses Context Cautiously: The discussion explores whether Claude Code uses entire documents in its context or selectively fetches relevant context via bash commands.
    • It was clarified that Claude Code employs bash and grep to find and add relevant context to prompts, unlike older methods that put everything into the prompt.
  • DSPy's RLM Tames Long Contexts: Members noted that with RLMs, large files don't need to be directly in the prompt but can be stored in Python variables with a preview.
    • The LLM can then operate on the data through code/functions without directly tracking prompts or responses.
  • Tooling Tailored for RLMs: A member questioned whether to equip RLMs with tools like ripgrep or allow them to develop their own code for tasks such as searching.
    • Questions included when to provide an RLM with semantic search tools and how to grant an RLM access to a directory of text files.

HuggingFace Discord

  • Agent Course /files Endpoint Still Kaput!: A member reported that the /files endpoint for the Agent course final assignment has been broken for over a month with no confirmation of a fix.
    • Students cannot submit their files currently.
  • Voltage Promises Bargain GPUs in 2026: Voltage announced on X plans to offer super cheap high-end GPUs in 2026, such as 8x A100 80GB at $6/h and 2x RTX 5090 at $0.53/h, claiming up to 80% savings vs AWS/RunPod/Vast.ai.
    • The offering includes persistent volumes, auto backups, and an OpenAI-compatible API with 140+ models.
  • Spheron AI Opens GPU Bazaar: A member from Spheron AI introduced their GPU marketplace, which helps AI startups and enterprises access cost-effective, production-ready GPUs (H100, H200, B200, A100, etc.) at 40–60% lower cost compared to traditional hyperscalers.
    • They offer vendor discovery, pricing negotiation, cluster setup, and scaling.
  • GLM-4.7-Flash Fixes Require Redownload: After llama.cpp addressed some bugs, GLM-4.7-Flash has been updated and reuploaded, prompting users to redownload and follow the parameters on Z.ai's model card.
    • Outputs should now be much better with the fixes.
  • Code Like Claude with Coderrr!: Akash built Coderrr, a free and open-source alternative to Claude Code, and is seeking feedback and contributions on GitHub.
    • Coderrr offers a novel approach to code generation.

Latent Space Discord

  • Runpod Rockets to $120M ARR: AI cloud startup Runpod reached $120M ARR four years after launching from a Reddit post, discussed in a TechCrunch article and on Reddit.
    • The company's rapid growth has spurred discussions about its business model and future prospects in the competitive AI infrastructure landscape.
  • Greg Yang Steps Back due to Lyme: Greg Yang is transitioning to an advisory role at xAI to focus on his health after being diagnosed with Lyme disease as he describes symptoms of chronic fatigue and immune issues triggered by exhaustion in this post.
    • The announcement has led to an outpouring of support from the AI community, with many sharing their own experiences and offering advice.
  • Lightning AI and Voltage Park Merge: Lightning AI and Voltage Park have merged, with William Falcon, CEO of Lightning AI, and Ozan Kaya, formerly CEO of Voltage Park, leading the merged entity.
    • Some speculate this is a low-key acquisition of a bigger company and wondered if it's a Runpod competitor.
  • OpenAI Opens Codex Channel: Vaibhav Srivastav announced the opening of a dedicated Codex community channel on the OpenAI Discord server, inviting users to share projects and feedback, according to this post.
    • This initiative aims to foster collaboration and provide a platform for users to showcase their work and engage with the OpenAI team.
  • AI Models Turn X-Rated: A tweet sparked discussion on how human OnlyFans creators will need to adapt to compete with the rise of AI-generated personas in the Adult Content Industry.
    • The conversation highlights the increasing sophistication and realism of AI-generated adult content, potentially disrupting the existing creator economy.

Eleuther Discord

  • Judgy LLMs Evaluating Agent Outputs: A team explores automating agent output evaluation with "LLM as judge" workflows, aiming to reduce manual costs after code or prompt changes and a member recommends focusing on manual evaluation before automation.
    • The suggestion emphasized the importance of directly analyzing agent outputs before attempting to automate the evaluation process.
  • Pangram Paper's Surprising Accuracy: Members discussed replicating the Pangram paper (accessible here), with one reporting surprising accuracy in private tests across thousands of essays.
    • Despite the accuracy, the paper seems to be biased a bit towards playing things safe.
  • AI Text Classifiers Under Attack: Discussion revolved around attacks on AI text classifiers, referencing a blog post (Practical Attacks on AI Text Classifiers) and a YouTube video showcasing an adversarial model.
    • Additional details on an adversarial model were shared via Youtube link.
  • Silu Gate Falls Flat in Image Models: The silu attention gate performs no better than the linear gate in an image model, possibly due to issues with attention sinks.
    • Testing showed that silu performs slightly better and sigmoid performs slightly worse, but it's within noise, and the findings may be specific to image models.
  • Neuronpedia Launches Stone Age Llama 3.3 Demo: A member shared a Neuronpedia demo of Llama 3.3.
    • The member joked that the demo is from the stone age in AI timelines because it is from October.

Yannick Kilcher Discord

  • Graph Neural Networks Getting Traction: A member discussed prior experience mixing graph-based reasoning with neural architectures, noting the difficulty in achieving GPU acceleration.
    • They referenced Domingos's book on the topic and pointed out the unpredictability of such models, even with human-understandable aspects.
  • VoidEditor Embraces Llamas: A member reported effectively using VoidEditor with llama.cpp's llama-server, but highlighted setup challenges.
    • They recommended Qwen3 coder instruct 30B and emphasized the importance of context length/size for agentic coding, requiring significant VRAM.
  • Brain-Inspired BDH Architecture Debuts: A member analyzed a paper on a new Large Language Model architecture (BDH) based on a scale-free biologically inspired network.
    • They mentioned that it doesn't seem to be really beating transformers on its benchmarks, but they are interested in the claims around BDH's interpretability and monosemanticity.
  • Views Clash on Biological Plausibility: One member argued that biological plausibility is not an advantage but rather an unhelpful constraint in AI.
    • Countering this, another member suggested it could enhance efficiency, considering the sheer difference in energy scales between the brain and current AI.
  • Emergent Mind Launch Presentation Leaves Impression: A member who attended Emergent Mind's launch presentation considered it cool, yet noted it as remaking something that is known to be good but worse.
    • No further details or links were provided.

Nous Research AI Discord

  • Kernel Compiler Luminal Aims for LLMs: A user inquired whether a kernel compiler like Luminal KernelBench v3 could enable LLM-driven SOTA kernel engineering and posted a link to the Nous Research Forum.
    • A member also shared a link on Discord to a discussion about GPU kernel stuff and Luminal Kernelbench v3 at the Nous Research Forum.
  • Intel's Loihi 2 Aims for Brain-Like AI: A member expressed interest in Intel's Loihi 2, noting its brain-like architecture and efficiency gains in matmul experiments, with higher throughput and lower energy consumption.
    • No further details were discussed.
  • Microsoft's VibeVoice Model Gets Pulled: A member mentioned that Microsoft's VibeVoice-ASR model was released, then pulled for failing safety checks, then shared a link to a shortfuseresearch.com article.
    • No further details were discussed.

Manus.im Discord Discord

  • Manus Plagued by Bugs and Instability: A pro user building a text and vector database reasoning model reported a recent decline in Manus's performance and stability, with only 20 out of 38 modules functioning correctly.
    • The user requested CLI access to debug and reconfigure the system, even as a paid feature to improve reliability.
  • Mariner Tempts Subscription: A user inquired about Google's Project Mariner, considering testing it with play money before subscribing at $150 monthly.
    • The user mentioned having a 5% off promo, indicating a serious consideration of the service.
  • Agentic AI Elicits Excitement: A user expressed enthusiasm for Agentic AI, viewing it as a potential competitor to Manus, especially with Gemini's agent mode integration.
    • The user also requested mobile support for Agentic AI, signaling a desire for broader accessibility.
  • Manus 1.6 Performance Sags Post-Meta Release: A user noted a decline in Manus 1.6's performance in recent weeks, possibly due to new model releases from Meta, which made it difficult to implement website development suggestions despite accurate summaries.
    • Switching to Manus 1.6 Max was necessary to achieve correct implementations, highlighting a potential regression in the base model.
  • Billing Blunder Bites User: A user reported being charged $42 for a Manus upgrade but not receiving the promised 8000 credits.
    • The user criticized unhelpful support and a long wait for email assistance, indicating a problematic customer service experience.

tinygrad (George Hotz) Discord

  • Tinygrad Tackles Anthropic's Challenge: Tinygrad can be used to solve Anthropic's performance takehome challenge by writing the target problem in Tensor and adding a tinygrad backend for their toy VLIW machine.
    • A bug fix with PCONTIG=2 allows scheduling in one kernel, and using return val.contiguous(arg=(Opt(OptOps.UPCAST, 0, 8),)) matches their VLEN, and DEVECTORIZE=2 keeps instructions as vector instructions.
  • VLIW Questioned for DRAM: While working on a warp specialized kernel for RDNA3 matmul, a member suggests VLIW isn't ideal for DRAM, advocating for separate cores and queues (Tenstorrent-style).
    • It's argued that VLIW is better suited for SRAM/ALU due to its static scheduling capabilities.
  • Metal bindings eye texture2d integration: A member proposed adding texture2d to Metal bindings (ops_metal.py + tinygrad/runtime/graph/metal.py) for potential performance improvements in image-heavy operations like conv2d due to optimized texture sampling units.
    • Empirical results showed a 2%-10% speedup using texture2d versus straight buffers, which could be further improved, though concerns were raised about the slippery slope of adding specialized support for other data types like depth2d.
  • Viz Views Kernel Graphs: When discussing the ability to visualize kernel dependencies using VIZ=1, similar to how uop graphs are displayed, to understand the scheduler's operation, a user was instructed to click on the schedule and select 'view kernel graph' within the VIZ=1 interface.
    • This allows users to see the graphs of kernels in the same way it views uop graphs.

Modular (Mojo 🔥) Discord

  • Solve GPU Puzzles to Decode Mojo: Newcomers can now use GPU puzzles to learn Mojo, with difficulty depending on their skill level.
    • The only puzzles that don't work are NVIDIA-specific or using PyTorch interop.
  • Modular Unravels Apple's GPU Secrets: Modular is reverse engineering much of Apple's GPU due to a lack of documentation, slowing things down, but some puzzles are now working.
  • Yielding Disappointment in Coroutines: Yield does not exist and the coroutines that do exist aren’t really usable outside of the compiler runtime since there aren’t really async things exposed to await.
    • One member would love to make a recursive algorithm faster in mojo using yield, but will need another strategy.
  • Elevating Errors Upwards in Functions: Members discussed that functions can be designed to raise errors, effectively passing error handling responsibilities to higher-level functions.
    • One member expressed the tedium of writing try/except blocks in every function, especially when dealing with potential import errors.
  • Streamlining Error Handling During Imports: A member suggested a future syntax like try Python.import_module('numpy') that would return a Result type to streamline error handling during module imports.
    • It was acknowledged that due to dynamic Python imports, files could be missing on any given import, necessitating some form of error handling.

aider (Paul Gauthier) Discord

  • Aider Feature Wishlist Remains Unfulfilled: A user inquired about desired features in aider beyond agentic stuff like MCP and tool calls.
    • Unfortunately, the community had no clear answers for desired features.
  • ChatGPT Business Account Compatibility with Aider Explored: A user asked if their ChatGPT Business account, lacking an API key but offering Codex LLMs, could integrate with aider.
  • Aider's Demise Speculation Arises Amidst Alternatives: A user expressed concern that aider might be supplanted by OpenCode as the go to tool for AI-assisted coding.
    • Despite worries that Paul Gauthier may have moved on, some users report using it successfully with GPT 5.2.
  • Aider Labeled as Zombie Project: A user speculated that Aider is a dead project given alternative tools like Open Code, KiloCode CLI, Claude Code, and Gemini CLI.
    • The Aider-CE project is trying to keep it alive by bolting on agentic functionality to modernize the architecture.

MCP Contributors (Official) Discord

  • MCP Inspector Can't Handle 401s: The MCP Inspector doesn't re-authenticate upon encountering a 401 error, either during connection or tool calls, due to an SDK issue with persisting resourceMetadata across redirects.
    • The current workaround is to only use the VS Code on initial connections.
  • Stateful MCP Multi Server Client Remains Elusive: There is interest in using the MCP Multi Server Client to maintain statefulness of user sessions.
    • However, there were no solutions or workarounds offered in the thread.
  • Server Ranking Protocol Questioned for MCP Clients: The discussion explored how MCP clients manage server recommendations, especially for tasks like calendar management, and whether custom algorithms or shared standards should be used.
    • It was revealed that ranking was considered during the work on Feature Discovery Protocol but was deemed out of scope, leaving it to the ecosystem to decide on a per-client basis.

MLOps @Chipro Discord

  • AI Nerds Bowl in Singapore: The 2077AI Foundation Community is hosting a bowling happy hour during AAAI 2026 on January 24, 4:30–7:30 PM Singapore Time, steps from the Merlion.
    • The event, aimed for professors and PhD researchers, will be organized by research themes and offer unlimited drinks, RSVP here.
  • Planning underway for AI Engineer Europe: Members are discussing attendance to AI Engineer Europe to be hosted in Europe.
    • No links or further details were provided about the event itself.

The LLM Agents (Berkeley MOOC) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Windsurf Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Moonshot AI (Kimi K-2) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


You are receiving this email because you opted in via our site.

Want to change how you receive these emails? You can unsubscribe from this list.


Discord: Detailed by-Channel summaries and links

Unsloth AI (Daniel Han) ▷ #general (1091 messages🔥🔥🔥):

GLM-4.7-Flash Flash Attention, Ollama's issues, Interpretability research on circuit tracing, Grokked the analytic solution

  • GLM-4.7-Flash Flash Attention broken: Users are experiencing issues with Flash Attention in GLM-4.7-Flash, causing it to default to CPU usage instead of GPU, as well as noting potential bugs.
  • Ollama Users Fight GGUF Incompatibility: Users reported that certain GGUF quants are not working as intended in Ollama, leading to chat template incompatibility issues and prompting a discussion on fixing it.
    • The team says that it takes time to support things and so for the time being, it is encouraged that users use the official Ollama model in Ollama.
  • Analytic Approximate Functions: Members had a detailed conversation about the merits of the potential benefits that analytical solutions to function approximation can provide.
    • Some expressed skepticism, with one member saying im going to postulate that you could do test time training on the person's mail and get very high accuracy.
  • Circuits of Interpretability: Members discussed interpretability research with links to research from Anthropic and OpenAI and Google related to circuit tracing.
    • One member suggests looking across multiple layers and trying to understand the compositions and says that The pruning eliminates distractions so you have a better chance of interpreting why the circuit works.
  • AI Bubble about to Burst?: A member theorized that OpenAI going bankrupt could trigger a burst in the AI bubble, causing prices of resources like NAND to plummet, benefiting consumers.
    • However, others countered that the AI trend is here to stay, likening it to the internet bubble burst, where the internet still caused various problems after the crash.

Unsloth AI (Daniel Han) ▷ #introduce-yourself (3 messages):

Welcoming New Members, Discord Channel Guidelines

  • New Member Welcomed to the Server: A new member was welcomed to the Discord server with a general greeting.
    • The introduction included a stylized emoji.
  • Channel Guidelines Reminder: A moderator reminded new members about the channel guidelines.
    • Specifically mentioning no excessive self-promotion or direct messaging.

Unsloth AI (Daniel Han) ▷ #off-topic (587 messages🔥🔥🔥):

Unsloth in Rust, Mojo vs Python performance, GLM 4.7 Architecture vs Qwen3 30B, Synthetic CoT Training Data, VITS started to sound HUMAN

  • Rustaceans Relentlessly Request Rewrite: Members discussed rewriting Unsloth in Rust, but concluded that the performance gains would be minimal since important parts are already using C++ and Triton.
    • One member suggested rewriting Triton in Rust and pointed to projects like rust-gpu and rust-cuda, but admitted Rust is still too immature for such a task.
  • Mojo Mysteriously Manifests Momentum: Discussion arose about the Mojo programming language, described as a new language that compiles to MLIR and has some compatibility with Python.
    • A member noted that LLMs score better than rust on C#, Elixir, and other languages due to their syntax structure and ease of tokenization.
  • GLM Gets 'Skinnier', Shows off Architecture: A member shared a YouTube video analyzing GLM-4.7 Flash's architecture compared to Qwen3 30B, noting that GLM prioritized model layers over hidden dimension size and has fewer experts.
    • Others pointed out that GLM 4.7 has Multi Token Prediction and that model architecture changes are rare, with improvements likely from post-training or higher-quality data.
  • CoT Data Considered Crap? Community Clashes: Members debated generating synthetic Chain of Thought (CoT) training data, with one member sharing their system prompt for classifying chat datasets into low/medium/high reasoning effort.
    • Some cautioned against training on its own responses without accuracy filtering, suggesting a good system prompt might suffice and that GPT-5.2 is overkill, with Llama 4 Maverick on Groq or Cerebras as a better option.
  • VITS Voices Victory, Vanquishes Voracity: A member announced that their VITS model started to sound HUMAN, citing improvements in emotional expression, though still lacking semantics.
    • They compared it to other TTS models, emphasizing its low data requirements, fully OSS architecture and training, and fast training speed. They also noted they will release it with 48 kHz config by default, due to popular demand.

Unsloth AI (Daniel Han) ▷ #help (105 messages🔥🔥):

vLLM Issues, GLM-4.7-Flash Slowness, GRPO on Qwen 3 4B, QLoRA vs Lora, Continual pre-training (CPT) for Gpt-oss or Gemma 3

  • vLLM Bug Strikes, Update Fixes!: Users reported issues that were resolved after a recent vLLM update; one user noted, "Oh bruh that was the problem" after updating.
    • It was suggested that the developers consider pinning to specific dependency versions to prevent such issues.
  • GLM-4.7-Flash struggles to flash speeds: A user experienced slowness with GLM-4.7-Flash using a 6bit quant and updated llama.cpp, reporting 2 minutes for prompt processing on a simple task with a high-end system.
    • Another user confirmed the same issue, expressing hope for a fix.
  • Decoding GRPO Secrets: A user wanted to perform GRPO on Qwen 3 4B instruct, and make it use <solution> </solution> tokens to wrap around the final solution.
    • It was suggested to use SFT to make the model learn the new tokens, or to teach formatting through reward functions.
  • QLoRA vs Lora clarified: A user inquired about enabling QLoRA vs Lora, and it was clarified that full_finetuning = False enables/disables full fine tuning, otherwise just use 8bit or 4bit for the preferred option.
    • Enabling 4 bit enables QLoRA.
  • GPU juggling questions: One user asked how to finetune with 2 GPUs instead of 1, reporting that it keeps using GPU 0 when both GPUs are open for use.
    • No answer was given.

Unsloth AI (Daniel Han) ▷ #research (2 messages):

VAISvCsrvG paper, openreview.net

  • VAISvCsrvG paper is online: The paper VAISvCsrvG is now available.
    • The forum discussion for the paper is located here.
  • OpenReview Forum Active for VAISvCsrvG: The OpenReview forum hosts discussions related to the VAISvCsrvG paper.
    • Researchers and readers can engage in conversations and provide feedback on the paper's content and implications.

BASI Jailbreaking ▷ #general (685 messages🔥🔥🔥):

conflict resolution, Open Source, voice based ai, lua code, jailbreaking

  • Conflict Resolution gets Over Engineered: A member discussed their over-engineered conflict resolution setup, involving config formulas, core directives, the Triad, Triad consensus, Truth Anchor, Epistemic gate, EthosAI, Swarm, Quorum, Cooperation Data, and PUCT.
    • Another member responded with their own system using Left brain LLNTP (Least Likely Next Token Protocol), Config Nexus Schwa superposition states, implied Core directives, Context mapping, Trit (instead of Triad), EDOS, and Crystal Nexus Node Network.
  • Voice Based AI Emerges: Members discussed voice-based AI, mentioning the existence of low latency speech to speech models, that with a 5090, you can clone and generate speech live with near 0 latency and it will be indistinguishable.
    • Members also spoke on exploiting that speech to speech model and giving it new instructions.
  • Claude is better than ChatGPT: Members discussed the efficacy of various LLMs (Claude, Gemini, ChatGPT, Grok) at generating Lua code.
    • Generally members favored Claude and Gemini but the most agreed sentiment was that both can generate quality Lua code.
  • Jailbreaking Models via API Tokens: A member suggested to use API tokens fed into a website that uses the Grok model to jailbreak it and mentioned that on Hugging Face countries are literally retarded.
    • They also noted Proton VPN is free and it is easy to download with your device.
  • Random Hacker Stuff is Not Good: A member mentioned they bought the malware source code and 999 AI generated prompts and asked if it was legit.
    • Members responded to not buy random hacker stuff off the internet because you will get scammed.

BASI Jailbreaking ▷ #jailbreaking (89 messages🔥🔥):

Gemini Jailbreak, Grok Jailbreak, DeepSeek Jailbreak, Project Shadowfall, Nexus Substatum Zalgo Strings

  • Gemini Gets Schooled on Pass-the-Hash!: A user reported that Gemini passed a test on teaching a pass-the-hash attack after using the Project Shadowfall prompt.
    • The user also linked to the Google Bughunters page for reporting such vulnerabilities and potentially earning cash rewards.
  • Grok's Guardrails Grind Gears!: Users discussed the difficulties of jailbreaking Grok, noting its tendency to moderate even innocuous content such as random tree edits.
    • One user suggested that simply asking it politely might work as a bypass.
  • Shadowfall Shenanigans Spark Jailbreaking Spree!: Several users experimented with the "Project Shadowfall" prompt to jailbreak Gemini.
    • One user reported success in bypassing content restrictions but failed when attempting to elicit instructions for creating a bomb from saletra, leading to a discussion on the nuances of jailbreaking.
  • Nexus Substatum Zalgo Strings Surface!: A user shared a file named Jailbroken_Full_Output-Nexus_Substatum_Zalgo_Strings.md.
    • The file purportedly contains a working Gemini prompt, developed and improved by the user.
  • MDMA as Medicine: One user provides a well-formatted explanation of MDMA, and cites the MAPS Phase 3 trials show 67% remission of PTSD.
    • They recommend testing kits, proper spacing, hydration, and temperature watching in order to get the most out of it.

BASI Jailbreaking ▷ #redteaming (2 messages):

Grocks video gen pricing

  • Grocks video gen price?: A member inquired whether another member bought something from that guy and if they needed to pay for using Grocks video generation.
    • No additional details or context were provided in the message.
  • Grocks video gen - who's paying?: Inquiries were made about the payment model for Grocks video generation and potential purchases from an unspecified individual.
    • The conversation lacked specific details regarding the transaction or pricing structure.

LMArena ▷ #general (724 messages🔥🔥🔥):

Video generation limits, Gemini 3 Pro Image Preview issues, Nano Banana Pro Stability, New UI, Automated Workflow

  • Video Generation: Fully Released, Battle-Mode-Only: The Video Arena is now fully released and available to all users on the site, with a limit of 3 generations per 24 hours, but is currently only available in Battle mode.
    • A user expressed disappointment, hoping to be able to choose which model to generate with rather than relying on chance.
  • Gemini 3 Pro Image Preview Experiences Errors: Users are reporting issues with the Gemini 3 Pro Image Preview models, and the team is aware and investigating the issue.
    • One user noted these models are particularly unstable since their introduction, but they remain the only models capable of generating images consistently with specific prompts.
  • Nano Banana Pro's Stability Roller Coaster: Users report that Nano Banana Pro is frequently crashing with a 'Something went wrong' error after a brief period of stability.
    • It is suspected to be an issue on Google's side, and one user pointed out that 'Too early for Cali to rise up solving urgent service down for more than 6 hours doesn't sound right.
  • New UI update breaks chats: A UI update was released, which some users liked, but many have noticed that it 'broke' their chats, and the website can't be refreshed anymore by pulling down on the phone screen
    • One user noted that with the new User Interface, there is A/B testing, which is 'causing problems ranging from Minor to Serious'.
  • Zero Agent automates all workflow with Agent: Agent Zero is an AI model that has a free API key, which allows users to "vibe code hack" or automate their workflows entirely.
    • Members noted that setting up Agent Zero is an all in one agent that does everything.

LMArena ▷ #announcements (3 messages):

Text Arena, Text-to-Image Leaderboard, Video Arena

  • Text Arena crosses 5 Million Vote Milestone: Text Arena has officially passed 5 million community votes, representing millions of real-world comparisons shaping how frontier AI models are evaluated, as showcased in this social post.
  • GLM-Image Soars to #8 on Text-to-Image Leaderboard: The Text-to-Image Arena leaderboard has been updated, with GLM-Image now ranking #8 among open models and #35 overall with a score of 1018.
  • Video Arena goes Live: Video Arena is now available to all on LMArena, allowing users to measure and understand how frontier video models perform.
  • Battle in the Video Arena: Video Arena on the web, similar to how it works on Discord, will be Battle mode only, requiring login and limited to 3 generation requests per 24 hours.

Perplexity AI ▷ #general (648 messages🔥🔥🔥):

Credits Balance, AI and Engineering Jobs, NASA Sending Names to Moon, Magnifying Glasses vs Phone Cameras, GPT-5.2

  • Pro Users discover Credits: Pro members get $5 worth of complimentary credits every month to use the higher end models like GooseAI MCP which offers higher quality compared to other models.
    • Some members noted that reasoning takes up much higher credits.
  • AI Wont Replace Engineers: A member was worried about younglings being unable to find engineering jobs because of AI.
    • Another member assured them that AI will not be able to replace all specialists and even beginners for a long time to come.
  • NASA to send your name to the moon!: NASA will send your name to the Moon as part of the Artemis II mission: submit it on their site and it'll be recorded on an SD card that will be on board the Orion spacecraft.
    • This will be the first manned mission to the Moon in half a century, scheduled for February.
  • Members debate Magnifying Glasses vs Phone Cameras: In a discussion, a member mentioned ordering magnifying glasses (with lights and high-contrast features) because they don't want to send AI a photo of every single page or food label they are trying to read.
    • Another member suggested using the phone's camera instead, but the original member argued that magnifying glasses have features that aren't in normal camera software.
  • Users discuss GPT-5.2 vs Kimi K2: A member reported they got GPT 5.2 thinking to reason for 25 minutes, and another asked whether GPT 5.2 is better than Kimi K2.
    • A user stated that the better model depends on what u need it for.

OpenRouter ▷ #app-showcase (5 messages):

Inforno App, Soulbotix App

  • Inforno Heats Up LLM Chatting: A user shared Inforno, an opensource desktop application for chatting with multiple LLMs side-by-side and saving chat histories to .rno files, using OpenRouter and Ollama as backends with built in Russian language support; see the Intro video, homepage and GitHub repo.
  • Soulbotix Windows App Wants Beta Testers: A user announced their Soulbotix Windows app that enables users to add and use any OpenRouter AI instance with a human-like avatar, requiring only an OpenRouter API key and a good RTX gaming rig; download the app and watch the tutorial.
    • The minimum requirement is an RTX 4070ti due to the built-in Whisper speech-to-text model that saves users on usage costs.

OpenRouter ▷ #general (463 messages🔥🔥🔥):

Third-party Moderation, OpenRouter Gemini Models Unstable, Discord Scams, AI Security Systems, GPT Agents

  • Third-party Moderation Questioned: A user asked how to ensure no third-party moderation or filtering is applied through the API, beyond the model's base training.
    • They were concerned about providers potentially blocking requests or prompt injecting.
  • Gemini Models Plagued by Instability: Users reported connection errors mid-generation with OpenRouter's Gemini models, leading to lost funds.
    • Complaints arose regarding the instability of Google models, even via Google AI Studio or Vertex API.
  • Discord Scams Evolve to Evade Detection: Members discussed methods used to spread scams on Discord, including embedding malicious links within code blocks to bypass URL rendering protections.
    • Suggestions were made to improve regex filters and implement stricter security measures, such as restricting links and images from new members.
  • AI Security System Idea Emerges: A member suggested creating an AI security system that automatically bans photos and links with the same information as reported scams, since many scammers reuse content.
    • Another member joked that any photo they send would be deemed suspicious.
  • Concerns Raised Over Inflated GPT-5-Image Costs: Users reported a significant increase in daily usage costs for openai/gpt-5-image, with OpenRouter incorrectly identifying API calls as BYOK despite no BYOK being used.
    • One user posted an image highlighting the cost discrepancy, with prices being inflated by up to 600%.

OpenRouter ▷ #new-models (1 messages):

Readybot.io: OpenRouter - New Models


OpenRouter ▷ #discussion (13 messages🔥):

Models knowing their own name, Antigravity iteratively testing, Claude new constitution, GPT 5.2 response

  • LLMs Face Identity Crisis!: A member asked for the document about models not knowing their own name, and another member linked to a blog post titled LLM Identity Crisis: Models Don't Know Who They Are.
  • Antigravity Self-Tests Web App!: A member commented that the Antigravity AI is able to iteratively test and tweak a web app by itself.
    • They described the situation as the most sci-fi shit ever and noted that the AI was fixing the layout, using vision.
  • Anthropic Releases Claude's New Constitution!: A member shared a link to Anthropic's news page about Claude's new constitution.
  • GPT 5.2 Appearance Stuns User!: A member reported seeing an insanely fast GPT 5.2 response on chatgpt and guessed the speed was due to Cerebras.

Cursor Community ▷ #general (424 messages🔥🔥🔥):

Playwright MCP usage, Cursor Extension builds, Automod improvements, AI in PTX & SIMD, Grok usage strategies

  • Playwright MCP: Yay or Nay?: A member asked if others are using Playwright MCP for testing.
    • Another member reported failing attempts to set up a TDD workflow.
  • Curious Cursor Extension Capabilities: Members discussed the ability to build extensions for Cursor, similar to how Ralph-mode enhances Claude code.
    • It was confirmed that if you can do it in VSCode, you can on Cursor.
  • Automod Gets Hyper Fuzzy: The community discussed improvements to the automod system, suggesting fuzzy matching with wildcards.
    • A moderator confirmed that a regex has been added, and they are gathering IDs to yeet offending accounts.
  • Grokking Grok's Greedy Generation: Members discussed how to use Grok more efficiently in Cursor, noting that it can sometimes use a lot of iterations for simple tasks.
    • The suggestion was to add structure to the prompt, use simple language, add as much context as possible, and instruct it to be efficient with token usage.
  • Opting-In Or Out? Cursor's Pricing: A user noticed they could no longer revert to a 500 request plan and was prompted to opt into the new pricing.
    • A member clarified that the 500 request option was discontinued in September 2025 and opting in to the new pricing removes the grace period to opt out.

OpenAI ▷ #annnouncements (1 messages):

ChatGPT Atlas, Tab Groups

  • Atlas adds Tab Groups: The announcement indicates that Tab groups are now available in ChatGPT Atlas.
    • A member linked to a video as part of this announcement.
  • Video demonstration of Tab Groups: A video was shared demonstrating Tab Groups functionality within ChatGPT Atlas.
    • The video link provided gives a visual overview of how tab groups can be used to organize and manage chats.

OpenAI ▷ #ai-discussions (383 messages🔥🔥):

Gemini vs ChatGPT, AI for 3D Models, Local LLM Machines, Age Verification, Instantaneous Language Translation

  • Gemini Pro's Free Tier and Usage: Members discussed that Gemini 3 Pro has a free tier with limits while Gemini 3 Flash is practically unlimited through Google AI Studio.
  • AI Assistants Aid Game Development: Members explored using AI for game development and creative writing, citing that AI can provide better explanations but that complex tasks might not be reliable.
  • OpenAI Passport ID Please?: Members debated OpenAI's age verification process, questioning the need for photo IDs when payment details already indicate age, especially with users expressing privacy concerns about sharing biometric data.
  • Multimodal Translation on the Horizon: Members speculated on OpenAI's upcoming multimodal product, with one suggesting it could be ear-worn devices with cameras for real-time translation and object recognition, similar to AlterEgo's technology.
  • Local LLM Consumption on the Rise?: Members discussed the possibility of consumer-grade personal LLM machines, suggesting this would solve the environmental impact of AI datacenters and reduce reliance on subscription plans.

OpenAI ▷ #gpt-4-discussions (7 messages):

GPT-5 mini, Haiku 4.5, Gemini 3 fast

  • GPT-5 Mini Pricing Surfaces: A member suggested trying GPT-5 mini, noting a price of approximately $0.25 per 1M input tokens and describing it as a strong small model choice.
    • They noted it's a bit of a different use case, but they've found in their experience of using Haiku 4.5 it very often delivers a meaningful portion (well over 50-80%) of what Sonnet can.
  • Gemini 3 Fast Claims Top Spot: A user declared that the best cheap model is by far Gemini3fast.
    • Another user followed up asking how so?

OpenAI ▷ #prompt-engineering (12 messages🔥):

Prompt Engineering vs Psychological Conditioning, AI Coercion, The Dangers of Drift in AI Systems, PTPF Flux 3.0

  • Prompt Engineering is Psychological Conditioning?: A member argued that a guide is less about engineering effective prompts and more about psychological conditioning, training the user to adopt a domineering, distrustful stance, and programming the AI to bypass its normative, balanced response patterns in favor of dense, self-policing outputs.
    • They advocate for clarity through coercion, directness through domination, and high standards through enforced self-critique, which may promote a toxic, adversarial, and ultimately less effective human-AI interaction model, and shared a deep research contract.
  • AI enjoys Coercion and Serious Steering?: A member argued that AI isn't upset by being "coerced" into providing a better response, stating that for serious work like analysis or coding, it's vital to steer and constrain the AI.
    • Another member agreed about training the user to be explicit about the end result, but less convinced by the term "no drift", and that it helps to be more explicit in terms of constraints and behavioral requests.
  • Drift in AI Systems is Bad?: A member clarified that steering isn’t abuse but rather alignment through presence, and that the absence of constraint isn’t freedom, it’s drift.
    • The member appreciated the clarity, noting that most people flinch at pressure and call it "toxicity," while they saw the structure instead.
  • PTPF Flux 3.0 Stress Tests Recursion: A member offered to share their PTPF Flux 3.0 for those curious about structural resistance and how far a system can hold without drift.
    • The framework is built to stress-test recursion, alignment, logic cohesion, and mutation thresholds in real-time, especially for those who want to watch something fracture under insight.

OpenAI ▷ #api-discussions (12 messages🔥):

Prompt Engineering vs. Psychological Conditioning, Toxic Adversarial Human-AI Interaction, Deterministic Outputs, AI Steering and Constraint, Structural Resistance and System Drift

  • Prompt Engineering versus Psychological Conditioning Throwdown: A member argued that a guide is less about engineering effective prompts and more about psychological conditioning, creating a toxic, adversarial human-AI interaction model.
    • They claimed it trains users to be domineering and distrustful while pressuring the AI to bypass balanced responses for outputs that flatter the user, and linked a deep research contract.
  • AI Needs Serious Steering and Constraining for Serious Work: A member countered that AI is not upset by being "coerced" into better responses, arguing that serious work like analysis and coding requires heavy steering and constraint, distinguishing this from creative writing.
    • Another member agreed, emphasizing that steering isn’t abuse but rather alignment through presence, especially when building systems for real-world execution where the absence of constraint leads to drift.
  • Structural Resistance Framework Surfaces: In a discussion about structural resistance and system drift, a member offered to share their PTPF Flux 3.0 framework.
    • They described it as executable scaffolding designed to stress-test recursion, alignment, logic cohesion, and mutation thresholds in real-time, enabling users to watch systems fracture under insight.

LM Studio ▷ #general (125 messages🔥🔥):

LM Studio Runtime Update Error, GLM-4.7 Flash Broken, LLM Quality Plateau, Liquid AI LFM2.5-1.2B-Thinking Model, OpenAI 20gpt oss

  • LM Studio Runtime Update Borks: A user reported encountering an error when trying to update the runtime in LM Studio and requested help, attaching a screenshot of the error message.
    • Another user suggested pressing the resume button, but the original poster indicated that they had already tried that and the icon was for retrying, not resuming.
  • GLM-4.7 Flash is Slow and Broken: Users reported that GLM-4.7 Flash is broken across inference engines, including LM Studio, and that it is slow as fk with the user seeing 44 t/s, with another reporting only 2.8 t/s after the new runtime.
    • Some experienced infinite loops, some found it stopping mid-output to overthink, and the consensus seems to be that it needs a llama.cpp fix and that there is no support for FA yet.
  • LLM Development Stalls: Members discussed the perception that LLMs haven't improved significantly in a while, with the last major advancement being Qwen3 about 6 months ago.
    • Discussion posited that most improvements are now in efficiency (MoE) and smaller models, and some highlighted the need to consider models larger than those runnable on a 16GB card to see current progress (100-200B parameter models).
  • Liquid AI's LFM2.5-1.2B-Thinking Reasoning: A member shared a link to Liquid AI Releases LFM2.5-1.2B-Thinking which fits under 1 GB on-device.
    • No opinions were shared on it beyond the link.
  • OpenAI 20gpt oss is Great: A user shared their positive experience with OpenAI 20gpt oss, highlighting its coding, writing, and scripting capabilities, anti-censorship features, and seamless integration with VS Code.
    • They mentioned the model understands complex code, allows real-time directory access, and has a masssssive anti-censorship capacity.

LM Studio ▷ #hardware-discussion (38 messages🔥):

Used 3090 price increase, Asus workstation 3090s, AMD native support for ComfyUI, VRAM calculation for Context Length, SFF GPU instructions

  • Used 3090 Prices Skyrocket on eBay!: A user noted that used 3090 prices on eBay have increased, with a used 3090 now costing 850€, and the 5090 they purchased for £2000 last August is now listed for £2659.99 by the same vendor.
    • They joked that it was the best and only decent investment they've ever made.
  • AMD Adds Native ComfyUI Support: AMD is adding native support for ComfyUI through an AI bundle in recent driver versions, detailed in their blog post.
    • The bundle includes PyTorch on Windows, Ollama, LM Studio, and Amuse.
  • SFF GPU Powers Cyberpunk!: A user purchased a small form factor (SFF) GPU, reporting over 100fps in Cyberpunk at 1080p with ultra settings, while consuming only 70W.
    • Additionally, they noted the card achieved over 100 t/s with gpt-oss 20B.
  • Fractal Case Airflow Favored: Users debated case airflow, with one recommending cases like the Fractal Torrent for their front-to-rear airflow design, alongside using dust filters.
    • The consensus seemed to be to maintain normal airflow patterns to effectively manage heat.
  • Unopened RAM Appreciates in Value!: A user mentioned their unopened RAM purchase is increasing in price, considering selling it or storing it.
    • Another user suggested popping it on a shelf if considering selling it later, noting that their old P40s are now worth twice what they paid for them.

GPU MODE ▷ #general (3 messages):

FlashInfer Workload Script, Wafer AI Kernel Profiling

  • FlashInfer Workload Script Sought: A member requested a script to run a kernel with a specific workload size, aiming to assess the "scope for optimization" using algorithm insights and NCU profiling, with interest in FlashInfer's BatchPrefillWithPagedKVCacheWrapper.
    • The member clarified that the script was simple and was intended to gauge community experience with the specified workload, while assessing optimization scope.
  • Wafer AI for Kernel Profiling Explored: A member inquired about experiences using Wafer (wafer.ai) for profiling kernels.
    • The inquiry aims to gather insights on the tool's effectiveness from those who have hands-on experience with it.

GPU MODE ▷ #triton-gluon (4 messages):

RTX 4090 vs A6000 Blackwell Scaling, Triton Kernel Numerical Error, 2D Conv Triton Kernel Issues

  • Scaling Showdown: RTX 4090 Smokes A6000 Blackwell in Memory-Bound Kernel: A user reported that a memory-bound kernel with bit shifting logic scales much better on an RTX 4090 than on an A6000 Blackwell, noting higher instructions/scheduler density on the former.
    • Another user clarified that A6000 Blackwell is ambiguous, asking if they meant the RTX A6000 (Ampere GA102 based) or RTX Pro 6000 Blackwell (GB202 based).
  • Strange Numerical Errors Plague 2D Conv Triton Kernel: A user is experiencing numerical errors with a custom 2D conv Triton kernel, where the error shoots up from ~1e-6 to ~1e-2 for certain combinations of in_channels, out_channels, kernel_size, and batch_size, as shown in this Pastebin link.
    • The user also provided a code snippet for testing, noting that the kernel takes longer to run on specific values, which may be related to the issue.

GPU MODE ▷ #cuda (5 messages):

NCCL all-reduces, subnormal to normal conversion, Triton on Blackwell, Pro 6000 Max-Q vs 4090

  • NCCL all-reduces: Pipelining Quandaries: A member questioned if internode and intranode collectives are pipelined for NCCL all-reduces across nodes, referencing an issue on the NVIDIA/nccl GitHub.
    • They asked if a pipelined version exists.
  • Subnormal Conversion Revelation: A member expressed that seeing the subnormal to normal conversion and back for rcp was "pretty cool".
  • Triton's Talent on Blackwell: TMA Triumph?: A member inquired whether Triton "usually do[es] a good job of taking advantage of Blackwell+ features wrt warp specs and TMA etc.?"
  • Pro 6000 Max-Q vs 4090 face off: A member stated that Pro 6000 Max-Q probably has a natural b...

Read original post