WeSearch
Hub / Search / coding model
SEARCH · CODING MODEL

Results for "coding model".

20 stories match your query across our 700+ source catalog. Ranked by relevance and recency.

20 results for "coding model"

REDDIT

Speculative Decoding Implementations: EAGLE-3, Medusa-1, PARD, Draft Models, N-gram and Suffix Decoding from scratch

I’ve been working on an educational implementation repo for speculative decoding: The goal is not to wrap existing libraries, but to implement several speculative decoding methods from scratch behind …

· 5 views
SIMON WILLISON'S WEBLOG

Qwen3.6-27B: Flagship-Level Coding in a 27B Dense Model

Qwen3.6-27B: Flagship-Level Coding in a 27B Dense Model Big claims from Qwen about their latest open weight model: Qwen3.6-27B delivers flagship-level agentic coding performance, surpassing the previo…

· 9 views
REDDIT

Speculative Decoding Implementations: EAGLE-3, Medusa-1, PARD, Draft Models, N-gram and Suffix Decoding from scratch [P]

I’ve been working on an educational implementation repo for speculative decoding: The goal is not to wrap existing libraries, but to implement several speculative decoding methods from scratch behind …

· 6 views
VENTUREBEAT

American AI startup Poolside launches free, high-performing open model Laguna XS.2 for local agentic coding

· 1 view
QUARTZ

OpenAI is bringing its AI models and coding agent to Amazon's cloud platform

· 2 views
ARTIFICIAL INTELLIGENCE (AI)

open models keep catching up and the frontier keeps moving. at some point one of those has to stop

a year ago there was a clear tier gap. now i'm less sure, but not in the way i expected. the tasks where open-weight models have genuinely caught up are real: coding assistance, summarization, instruc…

· 4 views
FIRETHERING

Xiaomi releases MiMo-v2.5 Family weights with strong coding and agent benchmarks

Peking University gives its computer science students a compiler project every semester. Build a complete SysY compiler in Rust including lexer, parser, abstract syntax tree, IR code generation, assem…

· 3 views
MACHINE LEARNING

I built a dataset of 120+ AI coding tools rated by free tier longevity, here's what the data actually shows (Completely free ) [P] [R]

background: i kept hitting paywalls i didn't see coming. tool says free, you integrate it into your workflow, three days later you're locked out. so i started keeping structured notes, then turned it …

· 4 views
JANE STREET BLOG

Using group theory to explore the space of positional encodings for attention

Attention is a computational primitive at the core of modern language models, allowing internal representations to reference and influence each other. It’s h...…

· 3 views
REDDIT

Speculative decoding with Gemma-4-31B + Gemma-4-E2B enables 120 - 200 tok/s output speed for specific tasks

So for my project I was using up until now either Gemini 3 / 2.5 Flash or Flash-lite. All my use cases are not agentic, simply LLM workflows for atomic tasks like extracting references from the law, c…

· 8 views
REDDIT

Field report: coding with Qwen 3.6 35B-A3B on an M2 Macbook Pro with 32GB RAM

TL;DR: I finally have this working and doing real work within the tight specs of my 32GB RAM Mac. So for those who would like to fly like Julien Chaumond , here's an updated HOW-TO, an explanation of …

· 6 views
ARXIV.ORG

LLMs Corrupt Your Documents When You Delegate

Large Language Models (LLMs) are poised to disrupt knowledge work, with the emergence of delegated work as a new interaction paradigm (e.g., vibe coding). Delegation requires trust - the expectation t…

· 5 views
ARXIV.ORG

StoryTR: Narrative-Centric Video Temporal Retrieval with Theory of Mind Reasoning

Current video moment retrieval excels at action-centric tasks but struggles with narrative content. Models can see \textit{what is happening} but fail to reason \textit{why it matters}. This semantic …

· 3 views
ARXIV.ORG

AdaMamba: Adaptive Frequency-Gated Mamba for Long-Term Time Series Forecasting

Accurate long-term time series forecasting (LTSF) requires the capture of complex long-range dependencies and dynamic periodic patterns. Recent advances in frequency-domain analysis offer a global per…

· 3 views
ARXIV.ORG

Vibe Medicine: Redefining Biomedical Research Through Human-AI Co-Work

With the emergence of large language models (LLMs) and AI agent frameworks, the human-AI co-work paradigm known as Vibe Coding is changing how people code, making it more accessible and productive. In…

· 3 views
SIMON WILLISON'S WEBLOG

Quoting Romain Huet

Since GPT-5.4, we’ve unified Codex and the main model into a single system, so there’s no separate coding line anymore. GPT-5.5 takes this further, with strong gains in agentic coding, computer use, a…

· 6 views
REDDIT

Qwen3.5/3.6 Coder?

With practically all of LocalLlama glazing Qwen 3.5/3.6 for it's coding skills. Along with the fact that Alibaba themselves are focusing on making Qwen a reliable coding agent, does this rule out the …

· 85 views
REDDIT

How do you decide which Claude Code tasks to run with Opus vs Sonnet vs Haiku?

Been vibe coding full-time for a few months. One workflow question I haven't nailed down yet: how do you decide which model to use for which task in Claude Code? I know Opus is overkill for simple edi…

· 8 views
REDDIT

Higher precision or higher parameter count

I’m wondering if we take models of the same family (e.g qwen3.5 moes). And we compared ggufs that are of different core counts different quantizations but similar sizes. Which model would be better fo…

· 7 views
REDDIT

How I personally deal with Claude's limits without giving up on Opus

I only use Sonnet as my main model. I instruct it to delegate indexing and similar grunt work to Haiku, and whenever something genuinely needs deeper thinking, I tell it to "consult Opus." Sonnet then…

· 5 views