WeSearch
Hub / Tags / Reasoning
TAG · #REASONING

Reasoning coverage.

Every story in the WeSearch catalog tagged with #reasoning, chronological, with view counts. Subscribe to the per-tag RSS feed to follow this topic in your reader of choice.

27 stories tagged with #reasoning, in publish-time order across the WeSearch catalog. Tag pages update as new stories ingest.

⌘ RSS feed for this tag →   or   search "Reasoning"

RELATED TAGS
#ai4#large-language-models3#vision-language-models3#scientific-reasoning2#ml2#autoformalisation1#formal-methods1#chain-of-thought-prompting1#adversarial-prompting1#reasoning-robustness1#prompt-optimization1#small-language-models1
DEV.TO (TOP)

Video Demo: How Does Model Compression Change AI Reasoning?

In this video, I benchmark Mistral-7B-Instruct-v0.2 on an NVIDIA H200 DigitalOcean GPU in three...…

2 views ·
#ai#model compression#quantization
ARXIV.ORG

Beyond 80/20: High-Entropy Minority Tokens Drive Effective RL for LLM Reasoning

Reinforcement Learning with Verifiable Rewards (RLVR) has emerged as a powerful approach to enhancing the reasoning capabilities of Large Language Models (LLMs), while its mechanis…

10 views ·
#reinforcement learning#language models
TOM'S GUIDE

I put ChatGPT-5.5 vs Gemini 3.1 Pro through 7 impossible tests — and the winner surprised me

We put OpenAI's new GPT-5.5 and Google's Gemini 3.1 Pro through 7 brutal real-world prompts. The winner of this ultimate AI showdown might surprise you…

10 views ·
#ai comparison#chatgpt#gemini
YAHOO SPORTS

Phillies reasoning for offensive struggles sounds crazier than it really is

What's wrong in Philly?…

6 views ·
#philadelphia phillies#mlb#offensive struggles
GITHUB

ChatGPT/Gemini can now draw on your screen to help you navigate complex software

SketchVLM: Vision-language models can annotate images to explain thoughts and guide users.…

5 views ·
#sketchvlm#vision-language models#svg overlays
LOCALLLAMA

Why isn’t LLM reasoning done in vector space instead of natural language?

Why don’t LLMs use explicit vector-based reasoning instead of language-based chain-of-thought? What would happen if they did? Most LLM reasoning we see is expressed through languag…

7 views ·
VENTUREBEAT

How to build custom reasoning agents with a fraction of the compute

Training AI reasoning models demands resources that most enterprise teams do not have. Engineering teams are often forced to choose between distilling knowledge from large, expensi…

17 views ·
NEWSWEEK

WNBA MVP A'ja Wilson Gives Perfect Reasoning for Wanting More Trophies

Las Vegas Aces All-Star A'ja Wilson is the reigning WNBA MVP, Defensive Player of the Year, and Finals MVP. She's not satisfied.…

15 views ·
#wnba#aja wilson#las vegas aces
REDDIT

Nemotron-3-Nano-Omni-30B-A3B-Reasoning, New model?

It is Audio-Image/vids-Text -> Text Original BF 16 GGUF:…

8 views ·
ARXIV.ORG

Does Point Cloud Boost Spatial Reasoning of Large Language Models?

3D Large Language Models (LLMs) leveraging spatial information in point clouds for 3D spatial reasoning attract great attention. Despite some promising results, the role of point c…

5 views ·
GITHUB

NARE: An LLM agent that amortizes reasoning into memory and executable rules

Contribute to starface77/Neuro-Adaptive-Reasoning-Engine development by creating an account on GitHub.…

5 views ·
#llm agent#reasoning amortization#skill compilation
REDDIT

Do the "*Claude-4.6-Opus-Reasoning-Distilled" really bring something new to the original models?

No offense to the fine-tune model providers, just curious. IMO the original models were already trained on massive amount of high quality data, so why bother with this fine-tune? J…

9 views ·
ARXIV.ORG

The Power of Power Law: Asymmetry Enables Compositional Reasoning

Natural language data follows a power-law distribution, with most knowledge and skills appearing at very low frequency. While a common intuition suggests that reweighting or curati…

6 views ·
ARXIV.ORG

FormalScience: Scalable Human-in-the-Loop Autoformalisation of Science with Agentic Code Generation in Lean

Formalising informal mathematical reasoning into formally verifiable code is a significant challenge for large language models. In scientific fields such as physics, domain-specifi…

7 views ·
#autoformalisation#artificial intelligence#formal methods
ARXIV.ORG

Analytica: Soft Propositional Reasoning for Robust and Scalable LLM-Driven Analysis

Large language model (LLM) agents are increasingly tasked with complex real-world analysis (e.g., in financial forecasting, scientific discovery), yet their reasoning suffers from …

6 views ·
ARXIV.ORG

StoryTR: Narrative-Centric Video Temporal Retrieval with Theory of Mind Reasoning

Current video moment retrieval excels at action-centric tasks but struggles with narrative content. Models can see \textit{what is happening} but fail to reason \textit{why it matt…

6 views ·
ARXIV.ORG

CAP-CoT: Cycle Adversarial Prompt for Improving Chain of Thoughts in LLM Reasoning

Chain-of-Thought (CoT) prompting has emerged as a simple and effective way to elicit step-by-step solutions from large language models (LLMs). However, CoT reasoning can be unstabl…

6 views ·
#chain-of-thought prompting#adversarial prompting#large language models
ARXIV.ORG

Constraint-Based Analysis of Reasoning Shortcuts in Neurosymbolic Learning

Neurosymbolic systems can satisfy logical constraints during learning without achieving the intended concept-label correspondence; this is a problem known as reasoning shortcuts. W…

6 views ·
ARXIV.ORG

Ulterior Motives: Detecting Misaligned Reasoning in Continuous Thought Models

Chain-of-Thought (CoT) reasoning has emerged as a key technique for eliciting complex reasoning in Large Language Models (LLMs). Although interpretable, its dependence on natural l…

6 views ·
ARXIV.ORG

Tandem: Riding Together with Large and Small Language Models for Efficient Reasoning

Recent advancements in large language models (LLMs) have catalyzed the rise of reasoning-intensive inference paradigms, where models perform explicit step-by-step reasoning before …

7 views ·
#large language models#small language models#efficient reasoning
ARXIV.ORG

Expert Evaluation of LLM's Open-Ended Legal Reasoning on the Japanese Bar Exam Writing Task

Large language models (LLMs) have shown strong performance on legal benchmarks, including multiple-choice components of bar exams. However, their capacity for generating open-ended…

6 views ·
ARXIV.ORG

PhysNote: Self-Knowledge Notes for Evolvable Physical Reasoning in Vision-Language Model

Vision-Language Models (VLMs) have demonstrated strong performance on textbook-style physics problems, yet they frequently fail when confronted with dynamic real-world scenarios th…

7 views ·
#vision-language models#physical reasoning#knowledge representation
ARXIV.ORG

Agentic clinical reasoning over longitudinal myeloma records: a retrospective evaluation against expert consensus

Multiple myeloma is managed through sequential lines of therapy over years to decades, with each decision depending on cumulative disease history distributed across dozens to hundr…

8 views ·
#artificial intelligence#clinical decision support#multiple myeloma
ARXIV.ORG

Beyond the Attention Stability Boundary: Agentic Self-Synthesizing Reasoning Protocols

As LLM agents transition to autonomous digital coworkers, maintaining deterministic goal-directedness in non-linear multi-turn conversations emerged as an architectural bottleneck.…

6 views ·
ARXIV.ORG

A systematic evaluation of vision-language models for observational astronomical reasoning tasks

Vision-language models (VLMs) are increasingly proposed as general-purpose tools for scientific data interpretation, yet their reliability on real astronomical observations across …

8 views ·
#astronomy#vision-language models#artificial intelligence
REDDIT

Structured CoT: Shorter Reasoning with a Grammar File

7 views ·
MACHINE LEARNING

Going from 3B/7B dense to Nemotron 3 Nano (hybrid Mamba-MoE) for multi-task reasoning — what changes in the fine-tuning playbook? [D]

Following up on something I posted a few days back about fine-tuning for multi-task reasoning. Read a lot since then, and I've moved past the dense 3B vs 7B question — landing on N…

10 views ·