We are finally there: Qwen3.6-27B + agentic search; 95.7% SimpleQA on a single 3090, fully local
·
0 reactions
·
0 comments
·
2 views
LDR maintainer here. Thanks to the strong support of r/LocalLLaMA community LDR got very far. I haven't reported in a while because I thought I was not ready for another prominent post in one of the leading outlets of Local LLM research. But I think the LDR community finally there again. I think it is finally time to report again. Setup RTX 3090, 24GB Ollama backend (qwen3.6:27b) LDR's langgraph_agent strategy — LangChain create_agent() with tool-calling, parallel subtopic decomposition, up to 5
Original article
LocalLlama
Anonymous · no account needed