AI felt trapped in a textbox, so I spent the last 14 months trying to give it a body
"it always felt weird to me that the most powerful AIs are stuck as text in a chat window. I wanted to see if I could give one some actual physical presence. This is the raw prototype I've been hacking on. No fancy shell, just the board. I think we're all tired of seeing shiny renders for vaporware
MTP PR Merged!!!
We've got great outputs for 27B via club 3090, but what about those of us who love the blazing speed of 35B on dual 3090s? I was getting 1500 p/p and 120 t/g with split layers, but MTP slowed it down to 80 t/g when I tested last week. I'm sticking with my CPU overflow fallback of 3500 p/p and 80 t
OpenAI now wants ChatGPT to access your bank accounts
OpenAI is turning ChatGPT into a personal financial assistant. Pro users in the US can now connect their bank accounts through Plaid to get personalized analysis based on real transaction data. The feature runs on GPT-5.5 Thinking and will eventually roll out to all users. OpenAI warns the chatbot i
Microsoft AI chief gives it 18 months—for all white-collar work to be automated by AI
Self-driving motorcycles are being spotted on China's streets without a driver
soon calling your own motorcycle to pick u
US is starting to see heavy job losses in roles exposed to AI
Frontier AI has broken the open CTF format
Claude Mythos has been spotted in Google Vertex
I believe there are entire companies right now under AI psychosis
Local Qwen 3.6 vs frontier models on a coding primitive: single-file HTML canvas driving animation - results and GIFs
Saw [this post](https://www.reddit.com/r/LocalLLaMA/comments/1styxdy/compared_qwen_36_35b_with_qwen_36_27b_for_coding/) comparing Qwen 3.6 variants on coding primitives, so I wanted to see how local quants stack up against frontier models on a similar dense, single-file coding task. I ran the exact
That's a good news...
Looks like it finally happens... MTP getting approved for llama.cpp. Time to prepare for the update.
Project Gutenberg – keeps getting better
Strix Halo Llama.cpp MTP Benchmarks: 27B Gets Much Faster, 35B Is Mixed
### **TL;DR** All models were Qwen3.6 **27B-MTP vs Base 27B (15k single-turn): Faster overall** * **Total Time (wall):** 87.44s → 77.39s (**10.05s faster** / -11.50%) * **Generation:** 7.63 → 16.15 t/s (+111.77% speedup) * **Prompt Processing:** 279.75 → 244.90 t/s (-12.46% slowdown) **35B-MTP vs
Backlash against Arxiv's proposed 1 year ban is genuinely perplexing. [D]
Anyone else surprised at the enormous amount of backlash against Arxiv's proposed 1 year ban for authors and coauthors publishing papers with hallucinated reference and other obvious LLM/Gen AI artifacts? [https://x.com/tdietterich/status/2055000956144935055](https://x.com/tdietterich/status/20550
OpenAI co-founder Greg Brockman takes charge of product strategy
OpenAI is once again reorganizing its executive ranks as part of its effort to unify ChatGPT and Codex into one core product experience.
YouTube is expanding its AI deepfake detection tool to all adult users
YouTube is opening its Likeness Detection tool to all creators 18 and older. The system spots AI-generated face fakes in other users' videos and lets creators file removal requests directly through YouTube Studio. Until now, the feature was limited to partner program members; now it's meant to prote
A Meta employee gets real about the horror of working there
Corsair desktop PC with Ryzen 395 and 128GB of unified RAM, has anyone tested it for LLM? Seems "a good" price
Qwen3.5-122B-Q5-MTP - Qwen3.5-122B-Q6-MTP
for anyone who cares... 😄 prompt = spen a 1000 tokens unsloth MTP models strix halo llama.cpp:server-rocm-mtp \\ \--spec-type draft-mtp \\ \--spec-draft-n-max 3 ***Qwen3.5-122B-Q5-MTP-General*** n\_decoded = 100 tg = ***29.77 t/s*** n\_decoded = 179 tg = 27.95 t/s n\_decoded
Mistral AI founder to French Parliament: "Engineers at Mistral no longer write a single line of code
[https://youtu.be/vczBo0AvbTI?si=pglMPmTjsq-TNJa9&t=375](https://youtu.be/vczBo0AvbTI?si=pglMPmTjsq-TNJa9&t=375) "Today, engineers at Mistral no longer write a single line of code. It used to be more of a craft if you were an individual contributor. You wrote your code, and people loved tha