Train a Reasoning-Capable LLM in One Weekend with NVIDIA NeMo - NVIDIA Developer
<a href="https://news.google.com/rss/articles/CBMinAFBVV95cUxQM2Vha244bG5yZ3Njbm1MUlk4dndsQzdyd2c0OHZ0a01HcVZLQUtsdGlfa1ZnLVRQTWJfOXJVRndpNUlmdU9QRUNlbXhqYkZKUDd1Tnk3RVhSWnlLMWlQdUFiUF8zY2h5WWZYNjQzbjh1YThkLWx2MVA1amt5WGpFeHZxWmJseEs2bVhSRjNkS2theUVYN1BBTzVrSGo?oc=5" target="_blank">Train a Reasoning-Capable LLM in One Weekend with NVIDIA NeMo</a> <font color="#6f6f6f">NVIDIA Developer</font>
Could not retrieve the full article text.
Read on GNews AI fine-tuning →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
reasoning
Show HN: Gemma Gem – AI model embedded in a browser – no API keys, no cloud
Gemma Gem is a Chrome extension that loads Google's Gemma 4 (2B) through WebGPU in an offscreen document and gives it tools to interact with any webpage: read content, take screenshots, click elements, type text, scroll, and run JavaScript. You get a small chat overlay on every page. Ask it about the page and it (usually) figures out which tools to call. It has a thinking mode that shows chain-of-thought reasoning as it works. It's a 2B model in a browser. It works for simple page questions and running JavaScript, but multi-step tool chains are unreliable and it sometimes ignores its tools entirely. The agent loop has zero external dependencies and can be extracted as a standalone library if anyone wants to experiment with it. Comments URL: https://news.ycombinator.com/item?id=47655367 Poi

Pre-1900 LLM Relativity Test
Wanted to share one of my personal projects, since similar work has been shared here. TLDR is that I trained an LLM from scratch on pre-1900 text to see if it could come up with quantum mechanics and relativity. The model was too small to do meaningful reasoning, but it has glimpses of intuition. When given observations from past landmark experiments, the model can declare that “light is made up of definite quantities of energy” and even suggest that gravity and acceleration are locally equivalent. I’m releasing the dataset + models and leave this as an open problem. You can play with one of the early instruction tuned models here (not physics post trained): gpt1900.com Blog post: https://michaelhla.com/blog/machina-mirabilis.html GitHub: https://github.com/michaelhla/gpt1900 submitted by
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

Quoting Chengpeng Mou
From anonymized U.S. ChatGPT data, we are seeing: ~2M weekly messages on health insurance ~600K weekly messages [classified as healthcare] from people living in “hospital deserts” (30 min drive to nearest hospital) 7 out of 10 msgs happen outside clinic hours Chengpeng Mou , Head of Business Finance, OpenAI Tags: ai-ethics , generative-ai , openai , chatgpt , ai , llms




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!