Time-Warping Recurrent Neural Networks for Transfer Learning
arXiv:2604.02474v1 Announce Type: cross Abstract: Dynamical systems describe how a physical system evolves over time. Physical processes can evolve faster or slower in different environmental conditions. We use time-warping as rescaling the time in a model of a physical system. This thesis proposes a new method of transfer learning for Recurrent Neural Networks (RNNs) based on time-warping. We prove that for a class of linear, first-order differential equations known as time lag models, an LSTM can approximate these systems with any desired accuracy, and the model can be time-warped while maintaining the approximation accuracy. The Time-Warping method of transfer learning is then evaluated in an applied problem on predicting fuel moisture content (FMC), an important concept in wildfire mod
View PDF HTML (experimental)
Abstract:Dynamical systems describe how a physical system evolves over time. Physical processes can evolve faster or slower in different environmental conditions. We use time-warping as rescaling the time in a model of a physical system. This thesis proposes a new method of transfer learning for Recurrent Neural Networks (RNNs) based on time-warping. We prove that for a class of linear, first-order differential equations known as time lag models, an LSTM can approximate these systems with any desired accuracy, and the model can be time-warped while maintaining the approximation accuracy. The Time-Warping method of transfer learning is then evaluated in an applied problem on predicting fuel moisture content (FMC), an important concept in wildfire modeling. An RNN with LSTM recurrent layers is pretrained on fuels with a characteristic time scale of 10 hours, where there are large quantities of data available for training. The RNN is then modified with transfer learning to generate predictions for fuels with characteristic time scales of 1 hour, 100 hours, and 1000 hours. The Time-Warping method is evaluated against several known methods of transfer learning. The Time-Warping method produces predictions with an accuracy level comparable to the established methods, despite modifying only a small fraction of the parameters that the other methods modify.
Subjects:
Machine Learning (cs.LG); Machine Learning (stat.ML)
Cite as: arXiv:2604.02474 [cs.LG]
(or arXiv:2604.02474v1 [cs.LG] for this version)
https://doi.org/10.48550/arXiv.2604.02474
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Jonathon Hirschi [view email] [v1] Thu, 2 Apr 2026 19:10:08 UTC (5,154 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelneural networktraining
Any RSS feeds for LLM related news?
I'm looking for RSS feeds that have relevant and interesting LLM related news, something to be able to keep up whenever a new interesting paper or model architecture comes out, or even new model family hits huggingface. Anybody has a few sources? submitted by /u/redblood252 [link] [comments]

Parallel prompting sessions across model sizes to detect gradient markers, has anyone tried this?
I run a 35b Qwen model on my own hardware (dual A4500, NVLinked) and have been thinking about a specific experiment I want to try, curious if anyone's done something similar. The hypothesis: there are specific markers that appear during generation that signal construction rather than retrieval, moments where the model is building something under constraint rather than pattern-matching to training data. These markers should be architectural properties of transformers, not size-dependent, so they should appear at roughly the same moments in a conversation whether you're running 35b or a much larger model. The content at those moments will differ in resolution, but the structural signal should be similar. The four markers I've identified through empirical conversation testing: - Convergence -

Trying to get a ChatGPT/Codex‑style autonomous experience with Hermes + Ollama, but it’s just not acting like it should — help?
Hey everyone, I’ve spent hours trying to get Hermes Agent working locally with Ollama, but I keep running into the same problem: Hermes runs and talks just fine, it connects to local models, but it almost never outputs the structured commands I need for automation — it just chats back with text, suggestions, or formatted output instead of real actions. What I really wanted was something like the old ChatGPT + Codex experience (where it reliably outputs run shell: ... or structured tool calls), so I could build autonomous workflows directly in my terminal (shell execution, scripting, multi‑step tasks, etc.). Instead I get stuff like: Current directory contents: /etc /usr /bin … Use `ls -la` for detailed listing …and nothing I can automatically parse or act on — even though the docs say Herm
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

Any RSS feeds for LLM related news?
I'm looking for RSS feeds that have relevant and interesting LLM related news, something to be able to keep up whenever a new interesting paper or model architecture comes out, or even new model family hits huggingface. Anybody has a few sources? submitted by /u/redblood252 [link] [comments]

What GPU is the best for my use case scenario?
TLDR: Medical student wondering whether they should buy a 5060Ti, 5070, 9070, or 9070 XT for a local LLM to help study using uploaded PDFs and documents. I’m a medical student and I used to have a ChatGPT Plus subscription. I have recently spent my allowance savings building a pc, mainly for gaming and study purposes. My specs include a Ryzen 7 7700 non-X CPU, and DDR5 32GB 6000 CL36 kit. The integrated graphics have been more than enough for study purposes, but I’d like to game soon too, so I was going to buy a graphics card. Coming to the crux of the issue, I will have saved enough by August/September to buy a GPU. I’m aiming for 1440p gaming, so my budget will range from NVIDIA RTX 5060Ti 16GB, 5070, AMD RX 9070 to AMD’s RX 9070 XT depending on pricing and availability. I know from a pu




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!