ROMAN: A Multiscale Routing Operator for Convolutional Time Series Models
arXiv:2604.02577v1 Announce Type: new Abstract: We introduce ROMAN (ROuting Multiscale representAtioN), a deterministic operator for time series that maps temporal scale and coarse temporal position into an explicit channel structure while reducing sequence length. ROMAN builds an anti-aliased multiscale pyramid, extracts fixed-length windows from each scale, and stacks them as pseudochannels, yielding a compact representation on which standard convolutional classifiers can operate. In this way, ROMAN provides a simple mechanism to control the inductive bias of downstream models: it can reduce temporal invariance, make temporal pooling implicitly coarse-position-aware, and expose multiscale interactions through channel mixing, while often improving computational efficiency by shortening th
View PDF HTML (experimental)
Abstract:We introduce ROMAN (ROuting Multiscale representAtioN), a deterministic operator for time series that maps temporal scale and coarse temporal position into an explicit channel structure while reducing sequence length. ROMAN builds an anti-aliased multiscale pyramid, extracts fixed-length windows from each scale, and stacks them as pseudochannels, yielding a compact representation on which standard convolutional classifiers can operate. In this way, ROMAN provides a simple mechanism to control the inductive bias of downstream models: it can reduce temporal invariance, make temporal pooling implicitly coarse-position-aware, and expose multiscale interactions through channel mixing, while often improving computational efficiency by shortening the processed time axis. We formally analyze the ROMAN operator and then evaluate it in two complementary ways by measuring its impact as a preprocessing step for four representative convolutional classifiers: MiniRocket, MultiRocket, a standard CNN-based classifier, and a fully convolutional network (FCN) classifier. First, we design synthetic time series classification tasks that isolate coarse position awareness, long-range correlation, multiscale interaction, and full positional invariance, showing that ROMAN behaves consistently with its intended mechanism and is most useful when class information depends on temporal structure that standard pooled convolution tends to suppress. Second, we benchmark the same models with and without ROMAN on long-sequence subsets of the UCR and UEA archives, showing that ROMAN provides a practically useful alternative representation whose effect on accuracy is task-dependent, but whose effect on efficiency is often favorable. Code is available at this https URL
Comments: 16 pages, appendix, 4 figures, 3 tables
Subjects:
Machine Learning (cs.LG)
Cite as: arXiv:2604.02577 [cs.LG]
(or arXiv:2604.02577v1 [cs.LG] for this version)
https://doi.org/10.48550/arXiv.2604.02577
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Gonzalo Uribarri Dr. [view email] [v1] Thu, 2 Apr 2026 23:19:19 UTC (1,901 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelbenchmarkannounce
Show HN: Ghost Pepper – 100% local hold-to-talk speech-to-text for macOS
I built this because I wanted to see how far I could get with a voice-to-text app that used 100% local models so no data left my computer. I've been using a ton for coding and emails. Experimenting with using it as a voice interface for my other agents too. 100% open-source MIT license, would love feedback, PRs, and ideas on where to take it. Comments URL: https://news.ycombinator.com/item?id=47666024 Points: 60 # Comments: 24

Are there Multiple Moral Endpoints?
This is a different approach to explaining an old idea . What is the deep future going to look like? Will we be proud of it, or ashamed of the choices that led to it? Lots of focus on the future is on the near future. How will ongoing wars go? What will the next AI model's capabilities be? Will this business succeed or fail? Let's zoom out and focus just on advanced artificial intelligence; my guess is that we'll have a "transition period" with many different relevant actors with different philosophies and moral considerations (we're in it) and then end up in an "equilibrium period" with much more homogeneity of philosophy and moral standpoints. [1] The transition period is hard to predict and involves many contingent facts, and might involve lots of dramatic turnovers; the equilibrium per
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

Show HN: Ghost Pepper – 100% local hold-to-talk speech-to-text for macOS
I built this because I wanted to see how far I could get with a voice-to-text app that used 100% local models so no data left my computer. I've been using a ton for coding and emails. Experimenting with using it as a voice interface for my other agents too. 100% open-source MIT license, would love feedback, PRs, and ideas on where to take it. Comments URL: https://news.ycombinator.com/item?id=47666024 Points: 60 # Comments: 24

OpenAI doesn’t expect to be profitable until at least 2030 as AI costs surge
As OpenAI and Anthropic move closer to their planned initial public offerings, more details about the finances of both artificial intelligence giants are starting to emerge. It was no secret these companies were bleeding cash, but seeing the actual numbers is still striking. Neither company has made its filings official. Both are in the process of recruiting investors and have recently closed funding rounds, which meant opening their books. The Wall Street Journal got a peek . According to internal estimates, OpenAI will not turn a profit until 2030, while Anthropic expects slight positive results this year, followed by another year of losses before staying in the green in 2028 and 2029. Spending on AI training will be staggering. In 2028, OpenAI projects spending $121 billion on computing

Qwen 4B/9B and Gemma E4B/26B A4B for multilingual entity extraction, summarisation and classification?
Hi, LLM newbie here. Has anyone benchmarked these smaller models on multilingual entity extraction, summarisation and classification? I'm particularly interested in your opinion when it comes to finetuning them to reach higher success rates and reliability. What is your general feeling of the performance and capabilities? I saw plenty posts here but rarely the ones that mention multilingual entity extraction, summarisation or classification submitted by /u/Creative-Fuel-2222 [link] [comments]



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!