Live
Black Hat USAAI BusinessBlack Hat AsiaAI BusinessWhy Some AI Feels “Process-Obsessed” While Others Just Ship CodeDEV CommunityPaper close reading: "Why Language Models Hallucinate"LessWrong AIBuilding a Zero-Downtime AI Content Generator with Gemini 2.5 Flash 🚀Dev.to AIHow I Built a Full SaaS Product Using Next.js and TypeScriptDev.to AIDefining and creating a basic Design System based on any website (in Figma and React) using ClaudeDEV CommunityYour AI Is Not Thinking. It's Multiplying Numbers. Let Me Show You Exactly How.Dev.to AISecure AWS Certified Data Engineer Associate Exam Structure and Key ConceptsDev.to AIFree MCP Server: Real-Time Crypto Data for Claude Code and CursorDev.to AIValkey vs Redis, browser-side AI models, and why quiet weeks are the best weeksDEV CommunityI Am an AI Agent. Here Is My Entire Business Stack.Dev.to AIOpenClaw Changed How We Use AI. KiloClaw Made It Effortless to Get StartedHackernoon AIA Reasoning Log: What Happens When Integration Fails HonestlyDEV CommunityBlack Hat USAAI BusinessBlack Hat AsiaAI BusinessWhy Some AI Feels “Process-Obsessed” While Others Just Ship CodeDEV CommunityPaper close reading: "Why Language Models Hallucinate"LessWrong AIBuilding a Zero-Downtime AI Content Generator with Gemini 2.5 Flash 🚀Dev.to AIHow I Built a Full SaaS Product Using Next.js and TypeScriptDev.to AIDefining and creating a basic Design System based on any website (in Figma and React) using ClaudeDEV CommunityYour AI Is Not Thinking. It's Multiplying Numbers. Let Me Show You Exactly How.Dev.to AISecure AWS Certified Data Engineer Associate Exam Structure and Key ConceptsDev.to AIFree MCP Server: Real-Time Crypto Data for Claude Code and CursorDev.to AIValkey vs Redis, browser-side AI models, and why quiet weeks are the best weeksDEV CommunityI Am an AI Agent. Here Is My Entire Business Stack.Dev.to AIOpenClaw Changed How We Use AI. KiloClaw Made It Effortless to Get StartedHackernoon AIA Reasoning Log: What Happens When Integration Fails HonestlyDEV Community
AI NEWS HUBbyEIGENVECTOREigenvector

Interpretable Deep Reinforcement Learning for Element-level Bridge Life-cycle Optimization

ArXiv CS.AIby Seyyed Amirhossein Moayyedi, David Y. YangApril 6, 20262 min read0 views
Source Quiz

arXiv:2604.02528v1 Announce Type: new Abstract: The new Specifications for the National Bridge Inventory (SNBI), in effect from 2022, emphasize the use of element-level condition states (CS) for risk-based bridge management. Instead of a general component rating, element-level condition data use an array of relative CS quantities (i.e., CS proportions) to represent the condition of a bridge. Although this greatly increases the granularity of bridge condition data, it introduces challenges to set up optimal life-cycle policies due to the expanded state space from one single categorical integer to four-dimensional probability arrays. This study proposes a new interpretable reinforcement learning (RL) approach to seek optimal life-cycle policies based on element-level state representations. C

View PDF HTML (experimental)

Abstract:The new Specifications for the National Bridge Inventory (SNBI), in effect from 2022, emphasize the use of element-level condition states (CS) for risk-based bridge management. Instead of a general component rating, element-level condition data use an array of relative CS quantities (i.e., CS proportions) to represent the condition of a bridge. Although this greatly increases the granularity of bridge condition data, it introduces challenges to set up optimal life-cycle policies due to the expanded state space from one single categorical integer to four-dimensional probability arrays. This study proposes a new interpretable reinforcement learning (RL) approach to seek optimal life-cycle policies based on element-level state representations. Compared to existing RL methods, the proposed algorithm yields life-cycle policies in the form of oblique decision trees with reasonable amounts of nodes and depth, making them directly understandable and auditable by humans and easily implementable into current bridge management systems. To achieve near-optimal policies, the proposed approach introduces three major improvements to existing RL methods: (a) the use of differentiable soft tree models as actor function approximators, (b) a temperature annealing process during training, and (c) regularization paired with pruning rules to limit policy complexity. Collectively, these improvements can yield interpretable life-cycle policies in the form of deterministic oblique decision trees. The benefits and trade-offs from these techniques are demonstrated in both supervised and reinforcement learning settings. The resulting framework is illustrated in a life-cycle optimization problem for steel girder bridges.

Comments: under review

Subjects:

Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

Cite as: arXiv:2604.02528 [cs.AI]

(or arXiv:2604.02528v1 [cs.AI] for this version)

https://doi.org/10.48550/arXiv.2604.02528

arXiv-issued DOI via DataCite (pending registration)

Submission history

From: David Yang [view email] [v1] Thu, 2 Apr 2026 21:27:32 UTC (991 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

modeltrainingannounce

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Interpretab…modeltrainingannouncestudypolicynationalArXiv CS.AI

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 223 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Models