A Simple Average-case Analysis of Recursive Randomized Greedy MIS
arXiv:2604.01462v1 Announce Type: new Abstract: We revisit the complexity analysis of the recursive version of the randomized greedy algorithm for computing a maximal independent set (MIS), originally analyzed by Yoshida, Yamamoto, and Ito (2009). They showed that, on average per vertex, the expected number of recursive calls made by this algorithm is upper bounded by the average degree of the input graph. While their analysis is clever and intricate, we provide a significantly simpler alternative that achieves the same guarantee. Our analysis is inspired by the recent work of Dalirrooyfard, Makarychev, and Mitrovi\'c (2024), who developed a potential-function-based argument to analyze a new algorithm for correlation clustering. We adapt this approach to the MIS setting, yielding a more di
View PDF HTML (experimental)
Abstract:We revisit the complexity analysis of the recursive version of the randomized greedy algorithm for computing a maximal independent set (MIS), originally analyzed by Yoshida, Yamamoto, and Ito (2009). They showed that, on average per vertex, the expected number of recursive calls made by this algorithm is upper bounded by the average degree of the input graph. While their analysis is clever and intricate, we provide a significantly simpler alternative that achieves the same guarantee. Our analysis is inspired by the recent work of Dalirrooyfard, Makarychev, and Mitrović (2024), who developed a potential-function-based argument to analyze a new algorithm for correlation clustering. We adapt this approach to the MIS setting, yielding a more direct and arguably more transparent analysis of the recursive randomized greedy MIS algorithm.
Subjects:
Data Structures and Algorithms (cs.DS)
Cite as: arXiv:2604.01462 [cs.DS]
(or arXiv:2604.01462v1 [cs.DS] for this version)
https://doi.org/10.48550/arXiv.2604.01462
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Slobodan Mitrović [view email] [v1] Wed, 1 Apr 2026 23:16:16 UTC (17 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
announceversionanalysis
Why I Built a Menu Bar App Instead of a Dashboard
Everyone who builds with AI eventually hits the same moment. You're deep in a coding session. Claude is flying. You're feeling productive. Then you open your API dashboard and the number hits you like a bucket of cold water. That happened to me. I don't want to talk about the exact number, but it was enough to make me stop and actually think about what I was doing. The problem wasn't that I was spending money. The problem was that I had no idea I was spending it. The dashboard problem My first instinct was what everyone does: open the Anthropic dashboard. Check the usage graphs. Try to correlate the spikes with what I was working on. But here's the thing about dashboards — they're designed for after-the-fact analysis, not real-time awareness. You go to a dashboard when something's already

OpenClaw CVE-2026-33579: Unauthorized Privilege Escalation via `/pair approve` Command Fixed
CVE-2026-33579: A Critical Analysis of OpenClaw’s Authorization Collapse The recently disclosed CVE-2026-33579 vulnerability in OpenClaw represents a catastrophic failure in its authorization framework, enabling trivial full instance takeovers. At the core of this issue lies the /pair approve command—a mechanism intended for secure device registration that, due to a fundamental design flaw, bypasses critical authorization checks. This analysis dissects the vulnerability’s root cause, exploitation process, and systemic failures, underscoring the urgency of patching and the ease of attack. Root Cause: Authorization Bypass via Implicit Trust OpenClaw’s pairing system is designed to facilitate temporary, low-privilege access for device registration. The /pair approve command, however, omits ex

How to Get Gemma 4 26B Running on a Mac Mini with Ollama
So you picked up a Mac mini with the idea of running local LLMs, pulled Gemma 4 26B through Ollama, and... it either crawls at 2 tokens per second or just refuses to load. I've been there. Let me walk you through what's actually going on and how to fix it. The Problem: "Why Is This So Slow?" The Mac mini with Apple Silicon is genuinely great hardware for local inference. Unified memory means the GPU can access your full RAM pool — no separate VRAM needed. But out of the box, macOS doesn't allocate enough memory to the GPU for a 26B parameter model, and Ollama's defaults aren't tuned for your specific hardware. The result? The model either fails to load, gets killed by the OOM reaper, or runs painfully slowly because half the layers are falling back to CPU inference. Step 0: Check Your Hard
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Releases

OpenClaw CVE-2026-33579: Unauthorized Privilege Escalation via `/pair approve` Command Fixed
CVE-2026-33579: A Critical Analysis of OpenClaw’s Authorization Collapse The recently disclosed CVE-2026-33579 vulnerability in OpenClaw represents a catastrophic failure in its authorization framework, enabling trivial full instance takeovers. At the core of this issue lies the /pair approve command—a mechanism intended for secure device registration that, due to a fundamental design flaw, bypasses critical authorization checks. This analysis dissects the vulnerability’s root cause, exploitation process, and systemic failures, underscoring the urgency of patching and the ease of attack. Root Cause: Authorization Bypass via Implicit Trust OpenClaw’s pairing system is designed to facilitate temporary, low-privilege access for device registration. The /pair approve command, however, omits ex

How to Get Gemma 4 26B Running on a Mac Mini with Ollama
So you picked up a Mac mini with the idea of running local LLMs, pulled Gemma 4 26B through Ollama, and... it either crawls at 2 tokens per second or just refuses to load. I've been there. Let me walk you through what's actually going on and how to fix it. The Problem: "Why Is This So Slow?" The Mac mini with Apple Silicon is genuinely great hardware for local inference. Unified memory means the GPU can access your full RAM pool — no separate VRAM needed. But out of the box, macOS doesn't allocate enough memory to the GPU for a 26B parameter model, and Ollama's defaults aren't tuned for your specific hardware. The result? The model either fails to load, gets killed by the OOM reaper, or runs painfully slowly because half the layers are falling back to CPU inference. Step 0: Check Your Hard




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!