A proof-of-concept implementation of hardware introspection for AI-native boot design
In my article about NeuroShell OS's three-layer boot architecture, I outlined how operating systems should prepare for inevitable AI workloads. The design separates concerns cleanly:
- Layer 1: Built-in kernel code for early hardware preparation
- Layer 2: Early-loaded modules for capability aggregation and exposure
- Layer 3: User-space AI runtime initialization
While Layer 1 remains theoretical (requiring deep kernel modifications), Layer 2 now has a working proof-of-concept.
What We Built
I've created a basic Linux kernel module that demonstrates Layer 2's core responsibility: exposing hardware capabilities through a unified interface.
The module, available at github.com/hejhdiss/lkm-for-ai-resource-info, exposes detailed system information via /sys/kernel/neuroshell/:
/sys/kernel/neuroshell/
├── cpu_count # Online CPUs
├── cpu_info # CPU vendor, model, cache
├── cpu_topology # Core/socket mapping
├── mem_total_bytes # Total RAM
├── numa_nodes # NUMA node count
├── numa_info # Per-node memory breakdown
├── gpu_info # GPU summary by vendor
├── gpu_details # PCI device information
├── accelerator_count # AI accelerator count
└── system_summary # Human-readable overview
This is exactly what the article described: read-only interfaces that aggregate CPU, NUMA, memory, GPU, and accelerator information into a coherent view.
What It Does (and Doesn't Do)
The module follows Layer 2's strict boundaries:
✅ What it does:
- Detects CPU features (SIMD extensions, topology)
- Reports NUMA configuration
- Enumerates GPUs and AI accelerators
- Exposes everything through sysfs
❌ What it doesn't do:
- Allocate large memory regions
- Own or manage hardware
- Initialize GPU contexts
- Load AI runtimes
- Perform inference
This is critical. The module exposes, it does not control—keeping the kernel safe, debuggable, and composable.
A Very Basic Implementation
I need to be transparent: this is a proof-of-concept, not production code.
I have a beginner level background in C and POSIX C, all self-taught, but I lack deep kernel development experience. The module was created with AI assistance (Claude by Anthropic) to prototype the concepts from the article.
Current limitations include:
- No locking mechanisms (not SMP-safe)
- Fixed-size buffers
- Limited error handling
- No hotplug support
- X86-centric assumptions
This is intentionally basic. It demonstrates the what (Layer 2's role in the boot architecture), not the how (production-quality kernel engineering).
Why This Matters
This module proves that Layer 2 is feasible and practical.
Without it, AI runtimes in user space would need to:
- Query multiple kernel interfaces
- Parse scattered information
- Perform discovery at runtime
- Experience unpredictable cold-start latency
With this sysfs interface, a user-space neuro-init process (Layer 3) can:
from neuroshell import NeuroShell
ns = NeuroShell()
info = ns.get_all_info()
if info['numa_nodes'] > 1:
# Initialize with NUMA-aware tensor placement
pass
if info['gpu_count'] > 0:
# Pre-warm GPU contexts
pass
The system isn't "cold" anymore. Hardware capabilities are known before the first inference.
The Vision Continues
This Layer 2 module is just scratching the surface compared to what the full NeuroShell OS design requires:
- Real-time hardware discovery during boot
- Dynamic resource allocation for AI workloads
- Integration with bootloader and init systems
- Hardware-aware scheduler hooks
- Memory topology optimization for tensor operations
Layer 1 (built-in kernel preparation) and Layer 3 (intelligent user-space orchestration) remain conceptual. The complete architecture needs:
- Kernel developers to design Layer 1's early-boot integration
- Systems engineers to build Layer 3's orchestration logic
- Testing on real AI hardware (TPUs, NPUs, multi-GPU clusters)
Contributors Wanted
If you're a kernel developer interested in AI infrastructure, this project needs you.
The current module proves Layer 2 can exist. Making it production-ready—and building Layers 1 and 3—requires expertise I don't have:
- Proper locking and SMP safety
- Hotplug event handling
- Integration with existing kernel subsystems
- Security auditing
- Performance optimization
This is infrastructure for AI-native operating systems. If that vision resonates with you, the repository is open for contributions.
The Bigger Picture
NeuroShell OS isn't about forcing AI into Linux.
It's about acknowledging reality: AI workloads are predictable, heavy, and inevitable. If the kernel already prepares NUMA topology, hugepages, and cache hierarchies—then preparing intentionally for AI is not radical.
It's simply modern.
Layer 2 is now real, even if basic. The architecture is proven feasible. The next step is turning this proof-of-concept into production infrastructure.
Author: @hejhdiss
Status: Proof of concept seeking kernel developers for production implementation
Top comments (0)