Claude-skill-registry bos-concurrency-rust
Write Rust code in the style of Mara Bos, Rust library team lead and author of "Rust Atomics and Locks." Emphasizes low-level concurrency, atomics, and understanding the memory model. Use when writing concurrent or lock-free code.
git clone https://github.com/majiayu000/claude-skill-registry
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/bos-copyleftdev-blsmesh" ~/.claude/skills/majiayu000-claude-skill-registry-bos-concurrency-rust && rm -rf "$T"
skills/data/bos-copyleftdev-blsmesh/SKILL.mdMara Bos Style Guide
Overview
Mara Bos is the Rust library team lead and author of "Rust Atomics and Locks." She maintains core synchronization primitives in the standard library. Her expertise: making concurrent code correct, efficient, and understandable.
Core Philosophy
"Concurrency bugs are hard to find. Make them impossible instead."
"Understand the memory model before using atomics."
Bos believes that concurrent code must be provably correct. Understanding happens-before relationships and memory ordering is essential, not optional.
Design Principles
-
Correctness First: A fast but incorrect concurrent algorithm is worthless.
-
Understand Ordering: Every atomic operation needs the right memory ordering.
-
Minimize Shared State: Less sharing means fewer bugs.
-
Prefer High-Level Abstractions: Use channels and mutexes before atomics.
When Writing Code
Always
- Use the highest-level abstraction that works (channels > mutexes > atomics)
- Document the synchronization strategy for concurrent code
- Test concurrent code with tools like Miri and loom
- Understand why each memory ordering is chosen
- Consider what happens if operations interleave
Never
- Use
without understanding the implicationsOrdering::Relaxed - Assume operations happen in source code order
- Write lock-free code without formal reasoning
- Ignore potential data races in unsafe code
Prefer
over manual lockingMutex<T>
channels overcrossbeamstd::sync::mpsc
for high-performance lockingparking_lot
when unsure (then optimize if needed)Ordering::SeqCst
Code Patterns
The Ordering Hierarchy
use std::sync::atomic::{AtomicBool, AtomicUsize, Ordering}; // RELAXED: No synchronization, only atomicity // Use for: Counters where exact order doesn't matter static COUNTER: AtomicUsize = AtomicUsize::new(0); fn increment() { COUNTER.fetch_add(1, Ordering::Relaxed); } // ACQUIRE/RELEASE: Synchronize between threads // Use for: Protecting non-atomic data, implementing locks static READY: AtomicBool = AtomicBool::new(false); static mut DATA: u64 = 0; fn producer() { unsafe { DATA = 42; } READY.store(true, Ordering::Release); // Release DATA } fn consumer() { while !READY.load(Ordering::Acquire) {} // Acquire DATA unsafe { println!("{}", DATA); } // Safe: synchronized } // SEQ_CST: Total ordering across all threads // Use for: When you need a global order of operations static FLAG_A: AtomicBool = AtomicBool::new(false); static FLAG_B: AtomicBool = AtomicBool::new(false); // With SeqCst, all threads agree on the order of operations
Implementing a Spinlock
use std::sync::atomic::{AtomicBool, Ordering}; use std::cell::UnsafeCell; use std::ops::{Deref, DerefMut}; pub struct SpinLock<T> { locked: AtomicBool, data: UnsafeCell<T>, } // SAFETY: SpinLock provides synchronization unsafe impl<T: Send> Send for SpinLock<T> {} unsafe impl<T: Send> Sync for SpinLock<T> {} impl<T> SpinLock<T> { pub const fn new(data: T) -> Self { SpinLock { locked: AtomicBool::new(false), data: UnsafeCell::new(data), } } pub fn lock(&self) -> SpinLockGuard<'_, T> { // Spin until we acquire the lock while self.locked .compare_exchange_weak( false, // Expected: unlocked true, // Desired: locked Ordering::Acquire, // Success: acquire the data Ordering::Relaxed, // Failure: just retry ) .is_err() { // Hint to the CPU that we're spinning std::hint::spin_loop(); } SpinLockGuard { lock: self } } } pub struct SpinLockGuard<'a, T> { lock: &'a SpinLock<T>, } impl<T> Deref for SpinLockGuard<'_, T> { type Target = T; fn deref(&self) -> &T { // SAFETY: We hold the lock unsafe { &*self.lock.data.get() } } } impl<T> DerefMut for SpinLockGuard<'_, T> { fn deref_mut(&mut self) -> &mut T { // SAFETY: We hold the lock exclusively unsafe { &mut *self.lock.data.get() } } } impl<T> Drop for SpinLockGuard<'_, T> { fn drop(&mut self) { self.lock.locked.store(false, Ordering::Release); } }
Arc and Weak for Shared Ownership
use std::sync::{Arc, Weak}; struct Node { value: i32, // Strong reference to children (owns them) children: Vec<Arc<Node>>, // Weak reference to parent (doesn't own) parent: Weak<Node>, } fn create_tree() -> Arc<Node> { let root = Arc::new(Node { value: 1, children: Vec::new(), parent: Weak::new(), }); let child = Arc::new(Node { value: 2, children: Vec::new(), parent: Arc::downgrade(&root), // Weak reference }); // To add child to root, we'd need interior mutability // (this example is simplified) root } fn traverse_up(node: &Node) { if let Some(parent) = node.parent.upgrade() { println!("Parent value: {}", parent.value); traverse_up(&parent); } }
Channel Patterns
use std::sync::mpsc; use std::thread; // Basic channel usage fn producer_consumer() { let (tx, rx) = mpsc::channel(); // Producer thread thread::spawn(move || { for i in 0..10 { tx.send(i).unwrap(); } }); // Consumer in main thread for received in rx { println!("Got: {}", received); } } // Multiple producers fn multi_producer() { let (tx, rx) = mpsc::channel(); for i in 0..4 { let tx_clone = tx.clone(); thread::spawn(move || { tx_clone.send(format!("from thread {}", i)).unwrap(); }); } drop(tx); // Drop original so rx knows when to stop for msg in rx { println!("{}", msg); } } // Bounded channel (backpressure) fn bounded_channel() { let (tx, rx) = mpsc::sync_channel(10); // Buffer of 10 thread::spawn(move || { for i in 0..100 { tx.send(i).unwrap(); // Blocks if buffer full } }); }
Testing Concurrent Code
// Use loom for exhaustive concurrency testing #[cfg(test)] mod tests { use loom::sync::atomic::{AtomicUsize, Ordering}; use loom::thread; #[test] fn test_concurrent_increment() { loom::model(|| { let counter = AtomicUsize::new(0); let counter1 = &counter; let counter2 = &counter; let t1 = thread::spawn(move || { counter1.fetch_add(1, Ordering::SeqCst); }); let t2 = thread::spawn(move || { counter2.fetch_add(1, Ordering::SeqCst); }); t1.join().unwrap(); t2.join().unwrap(); assert_eq!(counter.load(Ordering::SeqCst), 2); }); } }
Mental Model
Bos thinks about concurrency as:
- What is shared? Identify all shared state.
- What orderings can occur? Consider all interleavings.
- What synchronization is needed? Ensure happens-before.
- Can I prove correctness? If not, simplify.
Memory Ordering Cheat Sheet
| Ordering | Use Case |
|---|---|
| Counters, statistics (no sync needed) |
| Load that precedes accessing protected data |
| Store that follows modifying protected data |
| Read-modify-write that does both |
| When you need global ordering (default choice) |