More work on the paper

This commit is contained in:
2025-08-13 21:44:55 -04:00
parent 192f72d61a
commit e5792bd597
3 changed files with 13 additions and 3 deletions

View File

@@ -6,9 +6,7 @@ This repository contains the academic paper exploring how Ryan Williams' 2025 th
**Title**: The Ubiquity of Space-Time Simulation in Modern Computing: From Theory to Practice **Title**: The Ubiquity of Space-Time Simulation in Modern Computing: From Theory to Practice
**Author**: David H. Friedel Jr., Founder, MarketAlly LLC (USA) & MarketAlly Pte. Ltd. (Singapore) **Author**: David H. Friedel Jr., Founder, MarketAlly LLC (USA) & MarketAlly Pte. Ltd. (Singapore)
**Status**: Submitted to arXiv (needing endorsement) **Status**: Ongoing research: Version 2 released with additional clarity, preliminary results, and planned further experiments and refinements.
## Abstract ## Abstract
@@ -21,6 +19,8 @@ Ryan Williams' 2025 result demonstrates that any time-bounded algorithm can be s
## Key Findings ## Key Findings
Note: All findings in this version are subject to refinement as additional experiments and cross-hardware validation are conducted.
1. **Experimental validation**: 1. **Experimental validation**:
- Checkpointed sorting: 375-627× slowdown for √n space reduction - Checkpointed sorting: 375-627× slowdown for √n space reduction
- Real LLM inference (Ollama): 18.3× slowdown for √n context chunking - Real LLM inference (Ollama): 18.3× slowdown for √n context chunking

View File

@@ -43,6 +43,16 @@ Founder \\ MarketAlly LLC (USA) \\ Founder \\ MarketAlly Pte. Ltd. (Singapore) \
\maketitle \maketitle
\begin{center}
\setlength{\fboxsep}{8pt}
\noindent\fbox{%
\parbox{0.96\linewidth}{%
\textit{This version is part of ongoing research. Future versions will include additional experiments, analysis, and refinements.}
}%
}
\end{center}
\vspace{1em}
\begin{abstract} \begin{abstract}
Ryan Williams' 2025 result demonstrates that any time-bounded algorithm can be simulated using only $O(\sqrt{t \log t})$ space, establishing a fundamental limit on the space-time relationship in computation~\cite{williams2025}. This paper bridges the gap between this theoretical breakthrough and practical computing systems. Through rigorous experiments with statistical validation, we demonstrate space-time tradeoffs in six domains: external sorting (375-627× slowdown for $\sqrt{n}$ space), graph traversal (5× slowdown), stream processing (30× speedup for sliding window quantile queries), SQLite databases, LLM attention mechanisms, and real LLM inference with Ollama (18.3× slowdown). Surprisingly, we find that modern hardware can invert theoretical predictions—our simulated LLM experiments show 21× speedup with minimal cache due to memory bandwidth bottlenecks, while real model inference shows the expected slowdown. We analyze production systems including SQLite (billions of deployments) and transformer models (Flash Attention), showing that the $\sqrt{n}$ pattern emerges consistently despite hardware variations. Our work validates Williams' theoretical insight while revealing that practical constant factors range from $5\times$ to over $1{,}000{,}000\times$, fundamentally shaped by cache hierarchies, memory bandwidth, and I/O systems. Ryan Williams' 2025 result demonstrates that any time-bounded algorithm can be simulated using only $O(\sqrt{t \log t})$ space, establishing a fundamental limit on the space-time relationship in computation~\cite{williams2025}. This paper bridges the gap between this theoretical breakthrough and practical computing systems. Through rigorous experiments with statistical validation, we demonstrate space-time tradeoffs in six domains: external sorting (375-627× slowdown for $\sqrt{n}$ space), graph traversal (5× slowdown), stream processing (30× speedup for sliding window quantile queries), SQLite databases, LLM attention mechanisms, and real LLM inference with Ollama (18.3× slowdown). Surprisingly, we find that modern hardware can invert theoretical predictions—our simulated LLM experiments show 21× speedup with minimal cache due to memory bandwidth bottlenecks, while real model inference shows the expected slowdown. We analyze production systems including SQLite (billions of deployments) and transformer models (Flash Attention), showing that the $\sqrt{n}$ pattern emerges consistently despite hardware variations. Our work validates Williams' theoretical insight while revealing that practical constant factors range from $5\times$ to over $1{,}000{,}000\times$, fundamentally shaped by cache hierarchies, memory bandwidth, and I/O systems.
\end{abstract} \end{abstract}

View File

Binary file not shown.