More work on the paper
This commit is contained in:
parent
192f72d61a
commit
e5792bd597
@ -6,9 +6,7 @@ This repository contains the academic paper exploring how Ryan Williams' 2025 th
|
|||||||
|
|
||||||
**Title**: The Ubiquity of Space-Time Simulation in Modern Computing: From Theory to Practice
|
**Title**: The Ubiquity of Space-Time Simulation in Modern Computing: From Theory to Practice
|
||||||
**Author**: David H. Friedel Jr., Founder, MarketAlly LLC (USA) & MarketAlly Pte. Ltd. (Singapore)
|
**Author**: David H. Friedel Jr., Founder, MarketAlly LLC (USA) & MarketAlly Pte. Ltd. (Singapore)
|
||||||
**Status**: Submitted to arXiv (needing endorsement)
|
**Status**: Ongoing research: Version 2 released with additional clarity, preliminary results, and planned further experiments and refinements.
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
## Abstract
|
## Abstract
|
||||||
|
|
||||||
@ -21,6 +19,8 @@ Ryan Williams' 2025 result demonstrates that any time-bounded algorithm can be s
|
|||||||
|
|
||||||
## Key Findings
|
## Key Findings
|
||||||
|
|
||||||
|
Note: All findings in this version are subject to refinement as additional experiments and cross-hardware validation are conducted.
|
||||||
|
|
||||||
1. **Experimental validation**:
|
1. **Experimental validation**:
|
||||||
- Checkpointed sorting: 375-627× slowdown for √n space reduction
|
- Checkpointed sorting: 375-627× slowdown for √n space reduction
|
||||||
- Real LLM inference (Ollama): 18.3× slowdown for √n context chunking
|
- Real LLM inference (Ollama): 18.3× slowdown for √n context chunking
|
||||||
|
|||||||
10
main.tex
10
main.tex
@ -43,6 +43,16 @@ Founder \\ MarketAlly LLC (USA) \\ Founder \\ MarketAlly Pte. Ltd. (Singapore) \
|
|||||||
|
|
||||||
\maketitle
|
\maketitle
|
||||||
|
|
||||||
|
\begin{center}
|
||||||
|
\setlength{\fboxsep}{8pt}
|
||||||
|
\noindent\fbox{%
|
||||||
|
\parbox{0.96\linewidth}{%
|
||||||
|
\textit{This version is part of ongoing research. Future versions will include additional experiments, analysis, and refinements.}
|
||||||
|
}%
|
||||||
|
}
|
||||||
|
\end{center}
|
||||||
|
\vspace{1em}
|
||||||
|
|
||||||
\begin{abstract}
|
\begin{abstract}
|
||||||
Ryan Williams' 2025 result demonstrates that any time-bounded algorithm can be simulated using only $O(\sqrt{t \log t})$ space, establishing a fundamental limit on the space-time relationship in computation~\cite{williams2025}. This paper bridges the gap between this theoretical breakthrough and practical computing systems. Through rigorous experiments with statistical validation, we demonstrate space-time tradeoffs in six domains: external sorting (375-627× slowdown for $\sqrt{n}$ space), graph traversal (5× slowdown), stream processing (30× speedup for sliding window quantile queries), SQLite databases, LLM attention mechanisms, and real LLM inference with Ollama (18.3× slowdown). Surprisingly, we find that modern hardware can invert theoretical predictions—our simulated LLM experiments show 21× speedup with minimal cache due to memory bandwidth bottlenecks, while real model inference shows the expected slowdown. We analyze production systems including SQLite (billions of deployments) and transformer models (Flash Attention), showing that the $\sqrt{n}$ pattern emerges consistently despite hardware variations. Our work validates Williams' theoretical insight while revealing that practical constant factors range from $5\times$ to over $1{,}000{,}000\times$, fundamentally shaped by cache hierarchies, memory bandwidth, and I/O systems.
|
Ryan Williams' 2025 result demonstrates that any time-bounded algorithm can be simulated using only $O(\sqrt{t \log t})$ space, establishing a fundamental limit on the space-time relationship in computation~\cite{williams2025}. This paper bridges the gap between this theoretical breakthrough and practical computing systems. Through rigorous experiments with statistical validation, we demonstrate space-time tradeoffs in six domains: external sorting (375-627× slowdown for $\sqrt{n}$ space), graph traversal (5× slowdown), stream processing (30× speedup for sliding window quantile queries), SQLite databases, LLM attention mechanisms, and real LLM inference with Ollama (18.3× slowdown). Surprisingly, we find that modern hardware can invert theoretical predictions—our simulated LLM experiments show 21× speedup with minimal cache due to memory bandwidth bottlenecks, while real model inference shows the expected slowdown. We analyze production systems including SQLite (billions of deployments) and transformer models (Flash Attention), showing that the $\sqrt{n}$ pattern emerges consistently despite hardware variations. Our work validates Williams' theoretical insight while revealing that practical constant factors range from $5\times$ to over $1{,}000{,}000\times$, fundamentally shaped by cache hierarchies, memory bandwidth, and I/O systems.
|
||||||
\end{abstract}
|
\end{abstract}
|
||||||
|
|||||||
BIN
ubiquity.pdf
BIN
ubiquity.pdf
Binary file not shown.
Loading…
Reference in New Issue
Block a user