Async Throughput Evidence Dashboard

All benchmark data from LLM Showdown #22 — click cards for details

Async Efficiency at 50 Concurrent Requests
SynapseKit
96.8%
967.5 req/s · 1.7ms overhead
LangChain
80.8%
808.3 req/s · 11.9ms overhead
LlamaIndex
92.7%
927.2 req/s · 3.9ms overhead
SynapseKit BaseTool.run(): Thin async wrapper. Validates input against JSON schema, calls the function, returns result. No middleware chain, no callback infrastructure. The 1.7ms overhead at n=50 is object construction and schema validation — nearly invisible.
LangChain RunnableLambda.ainvoke(): Every invocation traverses the LCEL Runnable protocol — input validation, callback manager, tracing hooks, output parsing. Powerful for composition (chain1 | chain2) but adds 11.9ms per batch at n=50. The overhead is 7x SynapseKit's.
LlamaIndex FunctionTool.acall(): Moderate overhead. Some validation and dispatch logic but no LCEL-style chain traversal. 3.9ms at n=50 puts it cleanly between the other two. The async path is cleaner than expected.
Throughput Scaling Curves
Raw Benchmark Data
ConcurrencyBaselineSynapseKitLangChainLlamaIndex
n=119.619.819.419.7
n=597.898.896.197.3
n=10194.9195.7184.2193.3
n=20391.3388.9360.5381.9
n=50986.6967.5808.3927.2
Scaling Factor (n=50 / n=1)
Frameworkrps n=1rps n=50Scalingvs Perfect (50x)
Baseline19.6986.650.4x100.9%
SynapseKit19.8967.548.9x97.7%
LlamaIndex19.7927.247.1x94.2%
LangChain19.4808.341.7x83.5%
Benchmark Verdict — Notebook #22
SynapseKit's thin async wrapper achieves 96.8% of theoretical throughput at 50 concurrent requests. LangChain's LCEL middleware costs 19.2% of theoretical throughput — a 7x overhead difference. LlamaIndex splits the difference at 92.7%. Winner: SynapseKit.
www.engineersofai.com