NEW: The Decision Factory — a novel about decisions under uncertainty. Get it on Amazon
Decision Science · · Adam DeJans Jr.

Simulation as a Laboratory for Decisions

Why building a simulator is the most underrated step in solving complex decision problems.

simulationdecision-sciencemethodology

The Most Underrated Tool in the Toolbox

If you ask most data science teams how they validate a model, they’ll tell you about train/test splits, cross-validation, and held-out datasets. These are fine for prediction problems. But decision problems aren’t prediction problems. You’re not trying to forecast the future — you’re trying to choose actions that lead to good outcomes across a range of possible futures.

How do you validate a decision policy? You can’t A/B test your way through supply chain strategy — running two different inventory policies simultaneously in a live warehouse is somewhere between impractical and insane. And backtesting on historical data only tells you what would have happened in scenarios that already occurred.

Simulation is the answer. Build a model of the system, generate thousands of plausible scenarios, run your decision policies against them, and measure what happens. It’s not a shortcut — it’s a laboratory.

What a Simulator Gives You

1. Policy Testing Before Deployment

You’ve designed a new replenishment policy. Does it actually work? Not “does the math check out” — does it work when demand is lumpy, lead times are volatile, and three suppliers go offline in the same week?

A simulator lets you stress-test decision policies against the full range of uncertainty before you bet real money on them. You find the failure modes in simulation, not in production. The cost of a simulated stockout is zero. The cost of a real one is not.

2. Tail Risk Visibility

Average-case performance is easy to estimate. What’s hard — and what matters — is understanding the tails. What happens in the worst 5% of scenarios? The worst 1%?

Simulation gives you the full distribution of outcomes. You can measure not just “expected cost” but “probability of exceeding budget by more than 20%” or “frequency of service level dropping below 90%.” These tail metrics are where the real risk lives, and they’re invisible without simulation.

3. Parameter Tuning

Most decision policies have tunable parameters. An (s, S) inventory policy has reorder points and order-up-to levels. A routing heuristic has priority weights and time windows. A staffing policy has trigger thresholds.

How do you set these parameters? You could use analytical formulas (which assume away most of the complexity). You could use gut feel (which doesn’t scale). Or you could search over parameter combinations in simulation, evaluating each configuration against thousands of scenarios. This is how you find parameter settings that are robust, not just theoretically optimal.

4. Apples-to-Apples Comparison

Want to compare a simple reorder-point policy against a machine-learning-driven policy? If you test them on different time periods or different data, the comparison is meaningless. A simulator lets you run both policies against the exact same set of scenarios, isolating the effect of the policy itself from the randomness of the environment.

This is how you answer the question: “Is the complex approach actually better, or did it just get lucky?”

A Concrete Example: Last-Mile Delivery

Let’s make this tangible. Imagine you’re running last-mile delivery operations for an e-commerce company in a metro area. Your decisions include:

  • How many drivers to schedule each day
  • How to assign packages to routes
  • When to dispatch (fixed waves vs. dynamic)
  • How to handle same-day delivery requests

You build a simulator that models:

  • Demand: Order volumes by zip code, time of day, with realistic day-to-day variability and seasonal patterns
  • Drivers: Availability, speed distributions, break times, shift constraints
  • Geography: Actual road network, traffic patterns by time of day, parking difficulty by zone
  • Disruptions: Weather events, vehicle breakdowns, driver no-shows, address errors

Now you can run experiments. You simulate 1,000 operating days under your current dispatch policy. Then you simulate 1,000 days under a proposed alternative — say, switching from fixed dispatch waves to dynamic dispatching.

The results aren’t a single number. They’re a distribution: average cost per package, on-time delivery rate, driver utilization, peak-hour capacity stress, and crucially, how these metrics look on the bad days, not just the average ones.

Maybe the dynamic dispatch policy saves 8% on average but has 3x the variance in driver overtime. That’s a tradeoff you’d never see without simulation. Now the decision-maker has real information to work with.

”Optimize Once and Hope” vs. “Simulate, Learn, Adapt”

The traditional workflow looks like this:

  1. Collect data
  2. Build optimization model
  3. Solve to optimality
  4. Deploy solution
  5. Hope for the best

The simulation-driven workflow looks like this:

  1. Collect data
  2. Build a simulator that captures the key dynamics and uncertainties
  3. Design candidate decision policies
  4. Evaluate each policy across thousands of simulated scenarios
  5. Analyze the full distribution of outcomes — averages, tails, failure modes
  6. Select and deploy the policy that best matches your risk tolerance and objectives
  7. Continue running simulation as a monitoring tool, feeding in new data to detect when the environment has shifted enough to warrant a policy update

The second approach is more work upfront. But it’s the difference between engineering a solution and gambling on one.

When You Don’t Need a Simulator

Let’s be honest: not every problem needs a full simulation framework. If your decisions are simple, your uncertainty is low, and the cost of being wrong is manageable, an analytical model or even a spreadsheet might be fine.

But if you’re making high-stakes, repeated decisions in a complex, uncertain environment — which describes most supply chain and logistics operations — simulation isn’t optional. It’s the only way to know if your decision policy actually works before you find out the hard way.

Build the lab. Run the experiments. Then go to production with confidence.