Invited Speaker — Foundation Conf 2025


I shared practical findings from benchmarking DeepSeek against other frontier models.

  • Focus: cost/performance trade‑offs, latency SLOs, and reliability under load.
  • Infra context: multi‑tenant LLM platform with autoscaling and request shaping.
  • Takeaway: pick models by task profile and end‑to‑end latency budgets, not leaderboards alone.