<?xml version="1.0" encoding="UTF-8"?><urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9" xmlns:news="http://www.google.com/schemas/sitemap-news/0.9" xmlns:xhtml="http://www.w3.org/1999/xhtml" xmlns:image="http://www.google.com/schemas/sitemap-image/1.1" xmlns:video="http://www.google.com/schemas/sitemap-video/1.1"><url><loc>https://llm-d.ai/blog</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/archive</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/authors</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/intelligent-inference-scheduling-with-llm-d</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/kvcache-wins-you-can-see</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-announce</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-community-update-june-2025</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-press-release</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-v0.2-our-first-well-lit-paths</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-v0.3-expanded-hardware-faster-perf-and-igw-ga</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-v0.4-achieve-sota-inference-across-accelerators</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-v0.5-sustaining-performance-at-scale</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/llm-d-week-1-round-up</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/native-kv-cache-offloading-to-any-file-system-with-llm-d</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/page/2</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/predicted-latency-based-scheduling-for-llms</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/announce</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/blog</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/community</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/hello</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/inference</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/kv-cache</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/llm-d</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/news-releases</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/releases</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/scheduling</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/sig-benchmarking</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/storage</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/updates</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/blog/tags/welcome</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/videos</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/benchmark</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/inference-scheduler</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/inference-sim</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/infra</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/kv-cache</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/modelservice</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/Components/workload-variant-autoscaler</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/architecture/latest-release</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/community</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/community/code-of-conduct</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/community/contribute</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/community/events</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/community/security</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/community/sigs</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/InfraProviders/aks</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/InfraProviders/digitalocean</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/InfraProviders/gke</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/asynchronous-processing</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/inference-scheduling</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/pd-disaggregation</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/precise-prefix-cache-aware</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/predicted-latency-based-scheduling</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/prerequisites</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/quickstart</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/simulated-accelerators</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/tiered-prefix-cache</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/tiered-prefix-cache/cpu</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/wide-ep-lws</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/workload-autoscaling</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/workload-autoscaling/hpa-igw</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/guide/Installation/workload-autoscaling/wva</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/usage/customizing-your-gateway</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/usage/getting-started-inferencing</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/usage/monitoring</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/docs/usage/readiness-probes</loc><changefreq>weekly</changefreq><priority>0.5</priority></url><url><loc>https://llm-d.ai/</loc><changefreq>weekly</changefreq><priority>0.5</priority></url></urlset>