Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion serverless/endpoints/model-caching.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -179,7 +179,7 @@ else:

The following sample applications demonstrate how you can integrate cached models into your custom workers:

- [Cached models + LLMs](https://github.com/runpod/model-store-cache-example): A custom worker that uses cached models to serve LLMs.
- [Cached models + LLMs](https://github.com/runpod-workers/model-store-cache-example): A custom worker that uses cached models to serve LLMs.

## Current limitations

Expand Down
2 changes: 1 addition & 1 deletion tutorials/serverless/model-caching-text.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ tag: "NEW"
---

<Tip>
You can download the finished code for this tutorial [on GitHub](https://github.com/runpod/model-store-cache-example).
You can download the finished code for this tutorial [on GitHub](https://github.com/runpod-workers/model-store-cache-example).
</Tip>

This tutorial demonstrates how to build a custom Serverless worker that leverages Runpod's [cached model](/serverless/endpoints/model-caching) feature to serve the Phi-3 language model. You'll learn how to create a handler function that locates and loads cached models in offline mode, which can significantly reduce costs and cold start times.
Expand Down