-
Notifications
You must be signed in to change notification settings - Fork 285
Fix broken links across examples #1502
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 1 commit
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change | ||||
|---|---|---|---|---|---|---|
|
|
@@ -4,8 +4,8 @@ | |||||
|
|
||||||
| # # Training a mathematical reasoning model using the verifiers library with sandboxed code execution | ||||||
|
|
||||||
| # This example demonstrates how to train mathematical reasoning models on Modal using the [verifiers library](https://github.com/willccbb/verifiers) with [Modal Sandboxes](https://modal.com/docs/guide/sandbox) for executing generated code. | ||||||
| # The [verifiers library](https://github.com/willccbb/verifiers) is a set of tools and abstractions for training LLMs with reinforcement learning in verifiable multi-turn environments via [GRPO](https://arxiv.org/abs/2402.03300). | ||||||
| # This example demonstrates how to train mathematical reasoning models on Modal using the [verifiers library](https://github.com/PrimeIntellect-ai/verifiers) with [Modal Sandboxes](https://modal.com/docs/guide/sandbox) for executing generated code. | ||||||
| # The [verifiers library](https://github.com/PrimeIntellect-ai/verifiers) is a set of tools and abstractions for training LLMs with reinforcement learning in verifiable multi-turn environments via [GRPO](https://arxiv.org/abs/2402.03300). | ||||||
|
|
||||||
| # This example demonstrates how to: | ||||||
| # - Launch a distributed GRPO training job on Modal with 4× H100 GPUs. | ||||||
|
|
@@ -15,7 +15,7 @@ | |||||
|
|
||||||
| # ## Setup | ||||||
| # We start by importing modal and the dependencies from the verifiers library. Then, we create a Modal App and an image with a NVIDIA CUDA base image. | ||||||
| # We install the dependencies for the `verifiers` and `flash-attn` libraries, following the verifiers [README](https://github.com/willccbb/verifiers?tab=readme-ov-file#getting-started). | ||||||
| # We install the dependencies for the `verifiers` and `flash-attn` libraries, following the verifiers [README](https://github.com/PrimeIntellect-ai/verifiers?tab=readme-ov-file#getting-started). | ||||||
|
|
||||||
| import modal | ||||||
|
|
||||||
|
|
@@ -80,8 +80,8 @@ | |||||
| """ | ||||||
|
|
||||||
| # ## Training | ||||||
| # Following the [verifiers example](https://github.com/willccbb/verifiers/blob/main/verifiers/examples/math_python.py), we will need a training script and a config file. | ||||||
| # For sandboxed code execution, we will use [this training script](/docs/examples/trainer_script_grpo) and the config file defined [here](https://github.com/willccbb/verifiers/blob/main/configs/zero3.yaml). | ||||||
| # Following the [verifiers example](https://github.com/PrimeIntellect-ai/verifiers), we will need a training script and a config file. | ||||||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. 🚩 learn_math.py line 83: verifiers link now points to repo root instead of specific example file The old link Was this helpful? React with 👍 or 👎 to provide feedback.
|
||||||
| # Following the [verifiers example](https://github.com/PrimeIntellect-ai/verifiers), we will need a training script and a config file. | |
| # Following the [verifiers example](https://github.com/PrimeIntellect-ai/verifiers/blob/main/verifiers/examples/math_python.py), we will need a training script and a config file. |
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -46,7 +46,7 @@ | |
| "synonyms": [["cls", "class"]], | ||
| }, | ||
| "stop_urls": [ | ||
| "https://modal.com/docs/reference/modal.Stub", | ||
| "https://modal.com/docs/reference/modal.App", | ||
| "https://modal.com/gpu-glossary", | ||
| "https://modal.com/docs/reference/changelog", | ||
| ], | ||
|
|
@@ -156,7 +156,7 @@ def crawl_webhook(): | |
|
|
||
| # The indexed contents can be found at https://www.algolia.com/apps/APP_ID/explorer/browse/, for your | ||
| # APP_ID. Once you're happy with the results, you can [set up the `docsearch` package with your | ||
| # website](https://docsearch.algolia.com/docs/docsearch-v3/), and create a search component that uses this index. | ||
| # website](https://docsearch.algolia.com/docs/legacy/run-your-own/), and create a search component that uses this index. | ||
devin-ai-integration[bot] marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
|
|
||
| # ## Entrypoint for development | ||
|
|
||
|
|
||
| Original file line number | Diff line number | Diff line change | ||||
|---|---|---|---|---|---|---|
|
|
@@ -11,7 +11,7 @@ | |||||
| # to the sheer size of the model, the cold start time on Modal is around 2 minutes. | ||||||
| # | ||||||
| # For faster cold start at the expense of inference speed, check out | ||||||
| # [Running Falcon-40B with AutoGPTQ](https://modal.com/docs/examples/falcon_gptq). | ||||||
| # [Running Falcon-40B with AutoGPTQ](https://github.com/modal-labs/modal-examples/blob/main/misc/falcon_gptq.py). | ||||||
|
||||||
| # [Running Falcon-40B with AutoGPTQ](https://github.com/modal-labs/modal-examples/blob/main/misc/falcon_gptq.py). | |
| # [Running Falcon-40B with AutoGPTQ](./falcon_gptq.py). |
| Original file line number | Diff line number | Diff line change | ||||
|---|---|---|---|---|---|---|
|
|
@@ -8,7 +8,7 @@ | |||||
| # cold start time on Modal is around 25s. | ||||||
| # | ||||||
| # For faster inference at the expense of a slower cold start, check out | ||||||
| # [Running Falcon-40B with `bitsandbytes` quantization](https://modal.com/docs/examples/falcon_bitsandbytes). You can also | ||||||
| # [Running Falcon-40B with `bitsandbytes` quantization](https://github.com/modal-labs/modal-examples/blob/main/misc/falcon_bitsandbytes.py). You can also | ||||||
|
||||||
| # [Running Falcon-40B with `bitsandbytes` quantization](https://github.com/modal-labs/modal-examples/blob/main/misc/falcon_bitsandbytes.py). You can also | |
| # [Running Falcon-40B with `bitsandbytes` quantization](https://modal.com/docs/examples/falcon_bitsandbytes). You can also |
Uh oh!
There was an error while loading. Please reload this page.