# Examples

## Common Workflow Examples

Try some of these ready-to-use TUI configurations for common workflows like test runs, production inference, prompt files, and more.

### Minimal: Test Model

Use a small test model to verify REE is working.&#x20;

{% hint style="info" %}
Note that test models have random weights and will produce nonsensical output. This is expected.
{% endhint %}

In the TUI, fill in the following parameters:

* **Model Name:** `hf-internal-testing/tiny-random-LlamaForCausalLM`
* **Prompt Text:** `Hello world`
* **Max New Tokens:** `24`
* Press `r` to run.

<figure><img src="/files/Rj2ScYvvEwWREIa2kfLN" alt=""><figcaption></figcaption></figure>

### Production: Reproducible Inference with a Real Model

* **Model Name:** `Qwen/Qwen3-0.6B`
* **Prompt Text:** `Explain quantum entanglement in simple terms.`
* **Max New Tokens:** `256`
* **Extra Args:** `--operation-set reproducible --temperature 0.7 --top-p 0.9`
* Press `r` to run.

<figure><img src="/files/BHvj0WsnMJoy4dwVMjEO" alt=""><figcaption></figcaption></figure>

### Using a Prompt File

Save your prompt to a .JSONL file:

```json
{"prompt": "What is 2 + 2? Show your reasoning step by step."}
```

In the TUI:

* **Model Name:** `Qwen/Qwen3-0.6B`
* **Prompt Text:** *(leave blank)*
* **Prompt File:** `/path/to/your/prompt.JSONL`
* **Max New Tokens:** `128`
* **Extra Args:**&#x20;

```bash
--operation-set reproducible
```

* Press `r` to run.

<figure><img src="/files/CO7YDP3LhWv6F4Xw0Hef" alt=""><figcaption></figcaption></figure>

### Short-Circuiting (Reasoning Models)

Short-circuiting forces the model to exit a generation phase early by injecting a specific token at a given step. This is useful for reasoning models (e.g., Qwen3) that have thinking/end-thinking phases, where you want to limit the token budget spent on "thinking."

Both `--short-circuit-length` and `--short-circuit-token` must be provided together in **Extra Args**.

* **Model Name:** `Qwen/Qwen3-14B`
* **Prompt Text:** `Solve this math problem.`
* **Max New Tokens:** `300`
* **Extra Args:**

```bash
--operation-set reproducible --short-circuit-length 100 --short-circuit-token 151668
```

* Press `r` to run.

<figure><img src="/files/BvM3rmQ1Go6hWP0I5Qbf" alt=""><figcaption></figcaption></figure>

### Large Models with Pipeline Parallelism

Split a large model across multiple GPUs using pipeline parallelism. This is the mode to use for models above \~32B parameters.

* **Model Name:** `Qwen/Qwen2.5-72B-Instruct`
* **Prompt Text:** `Summarize the key ideas behind pipeline parallelism.`
* **Max New Tokens:** `256`
* **Extra Args:** `--operation-set reproducible --n-partitions 4`
* Press `r` to run.

{% hint style="info" %}
Set `--n-partitions` to match how many GPUs you want to split the model across. If the model still doesn't fit, increase the partition count.
{% endhint %}

### Validating a Receipt

Validation ensures that a receipt remains internally consistent and that its hashes are untampered and uncorrupted, without requiring re-computation.

After a successful run, switch the TUI to validate mode:

* **Subcommand:** `validate`
* **Receipt Path:** Paste the path to your receipt JSON file (e.g., `~/.cache/gensyn/Qwen--Qwen3-0.6B/.../metadata/receipt_20260311_155048.json`)
* Press `r` to run.

<figure><img src="/files/OwZIlDpKH8maDNiqPyw5" alt=""><figcaption></figcaption></figure>

### Verifying a Receipt

Verification re-runs the entire inference pipeline and comparing the results with the receipt to ensure reproducibility.

To prove a receipt is reproducible by re-running the full inference pipeline:

* **Subcommand:** `verify`
* **Receipt Path:** Paste the path to the receipt JSON file
* Press `r` to run.&#x20;

REE will re-execute the computation and compare the output against the receipt. This is slower than `validate` since it runs the full pipeline, but it's the strongest proof that the result is reproducible.

<figure><img src="/files/CwAmCGVfq67pUF8VLhP4" alt=""><figcaption></figcaption></figure>

{% hint style="success" %}
Use `validate` for a quick integrity check or use `verify` when you need definitive proof.
{% endhint %}


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://docs.gensyn.ai/tech/ree/examples.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
