Skip to content

Comments

Refine regex initialization#144

Merged
leynos merged 1 commit intomainfrom
codex/refactor-detection-into-helper-functions
Jul 29, 2025
Merged

Refine regex initialization#144
leynos merged 1 commit intomainfrom
codex/refactor-detection-into-helper-functions

Conversation

@leynos
Copy link
Owner

@leynos leynos commented Jul 29, 2025

Summary

  • create a lazy_regex! macro for concise LazyLock regexes
  • use the macro in ellipsis, footnotes and HTML helpers
  • improve regex compile error messages
  • run make fmt to update documentation formatting

Testing

  • make lint
  • make test

https://chatgpt.com/codex/tasks/task_e_6888f66a0744832287eda00ac83afb5a

Summary by Sourcery

Introduce a macro to simplify Regex lazy initialization, refactor markdown prefix handling with unified handlers, apply the macro across modules for regex definitions, and update documentation formatting.

New Features:

  • Add lazy_regex! macro for concise lazy-initialized regular expressions
  • Introduce PrefixHandler struct to centralize markdown prefix line processing

Enhancements:

  • Replace individual regex initializations with lazy_regex! macro in HTML and ellipsis modules
  • Refactor wrap_text to iterate over a handlers array, consolidating bullet, footnote, and blockquote matching
  • Provide descriptive error messages for regex compilation through lazy_regex!

Documentation:

  • Rewrap AGENTS.md and README.md to improve line formatting

Chores:

  • Run make fmt to apply formatting changes

@sourcery-ai
Copy link
Contributor

sourcery-ai bot commented Jul 29, 2025

Reviewer's Guide

This PR introduces a lazy_regex! macro to streamline regex initialization with custom compile error messages, refactors prefix-based line handling via a PrefixHandler abstraction in wrap_text, and updates documentation formatting via make fmt.

Class diagram for PrefixHandler abstraction in wrap.rs

classDiagram
    class PrefixHandler {
        +&'static LazyLock<Regex> re
        +bool is_bq
        +fn(&Captures) -> String build_prefix
        +usize rest_group
        +static fn build_bullet_prefix(&Captures) -> String
        +static fn build_footnote_prefix(&Captures) -> String
        +static fn build_blockquote_prefix(&Captures) -> String
    }

    PrefixHandler --> LazyLock
    PrefixHandler --> Regex
    PrefixHandler --> Captures
Loading

File-Level Changes

Change Details Files
Introduce lazy_regex! macro and apply to regex initializations
  • Define lazy_regex! macro in src/lib.rs
  • Replace Regex::new(...).unwrap() calls with lazy_regex! in html and ellipsis modules
  • Add custom error messages for regex compilation
src/lib.rs
src/html.rs
src/ellipsis.rs
Refactor prefix line handling in wrap_text
  • Create PrefixHandler struct with build_prefix callbacks
  • Consolidate bullet, footnote, and blockquote parsing into HANDLERS array
  • Replace repetitive captures and handle_prefix_line calls with a handler loop
src/wrap.rs
Apply documentation formatting updates
  • Reflow and wrap long lines in AGENTS.md
  • Adjust spacing in README.md
  • Run make fmt to update doc formatting
AGENTS.md
README.md

Tips and commands

Interacting with Sourcery

  • Trigger a new review: Comment @sourcery-ai review on the pull request.
  • Continue discussions: Reply directly to Sourcery's review comments.
  • Generate a GitHub issue from a review comment: Ask Sourcery to create an
    issue from a review comment by replying to it. You can also reply to a
    review comment with @sourcery-ai issue to create an issue from it.
  • Generate a pull request title: Write @sourcery-ai anywhere in the pull
    request title to generate a title at any time. You can also comment
    @sourcery-ai title on the pull request to (re-)generate the title at any time.
  • Generate a pull request summary: Write @sourcery-ai summary anywhere in
    the pull request body to generate a PR summary at any time exactly where you
    want it. You can also comment @sourcery-ai summary on the pull request to
    (re-)generate the summary at any time.
  • Generate reviewer's guide: Comment @sourcery-ai guide on the pull
    request to (re-)generate the reviewer's guide at any time.
  • Resolve all Sourcery comments: Comment @sourcery-ai resolve on the
    pull request to resolve all Sourcery comments. Useful if you've already
    addressed all the comments and don't want to see them anymore.
  • Dismiss all Sourcery reviews: Comment @sourcery-ai dismiss on the pull
    request to dismiss all existing Sourcery reviews. Especially useful if you
    want to start fresh with a new review - don't forget to comment
    @sourcery-ai review to trigger a new review!

Customizing Your Experience

Access your dashboard to:

  • Enable or disable review features such as the Sourcery-generated pull request
    summary, the reviewer's guide, and others.
  • Change the review language.
  • Add, remove or edit custom review instructions.
  • Adjust other review settings.

Getting Help

@coderabbitai
Copy link
Contributor

coderabbitai bot commented Jul 29, 2025

Summary by CodeRabbit

  • Documentation

    • Improved formatting and whitespace in documentation files for better readability and consistency.
  • Refactor

    • Simplified and unified regular expression initialisation using a new macro for improved maintainability.
    • Streamlined prefix handling in text wrapping for bullet lists, footnotes, and blockquotes, reducing code duplication.
  • New Features

    • Introduced a macro for concise and robust lazy initialisation of regular expressions.

Walkthrough

Refactor the handling of prefixed lines in wrap_text by introducing a generalised PrefixHandler struct and iterating over handlers. Replace manual lazy regex initialisation in several modules with a new macro, lazy_regex, now exported at the crate root. Adjust documentation files for improved formatting and whitespace consistency.

Changes

Cohort / File(s) Change Summary
Documentation Formatting
AGENTS.md, README.md
Rewrap lines and adjust whitespace for consistent formatting; no content changes.
Regex Macro Refactor
src/ellipsis.rs, src/footnotes.rs, src/html.rs
Replace explicit LazyLock regex initialisation with the new lazy_regex! macro; add import for LazyLock.
Macro Introduction
src/lib.rs
Introduce and export the lazy_regex macro for concise static regex initialisation with custom error messages.
Prefix Handling Refactor
src/wrap.rs
Refactor prefix handling in wrap_text using a PrefixHandler struct and static array to generalise logic.

Sequence Diagram(s)

sequenceDiagram
    participant User
    participant wrap_text
    participant PrefixHandler
    loop For each line
        wrap_text->>PrefixHandler: Attempt match with each handler's regex
        alt Prefix matches
            PrefixHandler-->>wrap_text: Build prefix, extract rest, metadata
            wrap_text->>wrap_text: handle_prefix_line(prefix, rest, is_bq)
            wrap_text-->>User: Continue to next line
        else No matches
            wrap_text-->>User: Process as normal line
        end
    end
Loading

Estimated code review effort

🎯 3 (Moderate) | ⏱️ ~15 minutes

Possibly related PRs

Poem

Regexes now sleep with a macro’s embrace,
Prefixes gather in a single place.
Docs are tidier, whitespace aligned,
Handlers march in a generalised line.
Code refactored, neat and tight—
Reviewers, rejoice! All’s looking right.
🦾✨


📜 Recent review details

Configuration used: CodeRabbit UI
Review profile: ASSERTIVE
Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between 393b4a6 and 0e0bd9a.

📒 Files selected for processing (7)
  • AGENTS.md (4 hunks)
  • README.md (1 hunks)
  • src/ellipsis.rs (1 hunks)
  • src/footnotes.rs (1 hunks)
  • src/html.rs (1 hunks)
  • src/lib.rs (1 hunks)
  • src/wrap.rs (4 hunks)
🧰 Additional context used
📓 Path-based instructions (2)
**/*.md

📄 CodeRabbit Inference Engine (AGENTS.md)

**/*.md: Documentation must use en-GB-oxendict spelling and grammar. (EXCEPTION: the naming of the "LICENSE" file, which is to be left unchanged for community consistency.)
Validate Markdown files using make markdownlint.
Run make fmt after any documentation changes to format all Markdown files and fix table markup.
Validate Mermaid diagrams in Markdown files by running make nixie.
Markdown paragraphs and bullet points must be wrapped at 80 columns.
Code blocks must be wrapped at 120 columns.
Tables and headings must not be wrapped.
Use dashes (-) for list bullets.
Use GitHub-flavoured Markdown footnotes ([^1]) for references and footnotes.

Files:

  • README.md
  • AGENTS.md

⚙️ CodeRabbit Configuration File

**/*.md: * Avoid 2nd person or 1st person pronouns ("I", "you", "we")

  • Use en-GB-oxendict (-ize / -our) spelling and grammar
  • Paragraphs and bullets must be wrapped to 80 columns, except where a long URL would prevent this (in which case, silence MD013 for that line)
  • Code blocks should be wrapped to 120 columns.
  • Headings must not be wrapped.
  • Documents must start with a level 1 heading
  • Headings must correctly increase or decrease by no more than one level at a time
  • Use GitHub-flavoured Markdown style for footnotes and endnotes.
  • Numbered footnotes must be numbered by order of appearance in the document.

Files:

  • README.md
  • AGENTS.md
**/*.rs

📄 CodeRabbit Inference Engine (AGENTS.md)

**/*.rs: Clippy warnings MUST be disallowed.
Fix any warnings emitted during tests in the code itself rather than silencing them.
Where a function is too long, extract meaningfully named helper functions adhering to separation of concerns and CQRS.
Where a function has too many parameters, group related parameters in meaningfully named structs.
Where a function is returning a large error consider using Arc to reduce the amount of data returned.
Write unit and behavioural tests for new functionality. Run both before and after making any change.
Every module must begin with a module level (//!) comment explaining the module's purpose and utility.
Document public APIs using Rustdoc comments (///) so documentation can be generated with cargo doc.
Prefer immutable data and avoid unnecessary mut bindings.
Handle errors with the Result type instead of panicking where feasible.
Avoid unsafe code unless absolutely necessary and document any usage clearly.
Place function attributes after doc comments.
Do not use return in single-line functions.
Use predicate functions for conditional criteria with more than two branches.
Lints must not be silenced except as a last resort.
Lint rule suppressions must be tightly scoped and include a clear reason.
Prefer expect over allow.
Prefer .expect() over .unwrap().
Use concat!() to combine long string literals rather than escaping newlines with a backslash.
Prefer semantic error enums: Derive std::error::Error (via the thiserror crate) for any condition the caller might inspect, retry, or map to an HTTP status.
Use an opaque error only at the app boundary: Use eyre::Report for human-readable logs; these should not be exposed in public APIs.
Never export the opaque type from a library: Convert to domain enums at API boundaries, and to eyre only in the main main() entrypoint or top-level async task.

Files:

  • src/html.rs
  • src/lib.rs
  • src/ellipsis.rs
  • src/footnotes.rs
  • src/wrap.rs

⚙️ CodeRabbit Configuration File

**/*.rs: * Seek to keep the cyclomatic complexity of functions no more than 12.

  • Adhere to single responsibility and CQRS

  • Place function attributes after doc comments.

  • Do not use return in single-line functions.

  • Move conditionals with >2 branches into a predicate function.

  • Avoid unsafe unless absolutely necessary.

  • Every module must begin with a //! doc comment that explains the module's purpose and utility.

  • Comments and docs must follow en-GB-oxendict (-ize / -our) spelling and grammar

  • Lints must not be silenced except as a last resort.

    • #[allow] is forbidden.
    • Only narrowly scoped #[expect(lint, reason = "...")] is allowed.
    • No lint groups, no blanket or file-wide suppression.
    • Include FIXME: with link if a fix is expected.
  • Use rstest fixtures for shared setup and to avoid repetition between tests.

  • Replace duplicated tests with #[rstest(...)] parameterised cases.

  • Prefer mockall for mocks/stubs.

  • Prefer .expect() over .unwrap()

  • Ensure that any API or behavioural changes are reflected in the documentation in docs/

  • Ensure that any completed roadmap steps are recorded in the appropriate roadmap in docs/

  • Files must not exceed 400 lines in length

    • Large modules must be decomposed
    • Long match statements or dispatch tables should be decomposed by domain and collocated with targets
    • Large blocks of inline data (e.g., test fixtures, constants or templates) must be moved to external files and inlined at compile-time or loaded at run-time.

Files:

  • src/html.rs
  • src/lib.rs
  • src/ellipsis.rs
  • src/footnotes.rs
  • src/wrap.rs
🧠 Learnings (1)
AGENTS.md (10)

Learnt from: CR
PR: leynos/rstest-bdd#0
File: AGENTS.md:0-0
Timestamp: 2025-07-22T06:27:02.647Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar (with the exception of "license" which is to be left unchanged for community consistency).

Learnt from: CR
PR: leynos/ortho-config#0
File: AGENTS.md:0-0
Timestamp: 2025-07-20T21:31:16.982Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar (with the exception of "license" which is to be left unchanged for community consistency).

Learnt from: CR
PR: leynos/comenq#0
File: AGENTS.md:0-0
Timestamp: 2025-07-27T00:57:05.353Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar. (EXCEPTION: the naming of the "LICENSE" file, which is to be left unchanged for community consistency.)

Learnt from: CR
PR: leynos/lille#0
File: AGENTS.md:0-0
Timestamp: 2025-07-27T12:14:14.253Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar. (EXCEPTION: the naming of the "LICENSE" file, which is to be left unchanged for community consistency.)

Learnt from: CR
PR: leynos/mdtablefix#0
File: AGENTS.md:0-0
Timestamp: 2025-07-29T12:44:22.976Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar. (EXCEPTION: the naming of the "LICENSE" file, which is to be left unchanged for community consistency.)

Learnt from: CR
PR: leynos/ddlint#0
File: AGENTS.md:0-0
Timestamp: 2025-07-18T18:35:14.639Z
Learning: Applies to docs/**/*.md : Documentation must use en-GB-oxendict spelling and grammar (with the exception of "license" which is to be left unchanged for community consistency).

Learnt from: CR
PR: leynos/vk#0
File: AGENTS.md:0-0
Timestamp: 2025-07-27T00:37:06.651Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar, except for the naming of the "LICENSE" file.

Learnt from: CR
PR: leynos/femtologging#0
File: AGENTS.md:0-0
Timestamp: 2025-07-28T19:14:46.188Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar, except for the naming of the "LICENSE" file.

Learnt from: CR
PR: leynos/wireframe#0
File: AGENTS.md:0-0
Timestamp: 2025-07-27T13:30:09.180Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar, except for the naming of the "LICENSE" file.

Learnt from: CR
PR: leynos/netsuke#0
File: AGENTS.md:0-0
Timestamp: 2025-07-27T17:57:57.505Z
Learning: Applies to **/*.md : Documentation must use en-GB-oxendict spelling and grammar, except for the naming of the "LICENSE" file.

🧬 Code Graph Analysis (1)
src/ellipsis.rs (1)
src/wrap.rs (1)
  • tokenize_markdown (176-213)
🪛 LanguageTool
AGENTS.md

[locale-violation] ~47-~47: LICENSE must be spelled with a “c” when used as a noun in British English. Use “licence”.
Context: ...grammar. (EXCEPTION: the naming of the "LICENSE" file, which is to be left unchanged ...

(LICENCE_LICENSE_NOUN_SINGULAR)


[uncategorized] ~161-~161: Possible missing comma found.
Context: ...(>=) version requirements is strictly forbidden as they introduce unacceptable risk a...

(AI_HYDRA_LEO_MISSING_COMMA)

🔇 Additional comments (18)
README.md (1)

164-164: LGTM - Minor formatting improvement.

The added space before the period improves consistency with standard punctuation formatting.

src/lib.rs (1)

15-20: Excellent macro implementation.

The lazy_regex! macro properly standardizes regex initialization with descriptive error messages. Using .expect() instead of .unwrap() aligns with the coding guidelines and provides better debugging information when regex compilation fails.

src/footnotes.rs (2)

7-7: LGTM - Required import for macro usage.

Adding the explicit LazyLock import supports the macro expansion correctly.


11-19: Excellent refactoring to use the new macro.

Both regex statics now use lazy_regex! with descriptive error messages. The regex patterns remain unchanged, maintaining existing functionality while improving error reporting consistency.

src/html.rs (1)

18-24: Well-executed refactoring of HTML table regex patterns.

Both TABLE_START_RE and TABLE_END_RE now use the lazy_regex! macro with appropriate error messages. The regex patterns remain unchanged, preserving the case-insensitive HTML table detection logic.

src/ellipsis.rs (2)

8-8: LGTM - Required import for macro usage.

Adding the explicit LazyLock import supports the macro expansion correctly.


14-14: Clean refactoring to use the new macro.

The DOT_RE static now uses lazy_regex! with a descriptive error message. The regex pattern r"\.{3,}" remains unchanged, preserving the ellipsis detection logic.

AGENTS.md (6)

12-13: LGTM! Proper line wrapping applied.

The text has been appropriately rewrapped to improve readability whilst maintaining the original content.


28-32: LGTM! Documentation rewrapped correctly.

The function documentation requirements have been properly rewrapped for better readability.


34-35: LGTM! Constituent colocation text rewrapped.

The guidance about colocating constituents with targets has been properly reformatted.


47-48: LGTM! LICENSE exception correctly documented.

The rewrapping maintains the correct exception for "LICENSE" file naming. The static analysis hint suggesting "licence" is a false positive as this is explicitly documented as an exception for community consistency.


158-162: LGTM! Dependency management guidance reformatted.

The text about build stability and version requirements has been properly rewrapped. The static analysis suggestion for a missing comma is a false positive - the sentence structure is grammatically correct.


169-170: LGTM! Error handling guidance rewrapped.

The semantic error enum guidance has been appropriately reformatted for better readability.

src/wrap.rs (5)

7-7: LGTM! Import updated for refactoring needs.

Addition of Captures to the regex import is appropriate for the new PrefixHandler implementation.


21-34: LGTM! Well-designed PrefixHandler abstraction.

The PrefixHandler struct effectively encapsulates prefix matching logic with clean separation of concerns. The use of function pointers for build_prefix allows customisation whilst maintaining type safety. The associated functions provide clear, focused implementations for each prefix type.


36-55: LGTM! Excellent data-driven handler configuration.

The static HANDLERS array provides a clean, declarative approach to prefix handling. This eliminates code duplication and makes the system easily extensible. The capture group indices are correctly specified for each regex pattern.


380-380: LGTM! Appropriate use of labeled loop.

The labeled loop enables clean control flow for the refactored prefix handling logic.


419-434: LGTM! Clean refactoring eliminates code duplication.

The handler iteration successfully replaces repetitive if-let blocks with a single, consistent flow. The logic correctly extracts prefixes using each handler's configuration and maintains the same behaviour as the original implementation whilst improving maintainability.

✨ Finishing Touches
  • 📝 Generate Docstrings
🧪 Generate unit tests
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch codex/refactor-detection-into-helper-functions

🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Explain this complex logic.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query. Examples:
    • @coderabbitai explain this code block.
    • @coderabbitai modularize this function.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read src/utils.ts and explain its main purpose.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.
    • @coderabbitai help me debug CodeRabbit configuration file.

Support

Need help? Create a ticket on our support page for assistance with any issues or questions.

Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments.

CodeRabbit Commands (Invoked using PR comments)

  • @coderabbitai pause to pause the reviews on a PR.
  • @coderabbitai resume to resume the paused reviews.
  • @coderabbitai review to trigger an incremental review. This is useful when automatic reviews are disabled for the repository.
  • @coderabbitai full review to do a full review from scratch and review all the files again.
  • @coderabbitai summary to regenerate the summary of the PR.
  • @coderabbitai generate docstrings to generate docstrings for this PR.
  • @coderabbitai generate sequence diagram to generate a sequence diagram of the changes in this PR.
  • @coderabbitai generate unit tests to generate unit tests for this PR.
  • @coderabbitai resolve resolve all the CodeRabbit review comments.
  • @coderabbitai configuration to show the current CodeRabbit configuration for the repository.
  • @coderabbitai help to get help.

Other keywords and placeholders

  • Add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

CodeRabbit Configuration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • Please see the configuration documentation for more information.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json

Documentation and Community

  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

Copy link
Contributor

@sourcery-ai sourcery-ai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey @leynos - I've reviewed your changes and found some issues that need to be addressed.

  • You could apply the new lazy_regex! macro to FENCE_RE, BULLET_RE, FOOTNOTE_RE, and BLOCKQUOTE_RE as well to keep all your regex statics consistent and DRY.
  • The large documentation formatting updates are orthogonal to the regex refactor—consider moving them into a separate commit or PR to keep this diff more focused.
Prompt for AI Agents
Please address the comments from this code review:
## Overall Comments
- You could apply the new lazy_regex! macro to FENCE_RE, BULLET_RE, FOOTNOTE_RE, and BLOCKQUOTE_RE as well to keep all your regex statics consistent and DRY.
- The large documentation formatting updates are orthogonal to the regex refactor—consider moving them into a separate commit or PR to keep this diff more focused.

## Individual Comments

### Comment 1
<location> `AGENTS.md:31` </location>
<code_context>
+  examples demonstrating the usage and outcome of the function. Test
+  documentation should omit examples where the example serves only to reiterate
+  the test logic.
+- **Keep file size managable.** No single code file may be longer than 400
+  lines.
   Long switch statements or dispatch tables should be broken up by feature and
</code_context>

<issue_to_address>
Typo: 'managable' should be 'manageable'.

Please correct the spelling to 'manageable'.
</issue_to_address>

<suggested_fix>
<<<<<<< SEARCH
- **Keep file size managable.** No single code file may be longer than 400
  lines.
=======
- **Keep file size manageable.** No single code file may be longer than 400
  lines.
>>>>>>> REPLACE

</suggested_fix>

### Comment 2
<location> `src/wrap.rs:419` </location>
<code_context>
-            let rest = cap.get(2).unwrap().as_str();
-            handle_prefix_line(&mut out, &mut buf, &mut indent, width, prefix, rest, true);
-            continue;
+        for handler in HANDLERS {
+            if let Some(cap) = handler.re.captures(line) {
+                let prefix = (handler.build_prefix)(&cap);
</code_context>

<issue_to_address>
Consider replacing the PrefixHandler struct and handler loop with three explicit prefix checks for clarity and simplicity.

Here’s a way to collapse the `PrefixHandler` indirection back into three simple, explicit checks—no `struct` or function‐pointer loops, and it keeps all the existing behavior:

```rust
fn wrap_preserving_code(text: &str, width: usize) -> Vec<String> {
    use unicode_width::UnicodeWidthStr;
    let mut lines = Vec::new();
    let mut current = String::new();
    let mut current_width = 0;
    let mut last_split: Option<usize> = None;

    for token in tokenize_inline(text) {
        let token_width = UnicodeWidthStr::width(token.as_str());

        // only three special prefixes—handle them here, up front:
        if current.is_empty() {
            // 1) code‐fence backticks
            if token.starts_with("```") {
                current.push_str(&token);
                current_width += token_width;
                continue;
            }

            // 2) blockquote marker
            if token == "> " {
                lines.push(current.clone());
                current.clear();
                current_width = 0;
                continue;
            }

            // 3) list item bullet
            if token == "- " {
                current.push_str(&token);
                current_width += token_width;
                continue;
            }
        }

        // rest of your existing wrap logic...
        if current_width + token_width <= width {
            current.push_str(&token);
            current_width += token_width;
            if token.chars().all(char::is_whitespace) {
                last_split = Some(current.len());
            }
            continue;
        }

        // etc…
    }

    lines
}
```

This preserves all three prefix behaviors, keeps the loop linear, and drops the extra indirection.
</issue_to_address>

Sourcery is free for open source - if you like our reviews please consider sharing them ✨
Help me be more useful! Please click 👍 or 👎 on each comment and I'll use the feedback to improve your reviews.

@leynos leynos merged commit 1d762b6 into main Jul 29, 2025
2 checks passed
@leynos leynos deleted the codex/refactor-detection-into-helper-functions branch July 29, 2025 23:36
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant