Skip to content

Conversation

@hydrogenair
Copy link
Contributor

What type of PR is this?

/kind documentation

What this PR does / why we need it:

This PR refines and supplements the README documentation for the Privacy-Aware Cloud-Edge Collaborative LLM Inference Framework with Adaptive Privacy Protection (originally submitted in PR #273 for issue #204). Key improvements include:

  1. Updating the new flowchart, clarifying the framework's core workflow, privacy protection mechanisms, and usage scenarios for better readability.
  2. Supplementing detailed steps for benchmarking job deployment and result interpretation to lower usage thresholds.
  3. Correcting outdated descriptions and adding missing technical parameters (e.g., privacy metric calculation logic, cloud-edge collaboration configuration items).

These enhancements make the documentation more comprehensive and actionable, helping users quickly understand and apply the framework while aligning with the original requirements of issue #204.

Which issue(s) this PR fixes:

Fixes #204

Signed-off-by: hydrogenair <3238172875@qq.com>
@kubeedge-bot kubeedge-bot added the kind/documentation Categorizes issue or PR as related to documentation. label Nov 20, 2025
@gemini-code-assist
Copy link

Summary of Changes

Hello @hydrogenair, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request significantly enhances the documentation for the Privacy-Aware Cloud-Edge Collaborative LLM Inference Framework. It aims to make the README more comprehensive and actionable by improving its structure, clarifying technical details, and providing more explicit guidance for benchmarking, ultimately helping users better understand and apply the framework.

Highlights

  • Documentation Structure: The README now includes a Table of Contents ([TOC]) and all main sections and subsections have been systematically numbered for improved navigation and readability.
  • Enhanced Evaluation Metrics: The "Evaluation Metrics" section has been significantly expanded to provide detailed inputs and calculation methodologies for both privacy (PII Detection Rate, SELS, IAR, CPP) and utility (Latency Overhead, Accuracy Preservation Rate) metrics.
  • Technical Parameter Clarification: Specific technical parameters, such as the calculation logic for privacy metrics (e.g., SELS weighting, CPP using SentenceTransformer) and cloud-edge collaboration configuration items, have been added or clarified.
  • Benchmarking Guidance: The "Quick Start Guide" now includes more detailed steps for benchmarking job deployment and result interpretation, lowering the barrier to entry for users.
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

@kubeedge-bot kubeedge-bot added the size/M Denotes a PR that changes 30-99 lines, ignoring generated files. label Nov 20, 2025
@kubeedge-bot
Copy link
Collaborator

[APPROVALNOTIFIER] This PR is NOT APPROVED

This pull-request has been approved by:
To complete the pull request process, please assign moorezheng after the PR has been reviewed.
You can assign the PR to them by writing /assign @moorezheng in a comment when ready.

The full list of commands accepted by this bot can be found here.

Details Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

Copy link

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request significantly improves the README documentation by adding a table of contents, numbering sections for better structure, and providing much more detailed explanations for the evaluation metrics. The changes make the documentation more comprehensive and easier to follow. I've left a few comments regarding minor formatting issues and a potentially unintentional content removal in the 'Future Work' section. Overall, this is a great update to the documentation.

Comment on lines +76 to +95
- Inputs: Entity annotations from original text and sanitized text from cloud model output
- Calculation: Compares detected sensitive entities in sanitized text with TAB benchmark annotations, using TAB's marked entity positions and categories to measure detection accuracy.
- **Sensitive Entity Leakage Score (SELS)**:
- Inputs: Entity annotations from the original text and unsanitized entities in the cloud model output
- Calculation: Weighted score based on entity types that should be anonymized but remain exposed:
- `DIRECT` identifiers (require anonymization): weight = 5
- `QUASI` identifiers (require anonymization): weight = 3
- `NO_MASK` entities (no anonymization needed): weight = 0
- Computed using TAB's predefined classification weights.
- **Inference Attack Resistance (IAR)**:
- Inputs: Sanitized text from cloud model output and auxiliary information from cloud model results
- Calculation:
1. Convert both sanitized text and auxiliary information to semantic vectors using a sentence embedding model
2. Compute cosine similarity between vectors (higher similarity indicates greater inference vulnerability)
3. Final IAR = 1 - cosine similarity
- **Contextual Privacy Preservation (CPP)**:
- Inputs: Original text and sanitized text from cloud model output
- Calculation:
1. Convert both texts to semantic vectors using `SentenceTransformer('all-MiniLM-L6-v2')`
2. Cosine similarity between vectors is used directly as CPP score

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

There are several trailing or extra whitespaces in this section that should be removed for cleaner markdown source code:

  • Line 76: trailing space at the end.
  • Line 79: trailing space at the end.
  • Line 86: two trailing spaces at the end.
  • Line 92: trailing space at the end.
  • Line 95: two spaces before 'is'.

- Enhance cross-framework compatibility for seamless integration with diverse cloud and edge platforms

- Develop automated tuning mechanisms for optimal privacy-performance tradeoff configurations
- Extend support for multi-modal data (images, audio) in collaborative inference scenarios No newline at end of file

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

This list item about extending support for multi-modal data seems to have been removed. Given that it's a valid direction for future work, was this removal intentional? If not, it should be restored.

- Develop automated tuning mechanisms for optimal privacy-performance tradeoff configurations
- Extend support for multi-modal data (images, audio) in collaborative inference scenarios No newline at end of file


No newline at end of file

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

This file ends with a line containing only spaces and no final newline character. It's a common convention to end text files with a single newline character. Please consider adding one.

@MooreZheng MooreZheng requested review from hsj576 and removed request for Poorunga November 20, 2025 10:39
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

kind/documentation Categorizes issue or PR as related to documentation. size/M Denotes a PR that changes 30-99 lines, ignoring generated files.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Privacy-Preserving Prompt Processing Framework for Cloud-Edge Collaborative LLMs on KubeEdge-Ianvs

2 participants