Skip to content

[Train] Training Pipeline#1214

Merged
kylesayrs merged 11 commits intomainfrom
train
Mar 13, 2025
Merged

[Train] Training Pipeline#1214
kylesayrs merged 11 commits intomainfrom
train

Conversation

@horheynm
Copy link
Copy Markdown

@horheynm horheynm commented Feb 28, 2025

Order of reviews:
#1206
#1207
#1209
#1212
#1214 <-- Here

SUMMARY:

  • Refactor Training pipeline
  • Remove initialize, finalize from the session functions
  • Add training information on entrypoints/readme.md on the different types of training that can be carried out on llm-compressor
  • Decouple training from text_generation.py::main. The new logic loves in llmcompressor/entrypoints/train.py that takes the flow of pre-process, carry out training logic and then post-process
  • Delete outdated info on transformers/finetune/readme.md
  • Update session_mixin.py to use session().initialize or session().finalize.
  • Deprecate train.py in text_generation.py, raising deprecation message if used.
  • Update tests to use llmcompressor's train, not llmcompressor.transformers' train

TEST PLAN:

  • Pass tests

Signed-off-by: George Ohashi <george@neuralmagic.com>
@github-actions
Copy link
Copy Markdown

👋 Hi! Thank you for contributing to llm-compressor. Please add the ready label when the PR is ready for review.

Note: This is required to complete the testing suite, please only add the label once the PR is code complete and local testing has been performed.

dsikka pushed a commit that referenced this pull request Mar 3, 2025
Order of reviews:
#1206
#1207 <-- Here
#1209 
#1212
#1214 

SUMMARY:
* Decouple arg parser to be used for both oneshot and train

TEST PLAN:
* Pass tests
dsikka added a commit that referenced this pull request Mar 5, 2025
Order of reviews:
#1206  <-- Here
#1207
#1209 
#1212
#1214 

SUMMARY:
Rename data_args to dataset_args

TEST PLAN:
Pass tests
FInd `data_args` using `grep`

---------

Signed-off-by: George Ohashi <george@neuralmagic.com>
Co-authored-by: Dipika Sikka <dipikasikka1@gmail.com>
dsikka pushed a commit that referenced this pull request Mar 5, 2025
Order of reviews:
#1206
#1207
#1209 <-- Here
#1212
#1214 

SUMMARY:
* Move dataset logic out of transformers module
`src/llmcompressor/transformers/finetune/data/data_helpers.py`, add it
to `src/llmcompressor/datasets/utils.py`


TEST PLAN:
Pass tests
dsikka pushed a commit that referenced this pull request Mar 6, 2025
…ot (#1212)

Order of reviews:
#1206
#1207
#1209
#1212  <-- Here
#1214

SUMMARY:
* Move the preprocessing and postprocessing logic out of
`src/llmcompressor/transformers/finetune/text_generation.py` and into
`src/llmcompressor/entrypoints/utils.py`

TEST PLAN:
Pass tests
@horheynm horheynm added the ready When a PR is ready for review label Mar 6, 2025
@horheynm horheynm changed the title [Train] Main refac [Train] Training Pipeline Mar 6, 2025
@rahul-tuli
Copy link
Copy Markdown
Collaborator

LGTM pending tests!

brian-dellabetta pushed a commit that referenced this pull request Mar 10, 2025
…ot (#1212)

Order of reviews:
#1206
#1207
#1209
#1212  <-- Here
#1214

SUMMARY:
* Move the preprocessing and postprocessing logic out of
`src/llmcompressor/transformers/finetune/text_generation.py` and into
`src/llmcompressor/entrypoints/utils.py`

TEST PLAN:
Pass tests

Signed-off-by: Brian Dellabetta <bdellabe@redhat.com>
Copy link
Copy Markdown
Collaborator

@dsikka dsikka left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please write a more descriptive PR description, summarizing changes and test steps.
The current description isn't very helpful.

Copy link
Copy Markdown
Collaborator

@dsikka dsikka left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@kylesayrs kylesayrs merged commit d43ea79 into main Mar 13, 2025
8 checks passed
@kylesayrs kylesayrs deleted the train branch March 13, 2025 14:24
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready When a PR is ready for review

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants