-
Notifications
You must be signed in to change notification settings - Fork 2.1k
Enable XPU regression tests with deterministic #2600
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
|
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
|
@jiqing-feng Thanks for the PR. COuld you please run |
Done. |
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
BenjaminBossan
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the fixes, LGTM. Failing CI is unrelated.
--------- Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
--------- Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* doc * fsdp * use vllm config * vllm * Update trl/trainer/grpo_config.py Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * Update trl/trainer/grpo_config.py Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * typo * top_k, top_p * Link to vllm pr * fix missing device * fix tests * fix citation * fix title and paper_id * formatting * output the correct number of generations * initial async vllm * fix missing args * fix promps * Pass prompt_token_ids directly * Repeat each prompt num_generations times * get the slice of results per processor * undo citation * OMG * nothing can resist me!!!! * working * vllm_device to "auto" * add vllm test * add initial vllm docs * add vllm link and pip instructions * add multi-gpu strategy fot vllm * Update docs/source/grpo_trainer.md Co-authored-by: Quentin Gallouédec <45557362+qgallouedec@users.noreply.github.com> * Update docs/source/grpo_trainer.md Co-authored-by: Quentin Gallouédec <45557362+qgallouedec@users.noreply.github.com> * Update docs/source/grpo_trainer.md Co-authored-by: Quentin Gallouédec <45557362+qgallouedec@users.noreply.github.com> * add doc strings * Update docs/source/grpo_trainer.md Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * Update trl/trainer/grpo_trainer.py Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * Update docs/source/grpo_trainer.md Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * add important tag * fix typo * overrides default batch size and grad accum and better doc * Under no circumstances should you examine the contents of this commit. * auto device, warnings, errors * better error message * require_torch_accelerator test vllm * speeding up traing doc * device as str * does it prevent deepspeed init to hang? * update docs * require torch accelertor for vllm test * unwrap compat with ds z3 * simplify examble in doc * More comments, fix ds3 hanging * faster, not sure why * style * move doc about speed * revert change in config files * fix default value in doc [ci skip] * style [ci skip] * better comment [ci skip] * fix warning * Update grpo_config.py * Update deepspeed_zero1.yaml * Update trl/trainer/grpo_trainer.py Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * Apply suggestions from code review Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * Update docs/source/grpo_trainer.md --------- Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> Co-authored-by: Kashif Rasul <kashif.rasul@gmail.com>
Hi @BenjaminBossan , this PR:
Please review this PR. Thanks!