Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

R2.1 #5651

Closed
wants to merge 24 commits into from
Closed

R2.1 #5651

wants to merge 24 commits into from

Conversation

ManfeiBai
Copy link
Collaborator

No description provided.

vanbasten23 and others added 24 commits August 29, 2023 01:09
Update more places

Add torch_pin
Summary:
This change enables megacore_dense by default to allow asynchorous cc
ops especailly for GSPMD.

Test Plan:
CI

Co-authored-by: Jiewen Tan <jwtan@google.com>
* Add optiona to unbundle libtpu

* Add clarifying note
* Fix fsdp not freeing forzen full params

* add test

* formatting

* remove unnecessary env var in test

Co-authored-by: Liyang90 <liyanglu@google.com>
* Update project metadata and remove useless files

* Update README

* Add manylinux platform tag

* formatting
* Add resnet50-weight-only-quant colab notebook

* update notebook with llama blog link

Co-authored-by: Siyuan Liu <lsiyuan@google.com>
Co-authored-by: Jiewen Tan <jwtan@google.com>
* Change `pjrt://` init method to `xla://` (#5560)

* Update PJRT documentation for the 2.1 release (#5557)

* Update PJRT documentation for the 2.1 release

* clarify plugins

* clarify PJRT doc

* Update `pjrt://` to `xla://`
…posing LoweringContext… (#5431) (#5580)

* Adding more explicit HLO lowering control by exposing LoweringContext (and utilities) to python for Neuron

* fixing linter issues

* fixing spacing

* apply comments and fix compilation errors

* add test for new apis

* fix linter

* update test

* update test

* modify test

* reverse back to GetIrValue()

* update test inputs with random numbers

* skip unittest because it only fails in CI

---------

Co-authored-by: aws-kingrj <78175353+aws-kingrj@users.noreply.github.com>
Co-authored-by: Ubuntu <ubuntu@ip-172-31-3-186.us-west-2.compute.internal>
Co-authored-by: seanlatias <seanlatias@gmail.com>
Co-authored-by: aws-kingrj <78175353+aws-kingrj@users.noreply.github.com>
* Move where clear pending IR is called to avoid crash

* fix CI

* fix CI and add some debugging messages
* Allow downcasting RngUniform genenration for Bernoulli

Co-authored-by: Yeounoh Chung <yeounoh@google.com>
* Enable autocast for XLA:GPU

* linter fix

* XLA autocast test for GPU and TPU

* linter fix

* Ensure that xla autocast is properly enabled for GPU and does not crash when torch cuda is not available.

* linter fix

* Add tests

* Support bf16

* linter fix

* exclude unsupported test cases

* increase GPU test timeout to 300

Co-authored-by: Yeounoh Chung <yeounoh@google.com>
Summary:
The 20230826 version contains openxla/xla@3b8a539
whch regresses our LLaMA2 GSPMD training benchmark. Let's rollback to version
before it.

Test Plan:
CI.

Co-authored-by: Jiewen Tan <jwtan@google.com>
* Fix log spam when libtpu is loaded (#5619)

* fix conflict

* only cherry-pick

---------

Co-authored-by: Will Cromar <wcromar@google.com>
…5647)

* lower NativeDropoutBackward (#5642)

* lower NativeDropoutBackward

* fix lowering and add python test

* lower native_dropout (#5643)

* prototype version (compiling error)

* Add native_dropout manual lowering.

* fix to tensor IR and add a simple native_dropout test

* fix data type issue and update test case

* fix IR hash issue

* fix corner case when probability==0

* remove typo line

* add test case when probability=0

---------

Co-authored-by: JackCaoG <59073027+JackCaoG@users.noreply.github.com>
Co-authored-by: zpcore <piz@google.com>
@ManfeiBai ManfeiBai closed this Sep 26, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants