Skip to content

merge latest#8

Merged
Shixiaowei02 merged 645 commits intoShixiaowei02:developfrom
PaddlePaddle:develop
Mar 27, 2019
Merged

merge latest#8
Shixiaowei02 merged 645 commits intoShixiaowei02:developfrom
PaddlePaddle:develop

Conversation

@Shixiaowei02
Copy link
Owner

No description provided.

velconia and others added 30 commits March 14, 2019 10:20
* make timeline.py compatible
test=develop

* Follow comment
test=develop
PaddingRNN model memory optimize
2. Add binutils 2.27 in manylinux support

test=develop
test=develop
* Add cpu_quantize_squash_pass for C-API quantization

test=develop

* add cpu_quantize_squash_pass teste

* fix lint: add include memory unorderd_map and unordered_set

test=develop

* lint fix 2

* fixes

test=develop

* refactored

test=develop

* fix windows ci

test=develop
2. remove lod check in reshape op
test=develop
test=develop
Revert "PaddingRNN model memory optimize"
test=develop
test=develop
Fix cross_entropy2_op numeric error
test=develop
* Impl fp16 compute kernel for slice_op

test=develop

* Use data() to replace mutable_data()
* Support Sync Batch Norm.
* Note, do not enable it in one device.

Usage:

build_strategy = fluid.BuildStrategy()
build_strategy.sync_batch_norm = True
binary = fluid.compiler.CompiledProgram(tp).with_data_parallel(
        loss_name=loss_mean.name,
        build_strategy=build_strategy)
sneaxiy and others added 23 commits March 26, 2019 09:30
Move cpu_quantize_* passes into mkldnn subfolder
* test fix fetch bar place for ce

* fix ps mode dist train in develop test=develop

* fix style check test=develop

* update test=develop
* update jitkernel doc

test=develop

* follow comments

* follow comments

test=develop
Cherry-pick from PaddlePaddle:feature/anakin-engine: Anakin subgraph support.
polish deepCF model to support real dataset
* fix cdn issue, test=develop

* fix memory optimize bugs, test=develop

* fix memory optimize bugs, test=develop

* remove add/sub_2 op, test=develop

* disable memory_optimize by default, test=develop

* disable inplace activation in python, test=develop

* fix unittests, test=develop

* fix unittests, test=develop

* bug-fix, test=develop
* test=develop, add some Layers and tests

* test=develop, add more layers

* test=develop, add more layers

* test=develop, add force cpu option

* Update test_layers.py

remove pdb

* test=develop, refine code
* extend the slice function for python
test=develop
* test=develop, add install check

* test=develop, add install check scripts

* test=develop, refine language

* test=develop, add api spec

* test=develop, change cdn to bj to pass ci
Enhance gc to support deleting tensor buffer in advance
* Disable compare for accuracy issue.

test=develop

* Add todo comments to show more information.

test=develop
@Shixiaowei02 Shixiaowei02 merged this pull request into Shixiaowei02:develop Mar 27, 2019
Shixiaowei02 pushed a commit that referenced this pull request Jan 18, 2022
…addlePaddle#38914)

* Merged LoDTensor with Tensor,test=allcases

* Patched python level LoDTensor

* Patched python level LoDTensor

* Merge Tensor into DenseTensor

* Fixed namespace issues,test=allcases

* Fixed merge issues

* Fixed inference issues

* Fixed NPU test issues

* Fixed merge issues
Shixiaowei02 pushed a commit that referenced this pull request Apr 9, 2022
…ePaddle#41387)

* [Refactor] refactored eager_gen.py PR #2

* [DoubleGrad PR #1] Decoupled code generation logics for Dygraph ForwardFunctions and GradNodes

* Fixed minor issue

* Adjusted logics of GenerateNodeCreationCodes and GenerateForwardDefinition

* Fixed issues

* Supported higher-order grad node generation

* [DoubleGrad PR #4] Supported higher-order GradNode generation

* [DoubleGrad #4] Bug Fixes to Double Grad Node Generation

* Fixed yaml typo

* Fixed yaml typo

* fixed minor issues

* [DoubleGrad PR #5] Enabled gradient computations for grad_tensors passed to paddle.grad()

* Fixed minor issue

* Fixed CI-Inference issue

* Fixed CI-inference issues

* [DoubleGrad PR #7] paddle.grad() to copy backward graph before backward run

* Fixed minor issues

* Fixed issue with backward graph construction logic

* Fixed implementation issues with backward graph reconstruction

* Fixed unittest issue

* Fixed issues

* [DoubleGrad PR #8] Enabled triple grads for sigmoid and matmul

* Fixed issues with phi kernel

* Added triple grad test case

* Fixed minor issue
Shixiaowei02 pushed a commit that referenced this pull request Jun 2, 2022
Add Sigmoid triple grad tests
Shixiaowei02 pushed a commit that referenced this pull request Mar 22, 2023
* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* Pr 50885 (#7)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix code in a dy2static-friendly way.

* [dystatic] add hooker for prim

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix cast prim and vjp dtype mapping error bug

* [dy2static-ci] fix dy2static ci errors.

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>
Shixiaowei02 pushed a commit that referenced this pull request Mar 22, 2023
* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

* add unittest

* fix typo

* fix typo

* fix map.at

* fix find

* fix test

* fix cinn cache key structure realize

* using ordered map for attributes

* add test by review advice

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* Pr 50885 (#7)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

* add unittest

* fix typo

* fix typo

* fix map.at

* fix find

* fix test

* fix cinn cache key structure realize

* using ordered map for attributes

* add test by review advice

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix code in a dy2static-friendly way.

* [dystatic] add hooker for prim

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix cast prim and vjp dtype mapping error bug

* Cxx prim custom vjp (#8)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* Pr 50885 (#7)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix code in a dy2static-friendly way.

* [dystatic] add hooker for prim

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix cast prim and vjp dtype mapping error bug

* [dy2static-ci] fix dy2static ci errors.

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [Prim] enable whitelist and blacklist for custom_vjp

* support softmax grad

* remove additional code

* add test back

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>
Co-authored-by: xiongkun <[email protected]>
Shixiaowei02 pushed a commit that referenced this pull request Mar 22, 2023
* Add flatten composite rule

* get the right xshape and pass func test

* add cinn unit test

* Remove cinn test, wait for it to be added after repair

* add comp test to test_flatten_contiguous_range_op.py

* remove func test on composite_ops

* Add comments to maybe_wrap_dim func

* remove commented code

* fix the problem with 0D tensor case

* add flatten split rule comment

* fix syntax issues

* block flatten on resnet_prim_cinn

* init change

* tmp commit

* add layer_norm InferMeta check

* cast type modify

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

* add unittest

* fix typo

* fix typo

* fix map.at

* fix find

* fix test

* fix cinn cache key structure realize

* using ordered map for attributes

* add test by review advice

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* Pr 50885 (#7)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

* add unittest

* fix typo

* fix typo

* fix map.at

* fix find

* fix test

* fix cinn cache key structure realize

* using ordered map for attributes

* add test by review advice

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix code in a dy2static-friendly way.

* [dystatic] add hooker for prim

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix cast prim and vjp dtype mapping error bug

* recover

* big tol

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

* add unittest

* fix typo

* fix typo

* fix map.at

* fix find

* fix test

* fix cinn cache key structure realize

* using ordered map for attributes

* add test by review advice

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* Pr 50885 (#7)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

* add unittest

* fix typo

* fix typo

* fix map.at

* fix find

* fix test

* fix cinn cache key structure realize

* using ordered map for attributes

* add test by review advice

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix code in a dy2static-friendly way.

* [dystatic] add hooker for prim

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix cast prim and vjp dtype mapping error bug

* Cxx prim custom vjp (#8)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* Pr 50885 (#7)

* [CINN]Enhance CacheKey hash logic by considering input dtypes (PaddlePaddle#50557)

* [CINN]Enhance CacheKey hash logic by considering input dtypes

---------

Co-authored-by: jiangcheng <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix code in a dy2static-friendly way.

* [dystatic] add hooker for prim

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [prim] enable dygraph_to_static to support custom_vjp

* fix cast prim and vjp dtype mapping error bug

* [dy2static-ci] fix dy2static ci errors.

---------

Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>

* [Prim] enable whitelist and blacklist for custom_vjp

* debug log

* clear log

* fix

* nothing

* less memory

* recover utils

* fix

* modify threshold value

* skip layer_norm for test_bert

* back to bert success state

* add epsion

* delete unnecessary compute

* modify amp dtype

* modify * order

* delete sqrt check and fp16

---------

Co-authored-by: xuyongsheng <[email protected]>
Co-authored-by: xysheng-baidu <[email protected]>
Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: jiangcheng <[email protected]>
Co-authored-by: cxxly <[email protected]>
Co-authored-by: xiongkun <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.