Skip to content

merge to local#22

Merged
junjun315 merged 64 commits intojunjun315:developfrom
PaddlePaddle:develop
May 7, 2019
Merged

merge to local#22
junjun315 merged 64 commits intojunjun315:developfrom
PaddlePaddle:develop

Conversation

@junjun315
Copy link
Owner

No description provided.

wzzju and others added 30 commits April 22, 2019 19:42
fix runtime_context_cache bug when gpu model has an op runs only on cpu
…parameter passing (#17058)

* resolve #17057

Fixed the bug that fuse_relu/fuse_residual option couldn't be passed to class TestConv2dInt8Op.

test=develop

* Fix the bug of test_conv2d_int8_mkldnn case which raised by improper parameter passing.

test=develop
add checkpoint functions for graph. test=develop
* implement distributed transpiler with fleet
implement dygraph.parallel.DataParallel to hook reduce op.
* Init mixed precision training interface

* Add fp16 test script

test=develop

* All initializers support float16

test=develop

* Code cleanup & add more code annotations

test=develop

* Update API spec

test=develop

* Add usage example in doc

test=develop
…near interpolation in forward (#17090)

* Cache the information of linear interpolation in forward and use it in backward.
test=develop

* Fix cuda kernel.
test=develop
backward of backward: leaky_relu
* Detailed coordinate description for yolov3 loss

test=develop

* modified api.spec

test=develop

* modified loss name

* fix api.spec

test=develop

* polish description

test=develop

* modified api.spec

test=develop
* fix python/paddle/fluid/__init__.py detecting problems
1. Use CudnnWorkspaceHandle in exhaustive search of conv_cudnn.
2. For Ops using CudnnWorkspaceHandle in exhaustive search, release their GPU memory after exhaustive search.

test=develop
* refine_dropout_mem,test=develop

* # This is a combination of 14 commits.
# The first commit's message is:
remove ut test_dist_word2vec in mac ci, will fix it in private, test=develop (#17066)

# This is the 2nd commit message:

Fleet unify distributed training (#16791)

* implement distributed transpiler with fleet
# This is the 3rd commit message:

ParallelDyGraph with GPU collective mode (#16827)

implement dygraph.parallel.DataParallel to hook reduce op.

# This is the 4th commit message:

Init mixed precision training interface (#16856)

* Init mixed precision training interface

* Add fp16 test script

test=develop

* All initializers support float16

test=develop

* Code cleanup & add more code annotations

test=develop

* Update API spec

test=develop

* Add usage example in doc

test=develop

# This is the 5th commit message:

fix reference_count_pass,test=develop (#17060)

test=develop
# This is the 6th commit message:

Speedup roi_perspective_transform op by caching the information of linear interpolation in forward (#17090)

* Cache the information of linear interpolation in forward and use it in backward.
test=develop

* Fix cuda kernel.
test=develop

# This is the 7th commit message:

remove unnecessary prepare_data (#17080)

test=develop
# This is the 8th commit message:

fix interpolate cu. test=develop (#17101)

# This is the 9th commit message:

test=develop, double backward leaky_relu (#17067)

backward of backward: leaky_relu
# This is the 10th commit message:

fix fuse optimizer ops (#17102)

test=develop
# This is the 11th commit message:

truncated_gaussian_random supported in distributed training, test=develop (#17091)

# This is the 12th commit message:

 Detailed coordinate description for yolov3 loss (#17007)

* Detailed coordinate description for yolov3 loss

test=develop

* modified api.spec

test=develop

* modified loss name

* fix api.spec

test=develop

* polish description

test=develop

* modified api.spec

test=develop

# This is the 13th commit message:

fix test_weight_decay (#17109)

test=develop
# This is the 14th commit message:

Path flag (#17105)

* fix python/paddle/fluid/__init__.py detecting problems
* 1. move the API check into CPU process
2. adjust the check order
cvm without LoD.
fix runtimeerror : dictionary changed size during iteration when calling uniform_random in python3+
* polish the label_smooth

test=develop

* polish code

test=develop
Update the folder name generation mechanism for saving the quantized model and weights.
The folder name would be unique by adding the timestamp postfix.

test=develop
fix python3 run_time_error in layers.ops caused by locals()
guru4elephant and others added 29 commits May 1, 2019 21:47
* remove async executor python api
test=develop

* remove test_async_executor.py
add executor train_from_dataset demo
test=develop

* fix import bug
test=develop
* remove unnecessary set_devices
* test=develop

* test=deelop
* add use_cuda to inplace pass,test=develop

* add test softmax_with_xe_inplace test,test=develop
* fix tensor_py,test=develop

* change class name,test=develop
test_distillation_strategy always failed on a machine with 4 gpus only, disable temporarily and need to figure out the root cause and add it back later
* fix profiler and name_scope API examples

test=develop

* update API.spec

test=develop
* fix distribute fpn proposals, test=develop
* fix unexecutable API comments, test=develop

* add API.spec,test=develop
)

* call SetNumThreads everytime to avoid missing omp thread setting

resolve #17153
test=develop

* add paddle_num_threads into config for test_analyzer_pyramid_dnn

resolve #17153
test=develop
* refine api comment, test=develop
* cherry-pick commit from 8877054

* cherry-pick commit from 3f0b97d

* cherry-pick from 16691:Anakin subgraph support yolo_v3 and faster-rcnn

(cherry picked from commit 8643dbc)

* Cherry-Pick from 16662 : Anakin subgraph cpu support

(cherry picked from commit 7ad182e)

* Cherry-pick from 1662, 16797.. : add anakin int8 support

(cherry picked from commit e14ab18)

* Cherry-pick from 16813 : change singleton to graph RegistBlock
test=release/1.4

(cherry picked from commit 4b9fa42)

* Cherry Pick : 16837 Support ShuffleNet and MobileNet-v2

Support ShuffleNet and MobileNet-v2, test=release/1.4

(cherry picked from commit a6fb066)

* Cherry-pick : anakin subgraph add opt config layout argument #16846
test=release/1.4

(cherry picked from commit 8121b3e)

* 1. add shuffle_channel_detect

(cherry picked from commit 6efdea8)

* update shuffle_channel op convert, test=release/1.4

(cherry picked from commit e4726a0)

* Modify symbol export rules

test=develop
* optimize sum op

fuse multi eigen kernel calls into one cuda kernel.
refine code

test=develop.

Signed-off-by: zhaoyuchen <[email protected]>

* Refine code.

test=develop

Signed-off-by: zhaoyuchen <[email protected]>

* Refine code according to comments.

test=develop

* refine code

delete sum_op_gpu.h
test=develop

* Fix test error.

test=develop

Signed-off-by: zhaoyuchen <[email protected]>

* refine code in format.

test=develop.

* refine code

test=develop

Signed-off-by: zhaoyuchen <[email protected]>

* refine code

test=develop

Signed-off-by: zhaoyuchen <[email protected]>
* Add MovingAverageAbsMaxScale operator which is only used for calculating the quantization scale.

* test=develop

* change the output into inplace. test=develop

* Revert "test=develop"

This reverts commit 696cf62.

* Revert "change the output into inplace. test=develop"

This reverts commit a19acd2.

* test=develop.

* update the MovingAverageAbsMaxScaleOp test. test=develop
* add attr axis infershape. test=develop

* add CUDA kernel. test=develop

* fix unittest. test=develop

* fix unittest for soft_label. test=develop

* fix fp16 unittest. test=develop

* remove comment code. test=develop

* refine test for axis. test=develop

* add python api. test=develop

* fix doc. test=develop

* fix fp16 unittest. test=develop

* fix ngraph test. test=develop

* fix ENFORCE for test_imperative_transformer. test=develop

* fit for ngraph test. test=develop

* fix after rebase develop. test=develop

* fix doc. test=develop

* fix API.spec. test=develop

* fix test_layers. test=develop

* fix format. test=develop
* remove unused FLAGS_warpctc_dir

test=develop

* remove FLAGS_warpctc_dir

test=develop
…tivations (#17235)

* fix api doc of hash, relu, concat, argmin, argmax, argsoft and all activations funcs with no attrs

test=develop

* refine doc example code

test=develop

* remove >>> in doc example

test=develop

* refine python code block

test=develop

* update API spec

test=develop
…p inplace (#17225)

* add use_cuda to inplace pass,test=develop

* add test softmax_with_xe_inplace test,test=develop

* fix potential inplace bug
test=develop

* add more skip vars in mem opt pass,test=develop

* follow comment,test=develop

* follow comments,move duplicate out arg check to program->graph,test=develop
@junjun315 junjun315 merged commit 0c73630 into junjun315:develop May 7, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.