Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

update #22

Merged
merged 57 commits into from
Aug 2, 2021
Merged

update #22

merged 57 commits into from
Aug 2, 2021

Conversation

AnnaTrainingG
Copy link
Owner

PR types

PR changes

Describe

thisjiang and others added 30 commits July 28, 2021 10:14
When Graph has sub-graph, apply pass to it and all sub-graph. And add single test script .
* [NPU] add NPU ops&uts of compare, test=develop

* testing

* try style-format

* [NPU] update compare_op_npu uts

* [NPU] fix code sytle of test_compare_op_npu.py
This PR added optional boolean is_parameter and stop_gradient in the VarDesc proto, and remove them during save_inference_model
* trt buildEngineWithConfig is deprecated

* add trt version control
* support ScaleTensor for scale npu kernel

* add more tests for adam npu

* fix compile

* fix unittest

* refine adam optimizer
* Support C++ import python on windows for paddle

* Support C++ import python on windows for paddle
* Add build_strategy in @to_static to support open pass

* fix os.environ

* add timeout

* disable test_build_strategy on openblas
* tile op

* more uts

* disable tile if trt6.0

* typo

* fix timeout issue

* opteller

* opteller remove duplicate code

* comments.	test=document_fix

* modify PADDLE_ENFORCE.

* fix reduce_mean issue
* add input option in model.summary
* add persistent_workers. test=develop
* fix paddle.summary's bug when output contains non-tensor
* graph engine demo

* upload unsaved changes

* fix dependency error

* fix shard_num problem

* py client

* remove lock and graph-type

* add load direct graph

* add load direct graph

* add load direct graph

* batch random_sample

* batch_sample_k

* fix num_nodes size

* batch brpc

* batch brpc

* add test

* add test

* add load_nodes; change add_node function

* change sample return type to pair

* resolve conflict

* resolved conflict

* resolved conflict

* separate server and client

* merge pair type

* fix

* resolved conflict

* fixed segment fault; high-level VLOG for load edges and load nodes

* random_sample return 0

* rm useless loop

* test:load edge

* fix ret -1

* test: rm sample

* rm sample

* random_sample return future

* random_sample return int

* test fake node

* fixed here

* memory leak

* remove test code

* fix return problem

* add common_graph_table

* random sample node &test & change data-structure from linkedList to vector

* add common_graph_table

* sample with srand

* add node_types

* optimize nodes sample

* recover test

* random sample

* destruct weighted sampler

* GraphEdgeBlob

* WeightedGraphEdgeBlob to GraphEdgeBlob

* WeightedGraphEdgeBlob to GraphEdgeBlob

* pybind sample nodes api

* pull nodes with step

* fixed pull_graph_list bug; add test for pull_graph_list by step

* add graph table;name

* add graph table;name

* add pybind

* add pybind

* add FeatureNode

* add FeatureNode

* add FeatureNode Serialize

* add FeatureNode Serialize

* get_feat_node

* avoid local rpc

* fix get_node_feat

* fix get_node_feat

* remove log

* get_node_feat return  py:bytes

* merge develop with graph_engine

* fix threadpool.h head

* fix

* fix typo

* resolve conflict

* fix conflict

* recover lost content

* fix pybind of FeatureNode

* recover cmake

* recover tools

* resolve conflict

* resolve linking problem

* code style

* change test_server port

* fix code problems

* remove shard_num config

* remove redundent threads

* optimize start server

* remove logs

* fix code problems by reviewers' suggestions

* move graph files into a folder

* code style change

* remove graph operations from base table

* optimize get_feat function of graph engine

* fix long long count problem

* remove redandunt graph files

* remove unused shell

* recover dropout_op_pass.h

* fix potential stack overflow when request number is too large & node add & node clear & node remove

* when sample k is larger than neigbor num, return directly

* using random seed generator of paddle to speed up

* fix bug of random sample k

* fix code style

* fix code style

* fix blocking_queue problem

* fix style

* fix

* recover capacity check

Co-authored-by: Huang Zhengjie <270018958@qq.com>
Co-authored-by: Weiyue Su <weiyue.su@gmail.com>
Co-authored-by: suweiyue <suweiyue@baidu.com>
Co-authored-by: luobin06 <luobin06@baidu.com>
Co-authored-by: liweibin02 <liweibin02@baidu.com>
Co-authored-by: tangwei12 <tangwei12@baidu.com>
qili93 and others added 27 commits July 29, 2021 20:50
* add fix op run order pass

* add ut for fix_op_run_order

* fix ci error

* improve coverage

* improve coverge again and fix cpu test case

* follow some comments
* fix lr in param group

* add unittest for adamw
* Support setitem by None index

* remove unreachable code

* Add Checkpoint for set_value_op because add a new attribute
* fix force kill for elastic
* [NPU] add clip and clip_grad on NPU, test=develop

* address review comments, test=develop

* update, test=develop
* added expand_v2 bf16/fp32 kernel

* minor change

* CI fix

* added missing test file

* added formatting

* reduced binary size

* CI fix
* add trainer desc config to distributed strategy

* code style modified
…ls (#34219)

* test version of matmul_v2

* added matmul_v2 grad kernel

* minor changes

* minor changes

* minor change for CI approval

* CI fix

* CI fix

* added squeeze and squeeze2 kernels

* CI fix

* CI fix

* CI fix

* disabled tests when compiled with cuda

* added setting format_tag by strides

* added sigmoid BF16 FWD/BWD and gelu BF16 BWD

* changes after review

* Revert "added sigmoid BF16 FWD/BWD and gelu BF16 BWD"

This reverts commit 6e3f767.

* Revert "Merge branch 'matmul_v2_grad' into squeeze2_op"

This reverts commit 06fcf67, reversing
changes made to 6e3f767.

* minor change

* added reshape1/2 kernels

* moved some functions into private block

* CI fix

* CI fix

* CI fix
* test version of matmul_v2

* added matmul_v2 grad kernel

* minor changes

* minor changes

* minor change for CI approval

* CI fix

* CI fix

* trigger CI

* changes after review, not working yet

* moved ops to anonymous namespaces

* changes after review
* add resnet50 trt test in pr-ci-inference test
The comment background message is too long, see details at #34521
* [NPU] add reduce_max

* [NPU] delete skipIf

* [NPU] add atrrs support or check

* [NPU] add attr out_dtype

* [NPU] delete debug codes
* notest;test=cpu-benchmark

* benchmark-cpu

* notest;test=cpu-benchmark

* notest;benchmark-cpu

* notest;benchmark-cpu

* notest;benchmark-cpu

* notest;benchmark-cpu

* notest;benchmark-cpu

* notest;benchmark-cpu

* fix

* fix

* add test_ci_model_benchmark.sh
* test=develop

* update identity

* add unittest

* notest,test=mac_py3

* modify comment & testname

* test=document_fix

* update comment

* test=document_fix

* activate all of the CI
@AnnaTrainingG AnnaTrainingG merged commit c1e59cf into AnnaTrainingG:develop Aug 2, 2021
AnnaTrainingG pushed a commit that referenced this pull request Sep 19, 2022
Refine print log and add args
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.