Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

develop #1

Merged
merged 305 commits into from
Aug 18, 2021
Merged

develop #1

merged 305 commits into from
Aug 18, 2021
This pull request is big! We’re only showing the most recent 250 commits.

Commits on Jul 27, 2021

  1. Configuration menu
    Copy the full SHA
    ede001f View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    6ab0a6a View commit details
    Browse the repository at this point in the history
  3. add module api files (#34208)

    This PR add some primitives api related files. Those files are almost empty for the moment, but will be filled later.
    xingfeng01 authored Jul 27, 2021
    Configuration menu
    Copy the full SHA
    563a669 View commit details
    Browse the repository at this point in the history

Commits on Jul 28, 2021

  1. apply pass strategy to sub graph (#34158)

    When Graph has sub-graph, apply pass to it and all sub-graph. And add single test script .
    thisjiang authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    5e27d16 View commit details
    Browse the repository at this point in the history
  2. [NPU] add NPU ops of compare, test=develop (#34365)

    * [NPU] add NPU ops&uts of compare, test=develop
    
    * testing
    
    * try style-format
    
    * [NPU] update compare_op_npu uts
    
    * [NPU] fix code sytle of test_compare_op_npu.py
    Aganlengzi authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    68b4a2c View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    a59f215 View commit details
    Browse the repository at this point in the history
  4. graph_to_program save parameter and stop_gradient information (#33771)

    This PR added optional boolean is_parameter and stop_gradient in the VarDesc proto, and remove them during save_inference_model
    thisjiang authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    8a7dee3 View commit details
    Browse the repository at this point in the history
  5. [CPU-PSLIB] Fix bug for consistency insepection of op's embedding nam…

    …e and sparse table name in config_fleet.py (#34441)
    WorgenZhang authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    f1654de View commit details
    Browse the repository at this point in the history
  6. graph_to_program topology sort (#33949)

    See #33949 for details
    thisjiang authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    167523e View commit details
    Browse the repository at this point in the history
  7. [Paddle-TRT] Fix TRT8 cuda error before program exit (#34403)

    * trt buildEngineWithConfig is deprecated
    
    * add trt version control
    zlsh80826 authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    995195f View commit details
    Browse the repository at this point in the history
  8. fix optimizer.py (#34431)

    sljlp authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    0fb15d9 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    9f60492 View commit details
    Browse the repository at this point in the history
  10. [NPU] Support ScaleTensor for scale npu kernel (#34418)

    * support ScaleTensor for scale npu kernel
    
    * add more tests for adam npu
    
    * fix compile
    
    * fix unittest
    
    * refine adam optimizer
    zhiqiu authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    f17ba93 View commit details
    Browse the repository at this point in the history
  11. fix ci bug (#34445)

    sneaxiy authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    0b2e510 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    a83a368 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    54cc065 View commit details
    Browse the repository at this point in the history
  14. Support C++ import python on windows for paddle (#34312)

    * Support C++ import python on windows for paddle
    
    * Support C++ import python on windows for paddle
    zhwesky2010 authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    cf12ea5 View commit details
    Browse the repository at this point in the history
  15. [Dy2Stat]Add build_strategy in @to_static to support open pass (#34347)

    * Add build_strategy in @to_static to support open pass
    
    * fix os.environ
    
    * add timeout
    
    * disable test_build_strategy on openblas
    Aurelius84 authored Jul 28, 2021
    Configuration menu
    Copy the full SHA
    eb27d8b View commit details
    Browse the repository at this point in the history

Commits on Jul 29, 2021

  1. Configuration menu
    Copy the full SHA
    92d8fed View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    e958316 View commit details
    Browse the repository at this point in the history
  3. Tile supported (#34388)

    * tile op
    
    * more uts
    
    * disable tile if trt6.0
    
    * typo
    
    * fix timeout issue
    
    * opteller
    
    * opteller remove duplicate code
    
    * comments.	test=document_fix
    
    * modify PADDLE_ENFORCE.
    
    * fix reduce_mean issue
    b3602sss authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    cffa15c View commit details
    Browse the repository at this point in the history
  4. fix test scope (#34450)

    kuizhiqing authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    d3dae0c View commit details
    Browse the repository at this point in the history
  5. add parameter of input in model.summary (#34165)

    * add input option in model.summary
    wangna11BD authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    40bd7a7 View commit details
    Browse the repository at this point in the history
  6. fix lrn bug when shape=0

    huangjun12 authored and chajchaj committed Jul 29, 2021
    Configuration menu
    Copy the full SHA
    104c82b View commit details
    Browse the repository at this point in the history
  7. fix unit test bug

    huangjun12 authored and chajchaj committed Jul 29, 2021
    Configuration menu
    Copy the full SHA
    9e1af38 View commit details
    Browse the repository at this point in the history
  8. fix ci coverage bug

    huangjun12 authored and chajchaj committed Jul 29, 2021
    Configuration menu
    Copy the full SHA
    b451ff2 View commit details
    Browse the repository at this point in the history
  9. add persistent_workers (#34017)

    * add persistent_workers. test=develop
    heavengate authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    76710e5 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    d71b9ba View commit details
    Browse the repository at this point in the history
  11. Enable FLAGS_convert_all_blocks (#34452)

    As the title
    zhhsplendid authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    76f94f8 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    b56dbe0 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    02cc3c5 View commit details
    Browse the repository at this point in the history
  14. fix paddle.summary's bug when outputs contains non-tensor (#34160)

    * fix paddle.summary's bug when output contains non-tensor
    HydrogenSulfate authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    b7fac0f View commit details
    Browse the repository at this point in the history
  15. recover capacity check (#34478)

    * graph engine demo
    
    * upload unsaved changes
    
    * fix dependency error
    
    * fix shard_num problem
    
    * py client
    
    * remove lock and graph-type
    
    * add load direct graph
    
    * add load direct graph
    
    * add load direct graph
    
    * batch random_sample
    
    * batch_sample_k
    
    * fix num_nodes size
    
    * batch brpc
    
    * batch brpc
    
    * add test
    
    * add test
    
    * add load_nodes; change add_node function
    
    * change sample return type to pair
    
    * resolve conflict
    
    * resolved conflict
    
    * resolved conflict
    
    * separate server and client
    
    * merge pair type
    
    * fix
    
    * resolved conflict
    
    * fixed segment fault; high-level VLOG for load edges and load nodes
    
    * random_sample return 0
    
    * rm useless loop
    
    * test:load edge
    
    * fix ret -1
    
    * test: rm sample
    
    * rm sample
    
    * random_sample return future
    
    * random_sample return int
    
    * test fake node
    
    * fixed here
    
    * memory leak
    
    * remove test code
    
    * fix return problem
    
    * add common_graph_table
    
    * random sample node &test & change data-structure from linkedList to vector
    
    * add common_graph_table
    
    * sample with srand
    
    * add node_types
    
    * optimize nodes sample
    
    * recover test
    
    * random sample
    
    * destruct weighted sampler
    
    * GraphEdgeBlob
    
    * WeightedGraphEdgeBlob to GraphEdgeBlob
    
    * WeightedGraphEdgeBlob to GraphEdgeBlob
    
    * pybind sample nodes api
    
    * pull nodes with step
    
    * fixed pull_graph_list bug; add test for pull_graph_list by step
    
    * add graph table;name
    
    * add graph table;name
    
    * add pybind
    
    * add pybind
    
    * add FeatureNode
    
    * add FeatureNode
    
    * add FeatureNode Serialize
    
    * add FeatureNode Serialize
    
    * get_feat_node
    
    * avoid local rpc
    
    * fix get_node_feat
    
    * fix get_node_feat
    
    * remove log
    
    * get_node_feat return  py:bytes
    
    * merge develop with graph_engine
    
    * fix threadpool.h head
    
    * fix
    
    * fix typo
    
    * resolve conflict
    
    * fix conflict
    
    * recover lost content
    
    * fix pybind of FeatureNode
    
    * recover cmake
    
    * recover tools
    
    * resolve conflict
    
    * resolve linking problem
    
    * code style
    
    * change test_server port
    
    * fix code problems
    
    * remove shard_num config
    
    * remove redundent threads
    
    * optimize start server
    
    * remove logs
    
    * fix code problems by reviewers' suggestions
    
    * move graph files into a folder
    
    * code style change
    
    * remove graph operations from base table
    
    * optimize get_feat function of graph engine
    
    * fix long long count problem
    
    * remove redandunt graph files
    
    * remove unused shell
    
    * recover dropout_op_pass.h
    
    * fix potential stack overflow when request number is too large & node add & node clear & node remove
    
    * when sample k is larger than neigbor num, return directly
    
    * using random seed generator of paddle to speed up
    
    * fix bug of random sample k
    
    * fix code style
    
    * fix code style
    
    * fix blocking_queue problem
    
    * fix style
    
    * fix
    
    * recover capacity check
    
    Co-authored-by: Huang Zhengjie <270018958@qq.com>
    Co-authored-by: Weiyue Su <weiyue.su@gmail.com>
    Co-authored-by: suweiyue <suweiyue@baidu.com>
    Co-authored-by: luobin06 <luobin06@baidu.com>
    Co-authored-by: liweibin02 <liweibin02@baidu.com>
    Co-authored-by: tangwei12 <tangwei12@baidu.com>
    7 people authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    b9d6c98 View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    9d985ca View commit details
    Browse the repository at this point in the history
  17. add fix op run order pass (#34427)

    * add fix op run order pass
    
    * add ut for fix_op_run_order
    
    * fix ci error
    
    * improve coverage
    
    * improve coverge again and fix cpu test case
    
    * follow some comments
    sneaxiy authored Jul 29, 2021
    Configuration menu
    Copy the full SHA
    79e758c View commit details
    Browse the repository at this point in the history

Commits on Jul 30, 2021

  1. Configuration menu
    Copy the full SHA
    423ea97 View commit details
    Browse the repository at this point in the history
  2. fix lr in param group (#34468)

    * fix lr in param group
    
    * add unittest for adamw
    jerrywgz authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    3041605 View commit details
    Browse the repository at this point in the history
  3. Support setitem by None index (#34442)

    * Support setitem by None index
    
    * remove unreachable code
    
    * Add Checkpoint for set_value_op because add a new attribute
    zyfncg authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    f775bfc View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    6c09496 View commit details
    Browse the repository at this point in the history
  5. fix force kill for elastic (#34488)

    * fix force kill for elastic
    kuizhiqing authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    ba19398 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    5571c98 View commit details
    Browse the repository at this point in the history
  7. [NPU] add clip and clip_grad on NPU, test=develop (#34429)

    * [NPU] add clip and clip_grad on NPU, test=develop
    
    * address review comments, test=develop
    
    * update, test=develop
    qili93 authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    b68e36d View commit details
    Browse the repository at this point in the history
  8. Added expand_v2 BF16/FP32 FWD/BWD kernels (#34284)

    * added expand_v2 bf16/fp32 kernel
    
    * minor change
    
    * CI fix
    
    * added missing test file
    
    * added formatting
    
    * reduced binary size
    
    * CI fix
    jakpiase authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    41c4f72 View commit details
    Browse the repository at this point in the history
  9. add trainer desc config to distributed strategy (#34457)

    * add trainer desc config to distributed strategy
    
    * code style modified
    esythan authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    e6aacd1 View commit details
    Browse the repository at this point in the history
  10. Added reshape, reshape2, squeeze and squeeze2 BF16/FP32 FWD/BWD kerne…

    …ls (#34219)
    
    * test version of matmul_v2
    
    * added matmul_v2 grad kernel
    
    * minor changes
    
    * minor changes
    
    * minor change for CI approval
    
    * CI fix
    
    * CI fix
    
    * added squeeze and squeeze2 kernels
    
    * CI fix
    
    * CI fix
    
    * CI fix
    
    * disabled tests when compiled with cuda
    
    * added setting format_tag by strides
    
    * added sigmoid BF16 FWD/BWD and gelu BF16 BWD
    
    * changes after review
    
    * Revert "added sigmoid BF16 FWD/BWD and gelu BF16 BWD"
    
    This reverts commit 6e3f767.
    
    * Revert "Merge branch 'matmul_v2_grad' into squeeze2_op"
    
    This reverts commit 06fcf67, reversing
    changes made to 6e3f767.
    
    * minor change
    
    * added reshape1/2 kernels
    
    * moved some functions into private block
    
    * CI fix
    
    * CI fix
    
    * CI fix
    jakpiase authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    22c4c18 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    87148a5 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    2ad1e4c View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    06b55ea View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    44e4d57 View commit details
    Browse the repository at this point in the history
  15. Added matmul_v2 BF16/FP32 BWD kernel (#34192)

    * test version of matmul_v2
    
    * added matmul_v2 grad kernel
    
    * minor changes
    
    * minor changes
    
    * minor change for CI approval
    
    * CI fix
    
    * CI fix
    
    * trigger CI
    
    * changes after review, not working yet
    
    * moved ops to anonymous namespaces
    
    * changes after review
    jakpiase authored Jul 30, 2021
    Configuration menu
    Copy the full SHA
    0be7157 View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    72a9c8f View commit details
    Browse the repository at this point in the history

Commits on Jul 31, 2021

  1. add resnet50 trt tests in pr-ci-inference (#34465)

    * add resnet50 trt test in pr-ci-inference test
    OliverLPH authored Jul 31, 2021
    Configuration menu
    Copy the full SHA
    8b72a1a View commit details
    Browse the repository at this point in the history

Commits on Aug 1, 2021

  1. Configuration menu
    Copy the full SHA
    41e2d41 View commit details
    Browse the repository at this point in the history

Commits on Aug 2, 2021

  1. Configuration menu
    Copy the full SHA
    a6f55e4 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    393a0b1 View commit details
    Browse the repository at this point in the history
  3. Fix Inference CE Error by Topo Order (#34521)

    The comment background message is too long, see details at #34521
    zhhsplendid authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    508b40e View commit details
    Browse the repository at this point in the history
  4. [NPU] add reduce_max (#34179)

    * [NPU] add reduce_max
    
    * [NPU] delete skipIf
    
    * [NPU] add atrrs support or check
    
    * [NPU] add attr out_dtype
    
    * [NPU] delete debug codes
    windstamp authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    de53f2b View commit details
    Browse the repository at this point in the history
  5. Change Bbenchmark-CI CPU/GPU (#34305)

    * notest;test=cpu-benchmark
    
    * benchmark-cpu
    
    * notest;test=cpu-benchmark
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * fix
    
    * fix
    
    * add test_ci_model_benchmark.sh
    tianshuo78520a authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    9d54a53 View commit details
    Browse the repository at this point in the history
  6. Add Identity OP (#34420)

    * test=develop
    
    * update identity
    
    * add unittest
    
    * notest,test=mac_py3
    
    * modify comment & testname
    
    * test=document_fix
    
    * update comment
    
    * test=document_fix
    
    * activate all of the CI
    shiyutang authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    80f7f7e View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    c7cc5ac View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    af88699 View commit details
    Browse the repository at this point in the history
  9. Add basic functions of Program Pass (#34524)

    * add basic APIs
    
    * add attr_types
    
    * follow comments
    
    * change pass attr types
    
    * add set pass attribute codes
    
    * refine PADDLE_THROW
    sneaxiy authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    145cdb5 View commit details
    Browse the repository at this point in the history
  10. Change formula error in paddle.optimizer (#34539)

    * fix paddle.optimizer test=document_fix
    
    * fix paddle.optimizer test=document_fix
    sunzhongkai588 authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    3b5fc2a View commit details
    Browse the repository at this point in the history
  11. [HybridParallel]Support 1f1b for PipelineParallel (#34483)

    * support 1f1b for pipeline
    
    * add utest
    
    * add send_partial/recv_partial
    
    * support amp for pp
    
    * fix logger
    ForFishes authored Aug 2, 2021
    Configuration menu
    Copy the full SHA
    9e0bb91 View commit details
    Browse the repository at this point in the history

Commits on Aug 3, 2021

  1. Configuration menu
    Copy the full SHA
    61e51c1 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    e7dcdb7 View commit details
    Browse the repository at this point in the history
  3. support more dim for mul op npu (#34546)

    * support more dim for mul op npu
    
    * update unit test according to reviewer's comment.
    houj04 authored Aug 3, 2021
    Configuration menu
    Copy the full SHA
    2aedf16 View commit details
    Browse the repository at this point in the history
  4. support Kunlun2 (#34459)

    * support Kunlun2
    
    * support KL2
    
    * support KL2
    QingshuChen authored Aug 3, 2021
    Configuration menu
    Copy the full SHA
    2d0f3d9 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    d7493df View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    45fa14f View commit details
    Browse the repository at this point in the history
  7. [docker] delete install_trt in release docker (#34266)

    * delete install_trt in release docker
    
    * fix whl package path
    pangyoki authored Aug 3, 2021
    Configuration menu
    Copy the full SHA
    2714fc7 View commit details
    Browse the repository at this point in the history
  8. [HybridParallel] Support segment for PipelineParallel (#34529)

    * add layer segment
    
    * add segement for transformer
    
    * add utest
    ForFishes authored Aug 3, 2021
    Configuration menu
    Copy the full SHA
    9b6c7eb View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    56b7ebb View commit details
    Browse the repository at this point in the history

Commits on Aug 4, 2021

  1. Configuration menu
    Copy the full SHA
    c79fa1c View commit details
    Browse the repository at this point in the history
  2. supplement the function of slice. (#34172)

    * supplement the function of slice
    
    * edit unittest
    
    * strided_slice_op support .
    
    * polish error message.
    
    * polish error message.
    
    * polish code.
    
    * polish unittest.
    
    * polish code.
    
    * polish code
    
    * polish error message.
    hbwx24 authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    1f0f5d3 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    54b6c39 View commit details
    Browse the repository at this point in the history
  4. Fix backward bug (#34582)

    * fix backward bug
    
    * format code style
    
    * add test case for grad tensor accumulator
    MingMingShangTian authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    a7c3836 View commit details
    Browse the repository at this point in the history
  5. Support npu kernel for Crop op (#34542)

    * add crop npu op
    
    * code style check
    
    * update code to remove usless headers
    
    * add testcase and upsate code
    yeliang2258 authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    3ce14a3 View commit details
    Browse the repository at this point in the history
  6. [NPU] Support npu kernel for assign_value op (#34568)

    * [NPU] Support npu kernel for assign_value op
    
    * move test_assign_value_op_npu.py into unittests/npu folder
    
    * correce copyright year; add TestAssignApi class using NPUplace in test files
    betterpig authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    f39c3a5 View commit details
    Browse the repository at this point in the history
  7. Update benchmark-cpu (#34538)

    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * notest;benchmark-cpu
    
    * fix
    tianshuo78520a authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    ee60e82 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    91be876 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    1f76a2f View commit details
    Browse the repository at this point in the history
  10. Add gradient with optimizer API (#34395)

    * add gradients_with_optimizer api
    
    * modify gradients_with_optimizer
    
    * add gradients_with_optimizer api into paddle.auto.backward_mode
    
    * add gradients_with_optimizer test case
    
    * add doc for gradients_with_optimizer
    
    * add doc for gradients_with_optimizer
    MingMingShangTian authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    d9e63a8 View commit details
    Browse the repository at this point in the history
  11. paddle/nn/functional docs' bug fix (#34580)

    * fix paddle.optimizer test=document_fix
    
    * fix paddle.optimizer test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    sunzhongkai588 authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    420570c View commit details
    Browse the repository at this point in the history
  12. Revert pull request 34212 (#34558)

    * revert commit id 34212
    youth123 authored Aug 4, 2021
    Configuration menu
    Copy the full SHA
    0989211 View commit details
    Browse the repository at this point in the history

Commits on Aug 5, 2021

  1. Configuration menu
    Copy the full SHA
    090c863 View commit details
    Browse the repository at this point in the history
  2. [HybridParallel]Fix bug of p2p for partial_send/recv (#34615)

    * fix bug of p2p for partial
    
    * fix error
    ForFishes authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    4cc3d9a View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    8144a73 View commit details
    Browse the repository at this point in the history
  4. add not_equal NPU op (#34560)

    * add not_equal NPU op
    
    * add not_equal NPU op
    
    * add not_equal NPU op
    
    * add not_equal NPU op
    baoachun authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    7e707ce View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    4d6f8f2 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    e47d8a5 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    7a38b76 View commit details
    Browse the repository at this point in the history
  8. add NPU support for zero_copy_tensor. (#34629)

    * add NPU support for zero_copy_tensor.
    
    * revert unnesessary codes.
    
    * revert unnesessary codes.
    houj04 authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    a68709d View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    1d7b75d View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    911c859 View commit details
    Browse the repository at this point in the history
  11. remove boost::algorithm::ends_with ,boost macro and boost::lexical_ca…

    …st apis (#34310)
    
    * replace boost::algorithm::ends_with with self define ends_with function
    
    * remove BOOST macro in certain operators
    
    * remove boost::lexical_cast
    
    * add test for string_helper
    
    * add more test case for string_helper
    
    * modify join_string func and test case
    
    * fix build_strategy_test failed bug
    
    * remove string_helper_test from parallel_UT_rule.py
    MingMingShangTian authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    bb7b4c0 View commit details
    Browse the repository at this point in the history
  12. [Dy2Stat]Support Mixed Precision training in @to_static (#34562)

    * Support Mixed Precision training in @to_static
    
    * fix block.vars logic
    
    * fix GPU training loss diff
    
    * remove unused code
    Aurelius84 authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    a842828 View commit details
    Browse the repository at this point in the history
  13. fix output dtype for paddle.sum (#34313)

    * support bool dtype for paddle.sum
    GuoxiaWang authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    ff062a4 View commit details
    Browse the repository at this point in the history
  14. [Dy2st]Integrated gast library to fix compatibility problem permanent…

    …ly (#34556)
    
    * integrated gast library
    
    * integrated gast library
    
    * fix unittest and remove ast2.py
    
    * remove 'gast' from __all__ in __init__.py
    
    * add copyright in other files
    
    * fix copyright
    0x45f authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    a9ee383 View commit details
    Browse the repository at this point in the history
  15. New executor dev (#34407)

    * first test version
    
    * add test exec;
    
    * add data transfer; test=develop
    
    * add new exec head;
    
    * add memcpy; test=develop
    
    * add python fetch
    
    * add new test
    
    * add graph node; test=develop
    
    * remove useless new executor test; test=develop
    
    * remove gperf dependency; test=develop
    
    * fix compile bugs; test=develop
    
    * remove useless code; test=develop
    
    * remove useless code; test=develop
    
    * add uni test; test=develop
    
    * polish code; test=develop
    
    * polish code; test=develop
    
    * add interpreter cmakefile; test=develop
    
    * remove useless code; test=develop
    phlrain authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    012d12b View commit details
    Browse the repository at this point in the history
  16. [NPU] Add relu6 and relu6_grad npu op (#34596)

    * Add relu6 and relu6_grad npu op
    
    * fixed pre-commit-config.yaml
    
    * fixed for CI
    wjj19950828 authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    6839994 View commit details
    Browse the repository at this point in the history
  17. [NPU] Support npu op: (1) cos (2) cos_grad (#34573)

    * [NPU] Support npu op: (1) cos (2) cos_grad
    
    * Update test_cos_op_npu.py
    
    * Update activation_op_npu.cc
    
    * rm redundant {1}
    veyron95 authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    6151ccd View commit details
    Browse the repository at this point in the history
  18. rm detach (#34644)

    ForFishes authored Aug 5, 2021
    Configuration menu
    Copy the full SHA
    6c8a10a View commit details
    Browse the repository at this point in the history
  19. Configuration menu
    Copy the full SHA
    4a52c0c View commit details
    Browse the repository at this point in the history
  20. Configuration menu
    Copy the full SHA
    68377b4 View commit details
    Browse the repository at this point in the history

Commits on Aug 6, 2021

  1. Fix ut test_pe_fix_op_run_order by using smaller model and batch size (

    …#34603)
    
    * fix ut
    
    * decrease gpu memory consumption
    
    * remove exclusive
    sneaxiy authored Aug 6, 2021
    Configuration menu
    Copy the full SHA
    06651c4 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    436a9f1 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    c91b1e0 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    ce73349 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    c16421c View commit details
    Browse the repository at this point in the history
  6. Support npu kernel for eye op (#34543)

    * add eye npu op
    
    * remove useless headers
    
    * code style
    
    * Update eye_op_npu.cc
    
    * Update eye_op_npu.cc
    
    * remove useless code in test file
    
    * code style check
    
    * change Copyright to 2021
    
    * add test case and do some fix
    
    * fix
    
    * update code
    
    * fix for CI
    
    * return
    
    * fix
    yeliang2258 authored Aug 6, 2021
    Configuration menu
    Copy the full SHA
    6e442e6 View commit details
    Browse the repository at this point in the history
  7. paddle/nn fix formula bugs (#34643)

    * fix paddle.optimizer test=document_fix
    
    * fix paddle.optimizer test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix bugs in paddle.nn.functional document test=document_fix
    
    * fix nn formula bugs test=document_fix
    
    * fix nn formula bugs test=document_fix
    
    * fix nn formula bugs test=document_fix
    sunzhongkai588 authored Aug 6, 2021
    Configuration menu
    Copy the full SHA
    0f19ac7 View commit details
    Browse the repository at this point in the history
  8. [NPU]add reduce_prod (#34182)

    * [NPU] add reduce_prod
    
    * [NPU] delete check_dygraph=False
    
    * [NPU] delete skipIf
    
    * add attrs support or check
    
    * [NPU] delete extra codes for test_reduce_max_op_npu
    
    * [NPU] add attr out_dtype
    windstamp authored Aug 6, 2021
    Configuration menu
    Copy the full SHA
    47d81b0 View commit details
    Browse the repository at this point in the history
  9. [NPU] Support npu kernel for atan and atan_grad op, test=develop (#34658

    )
    
    * fix npu compile error, test=develop
    
    * [NPU] Support npu kernel for atan and atan_grad op, test=develop
    
    * [NPU] Support npu kernel for atan and atan_grad op, test=develop
    
    Co-authored-by: qili93 <qili93@qq.com>
    Liu-xiandong and qili93 authored Aug 6, 2021
    Configuration menu
    Copy the full SHA
    cabfb4a View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    8a9dc5d View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    fa16c21 View commit details
    Browse the repository at this point in the history
  12. support kunlun black list and add kl1 op (#34605)

    * support kunlun black list and add kl1 op
    
    * xpu_op_list add device_context dependence
    QingshuChen authored Aug 6, 2021
    Configuration menu
    Copy the full SHA
    21beef9 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    4caf60d View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    52e38a0 View commit details
    Browse the repository at this point in the history

Commits on Aug 7, 2021

  1. Configuration menu
    Copy the full SHA
    46808af View commit details
    Browse the repository at this point in the history

Commits on Aug 8, 2021

  1. Configuration menu
    Copy the full SHA
    338f9e0 View commit details
    Browse the repository at this point in the history

Commits on Aug 9, 2021

  1. [NPU] add broadcast supporting for elementwise_add_op_npu (#34057)

    * add broadcast supporting for elementwise_add
    
    * add broadcast supporting for elementwise_add
    
    * add more tests
    
    * remove the redundant code
    
    * update
    
    * fix place error in unittest
    
    * remove skip.If
    ronny1996 authored Aug 9, 2021
    Configuration menu
    Copy the full SHA
    b7355d8 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    0dff82c View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    898acb1 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    a3cc2d0 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    56759ff View commit details
    Browse the repository at this point in the history
  6. [NPU] add one_hot_op_npu and tests (#34258)

    * add one_hot_op and tests
    
    * update
    
    * make code clear
    ronny1996 authored Aug 9, 2021
    Configuration menu
    Copy the full SHA
    4c1ba73 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    aab4d6e View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    3380778 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    7afd31b View commit details
    Browse the repository at this point in the history
  10. fix_trt_int8 (#34704)

    Wangzheee authored Aug 9, 2021
    Configuration menu
    Copy the full SHA
    8009257 View commit details
    Browse the repository at this point in the history
  11. [NPU] add lock for npu_pinned_allocator (#34700)

    * add lock
    
    * fix typo
    zhiqiu authored Aug 9, 2021
    Configuration menu
    Copy the full SHA
    e285258 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    bf54534 View commit details
    Browse the repository at this point in the history

Commits on Aug 10, 2021

  1. Fix error of HSigmoidLoss (#34719)

    * Fix error of HSigmoidLoss
    
    * update unittest
    
    * update unittest
    linjieccc authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    3f32b73 View commit details
    Browse the repository at this point in the history
  2. Support npu kernel for expand_as_v2 op (#34620)

    * Support npu kernel for expand_as_v2 op
    
    * mofify the registry data type name
    
    * fix test unit
    
    * fix npu compile error, test=develop
    
    * fix compute function
    
    Co-authored-by: qili93 <qili93@qq.com>
    rainyfly and qili93 authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    202c240 View commit details
    Browse the repository at this point in the history
  3. Support npu kernel for tile op (#34606)

    * Support npu kernel for tile op
    
    * modify according to the comments
    
    * fix compute function
    rainyfly authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    8a6aa59 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    84eb675 View commit details
    Browse the repository at this point in the history
  5. fix for div zero (#34724)

    * fix for div zero
    
    * fix err;test=develop
    
    * fix lod
    zh794390558 authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    d86c26d View commit details
    Browse the repository at this point in the history
  6. copy boost/any.hpp to utils and replace boost::any with self defined …

    …any (#34613)
    
    * add any.hpp to utils and replace boost::any with self defined paddle::any
    
    * add copy any.hpp to custom op depends
    
    * modify any.hpp include path
    
    * remove boost from setup.py.in
    
    * add copy any.hpp to custom op depends
    
    * move any.hpp to paddle/utils/ dirs
    
    * move any.h to extension/include direction
    
    * copy utils to right directions
    MingMingShangTian authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    1289292 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    f30a5c4 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    a160379 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    4f4662b View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    cfd49ac View commit details
    Browse the repository at this point in the history
  11. [NPU] Support op kernel for Fill constant batch size like op (#34721)

    * fix npu compile error, test=develop
    
    * add fill constant batch size lilke op npu,test=develop
    
    Co-authored-by: qili93 <qili93@qq.com>
    andyjiang1116 and qili93 authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    ed2641c View commit details
    Browse the repository at this point in the history
  12. Support npu op fill_any_like (#34518)

    * Support npu kernel for fill_any_like op
    
    * modify the description of exception
    
    * remove useless template element
    
    * remove useless decorator
    
    * fix the code format error
    zyfncg authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    e8df322 View commit details
    Browse the repository at this point in the history
  13. [NPU] add squared_l2_norm squared_l2_norm_grad and tests (#34708)

    * [NPU] add squared_l2_norm squared_l2_norm and tests
    
    * [NPU] replace Square&ReduceSumD with SquareSumV1
    Aganlengzi authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    b64312f View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    8b9bd16 View commit details
    Browse the repository at this point in the history
  15. Kernel primitives api (#34672)

    添加Kernel primitives api: ReadData, WriteData ComputeFunctor
    AnnaTrainingG authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    8f9d573 View commit details
    Browse the repository at this point in the history
  16. [NPU] Support npu kernel for flatten_contiguous_range op, test=develop (

    #34642)
    
    * fix npu compile error, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * [NPU] Support npu kernel for flatten_contiguous_range op, test=develop
    
    * Update flatten_op_npu.cc
    
    * Update flatten_op_npu.cc
    
    Co-authored-by: qili93 <qili93@qq.com>
    Liu-xiandong and qili93 authored Aug 10, 2021
    Configuration menu
    Copy the full SHA
    79be842 View commit details
    Browse the repository at this point in the history

Commits on Aug 11, 2021

  1. Add no need output to gc check list (#34754)

    * add not used output var to gc_check_list; test=develop
    
    * add useless output to gc check list; test=develop
    phlrain authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    17c1dae View commit details
    Browse the repository at this point in the history
  2. [NPU] Support NPU kernel for TopKV2 op (#34599)

    * Add NPU kernel for TopKV2 op
    
    * deleted unnecessary cache file static_mode_white_list.cpython-37.pyc
    
    * A draft for error checking
    
    * A commit with accuracy error for float32 data
    
    * Modify codes according to the review comments
    
    * Modify codes according to the review comments
    From00 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    bb01b12 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    6a9fac1 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    4d2994c View commit details
    Browse the repository at this point in the history
  5. miss format (#34771)

    b3602sss authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    addd5fc View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    45af4f2 View commit details
    Browse the repository at this point in the history
  7. [NPU] Add exp and exp_grad npu op (#34612)

    * add exp and exp_grad npu op
    
    * modify support register type
    
    * remove empty line and remove exp_grad support data type int/int64
    
    * move exp and epx_grad kernel to activation_op_npu.cc, delete attrs
    
    * move code to activation_op_npu.cc
    0x45f authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    b5ec65e View commit details
    Browse the repository at this point in the history
  8. [HybridParallel] Support save/load for PipeLineParallel (#34768)

    * add save/load for pipelineparallel
    
    * add save/load
    ForFishes authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    88f2f4a View commit details
    Browse the repository at this point in the history
  9. add the basic apis for auto_parallel (#33804)

    * add auto_parallel apis
    lilong12 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    3f962e7 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    4d7af37 View commit details
    Browse the repository at this point in the history
  11. Add ext_tensor.slice() API (#34227)

    * Add ext_tensor.slice() API, test=develop
    
    * Call Tensor::mutable_data first to fix bugs and add test for writing to sliced tensor
    
    * Fix unit test bug
    
    * Fix code format problem, test=develop
    
    * Fix code format problem
    
    * Fix code format problem
    
    * strengthen unit test
    
    * Use CustomTensorUtils::ShareDataFrom to simplify codes
    Hao Lin authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    3f011d8 View commit details
    Browse the repository at this point in the history
  12. [NPU] add batch_norm_op_npu and test (#34056)

    * add batch_norm_op_npu and tests
    
    * remove skip.If
    
    * fix bug
    ronny1996 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    9ed5db2 View commit details
    Browse the repository at this point in the history
  13. [NPU] add reduce_mean_op_npu and test (#34053)

    * add reduce_mean_op_npu and test
    
    * remove skip.If
    
    * update
    ronny1996 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    f6fab55 View commit details
    Browse the repository at this point in the history
  14. [NPU] add momentum_op_npu and test (#34082)

    * add momentum_op_npu and test
    
    * update
    
    * fix hang
    ronny1996 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    9e3e08f View commit details
    Browse the repository at this point in the history
  15. split_op for npu (#34699)

    sljlp authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    d45d311 View commit details
    Browse the repository at this point in the history
  16. [NPU] add while, read_from_array and write_to_array npu op (#34755)

    * add while read_from_array write_to_array npu op
    
    * optimize unittest
    pangyoki authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    234c21a View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    fc537d4 View commit details
    Browse the repository at this point in the history
  18. [Paddle TRT]fix_fc_int8_convert; fix_reshape_convert (#34787)

    * fix_fc_reshape_convert
    
    * fix
    Wangzheee authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    3429c04 View commit details
    Browse the repository at this point in the history
  19. set_value_grad propagate gradients to Input and TensorValue (#3…

    …4304)
    
    * add set_value_grad op
    
    * add unittest.
    
    * polish unittest.
    
    * polish code.
    
    * support cuda kernel
    
    * polish code according to CI
    
    * polish code.
    
    * polish code
    
    * remove *.pyc
    
    * polish code.
    
    * add unittest to improve coverage.
    
    * polish code.
    hbwx24 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    9d02313 View commit details
    Browse the repository at this point in the history
  20. [AMP] add state_dict and load_state_dict and unittest for class GradS…

    …caler (#34300)
    
    * add state_dict and load_state_dict and unittest for class GradScaler
    
    * refine unittest for coverage of load_state_dict
    
    * refine comments of code-block
    
    * refine some comments
    
    * refine state_dict code and unittest
    
    * add #require gpu, xpu for GradScaler get/set example code
    
    * add #require gpu, xpu for GradScaler get/set example code
    
    * refine example code
    
    * refine unittest for state_dict
    
    * refine unittest for state_dict
    
    * fix bug of DataLoader in TestGradScalerStateDict
    
    * add flag FLAGS_cudnn_deterministic
    zhangbo9674 authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    99f8f5c View commit details
    Browse the repository at this point in the history
  21. [oneDNN] Fix to issue #34554 (#34623)

    * - Added softmax without caching
    
    * - Binary is no longer manually cached
    
    * - Activation onednn caching removed
    
    * - Removed manual caching of activation
    
    * - modified UT
    
    * - fix
    
    * - fix
    
    * - fixes to building
    
    * - fix
    
    * - fix
    
    * - fix to UT
    
    * - Faulty UT workaround
    
    * - approval workaround
    
    * - Fixes after review
    
    * - compilation fixes
    
    * - more lint fixes
    
    * - more fixes after review
    
    * - fixes after another round of review
    jczaja authored Aug 11, 2021
    Configuration menu
    Copy the full SHA
    0a5c99e View commit details
    Browse the repository at this point in the history

Commits on Aug 12, 2021

  1. add det_mv3_db & LeViT test case in pr-ci-inference (#34803)

    * add det_mv3_db & LeViT test case in pr-ci-inference
    
    * fix LeViT model dir bugs
    
    * fix grammar error
    OliverLPH authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    1c31d9d View commit details
    Browse the repository at this point in the history
  2. [NPU] Support npu op expand_v2 and expand_v2_grad (#34764)

    * [NPU] Support npu op expand_v2 and expand_v2_grad
    
    * [NPU] Support npu op expand_v2 and expand_v2_grad
    
    * [NPU] Support npu op expand_v2 and expand_v2_grad
    
    * update test_expand_v2_op_npu.py
    
    * update test_expand_v2_op_npu.py
    
    * modify expand_v2_op_npu.cc
    
    * modify expand_v2_op_npu.cc
    WorgenZhang authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    bc543e3 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    cfa6913 View commit details
    Browse the repository at this point in the history
  4. [HybridParallel]Add Recompute for PipeLineParallel (#34607)

    * add recompute for pp
    
    * add recompute offload
    
    * add recompute partition
    ForFishes authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    589d13c View commit details
    Browse the repository at this point in the history
  5. Fix safety-bug of functional.linear (#34696)

    * Fix safety-bug of functional.linear
    
    * Fix safety-bug of functional.linear
    
    * Fix safety-bug of functional.linear
    
    * Fix safety-bug of functional.linear
    Ray2020BD authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    0e28c8b View commit details
    Browse the repository at this point in the history
  6. transformer c files (#34706)

    This PR adds fused transformer related files defining c interface including class, function etc..
    xingfeng01 authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    016cc56 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    6326c3e View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    dffb0b2 View commit details
    Browse the repository at this point in the history
  9. Revert "[oneDNN] Fix to issue #34554 (#34623)" (#34838)

    This reverts commit 0a5c99e.
    chenwhql authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    dc62a22 View commit details
    Browse the repository at this point in the history
  10. Remove incorrect signal error stack trace (#34842)

    * remove unmatched signal error stack
    
    * fix error writing for cond
    chenwhql authored Aug 12, 2021
    Configuration menu
    Copy the full SHA
    572adcc View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    3f71e8d View commit details
    Browse the repository at this point in the history

Commits on Aug 13, 2021

  1. Configuration menu
    Copy the full SHA
    2164ad6 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    e92f038 View commit details
    Browse the repository at this point in the history
  3. [Bug-Fix]fix bug of py36 import utils (#34873)

    * fix bug of py36 import
    ForFishes authored Aug 13, 2021
    Configuration menu
    Copy the full SHA
    507ea06 View commit details
    Browse the repository at this point in the history
  4. fix npu_finalize (#34857)

    ronny1996 authored Aug 13, 2021
    Configuration menu
    Copy the full SHA
    17a9976 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    5b86b99 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    8bc4d85 View commit details
    Browse the repository at this point in the history
  7. Add EmptyGradOpMaker CI Approval (#34810)

    * Add EmptyGradOpMaker CI Approval, test=develop
    
    * Fix typo in echo_line
    Hao Lin authored Aug 13, 2021
    Configuration menu
    Copy the full SHA
    ac56d54 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    f421741 View commit details
    Browse the repository at this point in the history
  9. Bug fix : Can't load multiple modules of custom c++ op (#34505)

    * Fix a bug : can't load more than one custom op module
    
    * Fix a bug : can't load more than one custom op module
    
    * add test for load multiple modules of custom c++ op
    
    * add config for Coverage CI
    zyfncg authored Aug 13, 2021
    Configuration menu
    Copy the full SHA
    fc6b4a5 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    ff4bdac View commit details
    Browse the repository at this point in the history
  11. New Einsum API (#33821)

    * OP dot: refactor CPU kernels and get better loop performance.
    
    * Minor fix on code format.
    
    * Fixed minor errors.
    
    * Add new API: einsum
    
    * Update the Einsum unit test.
    
    One case failed with matmul_v2, where the dtype is int64:
    
    a = np.arange(2 * 3 * 1).reshape(2, 3, 1)
    b = np.arange(1)
    paddle.einsum("...i, ...i", a, b)
    
    * Test cases in test_einsum test floating point dtypes only.
    
    As of now Paddle only supports float/double dtypes in matmul, which is
    one of building blocks of this Einsum implementation. We decide not to
    test einsum against other dtypes.
    
    * Polish format.
    
    * More formatting.
    
    * Format...
    
    * Einsum: improve test coverage.
    
    * Einsum: bug fixes and more testcases for testing error messages
    
    * Einsum: fix format..
    
    * Einsum: fixed typo and format.
    
    * Einsum: format again...
    
    * Einsum: applied suggested changes.
    
    * Einsum API: improve API documentation.
    
    * Einsum API: apply suggested changes.
    
    * Einsum API: Add dygraph only note.
    
    * Einsum API: Add dygraph only note.
    
    * Einsum API: fixed unittest.
    tongxin authored Aug 13, 2021
    Configuration menu
    Copy the full SHA
    8c8667f View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    2cd05d5 View commit details
    Browse the repository at this point in the history

Commits on Aug 16, 2021

  1. [NPU] add p_norm_op_npu (#34695)

    * add p_norm_op_npu
    
    * remove p_norm_grad op
    
    * update
    ronny1996 authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    7316018 View commit details
    Browse the repository at this point in the history
  2. Op-benchmark CI cpu and gpu (#34631)

    * notest;pm-op-benchmark
    
    * notest;pm-op-benchmark
    
    * notest;pm-op-benchmark
    
    * notest;pm-op-benchmark
    
    * notest;pm-op-benchmark
    
    * notest;pm-op-benchmark
    
    * notest;test=op_benchmark
    
    * notest;test=op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * notest;test=op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * notest;op_benchmark
    
    * fix
    
    * fix
    tianshuo78520a authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    8fb17fc View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    34d188b View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    28279f6 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    e84b2e9 View commit details
    Browse the repository at this point in the history
  6. [dev] fix dice_loss bug (#34757)

    * fix dice_loss bug
    ghostxsl authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    ad6c3b9 View commit details
    Browse the repository at this point in the history
  7. Support npu op hard_swish and hard_swish_grad (#34608)

    * Support NPU OP hard_swish and hard_swish_grad
    
    * Support NPU OP hard_swish and hard_swish_grad
    
    * add the unittest to compare the result between npu ans cpu
    
    * format the prompt of exception
    
    * replace Min and Max op by ClipByValue op
    
    * fix the precision problem for fp16
    
    * Using HardtanhGrad to improve performace
    zyfncg authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    fd92d94 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    dc439a1 View commit details
    Browse the repository at this point in the history
  9. support margin loss (arcface, cosface, sphereface) for single GPU and…

    … cross GPUs (#34247)
    
    * support margin loss (arcface, cosface, sphereface)
    GuoxiaWang authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    b0cb414 View commit details
    Browse the repository at this point in the history
  10. [amp] dygraph amp support param_group (#34899)

    * dygraph amp support param_group
    
    * remove unused code
    
    * fix doc
    zhiqiu authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    e29c2d1 View commit details
    Browse the repository at this point in the history
  11. add unique_consecutive_op (#34334)

    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * add unique_consecutive_op
    
    * remove unity build
    
    * add unique_consecutive op
    
    * add unique_consecutive op
    
    * add enable static
    
    * add noqa
    
    * add space line
    
    * add default case.
    
    * add comma
    
    * add space line
    
    * modify unique_consecutive unittest
    
    * optimize ut coverage
    
    * rebase develop
    
    * improve coverage
    
    * update en docs
    
    * update en docs
    
    * update en docs
    
    * update en docs
    
    * update en docs
    
    * update en doc
    firestonelib authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    875cfd5 View commit details
    Browse the repository at this point in the history
  12. [NPU] Support NPU kernel for nearest_interp and nearest_interp_grad op (

    #34881)
    
    * Add NPU kernel for nearest_interp op
    
    * Add grad op
    
    * Modify codes according to the review comments
    
    * Modify codes according to the review comments
    From00 authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    e4e8cc9 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    3b9f040 View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    f6d8ab5 View commit details
    Browse the repository at this point in the history
  15. [oneDNN] Fix to 34554 (same as previous PR but should build with GPU) (

    …#34859)
    
    * - Added softmax without caching
    
    * - Binary is no longer manually cached
    
    * - Activation onednn caching removed
    
    * - Removed manual caching of activation
    
    * - modified UT
    
    * - fix
    
    * - fix
    
    * - fixes to building
    
    * - fix
    
    * - fix
    
    * - fix to UT
    
    * - Faulty UT workaround
    
    * - approval workaround
    
    * - Fixes after review
    
    * - compilation fixes
    
    * - more lint fixes
    
    * - more fixes after review
    
    * - fixes after another round of review
    
    * - hopefully compilation fix
    
    - compilation fix
    jczaja authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    9cb6565 View commit details
    Browse the repository at this point in the history
  16. Fix elementwise_add quantization (#34820)

    * Remove force_fp32_output from elementwise_add quantization
    
    * Fix cpu_quantize_placement test
    
    * Review related changes
    wozna authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    ae80df9 View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    d028214 View commit details
    Browse the repository at this point in the history
  18. Change the invoking method of settiem by Ellipsis and None index from…

    … numpy to set_value op (#34911)
    
    * Change invoking mathod of the settiem by Ellipsis and None index from numpy to set_value op
    
    * add none_axes into attr of set_value_op in dygraph mode
    zyfncg authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    2e30134 View commit details
    Browse the repository at this point in the history
  19. [NPU] Add size npu op (#34636)

    * add size npu op
    
    * modify support data type
    
    * no longer use NPU size OP
    
    * remove useless comments, add test case
    
    * fix copyright, remove useless include
    0x45f authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    4981894 View commit details
    Browse the repository at this point in the history
  20. hccl init sync (#34918)

    Baibaifan authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    6b4b9fe View commit details
    Browse the repository at this point in the history
  21. Jetson nano bilinear (#34751)

    * change bilinear thread for nano and tx2
    
    * change bilinear thread for nano and tx2
    feng_shuai authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    2a4ed08 View commit details
    Browse the repository at this point in the history
  22. [NPU] Support npu op:(1)arg_min (2)arg_max (#34867)

    * [NPU] Support npu op:(1)arg_min (2)arg_max
    
    * Modify and add unit test cases
    
    * Modify unit test cases
    veyron95 authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    b1cc4a4 View commit details
    Browse the repository at this point in the history
  23. Fix typos in English docs for diag and diagflat. (#34869)

    * Fix typos in english docs for diag and diagflat.
    limin2021 authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    35ef418 View commit details
    Browse the repository at this point in the history
  24. concurrent (#34908)

    lelelelelez authored Aug 16, 2021
    Configuration menu
    Copy the full SHA
    ed6624a View commit details
    Browse the repository at this point in the history

Commits on Aug 17, 2021

  1. Configuration menu
    Copy the full SHA
    181f7ce View commit details
    Browse the repository at this point in the history
  2. fix drop_last not work on IterableDataset (#34801)

    * fix drop_last not work in IterableDataset. test=develop
    heavengate authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    1614608 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    5de576b View commit details
    Browse the repository at this point in the history
  4. Add some passes which can be applied to Program (#34730)

    * add inplace passes and tests
    
    * update
    
    * fix use_cuda undefined
    fix compile error of op compat
    
    * add more ut
    
    * fix CPU CI error
    
    * check adam unique
    
    * fix mac/windows ci, improve coverage
    
    * fix ci error
    
    * follow weihang's comment
    
    * fix BlockDesc::MoveFrom
    
    * follow qiuliang's comment
    
    * update
    
    * follow huihuang's comments
    sneaxiy authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    8046e33 View commit details
    Browse the repository at this point in the history
  5. Align CTC grad scale same with ESPNet (#34729)

    * dygraph support more ctc grad scale
    
    * scale for 1.x
    
    * fix unitest
    
    * fix unitest
    
    * format code
    
    * fix unittest
    
    * fix log info
    
    * unittest cov
    
    * fix format;notest,test=cpu,coverage
    
    * skip ctc_loss egs;test=cpu
    
    * warpctc grad cov;test=coverage
    
    * add dygraph test;test=coverage
    
    * format;test=cpu,coverage
    
    * format;test=cpu
    
    * add api compat;test=cpu
    
    * add cpu test
    
    * rename
    
    * rename
    
    * fix
    
    * fix test
    
    * format
    
    * eigen cpu
    
    * eigen gpu grad pass
    
    * cuda gpu pass
    
    * format
    
    * fix ci
    zh794390558 authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    10f9644 View commit details
    Browse the repository at this point in the history
  6. add mkl multi-thread test cases in PR-CI-INFERENCE (#34946)

    * add mkl multi-thread test cases
    
    * fix codestyle
    
    * fix codestyle & enable ernie mkl test
    OliverLPH authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    9d4f00b View commit details
    Browse the repository at this point in the history
  7. [bug fix] fix unfold negative_size_param (#34943)

    * [bug fix] fix unfold negative_size_param
    ghostxsl authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    8ef1bf8 View commit details
    Browse the repository at this point in the history
  8. Modify the name of class in unittest with the same name (#34952)

    * polish unittest.
    
    * polish code
    
    * polish code
    hbwx24 authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    01a3a2e View commit details
    Browse the repository at this point in the history
  9. add exclude rules of pre-commit for paddle/utils and third_party (#34880

    )
    
    * add exclude rules of pre-commit to paddle/utils and third_party
    
    * remove exclude direction distributed/third_party
    
    * remove exclude of paddle/utils for format cpplint check
    MingMingShangTian authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    7b3295a View commit details
    Browse the repository at this point in the history
  10. [oneDNN ] disabling more ops caching (#34830)

    * - disabled caching of layer norm
    
    - fix in compilation
    
    - compilation fix
    
    - transpose caching disabled
    
    - compilation fix
    
    - more compilation fixes
    
    - sum caching disabled
    
    - compilation fix
    
    * - LRN with disabled cache
    
    * lint fixes
    jczaja authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    f1c1d9e View commit details
    Browse the repository at this point in the history
  11. Copy boost optional to Paddle (#34780)

    * copy boost optional.hpp to paddle
    
    * copy boost optional.hpp to paddle
    
    * move directions
    
    * del fluid/utils
    
    * modify .hpp to .h
    
    * move directions
    
    * modify to paddle::optional
    
    * add modification description
    
    * format code stype for the files in paddle/utils
    
    * format code stype
    MingMingShangTian authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    9be4144 View commit details
    Browse the repository at this point in the history
  12. Update op-benchmark CI (#34962)

    * fix op-benchmark
    
    * test=document_fix
    tianshuo78520a authored Aug 17, 2021
    Configuration menu
    Copy the full SHA
    690f583 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    1ef2185 View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    b4474fb View commit details
    Browse the repository at this point in the history

Commits on Aug 18, 2021

  1. [Paddle-TRT] unitest_quant_dequant (#34929)

    * unitest_quant_dequant
    
    * fix
    
    * fix
    
    * deleted: test_trt_quant_conv2d_dequant_fuse_pass.py
    
    * fix
    Wangzheee authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    c7070cb View commit details
    Browse the repository at this point in the history
  2. support class center sample of PartialFC (#34106)

    * support class center sample of PartialFC
    GuoxiaWang authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    100db44 View commit details
    Browse the repository at this point in the history
  3. [NPU] add retry on HcclGetRootInfo to fix "bind fail" (#34977)

    * add retry for HcclGetRootInfo
    
    * refine code
    
    * reduce retry interval
    zhiqiu authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    52a7b0c View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    12bf046 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    a137371 View commit details
    Browse the repository at this point in the history
  6. Fix bug in alltoall (#34975)

    lilong12 authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    2e9a31e View commit details
    Browse the repository at this point in the history
  7. Add function to disable paddle signal handler (#34577)

    * Add function to disable paddle signal handler
    
    Paddle used google::InstallFaultSignalHandler to handle selected system signals,
    mainly for debugging and bug report purposes.
    
    However, this can be conflicted with other python packages whoever captures similar signals.
    Such python package involves tvm and more
    
    To resolve this issue, we support a function to disable signal handler
    
    * Remove signal test from WIN32 platform
    
    * Remove redundant return from disable_signal_handler() function
    
    * Add detailed messages to en_doc
    jim19930609 authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    dd533dd View commit details
    Browse the repository at this point in the history
  8. [CustomOp] Fix ext_tensor.cast failed bug (#34884)

    * fix ext_tensor.cast failed bug
    
    * remove useless deps
    
    * fix windows cmake failed
    
    * try to fix windows make failed
    
    * fix make error on windwos
    chenwhql authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    4d88cdb View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    8967a66 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    209075a View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    51939c8 View commit details
    Browse the repository at this point in the history
  12. [Hybrid Performance] Move the cast op of AMP which cast fp32 param to…

    … fp16 param to the optimizer (#34965)
    wangxicoding authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    a9673b4 View commit details
    Browse the repository at this point in the history
  13. [NPU] Add leaky Relu (#34894)

    * test=develop
    
    * test=develop
    Jackwaterveg authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    40f6273 View commit details
    Browse the repository at this point in the history
  14. [NPU] Add square grad (#34889)

    * test=develop
    
    * test=develop
    Jackwaterveg authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    1b71a71 View commit details
    Browse the repository at this point in the history
  15. Configuration menu
    Copy the full SHA
    1b747de View commit details
    Browse the repository at this point in the history
  16. code refactoring for new executor (#34970)

    * code refactoring, test=develop
    
    * refine, test=develop
    
    * refine, test=develop
    
    * refine, test=develop
    wanghuancoder authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    40d4d83 View commit details
    Browse the repository at this point in the history
  17. fix pad outliers err (#34979)

    * fix pad outliers err
    
    * fix pad api input type and doc
    
    * fix example of pad
    
    * add unittest for pad3d
    
    * fix unittest
    
    * fix error format
    
    * fix pad doc
    littletomatodonkey authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    248e27b View commit details
    Browse the repository at this point in the history
  18. Add NPU kernel for norm Op: float16 and float32 (#34609)

    * Add NPU kernel for norm Op: float16 and float32
    
    * fix code for code review
    
    * fix for code review
    
    * add type for paddle_throw
    
    * remove unnecessary head file.\nAdd more testcase
    
    * remove a broadcast
    2742195759 authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    755c8a1 View commit details
    Browse the repository at this point in the history
  19. [NPU]add rmsprop op (#34864)

    * [npu]add rmsprop op
    lzzyzlbb authored Aug 18, 2021
    Configuration menu
    Copy the full SHA
    9cbba97 View commit details
    Browse the repository at this point in the history