Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Flatten kernel refactor #8

Closed

Commits on Sep 18, 2021

  1. Configuration menu
    Copy the full SHA
    e761751 View commit details
    Browse the repository at this point in the history

Commits on Sep 19, 2021

  1. Configuration menu
    Copy the full SHA
    9f88d32 View commit details
    Browse the repository at this point in the history
  2. Optimization of pool2d grad (PaddlePaddle#35389)

    * Optimization of pool2d grad, first commit.
    
    * remove useless print codes
    
    * refine codes
    
    * refine codes
    
    * seal more operation into template specialization
    
    * fix template struct error in MaxPool2dGrad.
    
    * Fix header including error
    
    * refine code with comment
    
    * Seal the param-preparation codes into function for common use.
    
    * Seal the param-preparation codes into function for common use.
    
    * Seal the param-preparation into funciton and make it common for other kernels
    
    * polish code and erase useless template speicalization
    
    * Rerun triger
    
    * rerun trigger
    JamesLim-sy committed Sep 19, 2021
    Configuration menu
    Copy the full SHA
    8668519 View commit details
    Browse the repository at this point in the history

Commits on Sep 20, 2021

  1. Reuse OneDNN handler for SGD and SUM for SelectedRows input tensors. (P…

    …addlePaddle#35510)
    
    * Create stateful OneDNNAXPYHandler object.
    
    This makes it possible to call it multiple times without recreating the
    oneDNN primitives every time.
    
    * Prepare SGDOpKernel to reuse its implementation from OneDNN kernel.
    
    * OneDNN SGD kernel.
    
    * Update call to use new OneDNNAXPYHandler object api.
    
    * Setup seed in proper place.
    
    * Enable OneDNN kernel only for single case.
    
    * For dense param and sparse grad.
    
    * Small refactor.
    
    * Enable oneDNN by op attr or by cmd line flag.
    
    * Use int64_t type for number of elements.
    
    * Support dense param and grad from OneDNN kernel.
    
    * Enable SGD OneDNN kernel when use MP BF16 optimizer.
    
    * Force non-copyable/movable OneDNNAXPYHandler.
    
    * Reuse OneDNNAXPYHandler for spare tensors in SUM op.
    
    * Fix SFINAE rules.
    
    * Remove recording event inside AXPY.
    
    * Get rid of internal primitive caching.
    
    * Stop use PP cache mechanims to store mem and primitive obj.
    * Handler obj store and reuse needed desc & prim
    
    * Do not derive from MKLDNNHandlerT
    arogowie-intel committed Sep 20, 2021
    Configuration menu
    Copy the full SHA
    799f386 View commit details
    Browse the repository at this point in the history

Commits on Sep 21, 2021

  1. Configuration menu
    Copy the full SHA
    087c23a View commit details
    Browse the repository at this point in the history

Commits on Sep 22, 2021

  1. Configuration menu
    Copy the full SHA
    7713430 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    12ab017 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    3c0fea7 View commit details
    Browse the repository at this point in the history

Commits on Sep 28, 2021

  1. refactor flatten kernel

    YuanRisheng committed Sep 28, 2021
    Configuration menu
    Copy the full SHA
    4b498c4 View commit details
    Browse the repository at this point in the history