Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Wait for #2663][Mixed Precision] Fix gradient clipping logic #2749

Open
wants to merge 52 commits into
base: main
Choose a base branch
from

Commits on Aug 29, 2024

  1. [SWAP] Add swap mode property

    Signed-off-by: Jiho Chu <[email protected]>
    jihochu authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    a5d16a4 View commit details
    Browse the repository at this point in the history
  2. [SWAP] Add inference mode

    Signed-off-by: Jiho Chu <[email protected]>
    jihochu authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    23e40da View commit details
    Browse the repository at this point in the history
  3. [SWAP] Modify cache for inference mode

    This patch is for inference mode for swap device.
    It re-enable mmap feature, but writing time is controlled manually, due
    to the inference mode handling.
    
    Signed-off-by: Jiho Chu <[email protected]>
    
    asdfsadf
    
    Describe a commit content (Until 80 colums per line) in detail ASAP.
    
    **Changes proposed in this PR:**
    - Added TOC generator for README.md
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    jihochu authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    ba67088 View commit details
    Browse the repository at this point in the history
  4. [ hgemm ] Add hgemm experimental kernel

    - According to current paper, accumulating up to 64 ~ 128 w.r.t. K-direction is fine.
    - Since conventional error metric, and newly introduced metric (max component relative error) is fine as well, introduce experiemntal kernel.
    - using build option -Dhgemm-precision-level=low can enable such kernel when android build
    
    **Self evaluation:**
    1. Build test:     [X]Passed [ ]Failed [ ]Skipped
    2. Run test:     [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: skykongkong8 <[email protected]>
    skykongkong8 authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    a568c93 View commit details
    Browse the repository at this point in the history
  5. [ Weight ] Add Var32 Tensor in Weight.

    We will add Var32 Tensor if the Variable Weight is not Full
    precision (FP32). This eables the Weight Update with full precision
    and only Apply Gradient Process ueses this Tensor. Therefore, the
    lifespan of this tensor should be "ApplyGradient".
    
    . Modify TensorPool to generate Weigth considering Mixed Precsion.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    19da71d View commit details
    Browse the repository at this point in the history
  6. [ Mixed ] Create weight with var32 tensor

    This pr create the variable fp32 tensor when we create the Weight and
    Optimizer Weight.
    
    . update the manager to create Weight with  var32 tensor which
    requested to weight pool.
    . update the weight requests with Weight Spec and var, grad and var32
    tensors which created already.
    . add clone Tensor with specific type in tensor.h
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    5406860 View commit details
    Browse the repository at this point in the history
  7. [ Layers ] Update Layers to support FP16

    This PR enables the FP16 support for the layers below:
    
    . input layer
    . mse loss layer
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    b4c663e View commit details
    Browse the repository at this point in the history
  8. [ Test ] Mixed Precision Test Case

    This PR includes the mixed precision test case.
    
    . Input - FC - MSE
     : "batch_size=2", "model_tensor_type=FP16-FP16", "loss_scale=128"
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    5444fa0 View commit details
    Browse the repository at this point in the history
  9. [ Optimizer ] Update Optimizer / Adam to support Mixed training

    This commit modify apply gradient in optimizer.
    We do not need to save optimizer variables in weight type. Only
    Optimizer needs the optimizer variables and we should update the
    weight with full precision to maintain the accuracy. Therefore,
    remove the var32 tensors for optimizer variables.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    292eb71 View commit details
    Browse the repository at this point in the history
  10. [ Tensor ] add is_NaN check in Tensor

    This PR add is_NaN function to check if the tensor has NaN value. This
    is for the check NaN during mixed precision training.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    ae868ef View commit details
    Browse the repository at this point in the history
  11. [ Context ] Add loss scale in Context & using mse loss

    This PR add loss scale parameter in runcontext and use it to update
    mse loss.
    
    . Add Loss Scale Parameter in RunLayerContext Constructor
    . Add applyLossScale func to update return derivitive in Loss Layer
    . Change MSE Loss Layer to apply the loss scale to return derivitive
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    e0596ef View commit details
    Browse the repository at this point in the history
  12. [ Mixed Precision ] Enable Mixed Precision

    This PR enables the Mixed Precision Training. For now only FP16-FP32
    is considered. Additional Test cases will be added.
    
    . add getSortedLayerIdx to set the graph order for fowarding.
    . change clip_weights to lazy_apply_weights to use both cases.
    . add fowarding_op to run forwarding from that layer which has a
    gradient with nan.
    . add while loop for re-run backwarding after reset the loss scale.
    . add setLossScale in RunLayerContext
    . add check the gradient if mixed precsion enable.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    4b7a3ba View commit details
    Browse the repository at this point in the history
  13. [ Tensor ] Add inifinity check in Tensor

    This PR add inifinity value check in Tensor data.
    . rename the hasNaN to isValid
    . add infinity check in isValid Function and now it check NaN and Inf
    . modify to check the blas_avx and blas_neon
    . modify graph and model check is_valid rather than has_nan
    . add unittest of isValid Function
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    76efb04 View commit details
    Browse the repository at this point in the history
  14. [ MSE ] Fix for better MSE loss precision

    This PR chage the loss computation using full precsion rather than
    half precsion to maintain accuracy.
    
    **Changes proposed in this PR:**
    - Added TOC generator for README.md
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    376e67a View commit details
    Browse the repository at this point in the history
  15. [ TEST ] Add Torch Mixed Precision Model Test

    This PR enables the Mixed Precsion Unittest with Torch Model.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    d9242f1 View commit details
    Browse the repository at this point in the history
  16. [ TEST ] add torch input and output test data for mixed precision

    This PR add torch mixed precsion golden data generation and input and
    output for test.
    
    . some fixes to test.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    7d664ff View commit details
    Browse the repository at this point in the history
  17. [ TEST ] Add more unittest and fixes for mixed precsion

    This PR includes more unittest and fixes for mixed precsion.
    . Model Unittest
      . 2 fc layer which generate NaN or Inf Gradient from Troch.
      . MSE Loss and Check whole procedure of the mixed precsion training.
      . Even if the FC model only have one weight, but it is good enough
      to validate the mixed precsion.
      . Torch model also work similar way of NNTrainer.
      . Some fixes about the exeuction order of apply gradient when the
      mixed precision is on.
      . Update SGD to support Mixed Precision training
    
    **Changes proposed in this PR:**
    - Added TOC generator for README.md
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    afc2757 View commit details
    Browse the repository at this point in the history
  18. [ Layer ] Update Conv2D to support Mixed Precision

    This PR update the conv2D Layer to support Mixed Precision (FP16).
    It is based on the PR nnstreamer#2579
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    8c39c64 View commit details
    Browse the repository at this point in the history
  19. [ Layer ] enable Mixed Precision in LSTM Layer

    This commit enables mixed precision support for LSTM Layer.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    f38b831 View commit details
    Browse the repository at this point in the history
  20. [ Model ] Add Execution Mode in Compile

    This PR add Execution Mode parameter when we compile. The default is
    ml::train::ExeuctionMode::TRAIN. Currently we do not support compiler
    optimization for inference mode such as batch normalization fusing,
    etc. But we will add more optimization depending on the exeuction
    mode.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    d47102a View commit details
    Browse the repository at this point in the history
  21. [ Layer ] Mixed Precision support for BN Layer

    This PR includes Mixed Precision support for batch normalization
    layer. When the training, BN layer should run full precsion with FP16
    Weight data. Therefore, Reading the FP16 data read and data coversion
    of the current Weight and Activation are required.
    
    For the Inference, we do need compiler optimization like bn fusing. So
    it also includes execution mode parameters for compile.
    
    Because of compilcate data conversion of bn layer, test case
    generation also needs to update, so that taking the fp16 input,output
    tensors and weights and converting FP32 weight for computation.
    For veification, we do need convert FP32 to FP16.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    07b48ec View commit details
    Browse the repository at this point in the history
  22. [layer] enable mixed precision - reshape_layer

    enable mixed precision on reshape layer
    - reshape layer only change dim, so change dimensions and check datatype
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Donghak PARK <[email protected]>
    DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    b201807 View commit details
    Browse the repository at this point in the history
  23. [Layer] Enable mixed precision - pooling2d_layer

    Enable Mixed precision on Pooling 2D Layer
    - I modified it to properly cast for the case of FP16 so that the mixed precision function can be activated on the existing pooling 2d layer.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Donghak PARK <[email protected]>
    DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    825f07c View commit details
    Browse the repository at this point in the history
  24. [ Model ] Fix the gradient clipping for the FP16 or Low bit Gradient

    In this PR, when we compute the l2norm of gradient tensor, it converts
    to full precsion and computes the l2norm for gradient clipping.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    2184452 View commit details
    Browse the repository at this point in the history
  25. [ Layer ] Add mu and var backup up tensor.

    This PR add the mu and var backup tensor ( mu_b, var_b ) to restore
    the previous moving mean and moving variance for mixed precsion
    training.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    81fd9cb View commit details
    Browse the repository at this point in the history
  26. [ Layer ] prevent randomize when it restore the data

    In order to resotore previous iteration data, this pr disable
    randomnization of mask if it need restore previous data.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    003a5ce View commit details
    Browse the repository at this point in the history
  27. [ Context ] add check if it needs restore previous data

    This PR enable the check if it need restore previous data. By doing
    this, we can remove the NaN or Inf data in Tensor for the mixed
    precsion training.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    57a03ab View commit details
    Browse the repository at this point in the history
  28. [ Tensor ] remove sscal to set zero.

    We do need to remove the Nan or Inf value in Tensor by call setZero().
    However, if we using sscal, then Nan or Inf values are remain still.
    This PR change the sscal to memset.
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    b540642 View commit details
    Browse the repository at this point in the history
  29. [ Mixed ] set initialize gradient in layers and bugfixes

    This pr fixes some bugs when it runs as Mixed Precision Training
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    16e3a55 View commit details
    Browse the repository at this point in the history
  30. [ Mixed Training ] add is_mixed variable in weight

    Adding is_mixed variable to check if it is mixed precision training.
    It means that weight type of model is not full precision.
    
    **Changes proposed in this PR:**
    - Added TOC generator for README.md
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    aedb11c View commit details
    Browse the repository at this point in the history
  31. [ BUG FIX ] Fix bug for mixed precision

    For the mixed precision computation of bn layer, there is bug relate
    with f32 computation. Also Adam update has bug too.
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    cbefcd9 View commit details
    Browse the repository at this point in the history
  32. [ hgemm ] Use aligned memory allocation in transpose / padding gemm

    - Using unaligned memory may invoke SIGSEGV
    
    **Self evaluation:**
    1. Build test:     [X]Passed [ ]Failed [ ]Skipped
    2. Run test:     [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: skykongkong8 <[email protected]>
    skykongkong8 authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    d39f9d8 View commit details
    Browse the repository at this point in the history
  33. [TEST] using builddir/android_build_result to build test

    This PR includes changes in Android.mk to use
    builddir/android_build_result. In order to use, soft link of
    android_build_reuslt dir is necessary in upper dir (../)
    
    ln -s ../../buildir/android_build_result ../nntrainer
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    57e2759 View commit details
    Browse the repository at this point in the history
  34. [Mixed Precision] Fix mixed precsion to use Tensor V2

    This PR includes fixes to use TensorV2
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    jijoongmoon authored and DonghakPark committed Aug 29, 2024
    Configuration menu
    Copy the full SHA
    592253f View commit details
    Browse the repository at this point in the history

Commits on Oct 7, 2024

  1. temporary code for layer initialization

     - Temporary code for layer initialization
    
    Signed-off-by: hyeonseok lee <[email protected]>
    (cherry picked from commit fd0b6c3)
    lhs8928 authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    913a7fe View commit details
    Browse the repository at this point in the history
  2. [ SPEC ] chagne fp16

    Describe a commit content (Until 80 colums per line) in detail ASAP.
    
    **Changes proposed in this PR:**
    - Added TOC generator for README.md
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    (cherry picked from commit 846e2c0)
    jijoongmoon authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    1860cd2 View commit details
    Browse the repository at this point in the history
  3. [ NNStreamer ] disable nnstreamer trainer

    Describe a commit content (Until 80 colums per line) in detail ASAP.
    
    **Changes proposed in this PR:**
    - Added TOC generator for README.md
    
    Resolves:
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: jijoong.moon <[email protected]>
    (cherry picked from commit 80c9855)
    jijoongmoon authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    df4baa4 View commit details
    Browse the repository at this point in the history
  4. [Tizen7.0] Tizen7.0 Backporting

    - This commit adds some updates for Tizen7.0 backporting
    - Type mismatch bug is fixed.
    - Unused variable is removed.
    - Missing header files are added in spec file.
    - spec file is updated
    
    Self evaluation:
    
    Build test: [X]Passed [ ]Failed [ ]Skipped
    Run test: [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Eunju Yang <[email protected]>
    
    (cherry picked from commit 6d6e924)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    faddcdd View commit details
    Browse the repository at this point in the history
  5. [bugfix] fix coverity issues

    This PR resolves coverity issues in the ShortTensor class.
    Replace max_abs() implementation with maxValue() since the maximum absolute value of unsigned int equals to the maximum value.
    
    **Self-evaluation:**
    1. Build test: [X]Passed [ ]Failed [ ]Skipped
    2. Run test:   [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Donghyeon Jeong <[email protected]>
    (cherry picked from commit 656a231)
    djeong20 authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    ec06943 View commit details
    Browse the repository at this point in the history
  6. [ Tizen7.0 ] Include neuralnet.h in -dev header

    - Update the code to include `neuralnet.h` in -dev header.
    - Some applications, e.g., ReinforcementLearning uses `forwarding` and
    `backwarding` directly. To support it, this commit adds the header into
    dev package.
    
    **Self-evaluation:**
    1. Build test: [X]Passed [ ]Failed [ ]Skipped
    2. Run test:   [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Eunju Yang <[email protected]>
    (cherry picked from commit 77e56f1)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    145a726 View commit details
    Browse the repository at this point in the history
  7. [CI] Fix meson ubuntu ci build

    Fix build bug
    - Currently, there is a bug in the matrix used in CI where the first Meson build runs successfully but subsequent builds fail due to the presence of a 'build' folder. I would like to fix this issue.
    - Before running the Meson build, ensure that any existing folders named 'build' are deleted.
    - fix gcc version to 13
    
    Resolves:
    - nnstreamer#2715
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Donghak PARK <[email protected]>
    Co-authored-by: hyeonseok <[email protected]>
    Co-authored-by: Donghyeon Jeong <[email protected]>
    (cherry picked from commit b6fdba0)
    DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    e6830cc View commit details
    Browse the repository at this point in the history
  8. [Tizen7.0] Tizen7.0 Backporting

    - This commit adds some updates for Tizen7.0 backporting
    - Type mismatch bug is fixed.
    - Unused variable is removed.
    - Missing header files are added in spec file.
    - spec file is updated
    
    Self evaluation:
    
    Build test: [X]Passed [ ]Failed [ ]Skipped
    Run test: [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Eunju Yang <[email protected]>
    
    (cherry picked from commit e025352)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    19d7dc1 View commit details
    Browse the repository at this point in the history
  9. [ Tizen7.0 ] Include some headers in -dev header for neuralnet.h

    - In the previous PR (77e56f1), neuralnet.h was included in dev package.
    - However, some headers were missing used in nueralnet.h
    - This PR adds headers which have dependency with neuralnet.h
    - This PR is tested whether it supports ReinforcementLearning app on
    Tizen7.0
    
    Self evaluation:
    
    Build test: [X]Passed [ ]Failed [ ]Skipped
    Run test: [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Eunju Yang <[email protected]>
    
    (cherry picked from commit 9eb4c85)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    15cb248 View commit details
    Browse the repository at this point in the history
  10. [enhance] Registering OpenCL kernels at cl_context

    Register custom kernels as well as in-house kernels at cl_context initialization
    
    Signed-off-by: Debadri Samaddar <[email protected]>
    (cherry picked from commit b34fd53)
    s-debadri authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    2ab0371 View commit details
    Browse the repository at this point in the history
  11. [enhance/gpu] Removing layer_context dependency

    Removed layer_context dependency from blas OpenCL kernels.
    Temporarily commented out cl_layers to avoid build failure.
    
    Signed-off-by: Debadri Samaddar <[email protected]>
    (cherry picked from commit 7ee9294)
    s-debadri authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    1f6a2c0 View commit details
    Browse the repository at this point in the history
  12. [ FC ] update incremental_forwarding to support LoRA and multi-batch

    - This commit add some codes to support LoRA in incremental_forwarding.
    - This commit updates the incremental_forwarding to support multiple
    batch input. However, it is not the desirable way in that it cannot be
    parallelized across the batch axis. I left this issue on the comment.
    
    Self evaluation:
    
    Build test: [X]Passed [ ]Failed [ ]Skipped
    Run test: [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Eunju Yang <[email protected]>
    (cherry picked from commit 4ae1477)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    a182f84 View commit details
    Browse the repository at this point in the history
  13. [ LORA ] Bugfix in LoRA support in FC Layer

    - In the previous code, LoRA didn't work for the case batch_size > 1.
    - Tensors used in LoRA-related computation were not updated when the
    batch size is upsted.
    - `setBatch()` function is implemented for `FullyConnectedLayer`.
    - BugFix in Lifespan of loraTmp Tensor: FORWARD_DERIV_LIFESPANE ->
    FORWARD_GRAD_LIFESPAN
    
    Self evaluation:
    
    	Build test: [X]Passed [ ]Failed [ ]Skipped
    	Run test: [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Eunju Yang <[email protected]>
    (cherry picked from commit 8104cbe)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    710a160 View commit details
    Browse the repository at this point in the history
  14. [bugfix] Fix memcheck in CacheLoader unit tests

    This pull request fixes the issue of failing Continuous Integration.
    The new patch checks for nullptr after a flush operation on the cache pool.
    This adjustment is expected to rectify the previous failures encountered during the CI process.
    
    **Self-evaluation:**
    1. Build test: [X]Passed [ ]Failed [ ]Skipped
    2. Run test:   [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Donghyeon Jeong <[email protected]>
    (cherry picked from commit 119c60e)
    djeong20 authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    e0f5d3f View commit details
    Browse the repository at this point in the history
  15. [gpu/enhance] Utility for registering Blas kernels during initialization

    Default Blas kernel registration during cl_context initialization
    Remove RunLayerContext dependency from unit tests
    
    Signed-off-by: Debadri Samaddar <[email protected]>
    (cherry picked from commit 79a7c25)
    s-debadri authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    c43937c View commit details
    Browse the repository at this point in the history
  16. [ App ] Multi-Input Example Update

    - This commit is related to issue nnstreamer#2660
    - When using multi-inputs, users must feed the data in reverse order due
    to a known bug that needs fixing. In the current version, the input must
    be provided in reverse order, which was not shown in the previous
    example where random data with the same dimensions were used.
    - To provide a more accurate example to NNTrainer users, I have
    temporarily updated this example.
    - Once the issue is handled, further updates will be necessary.
    
    Signed-off-by: Eunju Yang <[email protected]>
    (cherry picked from commit 2807f69)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    f54120f View commit details
    Browse the repository at this point in the history
  17. [ Print ] Update print result of model summary

    - This commit updates the model summary print of the layer with multiple
    inputs.
    
    [ASIS]
                 concat0              concat            1:1:14:2              input0
                                                         1:1:4:2              input1
                                                         1:1:8:2              input2
    
    [TOBE]
                 concat0              concat            1:1:14:2              input0
                                                                              input1
                                                                              input2
    
    Signed-off-by: Eunju Yang <[email protected]>
    (cherry picked from commit f222ecf)
    EunjuYang authored and DonghakPark committed Oct 7, 2024
    Configuration menu
    Copy the full SHA
    0f043a4 View commit details
    Browse the repository at this point in the history

Commits on Oct 8, 2024

  1. [Mixed Precision] Fix gradient clipping logic

    update mixed precision - gradient clipping logic
    - when gradient clipping, gradient should unscale before calc l2norm
    
    **Self evaluation:**
    1. Build test:	 [X]Passed [ ]Failed [ ]Skipped
    2. Run test:	 [X]Passed [ ]Failed [ ]Skipped
    
    Signed-off-by: Donghak PARK <[email protected]>
    DonghakPark committed Oct 8, 2024
    Configuration menu
    Copy the full SHA
    2d4e347 View commit details
    Browse the repository at this point in the history