@@ -577,10 +577,10 @@ extern "C" {
577577
578578    // this tensor... 
579579    enum  ggml_tensor_flag  {
580-         GGML_TENSOR_FLAG_INPUT      =  1 , // ...is an input for the GGML compute graph 
581-         GGML_TENSOR_FLAG_OUTPUT     =  2 , // ...is an output for the GGML compute graph 
582-         GGML_TENSOR_FLAG_PARAM      =  4 , // ...contains trainable parameters 
583-         GGML_TENSOR_FLAG_LOSS       =  8 , // ...defines loss for numerical optimization (multiple loss tensors add up) 
580+         GGML_TENSOR_FLAG_INPUT   =    1 , // ...is an input for the GGML compute graph 
581+         GGML_TENSOR_FLAG_OUTPUT  =    2 , // ...is an output for the GGML compute graph 
582+         GGML_TENSOR_FLAG_PARAM   =    4 , // ...contains trainable parameters 
583+         GGML_TENSOR_FLAG_LOSS    =    8 , // ...defines loss for numerical optimization (multiple loss tensors add up) 
584584    };
585585
586586    // n-dimensional tensor 
@@ -1410,14 +1410,14 @@ extern "C" {
14101410    // supports 3D: a->ne[2] == b->ne[1] 
14111411    GGML_API  struct  ggml_tensor  *  ggml_get_rows (
14121412            struct  ggml_context  *  ctx ,
1413-             struct  ggml_tensor   *  a ,
1414-             struct  ggml_tensor   *  b );
1413+             struct  ggml_tensor   *  a ,   // data 
1414+             struct  ggml_tensor   *  b );  // row indices 
14151415
14161416    GGML_API  struct  ggml_tensor  *  ggml_get_rows_back (
14171417            struct  ggml_context  *  ctx ,
1418-             struct  ggml_tensor   *  a ,
1419-             struct  ggml_tensor   *  b ,
1420-             struct  ggml_tensor   *  c );
1418+             struct  ggml_tensor   *  a ,   // gradients of ggml_get_rows result 
1419+             struct  ggml_tensor   *  b ,   // row indices 
1420+             struct  ggml_tensor   *  c );  // data for ggml_get_rows, only used for its shape 
14211421
14221422    GGML_API  struct  ggml_tensor  *  ggml_diag (
14231423        struct  ggml_context      *  ctx ,
@@ -1568,9 +1568,9 @@ extern "C" {
15681568    // a - dy 
15691569    GGML_API  struct  ggml_tensor  *  ggml_rope_back (
15701570            struct  ggml_context  *  ctx ,
1571-             struct  ggml_tensor   *  a ,
1572-             struct  ggml_tensor   *  b ,
1573-             struct  ggml_tensor   *  c ,
1571+             struct  ggml_tensor   *  a ,  // gradients of ggml_rope result 
1572+             struct  ggml_tensor   *  b ,  // positions 
1573+             struct  ggml_tensor   *  c ,  // freq factors 
15741574            int                    n_dims ,
15751575            int                    mode ,
15761576            int                    n_ctx_orig ,
@@ -2036,15 +2036,15 @@ extern "C" {
20362036    // loss function 
20372037
20382038    GGML_API  struct  ggml_tensor  *  ggml_cross_entropy_loss (
2039-             struct  ggml_context           *  ctx ,
2040-             struct  ggml_tensor            *  a ,
2041-             struct  ggml_tensor            *  b );
2039+             struct  ggml_context  *  ctx ,
2040+             struct  ggml_tensor   *  a ,   // logits 
2041+             struct  ggml_tensor   *  b );  // labels 
20422042
20432043    GGML_API  struct  ggml_tensor  *  ggml_cross_entropy_loss_back (
2044-             struct  ggml_context           *  ctx ,
2045-             struct  ggml_tensor            *  a ,
2046-             struct  ggml_tensor            *  b ,
2047-             struct  ggml_tensor            *   c ); 
2044+             struct  ggml_context  *  ctx ,
2045+             struct  ggml_tensor   *  a ,   // logits 
2046+             struct  ggml_tensor   *  b ,   // labels 
2047+             struct  ggml_tensor   *   c );  // gradients of cross_entropy_loss result 
20482048
20492049    // AdamW optimizer step 
20502050    // Paper: https://arxiv.org/pdf/1711.05101v3.pdf 
@@ -2066,7 +2066,7 @@ extern "C" {
20662066    GGML_API  void  ggml_set_loss (struct  ggml_tensor  *  tensor );
20672067
20682068    GGML_API  void  ggml_build_forward_expand  (struct  ggml_cgraph  *  cgraph , struct  ggml_tensor  *  tensor );
2069-     GGML_API  void  ggml_build_backward_expand (struct  ggml_context  *  ctx , struct  ggml_cgraph  *  gf , struct  ggml_cgraph  *  gb , bool  accumulate ,  bool   keep );
2069+     GGML_API  void  ggml_build_backward_expand (struct  ggml_context  *  ctx , struct  ggml_cgraph  *  gf , struct  ggml_cgraph  *  gb , bool  accumulate );
20702070
20712071    GGML_API  void  ggml_build_opt_adamw (
20722072            struct  ggml_context  *  ctx ,
0 commit comments