Skip to content
Open
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions src/operators/batch-matrix-multiply-nc.c
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
#include "src/xnnpack/compute.h"
#include "src/xnnpack/config-types.h"
#include "src/xnnpack/config.h"
#include "src/xnnpack/internal.h"
#include "src/xnnpack/log.h"
#include "src/xnnpack/math.h"
#include "src/xnnpack/microfnptr.h"
Expand Down
6 changes: 3 additions & 3 deletions src/xnnpack/internal.h
Original file line number Diff line number Diff line change
Expand Up @@ -103,7 +103,7 @@ enum xnn_status xnn_setup_batch_matrix_multiply_nc_qp8_f32_qc8w(
xnn_operator_t batch_matrix_multiply_op, //
void* workspace, //
const int8_t* input_a, //
const int8_t* input_b, //
const float* input_b, //
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this doesnt look right. the input_b is qc8w which is int8_t

Copy link
Author

@cyyever cyyever Aug 26, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

But the function definition in src/operators/batch-matrix-multiply-nc.c uses const float* input_b.
PS: it looks like I should change the definition.

float* output);

enum xnn_status xnn_create_convert_nc_f32_qp8(
Expand Down Expand Up @@ -258,7 +258,7 @@ enum xnn_status xnn_create_batch_matrix_multiply_nc_pf16(
uint32_t flags, xnn_operator_t* batch_matrix_multiply_op_out);

enum xnn_status xnn_create_batch_matrix_multiply_nc_pf16_const_weights(
size_t batch_size_b, size_t k, size_t n, const void* data_b, uint32_t flags,
size_t batch_size_b, size_t k, size_t n, const xnn_float16* data_b, uint32_t flags,
xnn_operator_t* batch_matrix_multiply_op_out);

enum xnn_status xnn_reshape_batch_matrix_multiply_nc_pf16(
Expand Down Expand Up @@ -476,7 +476,7 @@ enum xnn_status xnn_reshape_batch_matrix_multiply_nc_qdu8_f32_qc8w(

enum xnn_status xnn_setup_batch_matrix_multiply_nc_qdu8_f32_qc8w(
xnn_operator_t batch_matrix_multiply_op, void* workspace,
const int8_t* input_a, const int8_t* input_b,
const int8_t* input_a, const float* input_b,
const struct xnn_quantization_params* quantization_params, float* output);

enum xnn_status xnn_create_fully_connected_nc_pf16(
Expand Down