Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion src/operators/batch-matrix-multiply-nc.c
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
#include "src/xnnpack/compute.h"
#include "src/xnnpack/config-types.h"
#include "src/xnnpack/config.h"
#include "src/xnnpack/internal.h"
#include "src/xnnpack/log.h"
#include "src/xnnpack/math.h"
#include "src/xnnpack/microfnptr.h"
Expand Down Expand Up @@ -1479,7 +1480,7 @@ enum xnn_status xnn_setup_batch_matrix_multiply_nc_qd8_f32_qc8w(

enum xnn_status xnn_setup_batch_matrix_multiply_nc_qp8_f32_qc8w(
xnn_operator_t batch_matrix_multiply_op, void* workspace,
const int8_t* input_a, const float* input_b, float* output) {
const int8_t* input_a, const int8_t* input_b, float* output) {
return setup_batch_matrix_multiply_nc(
batch_matrix_multiply_op,
xnn_operator_type_batch_matrix_multiply_nc_qp8_f32_qc8w, input_a,
Expand Down
6 changes: 3 additions & 3 deletions src/operators/fully-connected-nc.c
Original file line number Diff line number Diff line change
Expand Up @@ -3136,23 +3136,23 @@ enum xnn_status xnn_setup_fully_connected_nc_qdu8_f16_qc8w(
}

enum xnn_status xnn_setup_fully_connected_nc_qp8_f32_qc4w(
xnn_operator_t fully_connected_op, const float* input, float* output,
xnn_operator_t fully_connected_op, const int8_t* input, float* output,
void* workspace) {
return setup_fully_connected_nc(
fully_connected_op, xnn_operator_type_fully_connected_nc_qp8_f32_qc4w,
input, output, workspace, /*quantization_params=*/NULL);
}

enum xnn_status xnn_setup_fully_connected_nc_qp8_f32_qc8w(
xnn_operator_t fully_connected_op, const float* input, float* output,
xnn_operator_t fully_connected_op, const int8_t* input, float* output,
void* workspace) {
return setup_fully_connected_nc(
fully_connected_op, xnn_operator_type_fully_connected_nc_qp8_f32_qc8w,
input, output, workspace, /*quantization_params=*/NULL);
}

enum xnn_status xnn_setup_fully_connected_nc_qp8_f32_qb4w(
xnn_operator_t fully_connected_op, const float* input, float* output,
xnn_operator_t fully_connected_op, const int8_t* input, float* output,
void* workspace) {
return setup_fully_connected_nc(
fully_connected_op, xnn_operator_type_fully_connected_nc_qp8_f32_qb4w,
Expand Down
4 changes: 2 additions & 2 deletions src/xnnpack/internal.h
Original file line number Diff line number Diff line change
Expand Up @@ -258,7 +258,7 @@ enum xnn_status xnn_create_batch_matrix_multiply_nc_pf16(
uint32_t flags, xnn_operator_t* batch_matrix_multiply_op_out);

enum xnn_status xnn_create_batch_matrix_multiply_nc_pf16_const_weights(
size_t batch_size_b, size_t k, size_t n, const void* data_b, uint32_t flags,
size_t batch_size_b, size_t k, size_t n, const xnn_float16* data_b, uint32_t flags,
xnn_operator_t* batch_matrix_multiply_op_out);

enum xnn_status xnn_reshape_batch_matrix_multiply_nc_pf16(
Expand Down Expand Up @@ -476,7 +476,7 @@ enum xnn_status xnn_reshape_batch_matrix_multiply_nc_qdu8_f32_qc8w(

enum xnn_status xnn_setup_batch_matrix_multiply_nc_qdu8_f32_qc8w(
xnn_operator_t batch_matrix_multiply_op, void* workspace,
const int8_t* input_a, const int8_t* input_b,
const int8_t* input_a, const float* input_b,
const struct xnn_quantization_params* quantization_params, float* output);

enum xnn_status xnn_create_fully_connected_nc_pf16(
Expand Down