4 #ifndef FLATBUFFERS_GENERATED_MATMUL_TT_TARGET_TTNN_H_
5 #define FLATBUFFERS_GENERATED_MATMUL_TT_TARGET_TTNN_H_
7 #include "flatbuffers/flatbuffers.h"
11 static_assert(FLATBUFFERS_VERSION_MAJOR == 24 &&
12 FLATBUFFERS_VERSION_MINOR == 3 &&
13 FLATBUFFERS_VERSION_REVISION == 25,
14 "Non-compatible flatbuffers version included");
25 struct MatmulMultiCoreReuseProgramConfigBuilder;
28 struct MatmulMultiCoreReuseMultiCastProgramConfigBuilder;
31 struct MatmulMultiCoreReuseMultiCast1DProgramConfigBuilder;
34 struct MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfigBuilder;
37 struct MatmulOpBuilder;
40 struct LinearOpBuilder;
64 static const char *
const names[6] = {
66 "MatmulMultiCoreReuseProgramConfig",
67 "MatmulMultiCoreReuseMultiCastProgramConfig",
68 "MatmulMultiCoreReuseMultiCast1DProgramConfig",
69 "MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig",
77 const size_t index =
static_cast<size_t>(e);
102 bool VerifyMatmulProgramConfigVector(::flatbuffers::Verifier &verifier, const ::flatbuffers::Vector<::flatbuffers::Offset<void>> *values, const ::flatbuffers::Vector<MatmulProgramConfig> *types);
108 return "tt.target.ttnn.MatmulMultiCoreReuseProgramConfig";
110 enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
136 bool Verify(::flatbuffers::Verifier &verifier)
const {
137 return VerifyTableStart(verifier) &&
150 ::flatbuffers::FlatBufferBuilder &
fbb_;
174 ::flatbuffers::Offset<MatmulMultiCoreReuseProgramConfig>
Finish() {
176 auto o = ::flatbuffers::Offset<MatmulMultiCoreReuseProgramConfig>(end);
182 ::flatbuffers::FlatBufferBuilder &_fbb,
183 const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size =
nullptr,
184 uint64_t in0_block_w = 0,
185 uint64_t out_subblock_h = 0,
186 uint64_t out_subblock_w = 0,
187 uint64_t per_core_m = 0,
188 uint64_t per_core_n = 0) {
208 return "tt.target.ttnn.MatmulMultiCoreReuseMultiCastProgramConfig";
210 enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
256 bool Verify(::flatbuffers::Verifier &verifier)
const {
257 return VerifyTableStart(verifier) &&
268 verifier.VerifyTable(fused_activation()) &&
276 ::flatbuffers::FlatBufferBuilder &
fbb_;
315 ::flatbuffers::Offset<MatmulMultiCoreReuseMultiCastProgramConfig>
Finish() {
317 auto o = ::flatbuffers::Offset<MatmulMultiCoreReuseMultiCastProgramConfig>(end);
323 ::flatbuffers::FlatBufferBuilder &_fbb,
324 const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size =
nullptr,
325 uint64_t in0_block_w = 0,
326 uint64_t out_subblock_h = 0,
327 uint64_t out_subblock_w = 0,
328 uint64_t out_block_h = 0,
329 uint64_t out_block_w = 0,
330 uint64_t per_core_m = 0,
331 uint64_t per_core_n = 0,
332 bool transpose_mcast =
false,
333 ::flatbuffers::Offset<tt::target::ttnn::UnaryWithParam> fused_activation = 0,
334 bool fuse_batch =
false) {
359 return "tt.target.ttnn.MatmulMultiCoreReuseMultiCast1DProgramConfig";
361 enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
375 VT_NUM_GLOBAL_CB_RECEIVERS = 30
413 const tt::target::ttnn::CoreRangeSet *
hop_cores()
const {
414 return GetPointer<const tt::target::ttnn::CoreRangeSet *>(
VT_HOP_CORES);
417 return GetField<uint64_t>(VT_NUM_GLOBAL_CB_RECEIVERS, 0);
419 bool Verify(::flatbuffers::Verifier &verifier)
const {
420 return VerifyTableStart(verifier) &&
431 verifier.VerifyTable(fused_activation()) &&
435 verifier.VerifyTable(hop_cores()) &&
436 VerifyField<uint64_t>(verifier, VT_NUM_GLOBAL_CB_RECEIVERS, 8) &&
443 ::flatbuffers::FlatBufferBuilder &
fbb_;
481 void add_hop_cores(::flatbuffers::Offset<tt::target::ttnn::CoreRangeSet> hop_cores) {
485 fbb_.AddElement<uint64_t>(MatmulMultiCoreReuseMultiCast1DProgramConfig::VT_NUM_GLOBAL_CB_RECEIVERS, num_global_cb_receivers, 0);
491 ::flatbuffers::Offset<MatmulMultiCoreReuseMultiCast1DProgramConfig>
Finish() {
493 auto o = ::flatbuffers::Offset<MatmulMultiCoreReuseMultiCast1DProgramConfig>(end);
499 ::flatbuffers::FlatBufferBuilder &_fbb,
500 const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size =
nullptr,
501 uint64_t in0_block_w = 0,
502 uint64_t out_subblock_h = 0,
503 uint64_t out_subblock_w = 0,
504 uint64_t out_block_h = 0,
505 uint64_t out_block_w = 0,
506 uint64_t per_core_m = 0,
507 uint64_t per_core_n = 0,
508 bool fuse_batch =
false,
509 ::flatbuffers::Offset<tt::target::ttnn::UnaryWithParam> fused_activation = 0,
510 bool mcast_in0 =
false,
511 bool gather_in0 =
false,
512 ::flatbuffers::Offset<tt::target::ttnn::CoreRangeSet> hop_cores = 0,
513 uint64_t num_global_cb_receivers = 0) {
541 return "tt.target.ttnn.MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig";
543 enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
561 bool Verify(::flatbuffers::Verifier &verifier)
const {
562 return VerifyTableStart(verifier) &&
567 verifier.VerifyTable(fused_activation()) &&
574 ::flatbuffers::FlatBufferBuilder &
fbb_;
592 ::flatbuffers::Offset<MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig>
Finish() {
594 auto o = ::flatbuffers::Offset<MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig>(end);
600 ::flatbuffers::FlatBufferBuilder &_fbb,
601 uint64_t in0_block_w = 0,
602 uint64_t per_core_m = 0,
603 uint64_t per_core_n = 0,
604 ::flatbuffers::Offset<tt::target::ttnn::UnaryWithParam> fused_activation = 0) {
622 return "tt.target.ttnn.MatmulOp";
624 enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
631 VT_MATMUL_PROGRAM_CONFIG = 16
633 const tt::target::ttnn::TensorRef *
a()
const {
634 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_A);
636 const tt::target::ttnn::TensorRef *
b()
const {
637 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_B);
639 const tt::target::ttnn::TensorRef *
out()
const {
640 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_OUT);
652 return GetPointer<const void *>(VT_MATMUL_PROGRAM_CONFIG);
667 bool Verify(::flatbuffers::Verifier &verifier)
const {
668 return VerifyTableStart(verifier) &&
669 VerifyOffset(verifier,
VT_A) &&
670 verifier.VerifyTable(a()) &&
671 VerifyOffset(verifier,
VT_B) &&
672 verifier.VerifyTable(b()) &&
673 VerifyOffset(verifier,
VT_OUT) &&
674 verifier.VerifyTable(out()) &&
678 VerifyOffset(verifier, VT_MATMUL_PROGRAM_CONFIG) &&
684 template<>
inline const tt::target::ttnn::MatmulMultiCoreReuseProgramConfig *MatmulOp::matmul_program_config_as<tt::target::ttnn::MatmulMultiCoreReuseProgramConfig>()
const {
685 return matmul_program_config_as_MatmulMultiCoreReuseProgramConfig();
688 template<>
inline const tt::target::ttnn::MatmulMultiCoreReuseMultiCastProgramConfig *MatmulOp::matmul_program_config_as<tt::target::ttnn::MatmulMultiCoreReuseMultiCastProgramConfig>()
const {
689 return matmul_program_config_as_MatmulMultiCoreReuseMultiCastProgramConfig();
692 template<>
inline const tt::target::ttnn::MatmulMultiCoreReuseMultiCast1DProgramConfig *MatmulOp::matmul_program_config_as<tt::target::ttnn::MatmulMultiCoreReuseMultiCast1DProgramConfig>()
const {
693 return matmul_program_config_as_MatmulMultiCoreReuseMultiCast1DProgramConfig();
696 template<>
inline const tt::target::ttnn::MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig *MatmulOp::matmul_program_config_as<tt::target::ttnn::MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig>()
const {
697 return matmul_program_config_as_MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig();
702 ::flatbuffers::FlatBufferBuilder &
fbb_;
704 void add_a(::flatbuffers::Offset<tt::target::ttnn::TensorRef> a) {
707 void add_b(::flatbuffers::Offset<tt::target::ttnn::TensorRef> b) {
710 void add_out(::flatbuffers::Offset<tt::target::ttnn::TensorRef> out) {
723 fbb_.AddOffset(MatmulOp::VT_MATMUL_PROGRAM_CONFIG, matmul_program_config);
729 ::flatbuffers::Offset<MatmulOp>
Finish() {
731 auto o = ::flatbuffers::Offset<MatmulOp>(end);
737 ::flatbuffers::FlatBufferBuilder &_fbb,
738 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> a = 0,
739 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> b = 0,
740 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
741 bool transpose_a =
false,
742 bool transpose_b =
false,
744 ::flatbuffers::Offset<void> matmul_program_config = 0) {
765 return "tt.target.ttnn.LinearOp";
767 enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
775 const tt::target::ttnn::TensorRef *
a()
const {
776 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_A);
778 const tt::target::ttnn::TensorRef *
b()
const {
779 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_B);
781 const tt::target::ttnn::TensorRef *
bias()
const {
782 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_BIAS);
784 const tt::target::ttnn::TensorRef *
out()
const {
785 return GetPointer<const tt::target::ttnn::TensorRef *>(
VT_OUT);
793 bool Verify(::flatbuffers::Verifier &verifier)
const {
794 return VerifyTableStart(verifier) &&
795 VerifyOffset(verifier,
VT_A) &&
796 verifier.VerifyTable(a()) &&
797 VerifyOffset(verifier,
VT_B) &&
798 verifier.VerifyTable(b()) &&
799 VerifyOffset(verifier,
VT_BIAS) &&
800 verifier.VerifyTable(bias()) &&
801 VerifyOffset(verifier,
VT_OUT) &&
802 verifier.VerifyTable(out()) &&
811 ::flatbuffers::FlatBufferBuilder &
fbb_;
813 void add_a(::flatbuffers::Offset<tt::target::ttnn::TensorRef> a) {
816 void add_b(::flatbuffers::Offset<tt::target::ttnn::TensorRef> b) {
819 void add_bias(::flatbuffers::Offset<tt::target::ttnn::TensorRef> bias) {
822 void add_out(::flatbuffers::Offset<tt::target::ttnn::TensorRef> out) {
835 ::flatbuffers::Offset<LinearOp>
Finish() {
837 auto o = ::flatbuffers::Offset<LinearOp>(end);
843 ::flatbuffers::FlatBufferBuilder &_fbb,
844 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> a = 0,
845 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> b = 0,
846 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> bias = 0,
847 ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
848 bool transpose_a =
false,
849 bool transpose_b =
false) {
871 auto ptr =
reinterpret_cast<const tt::target::ttnn::MatmulMultiCoreReuseProgramConfig *
>(obj);
872 return verifier.VerifyTable(ptr);
875 auto ptr =
reinterpret_cast<const tt::target::ttnn::MatmulMultiCoreReuseMultiCastProgramConfig *
>(obj);
876 return verifier.VerifyTable(ptr);
879 auto ptr =
reinterpret_cast<const tt::target::ttnn::MatmulMultiCoreReuseMultiCast1DProgramConfig *
>(obj);
880 return verifier.VerifyTable(ptr);
883 auto ptr =
reinterpret_cast<const tt::target::ttnn::MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig *
>(obj);
884 return verifier.VerifyTable(ptr);
886 default:
return true;
890 inline bool VerifyMatmulProgramConfigVector(::flatbuffers::Verifier &verifier, const ::flatbuffers::Vector<::flatbuffers::Offset<void>> *values, const ::flatbuffers::Vector<MatmulProgramConfig> *types) {
891 if (!values || !types)
return !values && !types;
892 if (values->size() != types->size())
return false;
893 for (::flatbuffers::uoffset_t i = 0; i < values->size(); ++i) {
VT_TRANSPOSE_A
Definition: matmul_generated.h:628
VT_GATHER_IN0
Definition: matmul_generated.h:373
VT_TRANSPOSE_B
Definition: matmul_generated.h:629
VT_COMPUTE_WITH_STORAGE_GRID_SIZE
Definition: matmul_generated.h:111
VT_MCAST_IN0
Definition: matmul_generated.h:372
VT_PER_CORE_N
Definition: matmul_generated.h:218
VT_FUSED_ACTIVATION
Definition: matmul_generated.h:220
VT_PER_CORE_M
Definition: matmul_generated.h:115
VT_OUT_SUBBLOCK_H
Definition: matmul_generated.h:113
VT_TRANSPOSE_MCAST
Definition: matmul_generated.h:219
VT_OUT_BLOCK_H
Definition: matmul_generated.h:215
VT_OUT_SUBBLOCK_W
Definition: matmul_generated.h:114
VT_A
Definition: matmul_generated.h:625
VT_B
Definition: matmul_generated.h:626
VT_BIAS
Definition: matmul_generated.h:770
VT_OUT
Definition: matmul_generated.h:627
VT_IN0_BLOCK_W
Definition: matmul_generated.h:112
VT_HOP_CORES
Definition: matmul_generated.h:374
VT_OUT_BLOCK_W
Definition: matmul_generated.h:216
VT_FUSE_BATCH
Definition: matmul_generated.h:370
VT_MATMUL_PROGRAM_CONFIG_TYPE
Definition: matmul_generated.h:630
inline ::flatbuffers::Offset< MatmulMultiCoreReuseMultiCast1DProgramConfig > CreateMatmulMultiCoreReuseMultiCast1DProgramConfig(::flatbuffers::FlatBufferBuilder &_fbb, const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size=nullptr, uint64_t in0_block_w=0, uint64_t out_subblock_h=0, uint64_t out_subblock_w=0, uint64_t out_block_h=0, uint64_t out_block_w=0, uint64_t per_core_m=0, uint64_t per_core_n=0, bool fuse_batch=false, ::flatbuffers::Offset< tt::target::ttnn::UnaryWithParam > fused_activation=0, bool mcast_in0=false, bool gather_in0=false, ::flatbuffers::Offset< tt::target::ttnn::CoreRangeSet > hop_cores=0, uint64_t num_global_cb_receivers=0)
Definition: matmul_generated.h:498
inline ::flatbuffers::Offset< MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig > CreateMatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig(::flatbuffers::FlatBufferBuilder &_fbb, uint64_t in0_block_w=0, uint64_t per_core_m=0, uint64_t per_core_n=0, ::flatbuffers::Offset< tt::target::ttnn::UnaryWithParam > fused_activation=0)
Definition: matmul_generated.h:599
inline ::flatbuffers::Offset< MatmulMultiCoreReuseProgramConfig > CreateMatmulMultiCoreReuseProgramConfig(::flatbuffers::FlatBufferBuilder &_fbb, const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size=nullptr, uint64_t in0_block_w=0, uint64_t out_subblock_h=0, uint64_t out_subblock_w=0, uint64_t per_core_m=0, uint64_t per_core_n=0)
Definition: matmul_generated.h:181
const char *const * EnumNamesMatmulProgramConfig()
Definition: matmul_generated.h:63
MatmulProgramConfig
Definition: matmul_generated.h:42
@ MatmulMultiCoreReuseProgramConfig
@ MatmulMultiCoreReuseMultiCastProgramConfig
@ MatmulMultiCoreReuseMultiCast1DProgramConfig
@ MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig
inline ::flatbuffers::Offset< MatmulMultiCoreReuseMultiCastProgramConfig > CreateMatmulMultiCoreReuseMultiCastProgramConfig(::flatbuffers::FlatBufferBuilder &_fbb, const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size=nullptr, uint64_t in0_block_w=0, uint64_t out_subblock_h=0, uint64_t out_subblock_w=0, uint64_t out_block_h=0, uint64_t out_block_w=0, uint64_t per_core_m=0, uint64_t per_core_n=0, bool transpose_mcast=false, ::flatbuffers::Offset< tt::target::ttnn::UnaryWithParam > fused_activation=0, bool fuse_batch=false)
Definition: matmul_generated.h:322
bool VerifyMatmulProgramConfigVector(::flatbuffers::Verifier &verifier, const ::flatbuffers::Vector<::flatbuffers::Offset< void >> *values, const ::flatbuffers::Vector< MatmulProgramConfig > *types)
Definition: matmul_generated.h:890
const MatmulProgramConfig(& EnumValuesMatmulProgramConfig())[5]
Definition: matmul_generated.h:52
const char * EnumNameMatmulProgramConfig(MatmulProgramConfig e)
Definition: matmul_generated.h:75
inline ::flatbuffers::Offset< MatmulOp > CreateMatmulOp(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > a=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > b=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, bool transpose_a=false, bool transpose_b=false, tt::target::ttnn::MatmulProgramConfig matmul_program_config_type=tt::target::ttnn::MatmulProgramConfig::NONE, ::flatbuffers::Offset< void > matmul_program_config=0)
Definition: matmul_generated.h:736
bool VerifyMatmulProgramConfig(::flatbuffers::Verifier &verifier, const void *obj, MatmulProgramConfig type)
Definition: matmul_generated.h:865
inline ::flatbuffers::Offset< LinearOp > CreateLinearOp(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > a=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > b=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > bias=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, bool transpose_a=false, bool transpose_b=false)
Definition: matmul_generated.h:842
Definition: debug_info_generated.h:18
Definition: binary_generated.h:31
const tt::target::ttnn::MatmulMultiCoreReuseMultiCastProgramConfig * matmul_program_config_as_MatmulMultiCoreReuseMultiCastProgramConfig() const
Definition: matmul_generated.h:658
const tt::target::ttnn::MatmulMultiCoreReuseMultiCast1DProgramConfig * matmul_program_config_as_MatmulMultiCoreReuseMultiCast1DProgramConfig() const
Definition: matmul_generated.h:661
const tt::target::ttnn::TensorRef * bias() const
Definition: matmul_generated.h:781
uint64_t per_core_n() const
Definition: matmul_generated.h:133
const tt::target::ttnn::UnaryWithParam * fused_activation() const
Definition: matmul_generated.h:250
tt::target::ttnn::MatmulProgramConfig matmul_program_config_type() const
Definition: matmul_generated.h:648
bool transpose_mcast() const
Definition: matmul_generated.h:247
bool gather_in0() const
Definition: matmul_generated.h:410
const tt::target::ttnn::TensorRef * out() const
Definition: matmul_generated.h:639
uint64_t out_block_h() const
Definition: matmul_generated.h:235
uint64_t out_block_w() const
Definition: matmul_generated.h:238
const tt::target::ttnn::MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig * matmul_program_config_as_MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig() const
Definition: matmul_generated.h:664
const tt::target::ttnn::MatmulMultiCoreReuseProgramConfig * matmul_program_config_as_MatmulMultiCoreReuseProgramConfig() const
Definition: matmul_generated.h:655
bool fuse_batch() const
Definition: matmul_generated.h:253
MatmulMultiCoreReuseMultiCast1DProgramConfigBuilder Builder
Definition: matmul_generated.h:356
const void * matmul_program_config() const
Definition: matmul_generated.h:651
uint64_t out_subblock_w() const
Definition: matmul_generated.h:127
MatmulMultiCoreReuseProgramConfigBuilder Builder
Definition: matmul_generated.h:105
const tt::target::ttnn::CoreRangeSet * hop_cores() const
Definition: matmul_generated.h:413
bool transpose_a() const
Definition: matmul_generated.h:642
bool mcast_in0() const
Definition: matmul_generated.h:407
const T * matmul_program_config_as() const
const tt::target::ttnn::TensorRef * b() const
Definition: matmul_generated.h:636
const tt::target::ttnn::TensorRef * a() const
Definition: matmul_generated.h:633
uint64_t per_core_m() const
Definition: matmul_generated.h:130
MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfigBuilder Builder
Definition: matmul_generated.h:538
MatmulMultiCoreReuseMultiCastProgramConfigBuilder Builder
Definition: matmul_generated.h:205
bool Verify(::flatbuffers::Verifier &verifier) const
Definition: matmul_generated.h:136
const tt::target::ttnn::CoreCoord * compute_with_storage_grid_size() const
Definition: matmul_generated.h:118
bool transpose_b() const
Definition: matmul_generated.h:645
uint64_t num_global_cb_receivers() const
Definition: matmul_generated.h:416
static FLATBUFFERS_CONSTEXPR_CPP11 const char * GetFullyQualifiedName()
Definition: matmul_generated.h:107
uint64_t out_subblock_h() const
Definition: matmul_generated.h:124
LinearOpBuilder Builder
Definition: matmul_generated.h:762
uint64_t in0_block_w() const
Definition: matmul_generated.h:121
MatmulOpBuilder Builder
Definition: matmul_generated.h:619
Definition: matmul_generated.h:809
LinearOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: matmul_generated.h:831
void add_out(::flatbuffers::Offset< tt::target::ttnn::TensorRef > out)
Definition: matmul_generated.h:822
void add_a(::flatbuffers::Offset< tt::target::ttnn::TensorRef > a)
Definition: matmul_generated.h:813
void add_b(::flatbuffers::Offset< tt::target::ttnn::TensorRef > b)
Definition: matmul_generated.h:816
::flatbuffers::Offset< LinearOp > Finish()
Definition: matmul_generated.h:835
::flatbuffers::uoffset_t start_
Definition: matmul_generated.h:812
void add_transpose_a(bool transpose_a)
Definition: matmul_generated.h:825
void add_transpose_b(bool transpose_b)
Definition: matmul_generated.h:828
void add_bias(::flatbuffers::Offset< tt::target::ttnn::TensorRef > bias)
Definition: matmul_generated.h:819
::flatbuffers::FlatBufferBuilder & fbb_
Definition: matmul_generated.h:811
LinearOp Table
Definition: matmul_generated.h:810
Definition: matmul_generated.h:860
static constexpr auto Create
Definition: matmul_generated.h:862
LinearOp type
Definition: matmul_generated.h:861
Definition: matmul_generated.h:441
MatmulMultiCoreReuseMultiCast1DProgramConfigBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: matmul_generated.h:487
::flatbuffers::uoffset_t start_
Definition: matmul_generated.h:444
void add_per_core_n(uint64_t per_core_n)
Definition: matmul_generated.h:466
void add_gather_in0(bool gather_in0)
Definition: matmul_generated.h:478
void add_hop_cores(::flatbuffers::Offset< tt::target::ttnn::CoreRangeSet > hop_cores)
Definition: matmul_generated.h:481
::flatbuffers::Offset< MatmulMultiCoreReuseMultiCast1DProgramConfig > Finish()
Definition: matmul_generated.h:491
::flatbuffers::FlatBufferBuilder & fbb_
Definition: matmul_generated.h:443
void add_in0_block_w(uint64_t in0_block_w)
Definition: matmul_generated.h:448
void add_num_global_cb_receivers(uint64_t num_global_cb_receivers)
Definition: matmul_generated.h:484
void add_fused_activation(::flatbuffers::Offset< tt::target::ttnn::UnaryWithParam > fused_activation)
Definition: matmul_generated.h:472
void add_fuse_batch(bool fuse_batch)
Definition: matmul_generated.h:469
MatmulMultiCoreReuseMultiCast1DProgramConfig Table
Definition: matmul_generated.h:442
void add_out_block_h(uint64_t out_block_h)
Definition: matmul_generated.h:457
void add_mcast_in0(bool mcast_in0)
Definition: matmul_generated.h:475
void add_out_block_w(uint64_t out_block_w)
Definition: matmul_generated.h:460
void add_out_subblock_h(uint64_t out_subblock_h)
Definition: matmul_generated.h:451
void add_compute_with_storage_grid_size(const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size)
Definition: matmul_generated.h:445
void add_per_core_m(uint64_t per_core_m)
Definition: matmul_generated.h:463
void add_out_subblock_w(uint64_t out_subblock_w)
Definition: matmul_generated.h:454
Definition: matmul_generated.h:532
static constexpr auto Create
Definition: matmul_generated.h:534
MatmulMultiCoreReuseMultiCast1DProgramConfig type
Definition: matmul_generated.h:533
Definition: matmul_generated.h:572
void add_per_core_n(uint64_t per_core_n)
Definition: matmul_generated.h:582
MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig Table
Definition: matmul_generated.h:573
MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfigBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: matmul_generated.h:588
::flatbuffers::FlatBufferBuilder & fbb_
Definition: matmul_generated.h:574
void add_fused_activation(::flatbuffers::Offset< tt::target::ttnn::UnaryWithParam > fused_activation)
Definition: matmul_generated.h:585
void add_per_core_m(uint64_t per_core_m)
Definition: matmul_generated.h:579
::flatbuffers::uoffset_t start_
Definition: matmul_generated.h:575
::flatbuffers::Offset< MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig > Finish()
Definition: matmul_generated.h:592
void add_in0_block_w(uint64_t in0_block_w)
Definition: matmul_generated.h:576
Definition: matmul_generated.h:613
MatmulMultiCoreReuseMultiCastDRAMShardedProgramConfig type
Definition: matmul_generated.h:614
static constexpr auto Create
Definition: matmul_generated.h:615
Definition: matmul_generated.h:274
::flatbuffers::FlatBufferBuilder & fbb_
Definition: matmul_generated.h:276
void add_out_subblock_w(uint64_t out_subblock_w)
Definition: matmul_generated.h:287
void add_in0_block_w(uint64_t in0_block_w)
Definition: matmul_generated.h:281
void add_transpose_mcast(bool transpose_mcast)
Definition: matmul_generated.h:302
void add_out_block_h(uint64_t out_block_h)
Definition: matmul_generated.h:290
void add_per_core_n(uint64_t per_core_n)
Definition: matmul_generated.h:299
void add_out_block_w(uint64_t out_block_w)
Definition: matmul_generated.h:293
void add_compute_with_storage_grid_size(const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size)
Definition: matmul_generated.h:278
MatmulMultiCoreReuseMultiCastProgramConfig Table
Definition: matmul_generated.h:275
void add_out_subblock_h(uint64_t out_subblock_h)
Definition: matmul_generated.h:284
void add_fused_activation(::flatbuffers::Offset< tt::target::ttnn::UnaryWithParam > fused_activation)
Definition: matmul_generated.h:305
void add_fuse_batch(bool fuse_batch)
Definition: matmul_generated.h:308
void add_per_core_m(uint64_t per_core_m)
Definition: matmul_generated.h:296
::flatbuffers::Offset< MatmulMultiCoreReuseMultiCastProgramConfig > Finish()
Definition: matmul_generated.h:315
MatmulMultiCoreReuseMultiCastProgramConfigBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: matmul_generated.h:311
::flatbuffers::uoffset_t start_
Definition: matmul_generated.h:277
Definition: matmul_generated.h:350
MatmulMultiCoreReuseMultiCastProgramConfig type
Definition: matmul_generated.h:351
static constexpr auto Create
Definition: matmul_generated.h:352
Definition: matmul_generated.h:148
void add_compute_with_storage_grid_size(const tt::target::ttnn::CoreCoord *compute_with_storage_grid_size)
Definition: matmul_generated.h:152
void add_out_subblock_h(uint64_t out_subblock_h)
Definition: matmul_generated.h:158
MatmulMultiCoreReuseProgramConfigBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: matmul_generated.h:170
::flatbuffers::FlatBufferBuilder & fbb_
Definition: matmul_generated.h:150
::flatbuffers::Offset< MatmulMultiCoreReuseProgramConfig > Finish()
Definition: matmul_generated.h:174
void add_per_core_m(uint64_t per_core_m)
Definition: matmul_generated.h:164
MatmulMultiCoreReuseProgramConfig Table
Definition: matmul_generated.h:149
::flatbuffers::uoffset_t start_
Definition: matmul_generated.h:151
void add_out_subblock_w(uint64_t out_subblock_w)
Definition: matmul_generated.h:161
void add_in0_block_w(uint64_t in0_block_w)
Definition: matmul_generated.h:155
void add_per_core_n(uint64_t per_core_n)
Definition: matmul_generated.h:167
Definition: matmul_generated.h:199
static constexpr auto Create
Definition: matmul_generated.h:201
MatmulMultiCoreReuseProgramConfig type
Definition: matmul_generated.h:200
Definition: matmul_generated.h:700
::flatbuffers::uoffset_t start_
Definition: matmul_generated.h:703
void add_matmul_program_config_type(tt::target::ttnn::MatmulProgramConfig matmul_program_config_type)
Definition: matmul_generated.h:719
::flatbuffers::Offset< MatmulOp > Finish()
Definition: matmul_generated.h:729
void add_b(::flatbuffers::Offset< tt::target::ttnn::TensorRef > b)
Definition: matmul_generated.h:707
MatmulOp Table
Definition: matmul_generated.h:701
void add_transpose_a(bool transpose_a)
Definition: matmul_generated.h:713
void add_transpose_b(bool transpose_b)
Definition: matmul_generated.h:716
void add_out(::flatbuffers::Offset< tt::target::ttnn::TensorRef > out)
Definition: matmul_generated.h:710
void add_a(::flatbuffers::Offset< tt::target::ttnn::TensorRef > a)
Definition: matmul_generated.h:704
::flatbuffers::FlatBufferBuilder & fbb_
Definition: matmul_generated.h:702
void add_matmul_program_config(::flatbuffers::Offset< void > matmul_program_config)
Definition: matmul_generated.h:722
MatmulOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: matmul_generated.h:725
Definition: matmul_generated.h:756
static constexpr auto Create
Definition: matmul_generated.h:758
MatmulOp type
Definition: matmul_generated.h:757
Definition: matmul_generated.h:81
static const MatmulProgramConfig enum_value
Definition: matmul_generated.h:82