TT-MLIR
ccl_generated.h
Go to the documentation of this file.
1 // automatically generated by the FlatBuffers compiler, do not modify
2 
3 
4 #ifndef FLATBUFFERS_GENERATED_CCL_TT_TARGET_TTNN_H_
5 #define FLATBUFFERS_GENERATED_CCL_TT_TARGET_TTNN_H_
6 
7 #include "flatbuffers/flatbuffers.h"
8 
9 // Ensure the included flatbuffers.h is the same version as when this file was
10 // generated, otherwise it may not be compatible.
11 static_assert(FLATBUFFERS_VERSION_MAJOR == 24 &&
12  FLATBUFFERS_VERSION_MINOR == 3 &&
13  FLATBUFFERS_VERSION_REVISION == 25,
14  "Non-compatible flatbuffers version included");
15 
18 
19 namespace tt {
20 namespace target {
21 namespace ttnn {
22 
23 struct AllGatherOp;
24 struct AllGatherOpBuilder;
25 
26 struct CollectivePermuteOp;
27 struct CollectivePermuteOpBuilder;
28 
29 struct MeshShardOp;
30 struct MeshShardOpBuilder;
31 
32 struct ReduceScatterOp;
33 struct ReduceScatterOpBuilder;
34 
35 struct AllGatherOp FLATBUFFERS_FINAL_CLASS : private ::flatbuffers::Table {
37  struct Traits;
38  enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
39  VT_IN = 4,
40  VT_OUT = 6,
41  VT_DEVICE = 8,
44  VT_NUM_LINKS = 14
45  };
46  const tt::target::ttnn::TensorRef *in() const {
47  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_IN);
48  }
49  const tt::target::ttnn::TensorRef *out() const {
50  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_OUT);
51  }
52  const tt::target::DeviceRef *device() const {
53  return GetPointer<const tt::target::DeviceRef *>(VT_DEVICE);
54  }
55  int32_t all_gather_dim() const {
56  return GetField<int32_t>(VT_ALL_GATHER_DIM, 0);
57  }
58  uint32_t cluster_axis() const {
59  return GetField<uint32_t>(VT_CLUSTER_AXIS, 0);
60  }
61  uint32_t num_links() const {
62  return GetField<uint32_t>(VT_NUM_LINKS, 0);
63  }
64  bool Verify(::flatbuffers::Verifier &verifier) const {
65  return VerifyTableStart(verifier) &&
66  VerifyOffset(verifier, VT_IN) &&
67  verifier.VerifyTable(in()) &&
68  VerifyOffset(verifier, VT_OUT) &&
69  verifier.VerifyTable(out()) &&
70  VerifyOffset(verifier, VT_DEVICE) &&
71  verifier.VerifyTable(device()) &&
72  VerifyField<int32_t>(verifier, VT_ALL_GATHER_DIM, 4) &&
73  VerifyField<uint32_t>(verifier, VT_CLUSTER_AXIS, 4) &&
74  VerifyField<uint32_t>(verifier, VT_NUM_LINKS, 4) &&
75  verifier.EndTable();
76  }
77 };
78 
80  typedef AllGatherOp Table;
81  ::flatbuffers::FlatBufferBuilder &fbb_;
82  ::flatbuffers::uoffset_t start_;
83  void add_in(::flatbuffers::Offset<tt::target::ttnn::TensorRef> in) {
84  fbb_.AddOffset(AllGatherOp::VT_IN, in);
85  }
86  void add_out(::flatbuffers::Offset<tt::target::ttnn::TensorRef> out) {
87  fbb_.AddOffset(AllGatherOp::VT_OUT, out);
88  }
89  void add_device(::flatbuffers::Offset<tt::target::DeviceRef> device) {
90  fbb_.AddOffset(AllGatherOp::VT_DEVICE, device);
91  }
92  void add_all_gather_dim(int32_t all_gather_dim) {
93  fbb_.AddElement<int32_t>(AllGatherOp::VT_ALL_GATHER_DIM, all_gather_dim, 0);
94  }
95  void add_cluster_axis(uint32_t cluster_axis) {
96  fbb_.AddElement<uint32_t>(AllGatherOp::VT_CLUSTER_AXIS, cluster_axis, 0);
97  }
98  void add_num_links(uint32_t num_links) {
99  fbb_.AddElement<uint32_t>(AllGatherOp::VT_NUM_LINKS, num_links, 0);
100  }
101  explicit AllGatherOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
102  : fbb_(_fbb) {
103  start_ = fbb_.StartTable();
104  }
105  ::flatbuffers::Offset<AllGatherOp> Finish() {
106  const auto end = fbb_.EndTable(start_);
107  auto o = ::flatbuffers::Offset<AllGatherOp>(end);
108  return o;
109  }
110 };
111 
112 inline ::flatbuffers::Offset<AllGatherOp> CreateAllGatherOp(
113  ::flatbuffers::FlatBufferBuilder &_fbb,
114  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> in = 0,
115  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
116  ::flatbuffers::Offset<tt::target::DeviceRef> device = 0,
117  int32_t all_gather_dim = 0,
118  uint32_t cluster_axis = 0,
119  uint32_t num_links = 0) {
120  AllGatherOpBuilder builder_(_fbb);
121  builder_.add_num_links(num_links);
122  builder_.add_cluster_axis(cluster_axis);
123  builder_.add_all_gather_dim(all_gather_dim);
124  builder_.add_device(device);
125  builder_.add_out(out);
126  builder_.add_in(in);
127  return builder_.Finish();
128 }
129 
131  using type = AllGatherOp;
132  static auto constexpr Create = CreateAllGatherOp;
133 };
134 
135 struct CollectivePermuteOp FLATBUFFERS_FINAL_CLASS : private ::flatbuffers::Table {
137  struct Traits;
138  enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
139  VT_IN = 4,
140  VT_OUT = 6,
141  VT_DEVICE = 8,
142  VT_SOURCE_TARGET_PAIRS = 10
143  };
144  const tt::target::ttnn::TensorRef *in() const {
145  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_IN);
146  }
147  const tt::target::ttnn::TensorRef *out() const {
148  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_OUT);
149  }
150  const tt::target::DeviceRef *device() const {
151  return GetPointer<const tt::target::DeviceRef *>(VT_DEVICE);
152  }
153  const ::flatbuffers::Vector<int64_t> *source_target_pairs() const {
154  return GetPointer<const ::flatbuffers::Vector<int64_t> *>(VT_SOURCE_TARGET_PAIRS);
155  }
156  bool Verify(::flatbuffers::Verifier &verifier) const {
157  return VerifyTableStart(verifier) &&
158  VerifyOffset(verifier, VT_IN) &&
159  verifier.VerifyTable(in()) &&
160  VerifyOffset(verifier, VT_OUT) &&
161  verifier.VerifyTable(out()) &&
162  VerifyOffset(verifier, VT_DEVICE) &&
163  verifier.VerifyTable(device()) &&
164  VerifyOffset(verifier, VT_SOURCE_TARGET_PAIRS) &&
165  verifier.VerifyVector(source_target_pairs()) &&
166  verifier.EndTable();
167  }
168 };
169 
172  ::flatbuffers::FlatBufferBuilder &fbb_;
173  ::flatbuffers::uoffset_t start_;
174  void add_in(::flatbuffers::Offset<tt::target::ttnn::TensorRef> in) {
175  fbb_.AddOffset(CollectivePermuteOp::VT_IN, in);
176  }
177  void add_out(::flatbuffers::Offset<tt::target::ttnn::TensorRef> out) {
178  fbb_.AddOffset(CollectivePermuteOp::VT_OUT, out);
179  }
180  void add_device(::flatbuffers::Offset<tt::target::DeviceRef> device) {
181  fbb_.AddOffset(CollectivePermuteOp::VT_DEVICE, device);
182  }
183  void add_source_target_pairs(::flatbuffers::Offset<::flatbuffers::Vector<int64_t>> source_target_pairs) {
184  fbb_.AddOffset(CollectivePermuteOp::VT_SOURCE_TARGET_PAIRS, source_target_pairs);
185  }
186  explicit CollectivePermuteOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
187  : fbb_(_fbb) {
188  start_ = fbb_.StartTable();
189  }
190  ::flatbuffers::Offset<CollectivePermuteOp> Finish() {
191  const auto end = fbb_.EndTable(start_);
192  auto o = ::flatbuffers::Offset<CollectivePermuteOp>(end);
193  return o;
194  }
195 };
196 
197 inline ::flatbuffers::Offset<CollectivePermuteOp> CreateCollectivePermuteOp(
198  ::flatbuffers::FlatBufferBuilder &_fbb,
199  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> in = 0,
200  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
201  ::flatbuffers::Offset<tt::target::DeviceRef> device = 0,
202  ::flatbuffers::Offset<::flatbuffers::Vector<int64_t>> source_target_pairs = 0) {
203  CollectivePermuteOpBuilder builder_(_fbb);
204  builder_.add_source_target_pairs(source_target_pairs);
205  builder_.add_device(device);
206  builder_.add_out(out);
207  builder_.add_in(in);
208  return builder_.Finish();
209 }
210 
213  static auto constexpr Create = CreateCollectivePermuteOp;
214 };
215 
216 inline ::flatbuffers::Offset<CollectivePermuteOp> CreateCollectivePermuteOpDirect(
217  ::flatbuffers::FlatBufferBuilder &_fbb,
218  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> in = 0,
219  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
220  ::flatbuffers::Offset<tt::target::DeviceRef> device = 0,
221  const std::vector<int64_t> *source_target_pairs = nullptr) {
222  auto source_target_pairs__ = source_target_pairs ? _fbb.CreateVector<int64_t>(*source_target_pairs) : 0;
224  _fbb,
225  in,
226  out,
227  device,
228  source_target_pairs__);
229 }
230 
231 struct MeshShardOp FLATBUFFERS_FINAL_CLASS : private ::flatbuffers::Table {
233  struct Traits;
234  enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
235  VT_IN = 4,
236  VT_OUT = 6,
237  VT_DEVICE = 8,
241  VT_SHARD_DIMS = 16
242  };
243  const tt::target::ttnn::TensorRef *in() const {
244  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_IN);
245  }
246  const tt::target::ttnn::TensorRef *out() const {
247  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_OUT);
248  }
249  const tt::target::DeviceRef *device() const {
250  return GetPointer<const tt::target::DeviceRef *>(VT_DEVICE);
251  }
253  return static_cast<tt::target::ttnn::MeshShardDirection>(GetField<uint32_t>(VT_SHARD_DIRECTION, 0));
254  }
256  return static_cast<tt::target::ttnn::MeshShardType>(GetField<uint32_t>(VT_SHARD_TYPE, 0));
257  }
258  const ::flatbuffers::Vector<int64_t> *shard_shape() const {
259  return GetPointer<const ::flatbuffers::Vector<int64_t> *>(VT_SHARD_SHAPE);
260  }
261  const ::flatbuffers::Vector<int64_t> *shard_dims() const {
262  return GetPointer<const ::flatbuffers::Vector<int64_t> *>(VT_SHARD_DIMS);
263  }
264  bool Verify(::flatbuffers::Verifier &verifier) const {
265  return VerifyTableStart(verifier) &&
266  VerifyOffset(verifier, VT_IN) &&
267  verifier.VerifyTable(in()) &&
268  VerifyOffset(verifier, VT_OUT) &&
269  verifier.VerifyTable(out()) &&
270  VerifyOffset(verifier, VT_DEVICE) &&
271  verifier.VerifyTable(device()) &&
272  VerifyField<uint32_t>(verifier, VT_SHARD_DIRECTION, 4) &&
273  VerifyField<uint32_t>(verifier, VT_SHARD_TYPE, 4) &&
274  VerifyOffset(verifier, VT_SHARD_SHAPE) &&
275  verifier.VerifyVector(shard_shape()) &&
276  VerifyOffset(verifier, VT_SHARD_DIMS) &&
277  verifier.VerifyVector(shard_dims()) &&
278  verifier.EndTable();
279  }
280 };
281 
284  ::flatbuffers::FlatBufferBuilder &fbb_;
285  ::flatbuffers::uoffset_t start_;
286  void add_in(::flatbuffers::Offset<tt::target::ttnn::TensorRef> in) {
287  fbb_.AddOffset(MeshShardOp::VT_IN, in);
288  }
289  void add_out(::flatbuffers::Offset<tt::target::ttnn::TensorRef> out) {
290  fbb_.AddOffset(MeshShardOp::VT_OUT, out);
291  }
292  void add_device(::flatbuffers::Offset<tt::target::DeviceRef> device) {
293  fbb_.AddOffset(MeshShardOp::VT_DEVICE, device);
294  }
296  fbb_.AddElement<uint32_t>(MeshShardOp::VT_SHARD_DIRECTION, static_cast<uint32_t>(shard_direction), 0);
297  }
299  fbb_.AddElement<uint32_t>(MeshShardOp::VT_SHARD_TYPE, static_cast<uint32_t>(shard_type), 0);
300  }
301  void add_shard_shape(::flatbuffers::Offset<::flatbuffers::Vector<int64_t>> shard_shape) {
302  fbb_.AddOffset(MeshShardOp::VT_SHARD_SHAPE, shard_shape);
303  }
304  void add_shard_dims(::flatbuffers::Offset<::flatbuffers::Vector<int64_t>> shard_dims) {
305  fbb_.AddOffset(MeshShardOp::VT_SHARD_DIMS, shard_dims);
306  }
307  explicit MeshShardOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
308  : fbb_(_fbb) {
309  start_ = fbb_.StartTable();
310  }
311  ::flatbuffers::Offset<MeshShardOp> Finish() {
312  const auto end = fbb_.EndTable(start_);
313  auto o = ::flatbuffers::Offset<MeshShardOp>(end);
314  return o;
315  }
316 };
317 
318 inline ::flatbuffers::Offset<MeshShardOp> CreateMeshShardOp(
319  ::flatbuffers::FlatBufferBuilder &_fbb,
320  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> in = 0,
321  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
322  ::flatbuffers::Offset<tt::target::DeviceRef> device = 0,
325  ::flatbuffers::Offset<::flatbuffers::Vector<int64_t>> shard_shape = 0,
326  ::flatbuffers::Offset<::flatbuffers::Vector<int64_t>> shard_dims = 0) {
327  MeshShardOpBuilder builder_(_fbb);
328  builder_.add_shard_dims(shard_dims);
329  builder_.add_shard_shape(shard_shape);
330  builder_.add_shard_type(shard_type);
331  builder_.add_shard_direction(shard_direction);
332  builder_.add_device(device);
333  builder_.add_out(out);
334  builder_.add_in(in);
335  return builder_.Finish();
336 }
337 
339  using type = MeshShardOp;
340  static auto constexpr Create = CreateMeshShardOp;
341 };
342 
343 inline ::flatbuffers::Offset<MeshShardOp> CreateMeshShardOpDirect(
344  ::flatbuffers::FlatBufferBuilder &_fbb,
345  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> in = 0,
346  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
347  ::flatbuffers::Offset<tt::target::DeviceRef> device = 0,
350  const std::vector<int64_t> *shard_shape = nullptr,
351  const std::vector<int64_t> *shard_dims = nullptr) {
352  auto shard_shape__ = shard_shape ? _fbb.CreateVector<int64_t>(*shard_shape) : 0;
353  auto shard_dims__ = shard_dims ? _fbb.CreateVector<int64_t>(*shard_dims) : 0;
355  _fbb,
356  in,
357  out,
358  device,
359  shard_direction,
360  shard_type,
361  shard_shape__,
362  shard_dims__);
363 }
364 
365 struct ReduceScatterOp FLATBUFFERS_FINAL_CLASS : private ::flatbuffers::Table {
367  struct Traits;
368  enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE {
369  VT_IN = 4,
370  VT_OUT = 6,
371  VT_DEVICE = 8,
374  VT_CLUSTER_AXIS = 14,
375  VT_NUM_LINKS = 16
376  };
377  const tt::target::ttnn::TensorRef *in() const {
378  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_IN);
379  }
380  const tt::target::ttnn::TensorRef *out() const {
381  return GetPointer<const tt::target::ttnn::TensorRef *>(VT_OUT);
382  }
383  const tt::target::DeviceRef *device() const {
384  return GetPointer<const tt::target::DeviceRef *>(VT_DEVICE);
385  }
386  int32_t scatter_dim() const {
387  return GetField<int32_t>(VT_SCATTER_DIM, 0);
388  }
389  uint32_t reduce_type() const {
390  return GetField<uint32_t>(VT_REDUCE_TYPE, 0);
391  }
392  uint32_t cluster_axis() const {
393  return GetField<uint32_t>(VT_CLUSTER_AXIS, 0);
394  }
395  uint32_t num_links() const {
396  return GetField<uint32_t>(VT_NUM_LINKS, 0);
397  }
398  bool Verify(::flatbuffers::Verifier &verifier) const {
399  return VerifyTableStart(verifier) &&
400  VerifyOffset(verifier, VT_IN) &&
401  verifier.VerifyTable(in()) &&
402  VerifyOffset(verifier, VT_OUT) &&
403  verifier.VerifyTable(out()) &&
404  VerifyOffset(verifier, VT_DEVICE) &&
405  verifier.VerifyTable(device()) &&
406  VerifyField<int32_t>(verifier, VT_SCATTER_DIM, 4) &&
407  VerifyField<uint32_t>(verifier, VT_REDUCE_TYPE, 4) &&
408  VerifyField<uint32_t>(verifier, VT_CLUSTER_AXIS, 4) &&
409  VerifyField<uint32_t>(verifier, VT_NUM_LINKS, 4) &&
410  verifier.EndTable();
411  }
412 };
413 
416  ::flatbuffers::FlatBufferBuilder &fbb_;
417  ::flatbuffers::uoffset_t start_;
418  void add_in(::flatbuffers::Offset<tt::target::ttnn::TensorRef> in) {
419  fbb_.AddOffset(ReduceScatterOp::VT_IN, in);
420  }
421  void add_out(::flatbuffers::Offset<tt::target::ttnn::TensorRef> out) {
422  fbb_.AddOffset(ReduceScatterOp::VT_OUT, out);
423  }
424  void add_device(::flatbuffers::Offset<tt::target::DeviceRef> device) {
425  fbb_.AddOffset(ReduceScatterOp::VT_DEVICE, device);
426  }
427  void add_scatter_dim(int32_t scatter_dim) {
428  fbb_.AddElement<int32_t>(ReduceScatterOp::VT_SCATTER_DIM, scatter_dim, 0);
429  }
430  void add_reduce_type(uint32_t reduce_type) {
431  fbb_.AddElement<uint32_t>(ReduceScatterOp::VT_REDUCE_TYPE, reduce_type, 0);
432  }
433  void add_cluster_axis(uint32_t cluster_axis) {
434  fbb_.AddElement<uint32_t>(ReduceScatterOp::VT_CLUSTER_AXIS, cluster_axis, 0);
435  }
436  void add_num_links(uint32_t num_links) {
437  fbb_.AddElement<uint32_t>(ReduceScatterOp::VT_NUM_LINKS, num_links, 0);
438  }
439  explicit ReduceScatterOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
440  : fbb_(_fbb) {
441  start_ = fbb_.StartTable();
442  }
443  ::flatbuffers::Offset<ReduceScatterOp> Finish() {
444  const auto end = fbb_.EndTable(start_);
445  auto o = ::flatbuffers::Offset<ReduceScatterOp>(end);
446  return o;
447  }
448 };
449 
450 inline ::flatbuffers::Offset<ReduceScatterOp> CreateReduceScatterOp(
451  ::flatbuffers::FlatBufferBuilder &_fbb,
452  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> in = 0,
453  ::flatbuffers::Offset<tt::target::ttnn::TensorRef> out = 0,
454  ::flatbuffers::Offset<tt::target::DeviceRef> device = 0,
455  int32_t scatter_dim = 0,
456  uint32_t reduce_type = 0,
457  uint32_t cluster_axis = 0,
458  uint32_t num_links = 0) {
459  ReduceScatterOpBuilder builder_(_fbb);
460  builder_.add_num_links(num_links);
461  builder_.add_cluster_axis(cluster_axis);
462  builder_.add_reduce_type(reduce_type);
463  builder_.add_scatter_dim(scatter_dim);
464  builder_.add_device(device);
465  builder_.add_out(out);
466  builder_.add_in(in);
467  return builder_.Finish();
468 }
469 
472  static auto constexpr Create = CreateReduceScatterOp;
473 };
474 
475 } // namespace ttnn
476 } // namespace target
477 } // namespace tt
478 
479 #endif // FLATBUFFERS_GENERATED_CCL_TT_TARGET_TTNN_H_
VT_SHARD_SHAPE
Definition: ccl_generated.h:240
VT_CLUSTER_AXIS
Definition: ccl_generated.h:43
VT_ALL_GATHER_DIM
Definition: ccl_generated.h:42
VT_IN
Definition: ccl_generated.h:39
VT_SHARD_DIRECTION
Definition: ccl_generated.h:238
VT_DEVICE
Definition: ccl_generated.h:41
VT_REDUCE_TYPE
Definition: ccl_generated.h:373
VT_OUT
Definition: ccl_generated.h:40
VT_SCATTER_DIM
Definition: ccl_generated.h:372
VT_SHARD_TYPE
Definition: ccl_generated.h:239
MeshShardType
Definition: types_generated.h:170
inline ::flatbuffers::Offset< MeshShardOp > CreateMeshShardOp(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > in=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, ::flatbuffers::Offset< tt::target::DeviceRef > device=0, tt::target::ttnn::MeshShardDirection shard_direction=tt::target::ttnn::MeshShardDirection::FullToShardShape, tt::target::ttnn::MeshShardType shard_type=tt::target::ttnn::MeshShardType::Identity, ::flatbuffers::Offset<::flatbuffers::Vector< int64_t >> shard_shape=0, ::flatbuffers::Offset<::flatbuffers::Vector< int64_t >> shard_dims=0)
Definition: ccl_generated.h:318
inline ::flatbuffers::Offset< MeshShardOp > CreateMeshShardOpDirect(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > in=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, ::flatbuffers::Offset< tt::target::DeviceRef > device=0, tt::target::ttnn::MeshShardDirection shard_direction=tt::target::ttnn::MeshShardDirection::FullToShardShape, tt::target::ttnn::MeshShardType shard_type=tt::target::ttnn::MeshShardType::Identity, const std::vector< int64_t > *shard_shape=nullptr, const std::vector< int64_t > *shard_dims=nullptr)
Definition: ccl_generated.h:343
inline ::flatbuffers::Offset< CollectivePermuteOp > CreateCollectivePermuteOp(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > in=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, ::flatbuffers::Offset< tt::target::DeviceRef > device=0, ::flatbuffers::Offset<::flatbuffers::Vector< int64_t >> source_target_pairs=0)
Definition: ccl_generated.h:197
inline ::flatbuffers::Offset< CollectivePermuteOp > CreateCollectivePermuteOpDirect(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > in=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, ::flatbuffers::Offset< tt::target::DeviceRef > device=0, const std::vector< int64_t > *source_target_pairs=nullptr)
Definition: ccl_generated.h:216
inline ::flatbuffers::Offset< AllGatherOp > CreateAllGatherOp(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > in=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, ::flatbuffers::Offset< tt::target::DeviceRef > device=0, int32_t all_gather_dim=0, uint32_t cluster_axis=0, uint32_t num_links=0)
Definition: ccl_generated.h:112
inline ::flatbuffers::Offset< ReduceScatterOp > CreateReduceScatterOp(::flatbuffers::FlatBufferBuilder &_fbb, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > in=0, ::flatbuffers::Offset< tt::target::ttnn::TensorRef > out=0, ::flatbuffers::Offset< tt::target::DeviceRef > device=0, int32_t scatter_dim=0, uint32_t reduce_type=0, uint32_t cluster_axis=0, uint32_t num_links=0)
Definition: ccl_generated.h:450
MeshShardDirection
Definition: types_generated.h:140
Definition: debug_info_generated.h:18
Definition: debug_info_generated.h:36
Definition: ccl_generated.h:79
::flatbuffers::Offset< AllGatherOp > Finish()
Definition: ccl_generated.h:105
void add_num_links(uint32_t num_links)
Definition: ccl_generated.h:98
AllGatherOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: ccl_generated.h:101
void add_device(::flatbuffers::Offset< tt::target::DeviceRef > device)
Definition: ccl_generated.h:89
void add_cluster_axis(uint32_t cluster_axis)
Definition: ccl_generated.h:95
AllGatherOp Table
Definition: ccl_generated.h:80
void add_all_gather_dim(int32_t all_gather_dim)
Definition: ccl_generated.h:92
void add_in(::flatbuffers::Offset< tt::target::ttnn::TensorRef > in)
Definition: ccl_generated.h:83
::flatbuffers::uoffset_t start_
Definition: ccl_generated.h:82
::flatbuffers::FlatBufferBuilder & fbb_
Definition: ccl_generated.h:81
void add_out(::flatbuffers::Offset< tt::target::ttnn::TensorRef > out)
Definition: ccl_generated.h:86
Definition: ccl_generated.h:130
static constexpr auto Create
Definition: ccl_generated.h:132
AllGatherOp type
Definition: ccl_generated.h:131
Definition: ccl_generated.h:170
::flatbuffers::FlatBufferBuilder & fbb_
Definition: ccl_generated.h:172
void add_in(::flatbuffers::Offset< tt::target::ttnn::TensorRef > in)
Definition: ccl_generated.h:174
::flatbuffers::uoffset_t start_
Definition: ccl_generated.h:173
void add_source_target_pairs(::flatbuffers::Offset<::flatbuffers::Vector< int64_t >> source_target_pairs)
Definition: ccl_generated.h:183
void add_device(::flatbuffers::Offset< tt::target::DeviceRef > device)
Definition: ccl_generated.h:180
CollectivePermuteOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: ccl_generated.h:186
CollectivePermuteOp Table
Definition: ccl_generated.h:171
void add_out(::flatbuffers::Offset< tt::target::ttnn::TensorRef > out)
Definition: ccl_generated.h:177
::flatbuffers::Offset< CollectivePermuteOp > Finish()
Definition: ccl_generated.h:190
Definition: ccl_generated.h:211
CollectivePermuteOp type
Definition: ccl_generated.h:212
static constexpr auto Create
Definition: ccl_generated.h:213
Definition: binary_generated.h:31
uint32_t reduce_type() const
Definition: ccl_generated.h:389
const ::flatbuffers::Vector< int64_t > * shard_dims() const
Definition: ccl_generated.h:261
tt::target::ttnn::MeshShardType shard_type() const
Definition: ccl_generated.h:255
uint32_t num_links() const
Definition: ccl_generated.h:61
ReduceScatterOpBuilder Builder
Definition: ccl_generated.h:366
const tt::target::ttnn::TensorRef * out() const
Definition: ccl_generated.h:49
CollectivePermuteOpBuilder Builder
Definition: ccl_generated.h:136
uint32_t cluster_axis() const
Definition: ccl_generated.h:58
int32_t scatter_dim() const
Definition: ccl_generated.h:386
AllGatherOpBuilder Builder
Definition: ccl_generated.h:36
int32_t all_gather_dim() const
Definition: ccl_generated.h:55
const ::flatbuffers::Vector< int64_t > * source_target_pairs() const
Definition: ccl_generated.h:153
const tt::target::ttnn::TensorRef * in() const
Definition: ccl_generated.h:46
bool Verify(::flatbuffers::Verifier &verifier) const
Definition: ccl_generated.h:64
tt::target::ttnn::MeshShardDirection shard_direction() const
Definition: ccl_generated.h:252
const ::flatbuffers::Vector< int64_t > * shard_shape() const
Definition: ccl_generated.h:258
const tt::target::DeviceRef * device() const
Definition: ccl_generated.h:52
MeshShardOpBuilder Builder
Definition: ccl_generated.h:232
Definition: ccl_generated.h:282
MeshShardOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: ccl_generated.h:307
void add_in(::flatbuffers::Offset< tt::target::ttnn::TensorRef > in)
Definition: ccl_generated.h:286
void add_shard_dims(::flatbuffers::Offset<::flatbuffers::Vector< int64_t >> shard_dims)
Definition: ccl_generated.h:304
MeshShardOp Table
Definition: ccl_generated.h:283
void add_device(::flatbuffers::Offset< tt::target::DeviceRef > device)
Definition: ccl_generated.h:292
::flatbuffers::uoffset_t start_
Definition: ccl_generated.h:285
::flatbuffers::FlatBufferBuilder & fbb_
Definition: ccl_generated.h:284
void add_shard_type(tt::target::ttnn::MeshShardType shard_type)
Definition: ccl_generated.h:298
void add_shard_shape(::flatbuffers::Offset<::flatbuffers::Vector< int64_t >> shard_shape)
Definition: ccl_generated.h:301
void add_shard_direction(tt::target::ttnn::MeshShardDirection shard_direction)
Definition: ccl_generated.h:295
void add_out(::flatbuffers::Offset< tt::target::ttnn::TensorRef > out)
Definition: ccl_generated.h:289
::flatbuffers::Offset< MeshShardOp > Finish()
Definition: ccl_generated.h:311
Definition: ccl_generated.h:338
MeshShardOp type
Definition: ccl_generated.h:339
static constexpr auto Create
Definition: ccl_generated.h:340
Definition: ccl_generated.h:414
void add_num_links(uint32_t num_links)
Definition: ccl_generated.h:436
ReduceScatterOpBuilder(::flatbuffers::FlatBufferBuilder &_fbb)
Definition: ccl_generated.h:439
ReduceScatterOp Table
Definition: ccl_generated.h:415
void add_out(::flatbuffers::Offset< tt::target::ttnn::TensorRef > out)
Definition: ccl_generated.h:421
::flatbuffers::FlatBufferBuilder & fbb_
Definition: ccl_generated.h:416
::flatbuffers::uoffset_t start_
Definition: ccl_generated.h:417
void add_reduce_type(uint32_t reduce_type)
Definition: ccl_generated.h:430
void add_scatter_dim(int32_t scatter_dim)
Definition: ccl_generated.h:427
::flatbuffers::Offset< ReduceScatterOp > Finish()
Definition: ccl_generated.h:443
void add_in(::flatbuffers::Offset< tt::target::ttnn::TensorRef > in)
Definition: ccl_generated.h:418
void add_device(::flatbuffers::Offset< tt::target::DeviceRef > device)
Definition: ccl_generated.h:424
void add_cluster_axis(uint32_t cluster_axis)
Definition: ccl_generated.h:433
Definition: ccl_generated.h:470
static constexpr auto Create
Definition: ccl_generated.h:472
ReduceScatterOp type
Definition: ccl_generated.h:471