/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/ops/ |
H A D | Transpose.cpp | 30 const struct Block final { in transpose_4d() struct 31 ivec3 out_extents; in transpose_4d() 32 int32_t fill0; in transpose_4d() 33 ivec3 in_extents; in transpose_4d() 34 int32_t fill1; in transpose_4d() 35 uvec4 out_tensor_size; in transpose_4d() 36 uvec4 in_tensor_size; in transpose_4d() 37 uvec4 out_ndims; in transpose_4d() 38 uvec2 ch_info; in transpose_4d()
|
H A D | Random.cpp | 30 const struct Block final { in uniform_() struct 31 uvec3 extents; in uniform_() 32 float from; in uniform_() 33 float to; in uniform_() 91 const struct Block final { in normal_() struct 92 uvec3 extents; in normal_() 93 float mean; in normal_() 94 float std; in normal_()
|
H A D | Slice.cpp | 33 const struct Block final { in slice_4d() struct 34 ivec3 size; // output texture size in slice_4d() 35 int32_t fill_0; // dummy in slice_4d() 36 ivec3 isize; // input texture size in slice_4d() 37 int32_t fill_1; // dummy in slice_4d() 38 uvec4 tensor_size; // output tensor size in slice_4d() 39 uvec4 itensor_size; // input tensor size in slice_4d() 40 uvec4 args; // input arguments (dim, start, end, step) in slice_4d() 41 uvec2 c_info; // tensor channels aligned to 4 in slice_4d()
|
H A D | Concat.cpp | 81 const struct Block final { in cat_feature() struct 82 ivec3 outExtents; in cat_feature() 83 int32_t fill0; in cat_feature() 84 ivec3 inExtents; in cat_feature() 85 int32_t fill1; in cat_feature() 86 uvec2 outChInfo; in cat_feature() 87 uvec2 inChInfo; in cat_feature() 88 uvec4 appendedChInfo; in cat_feature()
|
H A D | UnaryOp.cpp | 28 const struct Block final { in unary_op() struct 29 uvec3 extents; in unary_op() 30 uint32_t fill0; in unary_op() 71 const struct Block final { in unary_op_() struct 72 uvec3 extents; in unary_op_() 73 uint32_t fill0; in unary_op_()
|
/aosp_15_r20/external/skia/src/base/ |
H A D | SkDeque.cpp | 14 struct SkDeque::Block { struct in SkDeque 15 Block* fNext; 16 Block* fPrev; 17 char* fBegin; // start of used section in this chunk 18 char* fEnd; // end of used section in this chunk 19 char* fStop; // end of the allocated chunk 21 char* start() { return (char*)(this + 1); } in start() 22 const char* start() const { return (const char*)(this + 1); } in start() 24 void init(size_t size) { in init()
|
/aosp_15_r20/external/perfetto/src/trace_processor/containers/ |
H A D | string_pool.h | 179 struct Block { struct 180 explicit Block(size_t size) in Block() argument 188 Block& operator=(Block&&) = default; argument 192 Block& operator=(const Block&) = delete; argument 194 uint8_t* Get(uint32_t offset) const { in Get() 220 // are the index of the Block in the pool, and the remaining 25 bits the argument
|
/aosp_15_r20/external/rust/android-crates-io/crates/ring/src/aead/ |
D | block.rs | 20 pub struct Block([BigEndian<u64>; 2]); struct 24 impl Block { implementation 45 impl From<[u64; 2]> for Block { implementation 59 impl BitXorAssign for Block { implementation 68 impl BitXor for Block { implementation 79 impl From<&'_ [u8; BLOCK_LEN]> for Block { implementation 86 impl AsRef<[u8; BLOCK_LEN]> for Block { implementation
|
/aosp_15_r20/external/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
H A D | SIMachineScheduler.cpp | 1254 SIScheduleBlock *Block = CurrentBlocks[i]; in createBlocksForVariant() local 1289 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1300 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1313 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1334 SIScheduleBlock *Block = CurrentBlocks[i]; in scheduleInsideBlocks() local 1350 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in scheduleInsideBlocks() local 1381 SIScheduleBlock *Block = CurrentBlocks[i]; in scheduleInsideBlocks() local 1411 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in fillStats() local 1426 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in fillStats() local 1459 SIScheduleBlock *Block = Blocks[i]; in SIScheduleBlockScheduler() local [all …]
|
/aosp_15_r20/external/llvm/lib/Target/AMDGPU/ |
H A D | SIMachineScheduler.cpp | 1090 SIScheduleBlock *Block = CurrentBlocks[i]; in createBlocksForVariant() local 1140 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1151 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1164 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1185 SIScheduleBlock *Block = CurrentBlocks[i]; in scheduleInsideBlocks() local 1201 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in scheduleInsideBlocks() local 1232 SIScheduleBlock *Block = CurrentBlocks[i]; in scheduleInsideBlocks() local 1264 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in fillStats() local 1279 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in fillStats() local 1314 SIScheduleBlock *Block = Blocks[i]; in SIScheduleBlockScheduler() local [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/core/platform/cloud/ |
H A D | ram_file_block_cache.h | 169 struct Block { struct 188 /// The block map is an ordered map from Key to Block. argument
|
/aosp_15_r20/external/tensorflow/tensorflow/c/experimental/filesystem/plugins/gcs/ |
H A D | ram_file_block_cache.h | 185 struct Block { struct 204 /// The block map is an ordered map from Key to Block. argument
|
/aosp_15_r20/external/rust/android-crates-io/crates/fixedbitset/src/block/ |
D | mod.rs | 69 impl Block { implementation 86 impl Eq for Block {} implementation 88 impl PartialOrd for Block { implementation 95 impl Ord for Block { implementation 102 impl Default for Block { implementation 109 impl Hash for Block { implementation
|
/aosp_15_r20/external/perfetto/include/perfetto/protozero/ |
H A D | message_arena.h | 78 struct Block { struct 81 Block() { PERFETTO_ASAN_POISON(storage, sizeof(storage)); } in Block() argument 90 std::forward_list<Block> blocks_; argument
|
/aosp_15_r20/prebuilts/runtime/mainline/platform/sdk/include/external/perfetto/include/perfetto/protozero/ |
H A D | message_arena.h | 78 struct Block { struct 81 Block() { PERFETTO_ASAN_POISON(storage, sizeof(storage)); } in Block() argument 90 std::forward_list<Block> blocks_; argument
|
/aosp_15_r20/external/swiftshader/third_party/llvm-16.0/llvm/lib/Target/AMDGPU/ |
H A D | SIMachineScheduler.cpp | 1257 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1268 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1281 SIScheduleBlock *Block = CurrentBlocks[i]; in topologicalSort() local 1302 SIScheduleBlock *Block = CurrentBlocks[i]; in scheduleInsideBlocks() local 1318 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in scheduleInsideBlocks() local 1349 SIScheduleBlock *Block = CurrentBlocks[i]; in scheduleInsideBlocks() local 1379 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in fillStats() local 1394 SIScheduleBlock *Block = CurrentBlocks[BlockIndice]; in fillStats() local 1455 SIScheduleBlock *Block = Blocks[i]; in SIScheduleBlockScheduler() local 1462 SIScheduleBlock *Block = Blocks[i]; in SIScheduleBlockScheduler() local [all …]
|
/aosp_15_r20/external/swiftshader/third_party/llvm-16.0/llvm/lib/Transforms/Utils/ |
H A D | SampleProfileInference.cpp | 625 auto &Block = Func.Blocks[I]; in joinIsolatedComponents() local 833 auto &Block = Func.Blocks[Queue.front()]; in findUnknownSubgraph() local 873 for (auto *Block : UnknownBlocks) { in canRebalanceSubgraph() local 926 auto fillInDegree = [&](const FlowBlock *Block) { in isAcyclicSubgraph() 934 for (auto *Block : UnknownBlocks) { in isAcyclicSubgraph() local 945 FlowBlock *Block = &Func.Blocks[Queue.front()]; in isAcyclicSubgraph() local 993 for (auto *Block : UnknownBlocks) { in rebalanceUnknownSubgraph() local 1008 const FlowBlock *Block, uint64_t BlockFlow) { in rebalanceBlock() 1075 auto &Block = Func.Blocks[B]; in initializeNetwork() local 1127 const FlowBlock &Block) { in assignBlockCosts() [all …]
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/impl/ |
H A D | Slice.cpp | 85 const struct Block final { in add_slice_tensor_copy_node() struct 86 int offset; in add_slice_tensor_copy_node() 87 int step; in add_slice_tensor_copy_node() 131 const struct Block final { in add_slice_tensor_copy_node() struct 132 int dim; in add_slice_tensor_copy_node() 133 int offset; in add_slice_tensor_copy_node() 134 int step; in add_slice_tensor_copy_node() 135 int stride; in add_slice_tensor_copy_node()
|
/aosp_15_r20/prebuilts/go/linux-x86/src/cmd/compile/internal/ssa/ |
D | branchelim.go | 115 func elimIf(f *Func, loadAddr *sparseSet, dom *Block) bool { 292 func isLeafPlain(b *Block) bool { 296 func clobberBlock(b *Block) { 309 func elimIfElse(f *Func, loadAddr *sparseSet, b *Block) bool { 396 func shouldElimIfElse(no, yes, post *Block, arch string) bool { 435 func canSpeculativelyExecute(b *Block) bool {
|
/aosp_15_r20/external/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Vectorize/ |
H A D | VPlan.cpp | 61 const VPBlockBase *Block = this; in getEntryBasicBlock() local 68 VPBlockBase *Block = this; in getEntryBasicBlock() local 76 const VPBlockBase *Block = this; in getExitBasicBlock() local 83 VPBlockBase *Block = this; in getExitBasicBlock() local 517 const Twine VPlanPrinter::getUID(const VPBlockBase *Block) { in getUID() 522 const Twine VPlanPrinter::getOrCreateName(const VPBlockBase *Block) { in getOrCreateName() 551 for (const VPBlockBase *Block : depth_first(Plan.getEntry())) in dump() local 557 void VPlanPrinter::dumpBlock(const VPBlockBase *Block) { in dumpBlock() 583 void VPlanPrinter::dumpEdges(const VPBlockBase *Block) { in dumpEdges() 645 for (const VPBlockBase *Block : depth_first(Region->getEntry())) in dumpRegion() local [all …]
|
/aosp_15_r20/external/deqp/framework/delibs/decpp/ |
H A D | deAppendList.hpp | 71 struct Block struct in de::AppendList 73 const size_t blockNdx; 74 ElementType *elements; 75 Block *volatile next; 77 Block(size_t blockNdx_, size_t size) in Block() function 85 ~Block(void) in ~Block()
|
/aosp_15_r20/external/executorch/backends/apple/coreml/runtime/sdk/ |
H A D | program_path.h | 39 struct Block { struct 41 static const char* kTypeName; 43 static const char* kIndexKeyName; 45 int64_t index; 47 Block(int64_t index) : index(index) { } in Block() argument
|
/aosp_15_r20/prebuilts/go/linux-x86/src/encoding/pem/ |
D | pem.go | 29 type Block struct { struct 30 Type string // The type, taken from the preamble (i.e. "RSA PRIVATE KEY"). 31 Headers map[string]string // Optional headers. 32 … []byte // The decoded bytes of the contents. Typically a DER encoded ASN.1 structure. 240 func Encode(out io.Writer, b *Block) error { 310 func EncodeToMemory(b *Block) []byte {
|
/aosp_15_r20/system/tools/hidl/host_utils/include/hidl-util/ |
D | Formatter.h | 37 struct Block { struct 43 bool populated = false; 44 bool printUnlessWrapped = false; 47 std::string content; 48 std::vector<Block> blocks; 50 Block* const parent;
|
/aosp_15_r20/hardware/google/aemu/host-common/ |
D | address_space_graphics.cpp | 59 struct Block { struct 60 char* buffer = nullptr; 61 uint64_t bufferSize = 0; 62 SubAllocator* subAlloc = nullptr; 63 uint64_t offsetIntoPhys = 0; // guest claimShared/mmap uses this 64 bool isEmpty = true; 65 std::optional<uint32_t> dedicatedContextHandle; 66 bool usesVirtioGpuHostmem = false; 67 uint64_t hostmemId = 0; 68 bool external = false;
|