/external/deqp/modules/gles31/functional/ |
es31fShaderAtomicOpTests.cpp | 68 ShaderAtomicOpCase (Context& context, const char* name, const char* funcName, AtomicOperandType operandType, DataType type, Precision precision, const UVec3& workGroupSize); 96 ShaderAtomicOpCase::ShaderAtomicOpCase (Context& context, const char* name, const char* funcName, AtomicOperandType operandType, DataType type, Precision precision, const UVec3& workGroupSize) 102 , m_workGroupSize (workGroupSize) 279 const int workGroupSize = (int)product(m_workGroupSize); 280 const int numWorkGroups = numValues/workGroupSize; 284 const int groupOffset = groupNdx*workGroupSize; 290 for (int localNdx = 0; localNdx < workGroupSize; localNdx++) 302 for (int localNdx = 0; localNdx < workGroupSize; localNdx++) 432 const int workGroupSize = (int)product(m_workGroupSize); 433 const int numWorkGroups = numValues/workGroupSize; [all...] |
es31fComputeShaderBuiltinVarTests.cpp | 85 virtual UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const = 0; 305 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 308 DE_UNREF(workGroupSize); 332 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 337 return workGroupSize; 355 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 358 DE_UNREF(workGroupSize); 381 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 384 DE_UNREF(workGroupSize); 406 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) cons [all...] |
es31fIndirectComputeDispatchTests.cpp | 87 glu::ProgramSources genVerifySources (const UVec3& workGroupSize) 105 args["LOCAL_SIZE_X"] = de::toString(workGroupSize.x()); 106 args["LOCAL_SIZE_Y"] = de::toString(workGroupSize.y()); 107 args["LOCAL_SIZE_Z"] = de::toString(workGroupSize.z()); 284 deUint32 computeInvocationCount (const UVec3& workGroupSize, const UVec3& numWorkGroups) 286 const int numInvocationsPerGroup = workGroupSize[0]*workGroupSize[1]*workGroupSize[2]; 395 SingleDispatchCase (Context& context, const char* name, const char* description, GenBuffer genBuffer, deUintptr bufferSize, deUintptr offset, const UVec3& workGroupSize, const UVec3& numWorkGroups) 399 m_workGroupSize = workGroupSize; [all...] |
es31fShaderSharedVarTests.cpp | 74 SharedBasicVarCase (Context& context, const char* name, DataType basicType, Precision precision, const tcu::UVec3& workGroupSize); 92 static std::string getBasicCaseDescription (DataType basicType, Precision precision, const tcu::UVec3& workGroupSize) 97 str << getDataTypeName(basicType) << ", work group size = " << workGroupSize; 101 SharedBasicVarCase::SharedBasicVarCase (Context& context, const char* name, DataType basicType, Precision precision, const tcu::UVec3& workGroupSize) 102 : TestCase (context, name, getBasicCaseDescription(basicType, precision, workGroupSize).c_str()) 105 , m_workGroupSize (workGroupSize) 375 const tcu::UVec3 workGroupSize (2,1,3); 376 basicTypeGroup->addChild(new SharedBasicVarCase(m_context, getDataTypeName(DataType(basicType)), DataType(basicType), PRECISION_LAST, workGroupSize)); 382 const tcu::UVec3 workGroupSize (2,1,3); 385 basicTypeGroup->addChild(new SharedBasicVarCase(m_context, name.c_str(), DataType(basicType), Precision(precision), workGroupSize)); [all...] |
es31fBasicComputeShaderTests.cpp | 609 const int workGroupSize = m_localSize[0]*m_localSize[1]*m_localSize[2]; 611 const int numValues = workGroupSize*workGroupCount; 621 << "shared uint offsets[" << workGroupSize << "];\n\n" 672 for (int localOffs = 0; localOffs < workGroupSize; localOffs++) 674 const int globalOffs = groupNdx*workGroupSize; 676 const int offs0 = localOffs-1 < 0 ? ((localOffs+workGroupSize-1)%workGroupSize) : ((localOffs-1)%workGroupSize); 677 const int offs1 = localOffs-2 < 0 ? ((localOffs+workGroupSize-2)%workGroupSize) : ((localOffs-2)%workGroupSize) [all...] |
/external/deqp/external/vulkancts/modules/vulkan/compute/ |
vktComputeShaderBuiltinVarTests.cpp | 155 virtual UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const = 0; 234 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 237 DE_UNREF(workGroupSize); 261 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 266 return workGroupSize; 285 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 288 DE_UNREF(workGroupSize); 311 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) const 314 DE_UNREF(workGroupSize); 336 UVec3 computeReference (const UVec3& numWorkGroups, const UVec3& workGroupSize, const UVec3& workGroupID, const UVec3& localInvocationID) cons [all...] |
vktComputeIndirectComputeDispatchTests.cpp | 103 const tcu::UVec3 workGroupSize, 108 , m_workGroupSize (workGroupSize) 124 const tcu::UVec3& workGroupSize, 162 const tcu::UVec3& workGroupSize, 172 , m_workGroupSize (workGroupSize) 441 const tcu::UVec3& workGroupSize, 443 : IndirectDispatchInstanceBufferUpload(context, name, bufferSize, workGroupSize, dispatchCommands) {}
|
vktComputeBasicComputeShaderTests.cpp | 152 const int workGroupSize = multiplyComponents(m_localSize); 154 const int numValues = workGroupSize * workGroupCount; 162 << "shared uint offsets[" << workGroupSize << "];\n\n" 198 const int workGroupSize = multiplyComponents(m_localSize); 203 const VkDeviceSize bufferSizeBytes = sizeof(deUint32) * workGroupSize * workGroupCount; 262 const int globalOffset = groupNdx * workGroupSize; 263 for (int localOffset = 0; localOffset < workGroupSize; ++localOffset) 266 const deUint32 ref = globalOffset + squared(workGroupSize - localOffset - 1); 323 const int workGroupSize = multiplyComponents(m_localSize); 325 const int numValues = workGroupSize * workGroupCount [all...] |
/external/deqp/external/vulkancts/modules/vulkan/sparse_resources/ |
vktSparseResourcesBufferMemoryAliasing.cpp | 116 const tcu::UVec3 workGroupSize = computeWorkGroupSize(numInvocations); 120 << "layout (local_size_x = " << workGroupSize.x() << ", local_size_y = " << workGroupSize.y() << ", local_size_z = " << workGroupSize.z() << ") in;\n" 326 const tcu::UVec3 workGroupSize = computeWorkGroupSize(numInvocationsLeft); 329 numInvocationsLeft -= workGroupSize.x()*workGroupSize.y()*workGroupSize.z();
|
vktSparseResourcesShaderIntrinsicsStorage.cpp | 85 << "OpDecorate %input_WorkGroupSize BuiltIn WorkgroupSize\n" 431 { 4u, 3u * (deUint32)sizeof(deUint32), sizeof(deUint32) }, // WorkGroupSize.x 432 { 5u, 4u * (deUint32)sizeof(deUint32), sizeof(deUint32) }, // WorkGroupSize.y 433 { 6u, 5u * (deUint32)sizeof(deUint32), sizeof(deUint32) }, // WorkGroupSize.z 441 const tcu::UVec3 workGroupSize = computeWorkGroupSize(gridSize); 442 const tcu::UVec3 specializationData[2] = { gridSize, workGroupSize }; 483 const deUint32 xWorkGroupCount = gridSize.x() / workGroupSize.x() + (gridSize.x() % workGroupSize.x() ? 1u : 0u); 484 const deUint32 yWorkGroupCount = gridSize.y() / workGroupSize.y() + (gridSize.y() % workGroupSize.y() ? 1u : 0u) [all...] |
vktSparseResourcesImageSparseResidency.cpp | 153 const tcu::UVec3 workGroupSize = computeWorkGroupSize(gridSize); 157 << "layout (local_size_x = " << workGroupSize.x() << ", local_size_y = " << workGroupSize.y() << ", local_size_z = " << workGroupSize.z() << ") in; \n" 516 const tcu::UVec3 workGroupSize = computeWorkGroupSize(gridSize); 518 const deUint32 xWorkGroupCount = gridSize.x() / workGroupSize.x() + (gridSize.x() % workGroupSize.x() ? 1u : 0u); 519 const deUint32 yWorkGroupCount = gridSize.y() / workGroupSize.y() + (gridSize.y() % workGroupSize.y() ? 1u : 0u); 520 const deUint32 zWorkGroupCount = gridSize.z() / workGroupSize.z() + (gridSize.z() % workGroupSize.z() ? 1u : 0u) [all...] |
/prebuilts/ndk/r16/sources/third_party/shaderc/third_party/spirv-tools/source/ |
name_mapper.cpp | 139 GLCASE2(WorkgroupSize, WorkGroupSize)
|
/external/mesa3d/src/amd/common/ |
amd_kernel_code_t.h | 284 * ((gridSize.x+workgroupSize.x-1)/workgroupSize.x). 290 * ((gridSize.y+workgroupSize.y-1)/workgroupSize.y). 298 * ((gridSize.z+workgroupSize.z-1)/workgroupSize.z).
|
/external/llvm/lib/Target/AMDGPU/ |
AMDKernelCodeT.h | 428 /// ((gridSize.x+workgroupSize.x-1)/workgroupSize.x). 434 /// ((gridSize.y+workgroupSize.y-1)/workgroupSize.y). 442 /// ((gridSize.z+workgroupSize.z-1)/workgroupSize.z). [all...] |
AMDGPUPromoteAlloca.cpp | 654 unsigned WorkGroupSize = AMDGPU::getMaximumWorkGroupSize(ContainingFunction); 669 uint32_t AllocSize = WorkGroupSize * DL.getTypeAllocSize(AllocaTy); 691 Type *GVTy = ArrayType::get(I.getAllocatedType(), WorkGroupSize);
|
/frameworks/rs/rsov/compiler/spirit/test_data/ |
greyscale.spt | 47 OpDecorate %gl_WorkGroupSize BuiltIn WorkgroupSize
|
greyscale2.spt | 45 OpDecorate %gl_WorkGroupSize BuiltIn WorkgroupSize
|
/frameworks/rs/rsov/compiler/spirit/ |
builder_test.cpp | 89 GSize->decorate(Decoration::BuiltIn)->addExtraOperand(BuiltIn::WorkgroupSize);
|
/prebuilts/ndk/r16/sources/third_party/shaderc/third_party/glslang/glslang/Include/ |
BaseTypes.h | 275 case EbvWorkGroupSize: return "WorkGroupSize";
|
/external/ImageMagick/MagickCore/ |
accelerate.c | [all...] |
/external/shaderc/spirv-headers/include/spirv/1.0/ |
spirv.hpp11 | 404 WorkgroupSize = 25,
|
spirv.json | 447 "WorkgroupSize": 25, [all...] |
spirv.lua | 376 WorkgroupSize = 25,
|
spirv.py | 376 'WorkgroupSize' : 25,
|
/external/shaderc/spirv-headers/include/spirv/1.1/ |
spirv.hpp11 | 409 WorkgroupSize = 25,
|