/external/deqp/external/vulkancts/modules/vulkan/draw/ |
vktDrawBufferObjectUtil.hpp | 65 vk::VkDevice m_device; member in class:vkt::Draw::Buffer
|
vktDrawImageObjectUtil.hpp | 255 vk::VkDevice m_device; member in class:vkt::Draw::Image
|
/external/deqp/external/vulkancts/modules/vulkan/api/ |
vktApiComputeInstanceResultBuffer.hpp | 69 const vk::VkDevice m_device; member in class:vkt::api::ComputeInstanceResultBuffer
|
vktApiBufferTests.cpp | 110 return *(m_sparseContext->m_device); 146 : m_device (device) 148 , m_deviceInterface (interface, *m_device) 152 Unique<VkDevice> m_device; member in struct:vkt::api::__anon16478::BufferTestInstance::SparseContext
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
TensorDevice.h | 29 TensorDevice(const DeviceType& device, ExpressionType& expression) : m_device(device), m_expression(expression) {} 35 internal::TensorExecutor<const Assign, DeviceType>::run(assign, m_device); 46 internal::TensorExecutor<const Assign, DeviceType>::run(assign, m_device); 57 internal::TensorExecutor<const Assign, DeviceType>::run(assign, m_device); 62 const DeviceType& m_device; member in class:Eigen::TensorDevice
|
TensorCustomOp.h | 101 : m_op(op), m_device(device), m_result(NULL) 114 m_device.allocate(dimensions().TotalSize() * sizeof(Scalar))); 122 m_device.deallocate(m_result); 147 m_op.func().eval(m_op.expression(), result, m_device); 152 const Device& m_device; member in struct:Eigen::TensorEvaluator 257 : m_op(op), m_device(device), m_result(NULL) 269 m_result = static_cast<Scalar *>(m_device.allocate(dimensions().TotalSize() * sizeof(Scalar))); 277 m_device.deallocate(m_result); 301 m_op.func().eval(m_op.lhsExpression(), m_op.rhsExpression(), result, m_device); 306 const Device& m_device; member in struct:Eigen::TensorEvaluator [all...] |
TensorEvalTo.h | 112 : m_impl(op.expression(), device), m_device(device), 168 const Device& device() const{return m_device;} 172 const Device& m_device; member in struct:Eigen::TensorEvaluator
|
TensorEvaluator.h | 49 : m_data(const_cast<typename internal::traits<Derived>::template MakePointer<Scalar>::Type>(m.data())), m_dims(m.dimensions()), m_device(device), m_impl(m) 58 m_device.memcpy((void*)dest, m_data, sizeof(Scalar) * m_dims.TotalSize()); 114 const Device& device() const{return m_device;} 119 const Device& m_device; member in struct:Eigen::TensorEvaluator 172 : m_data(m.data()), m_dims(m.dimensions()), m_device(device), m_impl(m) 179 m_device.memcpy((void*)data, m_data, m_dims.TotalSize() * sizeof(Scalar)); 213 const Device& device() const{return m_device;} 218 const Device& m_device; member in struct:Eigen::TensorEvaluator
|
TensorForcedEval.h | 111 : m_impl(op.expression(), device), m_op(op.expression()), m_device(device), m_buffer(NULL) 118 m_buffer = (CoeffReturnType*)m_device.allocate(numValues * sizeof(CoeffReturnType)); 128 internal::TensorExecutor<const EvalTo, typename internal::remove_const<Device>::type, PacketAccess>::run(evalToTmp, m_device); 132 m_device.deallocate(m_buffer); 156 const Device& device() const{return m_device;} 160 const Device& m_device; member in struct:Eigen::TensorEvaluator
|
TensorChipping.h | 153 : m_impl(op.expression(), device), m_dim(op.dim()), m_device(device) 306 const Device& m_device; member in struct:Eigen::TensorEvaluator
|
TensorScan.h | 107 m_device(device), 156 return m_device; 168 m_output = static_cast<CoeffReturnType*>(m_device.allocate(total_size * sizeof(Scalar))); 194 m_device.deallocate(m_output); 202 const Device& m_device; member in struct:Eigen::TensorEvaluator
|
TensorFFT.h | 143 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorEvaluator(const XprType& op, const Device& device) : m_fft(op.fft()), m_impl(op.expression(), device), m_data(NULL), m_device(device) { 174 m_data = (CoeffReturnType*)m_device.allocate(sizeof(CoeffReturnType) * m_size); 182 m_device.deallocate(m_data); 209 ComplexScalar* buf = write_to_out ? (ComplexScalar*)data : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * m_size); 220 ComplexScalar* line_buf = (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * line_len); 225 ComplexScalar* a = is_power_of_two ? NULL : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * good_composite); 226 ComplexScalar* b = is_power_of_two ? NULL : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * good_composite); 227 ComplexScalar* pos_j_base_powered = is_power_of_two ? NULL : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * (line_len + 1)); 283 m_device.deallocate(line_buf); 285 m_device.deallocate(a) 570 const Device& m_device; member in struct:Eigen::TensorEvaluator [all...] |
TensorMorphing.h | 322 : m_impl(op.expression(), device), m_device(device), m_dimensions(op.sizes()), m_offsets(op.startIndices()) 386 const MemcpyTriggerForSlicing<Index, Device> trigger(m_device); 391 m_device.memcpy((void*)(data+i), src+offset, contiguous_values * sizeof(Scalar)); 523 const Device& m_device; member in struct:Eigen::TensorEvaluator 709 : m_impl(op.expression(), device), m_device(device), m_strides(op.strides()) 842 const Device& m_device; member in struct:Eigen::TensorEvaluator [all...] |
TensorReduction.h | 412 : m_impl(op.expression(), device), m_reducer(op.reducer()), m_result(NULL), m_device(device), m_xpr_dims(op.dims()) 488 ((RunningOnGPU && (m_device.majorDeviceVersion() >= 3)) || 492 m_result = static_cast<CoeffReturnType*>(m_device.allocate(sizeof(CoeffReturnType))); 497 internal::FullReducer<Self, Op, Device>::run(*this, reducer, m_device, data); 504 data = static_cast<CoeffReturnType*>(m_device.allocate(sizeof(CoeffReturnType) * num_coeffs_to_preserve)); 508 internal::InnerReducer<Self, Op, Device>::run(*this, reducer, m_device, data, num_values_to_reduce, num_coeffs_to_preserve); 513 else if (RunningOnGPU && (m_device.majorDeviceVersion() >= 3)) { 528 data = static_cast<CoeffReturnType*>(m_device.allocate(sizeof(CoeffReturnType) * num_coeffs_to_preserve)); 536 if (internal::InnerReducer<Self, Op, Device>::run(*this, reducer, m_device, data, num_values_to_reduce, num_coeffs_to_preserve)) { 538 m_device.deallocate(m_result) 775 const Device& m_device; member in struct:Eigen::TensorEvaluator [all...] |
/external/deqp/external/vulkancts/modules/vulkan/ |
vktTestCase.hpp | 83 const de::UniquePtr<DefaultDevice> m_device; member in class:vkt::Context
|
vktTestCase.cpp | 257 VkDevice getDevice (void) const { return *m_device; } 279 const Unique<VkDevice> m_device; member in class:vkt::DefaultDevice 292 , m_device (createDefaultDevice(m_instanceInterface, 298 , m_deviceInterface (m_instanceInterface, *m_device) 308 return getDeviceQueue(m_deviceInterface, *m_device, m_universalQueueFamilyIndex, 0); 334 , m_device (new DefaultDevice(m_platformInterface, testCtx.getCommandLine())) 335 , m_allocator (createAllocator(m_device.get())) 343 const vector<string>& Context::getInstanceExtensions (void) const { return m_device->getInstanceExtensions(); } 344 vk::VkInstance Context::getInstance (void) const { return m_device->getInstance(); } 345 const vk::InstanceInterface& Context::getInstanceInterface (void) const { return m_device->getInstanceInterface(); [all...] |
/external/rmi4utils/rmi4update/ |
rmi4update.h | 29 RMI4Update(RMIDevice & device, FirmwareImage & firmwareImage) : m_device(device), 47 RMIDevice & m_device; member in class:RMI4Update
|
/external/deqp/external/vulkancts/framework/vulkan/ |
vkMemUtil.cpp | 52 const VkDevice m_device; member in class:vk::__anon16474::HostPtr 59 , m_device (device) 67 m_vkd.unmapMemory(m_device, m_memory); 163 , m_device (device) 172 Move<VkDeviceMemory> mem = allocateMemory(m_vk, m_device, &allocInfo); 176 hostPtr = MovePtr<HostPtr>(new HostPtr(m_vk, m_device, *mem, 0u, allocInfo.allocationSize, 0u)); 192 Move<VkDeviceMemory> mem = allocateMemory(m_vk, m_device, &allocInfo); 198 hostPtr = MovePtr<HostPtr>(new HostPtr(m_vk, m_device, *mem, 0u, allocInfo.allocationSize, 0u));
|
vkMemUtil.hpp | 132 const VkDevice m_device; member in class:vk::SimpleAllocator
|
vkRef.hpp | 79 , m_device (device) 84 , m_device (DE_NULL) 92 VkDevice m_device; member in class:vk::refdetails::Deleter 188 , m_device (device) 193 , m_device (DE_NULL) 197 void operator() (VkDescriptorSet obj) const { m_deviceIface->freeDescriptorSets(m_device, m_pool, 1, &obj); } 201 VkDevice m_device; member in class:vk::refdetails::Deleter 211 , m_device (device) 216 , m_device (DE_NULL) 220 void operator() (VkCommandBuffer obj) const { m_deviceIface->freeCommandBuffers(m_device, m_pool, 1, &obj); 224 VkDevice m_device; member in class:vk::refdetails::Deleter [all...] |
vkNullDriver.cpp | 338 : m_device(device) 346 const VkDevice m_device; member in class:vk::__anon16475::CommandPool 359 CommandBuffer* const impl = new CommandBuffer(m_device, VkCommandPool(reinterpret_cast<deUintptr>(this)), level); 402 : m_device (device) 416 const VkDevice m_device; member in class:vk::__anon16475::DescriptorPool 424 DescriptorSet* const impl = new DescriptorSet(m_device, VkDescriptorPool(reinterpret_cast<deUintptr>(this)), setLayout); [all...] |
/external/deqp/external/vulkancts/modules/vulkan/compute/ |
vktComputeIndirectComputeDispatchTests.cpp | 143 const vk::VkDevice m_device; member in class:vkt::compute::__anon16541::IndirectDispatchInstanceBufferUpload 168 , m_device (context.getDevice()) 198 vk::flushMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), m_bufferSize); 223 m_device_interface, m_device, m_allocator, 241 vk::flushMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), resultBufferSize); 246 m_device_interface, m_device, m_context.getBinaryCollection().get("indirect_dispatch_" + m_name + "_verify"), 0u)); 251 vk::Unique<vk::VkDescriptorSetLayout> descriptorSetLayout(layoutBuilder.build(m_device_interface, m_device)); 254 const vk::Unique<vk::VkPipelineLayout> pipelineLayout(makePipelineLayout(m_device_interface, m_device, *descriptorSetLayout)); 255 const vk::Unique<vk::VkPipeline> computePipeline(makeComputePipeline(m_device_interface, m_device, *pipelineLayout, *verifyShader)); 261 .build(m_device_interface, m_device, vk::VK_DESCRIPTOR_POOL_CREATE_FREE_DESCRIPTOR_SET_BIT, static_cast<deUint32>(m_dispatchCommands.s (…) [all...] |
vktComputeShaderBuiltinVarTests.cpp | 133 const VkDevice m_device; member in class:vkt::compute::__anon16543::ComputeBuiltinVarInstance 370 , m_device (m_context.getDevice()) 413 Buffer uniformBuffer(m_vki, m_device, m_context.getDefaultAllocator(), makeBufferCreateInfo(sizeOfUniformBuffer, VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT), MemoryRequirement::HostVisible); 414 Buffer resultBuffer(m_vki, m_device, m_context.getDefaultAllocator(), makeBufferCreateInfo(resultBufferSize, VK_BUFFER_USAGE_STORAGE_BUFFER_BIT), MemoryRequirement::HostVisible); 419 flushMappedMemoryRange(m_vki, m_device, alloc.getMemory(), alloc.getOffset(), sizeOfUniformBuffer); 427 .build(m_vki, m_device)); 429 const Unique<VkShaderModule> shaderModule(createShaderModule(m_vki, m_device, m_context.getBinaryCollection().get(program_name.str()), 0u)); 430 const Unique<VkPipelineLayout> pipelineLayout(makePipelineLayout(m_vki, m_device, *descriptorSetLayout)); 431 const Unique<VkPipeline> pipeline(makeComputePipeline(m_vki, m_device, *pipelineLayout, *shaderModule)); 437 .build(m_vki, m_device, VK_DESCRIPTOR_POOL_CREATE_FREE_DESCRIPTOR_SET_BIT, 1u)) [all...] |
/external/deqp/external/vulkancts/modules/vulkan/synchronization/ |
vktSynchronizationOperation.hpp | 134 vk::VkDevice getDevice (void) const { return m_device; } 144 const vk::VkDevice m_device; member in class:vkt::synchronization::OperationContext
|
vktSynchronizationCrossInstanceSharingTests.cpp | 93 const vk::VkDevice m_device; member in class:vkt::synchronization::__anon16752::SimpleAllocation 101 , m_device (device) 107 m_vkd.freeMemory(m_device, getMemory(), DE_NULL); [all...] |