HomeSort by relevance Sort by last modified time
    Searched defs:m_device (Results 1 - 25 of 40) sorted by null

1 2

  /external/deqp/external/vulkancts/modules/vulkan/draw/
vktDrawBufferObjectUtil.hpp 65 vk::VkDevice m_device; member in class:vkt::Draw::Buffer
vktDrawImageObjectUtil.hpp 255 vk::VkDevice m_device; member in class:vkt::Draw::Image
  /external/deqp/external/vulkancts/modules/vulkan/api/
vktApiComputeInstanceResultBuffer.hpp 69 const vk::VkDevice m_device; member in class:vkt::api::ComputeInstanceResultBuffer
vktApiBufferTests.cpp 110 return *(m_sparseContext->m_device);
146 : m_device (device)
148 , m_deviceInterface (interface, *m_device)
152 Unique<VkDevice> m_device; member in struct:vkt::api::__anon16478::BufferTestInstance::SparseContext
  /external/eigen/unsupported/Eigen/CXX11/src/Tensor/
TensorDevice.h 29 TensorDevice(const DeviceType& device, ExpressionType& expression) : m_device(device), m_expression(expression) {}
35 internal::TensorExecutor<const Assign, DeviceType>::run(assign, m_device);
46 internal::TensorExecutor<const Assign, DeviceType>::run(assign, m_device);
57 internal::TensorExecutor<const Assign, DeviceType>::run(assign, m_device);
62 const DeviceType& m_device; member in class:Eigen::TensorDevice
TensorCustomOp.h 101 : m_op(op), m_device(device), m_result(NULL)
114 m_device.allocate(dimensions().TotalSize() * sizeof(Scalar)));
122 m_device.deallocate(m_result);
147 m_op.func().eval(m_op.expression(), result, m_device);
152 const Device& m_device; member in struct:Eigen::TensorEvaluator
257 : m_op(op), m_device(device), m_result(NULL)
269 m_result = static_cast<Scalar *>(m_device.allocate(dimensions().TotalSize() * sizeof(Scalar)));
277 m_device.deallocate(m_result);
301 m_op.func().eval(m_op.lhsExpression(), m_op.rhsExpression(), result, m_device);
306 const Device& m_device; member in struct:Eigen::TensorEvaluator
    [all...]
TensorEvalTo.h 112 : m_impl(op.expression(), device), m_device(device),
168 const Device& device() const{return m_device;}
172 const Device& m_device; member in struct:Eigen::TensorEvaluator
TensorEvaluator.h 49 : m_data(const_cast<typename internal::traits<Derived>::template MakePointer<Scalar>::Type>(m.data())), m_dims(m.dimensions()), m_device(device), m_impl(m)
58 m_device.memcpy((void*)dest, m_data, sizeof(Scalar) * m_dims.TotalSize());
114 const Device& device() const{return m_device;}
119 const Device& m_device; member in struct:Eigen::TensorEvaluator
172 : m_data(m.data()), m_dims(m.dimensions()), m_device(device), m_impl(m)
179 m_device.memcpy((void*)data, m_data, m_dims.TotalSize() * sizeof(Scalar));
213 const Device& device() const{return m_device;}
218 const Device& m_device; member in struct:Eigen::TensorEvaluator
TensorForcedEval.h 111 : m_impl(op.expression(), device), m_op(op.expression()), m_device(device), m_buffer(NULL)
118 m_buffer = (CoeffReturnType*)m_device.allocate(numValues * sizeof(CoeffReturnType));
128 internal::TensorExecutor<const EvalTo, typename internal::remove_const<Device>::type, PacketAccess>::run(evalToTmp, m_device);
132 m_device.deallocate(m_buffer);
156 const Device& device() const{return m_device;}
160 const Device& m_device; member in struct:Eigen::TensorEvaluator
TensorChipping.h 153 : m_impl(op.expression(), device), m_dim(op.dim()), m_device(device)
306 const Device& m_device; member in struct:Eigen::TensorEvaluator
TensorScan.h 107 m_device(device),
156 return m_device;
168 m_output = static_cast<CoeffReturnType*>(m_device.allocate(total_size * sizeof(Scalar)));
194 m_device.deallocate(m_output);
202 const Device& m_device; member in struct:Eigen::TensorEvaluator
TensorFFT.h 143 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorEvaluator(const XprType& op, const Device& device) : m_fft(op.fft()), m_impl(op.expression(), device), m_data(NULL), m_device(device) {
174 m_data = (CoeffReturnType*)m_device.allocate(sizeof(CoeffReturnType) * m_size);
182 m_device.deallocate(m_data);
209 ComplexScalar* buf = write_to_out ? (ComplexScalar*)data : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * m_size);
220 ComplexScalar* line_buf = (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * line_len);
225 ComplexScalar* a = is_power_of_two ? NULL : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * good_composite);
226 ComplexScalar* b = is_power_of_two ? NULL : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * good_composite);
227 ComplexScalar* pos_j_base_powered = is_power_of_two ? NULL : (ComplexScalar*)m_device.allocate(sizeof(ComplexScalar) * (line_len + 1));
283 m_device.deallocate(line_buf);
285 m_device.deallocate(a)
570 const Device& m_device; member in struct:Eigen::TensorEvaluator
    [all...]
TensorMorphing.h 322 : m_impl(op.expression(), device), m_device(device), m_dimensions(op.sizes()), m_offsets(op.startIndices())
386 const MemcpyTriggerForSlicing<Index, Device> trigger(m_device);
391 m_device.memcpy((void*)(data+i), src+offset, contiguous_values * sizeof(Scalar));
523 const Device& m_device; member in struct:Eigen::TensorEvaluator
709 : m_impl(op.expression(), device), m_device(device), m_strides(op.strides())
842 const Device& m_device; member in struct:Eigen::TensorEvaluator
    [all...]
TensorReduction.h 412 : m_impl(op.expression(), device), m_reducer(op.reducer()), m_result(NULL), m_device(device), m_xpr_dims(op.dims())
488 ((RunningOnGPU && (m_device.majorDeviceVersion() >= 3)) ||
492 m_result = static_cast<CoeffReturnType*>(m_device.allocate(sizeof(CoeffReturnType)));
497 internal::FullReducer<Self, Op, Device>::run(*this, reducer, m_device, data);
504 data = static_cast<CoeffReturnType*>(m_device.allocate(sizeof(CoeffReturnType) * num_coeffs_to_preserve));
508 internal::InnerReducer<Self, Op, Device>::run(*this, reducer, m_device, data, num_values_to_reduce, num_coeffs_to_preserve);
513 else if (RunningOnGPU && (m_device.majorDeviceVersion() >= 3)) {
528 data = static_cast<CoeffReturnType*>(m_device.allocate(sizeof(CoeffReturnType) * num_coeffs_to_preserve));
536 if (internal::InnerReducer<Self, Op, Device>::run(*this, reducer, m_device, data, num_values_to_reduce, num_coeffs_to_preserve)) {
538 m_device.deallocate(m_result)
775 const Device& m_device; member in struct:Eigen::TensorEvaluator
    [all...]
  /external/deqp/external/vulkancts/modules/vulkan/
vktTestCase.hpp 83 const de::UniquePtr<DefaultDevice> m_device; member in class:vkt::Context
vktTestCase.cpp 257 VkDevice getDevice (void) const { return *m_device; }
279 const Unique<VkDevice> m_device; member in class:vkt::DefaultDevice
292 , m_device (createDefaultDevice(m_instanceInterface,
298 , m_deviceInterface (m_instanceInterface, *m_device)
308 return getDeviceQueue(m_deviceInterface, *m_device, m_universalQueueFamilyIndex, 0);
334 , m_device (new DefaultDevice(m_platformInterface, testCtx.getCommandLine()))
335 , m_allocator (createAllocator(m_device.get()))
343 const vector<string>& Context::getInstanceExtensions (void) const { return m_device->getInstanceExtensions(); }
344 vk::VkInstance Context::getInstance (void) const { return m_device->getInstance(); }
345 const vk::InstanceInterface& Context::getInstanceInterface (void) const { return m_device->getInstanceInterface();
    [all...]
  /external/rmi4utils/rmi4update/
rmi4update.h 29 RMI4Update(RMIDevice & device, FirmwareImage & firmwareImage) : m_device(device),
47 RMIDevice & m_device; member in class:RMI4Update
  /external/deqp/external/vulkancts/framework/vulkan/
vkMemUtil.cpp 52 const VkDevice m_device; member in class:vk::__anon16474::HostPtr
59 , m_device (device)
67 m_vkd.unmapMemory(m_device, m_memory);
163 , m_device (device)
172 Move<VkDeviceMemory> mem = allocateMemory(m_vk, m_device, &allocInfo);
176 hostPtr = MovePtr<HostPtr>(new HostPtr(m_vk, m_device, *mem, 0u, allocInfo.allocationSize, 0u));
192 Move<VkDeviceMemory> mem = allocateMemory(m_vk, m_device, &allocInfo);
198 hostPtr = MovePtr<HostPtr>(new HostPtr(m_vk, m_device, *mem, 0u, allocInfo.allocationSize, 0u));
vkMemUtil.hpp 132 const VkDevice m_device; member in class:vk::SimpleAllocator
vkRef.hpp 79 , m_device (device)
84 , m_device (DE_NULL)
92 VkDevice m_device; member in class:vk::refdetails::Deleter
188 , m_device (device)
193 , m_device (DE_NULL)
197 void operator() (VkDescriptorSet obj) const { m_deviceIface->freeDescriptorSets(m_device, m_pool, 1, &obj); }
201 VkDevice m_device; member in class:vk::refdetails::Deleter
211 , m_device (device)
216 , m_device (DE_NULL)
220 void operator() (VkCommandBuffer obj) const { m_deviceIface->freeCommandBuffers(m_device, m_pool, 1, &obj);
224 VkDevice m_device; member in class:vk::refdetails::Deleter
    [all...]
vkNullDriver.cpp 338 : m_device(device)
346 const VkDevice m_device; member in class:vk::__anon16475::CommandPool
359 CommandBuffer* const impl = new CommandBuffer(m_device, VkCommandPool(reinterpret_cast<deUintptr>(this)), level);
402 : m_device (device)
416 const VkDevice m_device; member in class:vk::__anon16475::DescriptorPool
424 DescriptorSet* const impl = new DescriptorSet(m_device, VkDescriptorPool(reinterpret_cast<deUintptr>(this)), setLayout);
    [all...]
  /external/deqp/external/vulkancts/modules/vulkan/compute/
vktComputeIndirectComputeDispatchTests.cpp 143 const vk::VkDevice m_device; member in class:vkt::compute::__anon16541::IndirectDispatchInstanceBufferUpload
168 , m_device (context.getDevice())
198 vk::flushMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), m_bufferSize);
223 m_device_interface, m_device, m_allocator,
241 vk::flushMappedMemoryRange(m_device_interface, m_device, alloc.getMemory(), alloc.getOffset(), resultBufferSize);
246 m_device_interface, m_device, m_context.getBinaryCollection().get("indirect_dispatch_" + m_name + "_verify"), 0u));
251 vk::Unique<vk::VkDescriptorSetLayout> descriptorSetLayout(layoutBuilder.build(m_device_interface, m_device));
254 const vk::Unique<vk::VkPipelineLayout> pipelineLayout(makePipelineLayout(m_device_interface, m_device, *descriptorSetLayout));
255 const vk::Unique<vk::VkPipeline> computePipeline(makeComputePipeline(m_device_interface, m_device, *pipelineLayout, *verifyShader));
261 .build(m_device_interface, m_device, vk::VK_DESCRIPTOR_POOL_CREATE_FREE_DESCRIPTOR_SET_BIT, static_cast<deUint32>(m_dispatchCommands.s (…)
    [all...]
vktComputeShaderBuiltinVarTests.cpp 133 const VkDevice m_device; member in class:vkt::compute::__anon16543::ComputeBuiltinVarInstance
370 , m_device (m_context.getDevice())
413 Buffer uniformBuffer(m_vki, m_device, m_context.getDefaultAllocator(), makeBufferCreateInfo(sizeOfUniformBuffer, VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT), MemoryRequirement::HostVisible);
414 Buffer resultBuffer(m_vki, m_device, m_context.getDefaultAllocator(), makeBufferCreateInfo(resultBufferSize, VK_BUFFER_USAGE_STORAGE_BUFFER_BIT), MemoryRequirement::HostVisible);
419 flushMappedMemoryRange(m_vki, m_device, alloc.getMemory(), alloc.getOffset(), sizeOfUniformBuffer);
427 .build(m_vki, m_device));
429 const Unique<VkShaderModule> shaderModule(createShaderModule(m_vki, m_device, m_context.getBinaryCollection().get(program_name.str()), 0u));
430 const Unique<VkPipelineLayout> pipelineLayout(makePipelineLayout(m_vki, m_device, *descriptorSetLayout));
431 const Unique<VkPipeline> pipeline(makeComputePipeline(m_vki, m_device, *pipelineLayout, *shaderModule));
437 .build(m_vki, m_device, VK_DESCRIPTOR_POOL_CREATE_FREE_DESCRIPTOR_SET_BIT, 1u))
    [all...]
  /external/deqp/external/vulkancts/modules/vulkan/synchronization/
vktSynchronizationOperation.hpp 134 vk::VkDevice getDevice (void) const { return m_device; }
144 const vk::VkDevice m_device; member in class:vkt::synchronization::OperationContext
vktSynchronizationCrossInstanceSharingTests.cpp 93 const vk::VkDevice m_device; member in class:vkt::synchronization::__anon16752::SimpleAllocation
101 , m_device (device)
107 m_vkd.freeMemory(m_device, getMemory(), DE_NULL);
    [all...]

Completed in 319 milliseconds

1 2