/external/deqp/external/vulkancts/modules/vulkan/tessellation/ |
vktTessellationWindingTests.cpp | 412 const VkQueue queue = m_context.getUniversalQueue(); local 563 submitCommandsAndWait(vk, device, queue, *cmdBuffer);
|
/external/emma/tools/java/com/vladium/tools/ |
ClassDep.java | 125 final LinkedList /* class VM name:String */ queue = new LinkedList (); local 128 queue.add (Descriptors.javaNameToVMName (m_rootSet [i])); 134 while (! queue.isEmpty ()) 136 final String classVMName = (String) queue.removeFirst (); 173 queue.addLast (classDepVMName);
|
/external/fio/ |
ioengines.h | 12 * io_ops->queue() return values 28 int (*queue)(struct thread_data *, struct io_u *); member in struct:ioengine_ops 50 FIO_SYNCIO = 1 << 0, /* io engine has synchronous ->queue */
|
/external/guava/guava/src/com/google/common/base/ |
FinalizableReferenceQueue.java | 36 * A reference queue with an associated background thread that dequeues references and invokes 111 * thread keeps an indirect strong reference to the queue in ReferenceMap, which keeps the 144 * The actual reference queue that our background thread will poll. 146 final ReferenceQueue<Object> queue; field in class:FinalizableReferenceQueue 156 * Constructs a new queue. 160 queue = new ReferenceQueue<Object>(); 161 frqRef = new PhantomReference<Object>(this, queue); 164 startFinalizer.invoke(null, FinalizableReference.class, queue, frqRef); 183 * Repeatedly dequeues references from the queue and invokes {@link 184 * FinalizableReference#finalizeReferent()} on them until the queue is empty. This method is [all...] |
/external/iproute2/ip/ |
ipntable.c | 54 " [ delay_probe MSEC ] [ queue LEN ]\n" 197 } else if (strcmp(*argv, "queue") == 0) { 198 __u32 queue; local 202 if (get_u32(&queue, *argv, 0)) 203 invarg("\"queue\" value is invalid", *argv); 206 NDTPA_QUEUE_LEN, queue); 511 __u32 queue = rta_getattr_u32(tpb[NDTPA_QUEUE_LEN]); local 513 fprintf(fp, "queue %u ", queue);
|
/external/javassist/src/main/javassist/util/ |
HotSwapper.java | 243 EventQueue queue = jvm.eventQueue(); local 244 return queue.remove();
|
/external/robolectric-shadows/robolectric/src/test/java/org/robolectric/shadows/ |
ShadowMessageQueueTest.java | 27 private MessageQueue queue; field in class:ShadowMessageQueueTest 60 queue = looper.getQueue(); 61 shadowQueue = shadowOf(queue); 75 return callInstanceMethod(queue, "enqueueMessage", 82 callInstanceMethod(queue, "removeMessages", 108 setField(queue, quitField, true); 114 setField(queue, quitField, true);
|
/external/skia/src/gpu/vk/ |
GrVkGpu.h | 52 VkQueue queue() const { return fQueue; } function in class:GrVkGpu 204 // Ends and submits the current command buffer to the queue and then creates a new command 206 // work in the queue to finish before returning. If this GrVkGpu object has any semaphores in 248 VkQueue fQueue; // Must be Graphics queue
|
/external/skia/tests/ |
SurfaceSemaphoreTest.cpp | 253 VkQueue queue = gpu->queue(); local 302 GR_VK_CALL_ERRCHECK(interface, QueueSubmit(queue, 1, &submitInfo, fence));
|
/external/skqp/src/gpu/vk/ |
GrVkGpu.h | 52 VkQueue queue() const { return fQueue; } function in class:GrVkGpu 219 // Ends and submits the current command buffer to the queue and then creates a new command 221 // work in the queue to finish before returning. If this GrVkGpu object has any semaphores in 267 VkQueue fQueue; // Must be Graphics queue
|
/external/skqp/tests/ |
SurfaceSemaphoreTest.cpp | 253 VkQueue queue = gpu->queue(); local 302 GR_VK_CALL_ERRCHECK(interface, QueueSubmit(queue, 1, &submitInfo, fence));
|
/external/squashfs-tools/squashfs-tools/ |
caches-queues-lists.h | 131 struct queue { struct 175 extern struct queue *queue_init(int); 176 extern void queue_put(struct queue *, void *); 177 extern void *queue_get(struct queue *); 178 extern int queue_empty(struct queue *); 179 extern void queue_flush(struct queue *); 180 extern void dump_queue(struct queue *);
|
/external/syslinux/core/lwip/src/core/ |
tcp_out.c | 318 LWIP_DEBUGF(TCP_OUTPUT_DEBUG | 3, ("tcp_write: too long queue %"U16_F" (max %"U16_F")\n", 325 LWIP_ASSERT("tcp_write: pbufs on queue => at least one queue non-empty", 328 LWIP_ASSERT("tcp_write: no pbufs on queue => both queues empty", 354 struct tcp_seg *last_unsent = NULL, *seg = NULL, *prev_seg = NULL, *queue = NULL; local 416 /* Find the tail of the unsent queue. */ 514 * The new segments are chained together in the local 'queue' 530 if ((p = tcp_pbuf_prealloc(PBUF_TRANSPORT, seglen + optlen, pcb->mss, &oversize, pcb, apiflags, queue == NULL)) == NULL) { 573 * length of the queue exceeds the configured maximum or 576 LWIP_DEBUGF(TCP_OUTPUT_DEBUG | 2, ("tcp_write: queue too long %"U16_F" (%"U16_F")\n", queuelen, TCP_SND_QUEUELEN)) [all...] |
/external/syslinux/gpxe/src/include/gpxe/ |
virtio-ring.h | 70 virtio_queue_t queue; member in struct:vring_virtqueue 85 unsigned int num, unsigned char *queue) 94 pa = virt_to_phys(queue);
|
/external/tensorflow/tensorflow/cc/framework/ |
gradients.cc | 169 std::deque<Node*> queue; local 173 queue.push_back(out.node()); 178 while (!queue.empty()) { 179 Node* n = queue.front(); 180 queue.pop_front(); 184 queue.push_back(e->src()); 199 // Initialize `queue` for BFS traversal. Nodes in `queue` hold upcoming nodes 202 std::deque<std::pair<Node*, Node*>> queue; local 205 queue.push_back(std::make_pair(nout.node(), static_cast<Node*>(nullptr))) 287 std::deque<Node*> queue; local [all...] |
/external/tensorflow/tensorflow/core/distributed_runtime/ |
scheduler.cc | 18 #include <queue> 83 std::deque<Node*> queue; local 85 queue.push_back(srcNode); 88 while (!queue.empty()) { 89 const Node* curr = queue.front(); 90 queue.pop_front(); 112 queue.push_back(out); 140 std::deque<Node*> queue; local 142 queue.push_back(sinkNode); 145 while (!queue.empty()) [all...] |
/external/tensorflow/tensorflow/core/graph/ |
gradients.cc | 225 std::deque<Node*> queue; local 227 queue.push_back(nout.node); 234 while (!queue.empty()) { 235 Node* n = queue.front(); 236 queue.pop_front(); 245 queue.push_back(e->dst());
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
shared_batch_scheduler_test.cc | 161 std::unique_ptr<BatchScheduler<FakeTask>> queue; local 162 TF_ASSERT_OK(scheduler->AddQueue(queue_options, callback, &queue)); 165 TF_ASSERT_OK(ScheduleTask(3, queue.get())); 166 TF_ASSERT_OK(ScheduleTask(5, queue.get())); 169 TF_ASSERT_OK(ScheduleTask(3 /* (3+5) + 3 > 10 */, queue.get())); 170 TF_ASSERT_OK(ScheduleTask(1, queue.get())); 171 TF_ASSERT_OK(ScheduleTask(6, queue.get())); 223 std::unique_ptr<BatchScheduler<FakeTask>> queue; local 224 TF_ASSERT_OK(scheduler->AddQueue(queue_options, callback, &queue)); 228 TF_ASSERT_OK(ScheduleTask(1, queue.get())) 279 std::unique_ptr<BatchScheduler<FakeTask>> queue; local 325 std::unique_ptr<BatchScheduler<FakeTask>> queue; local 429 std::unique_ptr<BatchScheduler<FakeTask>> queue; local 556 std::unique_ptr<BatchScheduler<FakeTask>> queue; local [all...] |
/external/testng/src/main/java/org/testng/internal/ |
Graph.java | 221 // already encountered. "queue" is the queue of items whose 226 LinkedList<T> queue = new LinkedList<>(); local 228 queue.addLast(o); 230 while (! queue.isEmpty()) { 231 for (T obj : getPredecessors(queue.removeFirst())) { 234 queue.addLast(obj);
|
/external/vulkan-validation-layers/layers/ |
core_validation.h | 112 VkQueue queue; member in class:QUEUE_STATE
|
object_tracker.h | 79 // Track Queue information 82 VkQueue queue; member in struct:object_tracker::ObjTrackQueueInfo 111 // Map of queue information structures, one per queue 141 void AddQueueInfo(VkDevice device, uint32_t queue_node_index, VkQueue queue); 142 void ValidateQueueFlags(VkQueue queue, const char *function);
|
/frameworks/av/media/libstagefright/rtsp/ |
AMPEG4ElementaryAssembler.cpp | 208 List<sp<ABuffer> > *queue = source->queue(); local 210 if (queue->empty()) { 215 List<sp<ABuffer> >::iterator it = queue->begin(); 216 while (it != queue->end()) { 221 it = queue->erase(it); 224 if (queue->empty()) { 229 sp<ABuffer> buffer = *queue->begin(); 368 queue->erase(queue->begin()) [all...] |
/frameworks/base/core/java/android/app/ |
QueuedWork.java | 36 * New work will be {@link #queue queued}. 79 /** Work queued via {@link #queue} */ 113 * Add a finisher-runnable to wait for {@link #queue asynchronously processed work}. 210 * Queue a work-runnable for processing asynchronously. 215 public static void queue(Runnable work, boolean shouldDelay) { method in class:QueuedWork 230 * @return True iff there is any {@link #queue async work queued}.
|
/frameworks/base/core/java/android/content/ |
BroadcastReceiver.java | 217 QueuedWork.queue(new Runnable() {
|
/frameworks/base/core/java/android/hardware/usb/ |
UsbRequest.java | 36 * or asynchronously via {@link #queue} and {@link UsbDeviceConnection#requestWait}. 57 /** The amount of data to read / write when using {@link #queue} */ 67 * Whether this buffer was {@link #queue(ByteBuffer) queued using the new behavior} or 68 * {@link #queue(ByteBuffer, int) queued using the deprecated behavior}. 78 * Lock for queue, enqueue and dequeue, so a queue operation can be finished by a dequeue 145 * maintaining state between calls to {@link #queue} and 158 * maintaining state between calls to {@link #queue} and 185 * @deprecated Use {@link #queue(ByteBuffer)} instead. 188 public boolean queue(ByteBuffer buffer, int length) method in class:UsbRequest 240 public boolean queue(@Nullable ByteBuffer buffer) { method in class:UsbRequest [all...] |