/external/chromium_org/third_party/mesa/src/src/gallium/state_trackers/d3d1x/d3d1xshader/src/ |
dxbc_dump.cpp | 36 struct dxbc_chunk_header* chunk = container.chunks[i]; local 38 memcpy(fourcc_str, &chunk->fourcc, 4); 40 out << "# DXBC chunk " << std::setw(2) << i << ": " << fourcc_str << " offset " << ((char*)chunk - (char*)container.data) << " size " << bswap_le32(chunk->size) << "\n";
|
/external/mesa3d/src/gallium/state_trackers/d3d1x/d3d1xshader/src/ |
dxbc_dump.cpp | 36 struct dxbc_chunk_header* chunk = container.chunks[i]; local 38 memcpy(fourcc_str, &chunk->fourcc, 4); 40 out << "# DXBC chunk " << std::setw(2) << i << ": " << fourcc_str << " offset " << ((char*)chunk - (char*)container.data) << " size " << bswap_le32(chunk->size) << "\n";
|
/external/chromium_org/content/browser/speech/ |
chunked_byte_buffer_unittest.cc | 19 0x00, 0x00, 0x00, 0x04, 0x01, 0x02, 0x03, 0x04, // Chunk 1: 4 bytes 20 0x00, 0x00, 0x00, 0x02, 0x05, 0x06, // Chunk 2: 2 bytes 21 0x00, 0x00, 0x00, 0x01, 0x07 // Chunk 3: 1 bytes 27 // Append partially chunk 1. 32 // Complete chunk 1. 37 // Append fully chunk 2. 42 // Remove and check chunk 1. 43 scoped_ptr<ByteVector> chunk; local 44 chunk = buffer.PopChunk(); 45 EXPECT_TRUE(chunk != NULL) [all...] |
audio_buffer.cc | 61 scoped_refptr<AudioChunk> chunk(chunks_.front()); 63 return chunk; 67 scoped_refptr<AudioChunk> chunk(new AudioChunk(bytes_per_sample_)); 75 chunk->data_string_.reserve(resulting_length); 77 chunk->data_string_.append((*it)->data_string_); 80 return chunk;
|
/external/chromium_org/gpu/command_buffer/client/ |
mapped_memory.cc | 40 MemoryChunk* chunk = *iter; local 41 cmd_buf->DestroyTransferBuffer(chunk->shm_id()); 53 MemoryChunk* chunk = chunks_[ii]; local 54 chunk->FreeUnused(); 55 total_bytes_in_use += chunk->bytes_in_use(); 56 if (chunk->GetLargestFreeSizeWithoutWaiting() >= size) { 57 void* mem = chunk->Alloc(size); 59 *shm_id = chunk->shm_id(); 60 *shm_offset = chunk->GetOffset(mem); 72 MemoryChunk* chunk = chunks_[ii] local 107 MemoryChunk* chunk = chunks_[ii]; local 118 MemoryChunk* chunk = chunks_[ii]; local 131 MemoryChunk* chunk = *iter; local [all...] |
/external/openssl/crypto/md4/ |
md4_one.c | 81 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD4_Update(&c,temp,chunk); 88 n -= chunk; 89 d += chunk;
|
/external/openssl/crypto/md5/ |
md5_one.c | 81 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD5_Update(&c,temp,chunk); 88 n -= chunk; 89 d += chunk;
|
/external/jemalloc/src/ |
arena.c | 56 static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, 58 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, 168 * Break ties by chunk address. For fragmented chunks, report lower 191 arena_avail_adjac_pred(arena_chunk_t *chunk, size_t pageind) 198 ret = (arena_mapbits_allocated_get(chunk, pageind-1) == 0); 199 assert(ret == false || arena_mapbits_dirty_get(chunk, 200 pageind-1) != arena_mapbits_dirty_get(chunk, pageind)); 206 arena_avail_adjac_succ(arena_chunk_t *chunk, size_t pageind, size_t npages) 214 ret = (arena_mapbits_allocated_get(chunk, pageind+npages) == 0); 215 assert(ret == false || arena_mapbits_dirty_get(chunk, pageind 320 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 430 arena_chunk_t *chunk; local 497 arena_chunk_t *chunk; local 542 arena_chunk_t *chunk; local 565 arena_chunk_t *chunk; local 623 arena_chunk_t *chunk; local 683 arena_chunk_t *chunk; local 784 arena_chunk_t *chunk; local 838 arena_chunk_t *chunk; local 1101 arena_chunk_t *chunk; local 1259 arena_chunk_t *chunk; local 1402 arena_chunk_t *chunk; local 1420 arena_chunk_t *chunk = CHUNK_ADDR2BASE(run); local 1432 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 1528 arena_chunk_t *chunk; local 1800 arena_chunk_t *chunk; local 1852 arena_chunk_t *chunk; local 2209 arena_chunk_t *chunk; local [all...] |
chunk.c | 26 /* Various chunk-related settings. */ 39 static void chunk_dalloc_core(void *chunk, size_t size); 56 * the current chunk allocation request is on behalf of the 87 /* Insert the leading space as a smaller chunk. */ 94 /* Insert the trailing space as a smaller chunk. */ 98 * base_node_alloc() can cause a new base chunk to be 182 chunk_register(void *chunk, size_t size, bool base) 185 assert(chunk != NULL); 186 assert(CHUNK_ADDR2BASE(chunk) == chunk); [all...] |
/external/apache-xml/src/main/java/org/apache/xml/dtm/ref/ |
ChunkedIntArray.java | 94 int[] chunk = chunks.elementAt(chunkpos); local 95 chunk[slotpos] = w0; 96 chunk[slotpos+1] = w1; 97 chunk[slotpos+2] = w2; 98 chunk[slotpos+3] = w3; 126 int[] chunk = chunks.elementAt(chunkpos); local 127 return chunk[slotpos + offset]; 148 int[] chunk = chunks.elementAt(chunkpos); local 153 ancestor = chunk[slotpos + 1]; 175 CIA; when only a single characters() chunk has been recieved, its inde 207 int[] chunk = chunks.elementAt(chunkpos); local 230 int[] chunk = chunks.elementAt(chunkpos); local 263 int[] chunk = chunks.elementAt(chunkpos); local [all...] |
/external/guava/guava-tests/test/com/google/common/io/ |
LineBufferTest.java | 73 for (int chunk : CHUNK_SIZES) { 74 chunk = Math.max(1, Math.min(chunk, input.length())); 75 assertEquals(expectProcess, bufferHelper(input, chunk)); 76 assertEquals(expectRead, readUsingJava(input, chunk)); 77 assertEquals(expectRead, readUsingReader(input, chunk, true)); 78 assertEquals(expectRead, readUsingReader(input, chunk, false)); 82 private static List<String> bufferHelper(String input, int chunk) 93 int len = Math.min(chars.length, off + chunk) - off; 101 private static List<String> readUsingJava(String input, int chunk) [all...] |
/external/libpng/contrib/libtests/ |
fakepng.c | 15 put_chunk(const unsigned char *chunk, uInt length) 21 fwrite(chunk, length, 1, stdout); 24 put_uLong(crc32(crc, chunk, length));
|
/external/jemalloc/include/jemalloc/internal/ |
arena.h | 57 /* Each element of the chunk map corresponds to one page within the chunk. */ 63 * chunk map overhead accounts for a percentage of memory, rather than 162 /* Arena chunk header. */ 164 /* Arena that owns the chunk. */ 186 * Map of pages within chunk that keeps track of free/large/small. The 187 * first map_bias entries are omitted, since the chunk header does not 189 * for common chunk sizes (e.g. 4 MiB). 318 * 3) Chunk- and run-related operations are protected by this mutex. 337 * In order to avoid rapid chunk allocation/deallocation when an aren 1027 arena_chunk_t *chunk; local 1049 arena_chunk_t *chunk; local 1099 arena_chunk_t *chunk; local [all...] |
chunk_mmap.h | 15 bool chunk_dalloc_mmap(void *chunk, size_t size);
|
/external/nanohttpd/core/src/test/java/fi/iki/elonen/ |
HttpChunkedResponseTest.java | 43 int chunk = 0; field in class:HttpChunkedResponseTest.ChunkedInputStream 53 for (int i = 0; i < chunks[chunk].length(); ++i) { 54 buffer[i] = (byte) chunks[chunk].charAt(i); 56 return chunks[chunk++].length();
|
/external/bison/lib/ |
obstack.c | 133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 147 register struct _obstack_chunk *chunk; /* points to new chunk */ 174 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 175 if (!chunk) 177 h->next_free = h->object_base = __PTR_ALIGN ((char *) chunk, chunk->contents, 179 h->chunk_limit = chunk->limit 180 = (char *) chunk + h->chunk_size 146 register struct _obstack_chunk *chunk; \/* points to new chunk *\/ local 193 register struct _obstack_chunk *chunk; \/* points to new chunk *\/ local [all...] |
/external/elfutils/0.153/libelf/ |
elf_getdata_rawchunk.c | 1 /* Return converted data from raw chunk of ELF file. 170 Elf_Data_Chunk *chunk = calloc (1, sizeof *chunk); local 171 if (chunk == NULL) 178 chunk->dummy_scn.elf = elf; 179 chunk->dummy_scn.flags = flags; 180 chunk->data.s = &chunk->dummy_scn; 181 chunk->data.d.d_buf = buffer; 182 chunk->data.d.d_size = size [all...] |
/external/mockito/src/org/mockito/internal/verification/checkers/ |
NumberOfInvocationsInOrderChecker.java | 35 List<Invocation> chunk = finder.findMatchingChunk(invocations, wanted, wantedCount, context); local 37 int actualCount = chunk.size(); 40 Location lastInvocation = finder.getLastLocation(chunk); 43 Location firstUndesired = chunk.get(wantedCount).getLocation(); 47 invocationMarker.markVerifiedInOrder(chunk, wanted, context);
|
/external/mockito/src/org/mockito/internal/invocation/ |
InvocationMarker.java | 25 public void markVerifiedInOrder(List<Invocation> chunk, CapturesArgumensFromInvocation wanted, InOrderContext context) {
26 markVerified(chunk, wanted);
28 for (Invocation i : chunk) {
|
/external/chromium_org/chrome/browser/safe_browsing/ |
chunk_range.cc | 60 for (int chunk = range.start(); chunk <= range.stop(); ++chunk) { 61 chunks->push_back(chunk); 70 // Crack the string into chunk parts, then crack each part looking for a 83 // atoi error, since chunk numbers are guaranteed to never be 0. 104 const ChunkRange& chunk = ranges[mid]; local 105 if ((chunk.stop() >= chunk_number) && (chunk.start() <= chunk_number)) 109 if (chunk.stop() < chunk_number [all...] |
/external/chromium_org/v8/tools/testrunner/server/ |
compression.py | 74 chunk = self.sock.recv(8192) 77 if not chunk: return None 78 self._AppendData(chunk) 83 chunk = self.sock.recv(8192) 86 if not chunk: return None 87 self._AppendData(chunk) 104 chunk = self.data.read(length) 111 return chunk
|
/external/jemalloc/android/test/ |
run_integration.sh | 9 $test_dir/chunk
|
/external/chromium_org/tools/grit/grit/gather/ |
skeleton_gatherer.py | 34 # True if a translatable chunk has been added 104 def _AddNontranslateableChunk(self, chunk): 105 '''Adds a nontranslateable chunk.''' 107 ph = tclib.Placeholder('XX%02dXX' % self.ph_counter_, chunk, chunk) 111 self.skeleton_.append(chunk) 113 def _AddTranslateableChunk(self, chunk): 114 '''Adds a translateable chunk. It will be unescaped before being added.''' 117 if chunk == '': 120 unescaped_text = self.UnEscape(chunk) [all...] |
/external/compiler-rt/lib/lsan/ |
lsan_allocator.cc | 139 uptr chunk = reinterpret_cast<uptr>(allocator.GetBlockBeginFastLocked(p)); local 140 if (!chunk) return 0; 141 // LargeMmapAllocator considers pointers to the meta-region of a chunk to be 143 if (addr < chunk) return 0; 144 ChunkMetadata *m = Metadata(reinterpret_cast<void *>(chunk)); 148 if (addr < chunk + m->requested_size) 149 return chunk; 150 if (IsSpecialCaseOfOperatorNew0(chunk, m->requested_size, addr)) 151 return chunk; 155 uptr GetUserBegin(uptr chunk) { 189 void *chunk = allocator.GetBlockBegin(p); local [all...] |
/external/fonttools/Lib/fontTools/ |
t1Lib.py | 169 raise T1Error('bad chunk code: ' + repr(code)) 186 chunk = f.read(chunklen) 187 assert len(chunk) == chunklen 188 data.append(chunk) 192 raise T1Error('bad chunk code: ' + repr(code)) 209 for isEncrypted, chunk in chunks: 210 if isEncrypted and isHex(chunk[:4]): 211 data.append(deHexString(chunk)) 213 data.append(chunk) 225 for isEncrypted, chunk in chunks [all...] |