shift 2880 MMgc/GC.cpp int shift = (itemNum&0x7)<<2; shift 2883 MMgc/GC.cpp if((bits2 & ((GCAlloc::kMark|GCAlloc::kQueued)<<shift)) == 0) shift 2892 MMgc/GC.cpp *pbits = bits2 | (GCAlloc::kQueued << shift); shift 2905 MMgc/GC.cpp *pbits = bits2 | (GCAlloc::kMark << shift); shift 219 MMgc/GCAlloc-inlines.h int index = (int)((((char*) item - block->items) * block->alloc->multiple) >> block->alloc->shift); shift 91 MMgc/GCAlloc.cpp ComputeMultiplyShift((uint16_t)m_itemSize, multiple, shift); shift 234 MMgc/GCAlloc.h uint16_t shift; shift 105 core/ArrayClass.cpp ar->shift(); AvmAssert(s->RefCount()==1); shift 336 core/ArrayClass.cpp return (a->m_denseArr.shift()); shift 90 core/ArrayClass.h inline Atom shift(Atom thisAtom) { return ArrayClass::generic_shift(toplevel(), thisAtom); } shift 64 core/AtomArray.h Atom shift(); shift 1495 core/WordcodeEmitter.cpp for ( uintptr_t i=0, limit=nextI-shift ; i < limit ; i++ ) { shift 1496 core/WordcodeEmitter.cpp I[i] = I[i+shift]; shift 1497 core/WordcodeEmitter.cpp O[i] = O[i+shift]; shift 1499 core/WordcodeEmitter.cpp nextI -= shift; shift 225 core/WordcodeEmitter.h void shiftBuffers(uint32_t shift); shift 140 nanojit/NativeARM.cpp uint32_t shift = leading_zeroes + try_shift; shift 141 nanojit/NativeARM.cpp if (((data << shift) >> shift) == data) { shift 142 nanojit/NativeARM.cpp leading_zeroes = shift; shift 133 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op); shift 134 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((b&8)>>3); shift 135 nanojit/NativeX64.cpp return rex != 0x40 ? op | rex << shift : op - 1; shift 140 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op); shift 141 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((x&8)>>2) | ((b&8)>>3); shift 142 nanojit/NativeX64.cpp return rex != 0x40 ? op | rex << shift : op - 1; shift 148 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op); shift 149 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((b&8)>>3); shift 150 nanojit/NativeX64.cpp return ((rex | (b & ~3)) != 0x40) ? (op | (rex << shift)) : op - 1; shift 156 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op) + 8; shift 157 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((b&8)>>3); shift 159 nanojit/NativeX64.cpp return rex != 0x40 ? op | rex << shift : shift 160 nanojit/NativeX64.cpp ((op & ~(255LL<<shift)) | (op>>(shift-8)&255) << shift) - 1; shift 684 nanojit/NativeX64.cpp int shift = ins->oprnd2()->imm32() & 63; shift 687 nanojit/NativeX64.cpp case LIR_qursh: SHRQI(rr, shift); break; shift 688 nanojit/NativeX64.cpp case LIR_qirsh: SARQI(rr, shift); break; shift 689 nanojit/NativeX64.cpp case LIR_qilsh: SHLQI(rr, shift); break; shift 690 nanojit/NativeX64.cpp case LIR_ush: SHRI( rr, shift); break; shift 691 nanojit/NativeX64.cpp case LIR_rsh: SARI( rr, shift); break; shift 692 nanojit/NativeX64.cpp case LIR_lsh: SHLI( rr, shift); break; shift 92 shell/swf.cpp int shift; shift 93 shell/swf.cpp for (shift = bitsLeft - bitPos; shift > 0; shift = bitsLeft - bitPos) { shift 94 shell/swf.cpp result |= bitBuf << shift; shift 99 shell/swf.cpp result |= bitBuf >> -shift; shift 108 shell/swf.cpp int shift = 32 - n; shift 109 shell/swf.cpp return (num << shift) >> shift; // sign extend