shift 2880 MMgc/GC.cpp int shift = (itemNum&0x7)<<2;
shift 2883 MMgc/GC.cpp if((bits2 & ((GCAlloc::kMark|GCAlloc::kQueued)<<shift)) == 0)
shift 2892 MMgc/GC.cpp *pbits = bits2 | (GCAlloc::kQueued << shift);
shift 2905 MMgc/GC.cpp *pbits = bits2 | (GCAlloc::kMark << shift);
shift 219 MMgc/GCAlloc-inlines.h int index = (int)((((char*) item - block->items) * block->alloc->multiple) >> block->alloc->shift);
shift 91 MMgc/GCAlloc.cpp ComputeMultiplyShift((uint16_t)m_itemSize, multiple, shift);
shift 234 MMgc/GCAlloc.h uint16_t shift;
shift 105 core/ArrayClass.cpp ar->shift(); AvmAssert(s->RefCount()==1);
shift 336 core/ArrayClass.cpp return (a->m_denseArr.shift());
shift 90 core/ArrayClass.h inline Atom shift(Atom thisAtom) { return ArrayClass::generic_shift(toplevel(), thisAtom); }
shift 64 core/AtomArray.h Atom shift();
shift 1495 core/WordcodeEmitter.cpp for ( uintptr_t i=0, limit=nextI-shift ; i < limit ; i++ ) {
shift 1496 core/WordcodeEmitter.cpp I[i] = I[i+shift];
shift 1497 core/WordcodeEmitter.cpp O[i] = O[i+shift];
shift 1499 core/WordcodeEmitter.cpp nextI -= shift;
shift 225 core/WordcodeEmitter.h void shiftBuffers(uint32_t shift);
shift 140 nanojit/NativeARM.cpp uint32_t shift = leading_zeroes + try_shift;
shift 141 nanojit/NativeARM.cpp if (((data << shift) >> shift) == data) {
shift 142 nanojit/NativeARM.cpp leading_zeroes = shift;
shift 133 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op);
shift 134 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((b&8)>>3);
shift 135 nanojit/NativeX64.cpp return rex != 0x40 ? op | rex << shift : op - 1;
shift 140 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op);
shift 141 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((x&8)>>2) | ((b&8)>>3);
shift 142 nanojit/NativeX64.cpp return rex != 0x40 ? op | rex << shift : op - 1;
shift 148 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op);
shift 149 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((b&8)>>3);
shift 150 nanojit/NativeX64.cpp return ((rex | (b & ~3)) != 0x40) ? (op | (rex << shift)) : op - 1;
shift 156 nanojit/NativeX64.cpp int shift = 64 - 8*oplen(op) + 8;
shift 157 nanojit/NativeX64.cpp uint64_t rex = ((op >> shift) & 255) | ((r&8)>>1) | ((b&8)>>3);
shift 159 nanojit/NativeX64.cpp return rex != 0x40 ? op | rex << shift :
shift 160 nanojit/NativeX64.cpp ((op & ~(255LL<<shift)) | (op>>(shift-8)&255) << shift) - 1;
shift 684 nanojit/NativeX64.cpp int shift = ins->oprnd2()->imm32() & 63;
shift 687 nanojit/NativeX64.cpp case LIR_qursh: SHRQI(rr, shift); break;
shift 688 nanojit/NativeX64.cpp case LIR_qirsh: SARQI(rr, shift); break;
shift 689 nanojit/NativeX64.cpp case LIR_qilsh: SHLQI(rr, shift); break;
shift 690 nanojit/NativeX64.cpp case LIR_ush: SHRI( rr, shift); break;
shift 691 nanojit/NativeX64.cpp case LIR_rsh: SARI( rr, shift); break;
shift 692 nanojit/NativeX64.cpp case LIR_lsh: SHLI( rr, shift); break;
shift 92 shell/swf.cpp int shift;
shift 93 shell/swf.cpp for (shift = bitsLeft - bitPos; shift > 0; shift = bitsLeft - bitPos) {
shift 94 shell/swf.cpp result |= bitBuf << shift;
shift 99 shell/swf.cpp result |= bitBuf >> -shift;
shift 108 shell/swf.cpp int shift = 32 - n;
shift 109 shell/swf.cpp return (num << shift) >> shift; // sign extend