R 168 MMgc/GC.cpp , R(R_INITIAL_VALUE) // bytes/second; will be updated on-line
R 280 MMgc/GC.cpp return P * R / W();
R 295 MMgc/GC.cpp R = double(bytesScannedTotal) / (double(adjustR_totalTime) / double(VMPI_getPerformanceFrequency()));
R 296 MMgc/GC.cpp if (R < R_LOWER_LIMIT)
R 297 MMgc/GC.cpp R = R_LOWER_LIMIT;
R 345 MMgc/GC.cpp R / (1024*1024),
R 656 MMgc/GC.h double R;
R 644 core/CdeclThunk.cpp #define ASM_FUNC_BEGIN(R, N, A) static R __declspec(naked) N A {
R 653 core/CdeclThunk.cpp #define ASM_FUNC_BEGIN(R, N, A) typedef R (* N##_type)A; static void* N##_container () { \
R 670 core/CdeclThunk.cpp typedef R (* N##_type)A; \
R 1228 core/WordcodeEmitter.cpp R[k++] = I[n][j];
R 1274 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1275 core/WordcodeEmitter.cpp int32_t offset = int32_t(R[i++]);
R 1288 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1290 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1298 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1301 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1302 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1305 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1306 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1307 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1310 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1311 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1312 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1313 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 1314 core/WordcodeEmitter.cpp *dest++ = R[i++];
R 216 core/WordcodeEmitter.h uintptr_t R[30]; // replacement data
R 205 core/peephole.cpp R[0] = NEW_OPCODE(WOP_getlocal0 + I[0][1]);
R 213 core/peephole.cpp R[0] = NEW_OPCODE(WOP_get2locals);
R 214 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 222 core/peephole.cpp R[0] = NEW_OPCODE(WOP_get3locals);
R 223 core/peephole.cpp R[1] = (I[2][1] << 20) | (I[1][1] << 10) | I[0][1];
R 231 core/peephole.cpp R[0] = NEW_OPCODE(WOP_get4locals);
R 232 core/peephole.cpp R[1] = (I[3][1] << 24) | (I[2][1] << 16) | (I[1][1] << 8) | I[0][1];
R 240 core/peephole.cpp R[0] = NEW_OPCODE(WOP_get5locals);
R 241 core/peephole.cpp R[1] = (I[4][1] << 24) | (I[3][1] << 18) | (I[2][1] << 12) | (I[1][1] << 6) | I[0][1];
R 249 core/peephole.cpp R[0] = NEW_OPCODE(WOP_add_ll);
R 250 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 258 core/peephole.cpp R[0] = NEW_OPCODE(WOP_add_set_lll);
R 259 core/peephole.cpp R[1] = (I[3][1] << 20) | (I[1][1] << 10) | I[0][1];
R 267 core/peephole.cpp R[0] = NEW_OPCODE(WOP_subtract_ll);
R 268 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 276 core/peephole.cpp R[0] = NEW_OPCODE(WOP_multiply_ll);
R 277 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 285 core/peephole.cpp R[0] = NEW_OPCODE(WOP_divide_ll);
R 286 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 294 core/peephole.cpp R[0] = NEW_OPCODE(WOP_modulo_ll);
R 295 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 303 core/peephole.cpp R[0] = NEW_OPCODE(WOP_bitand_ll);
R 304 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 312 core/peephole.cpp R[0] = NEW_OPCODE(WOP_bitor_ll);
R 313 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 321 core/peephole.cpp R[0] = NEW_OPCODE(WOP_bitxor_ll);
R 322 core/peephole.cpp R[1] = (I[1][1] << 16) | I[0][1];
R 331 core/peephole.cpp R[0] = NEW_OPCODE(WOP_iflt_ll);
R 332 core/peephole.cpp R[1] = I[2][1];
R 333 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 342 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifnlt_ll);
R 343 core/peephole.cpp R[1] = I[2][1];
R 344 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 353 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifle_ll);
R 354 core/peephole.cpp R[1] = I[2][1];
R 355 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 364 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifnle_ll);
R 365 core/peephole.cpp R[1] = I[2][1];
R 366 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 375 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifgt_ll);
R 376 core/peephole.cpp R[1] = I[2][1];
R 377 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 386 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifngt_ll);
R 387 core/peephole.cpp R[1] = I[2][1];
R 388 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 397 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifge_ll);
R 398 core/peephole.cpp R[1] = I[2][1];
R 399 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 408 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifnge_ll);
R 409 core/peephole.cpp R[1] = I[2][1];
R 410 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 419 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifeq_ll);
R 420 core/peephole.cpp R[1] = I[2][1];
R 421 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 430 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifne_ll);
R 431 core/peephole.cpp R[1] = I[2][1];
R 432 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 441 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifstricteq_ll);
R 442 core/peephole.cpp R[1] = I[2][1];
R 443 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 452 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifstrictne_ll);
R 453 core/peephole.cpp R[1] = I[2][1];
R 454 core/peephole.cpp R[2] = (I[1][1] << 16) | I[0][1];
R 462 core/peephole.cpp R[0] = NEW_OPCODE(WOP_add_lb);
R 463 core/peephole.cpp R[1] = I[0][1];
R 464 core/peephole.cpp R[2] = I[1][1];
R 472 core/peephole.cpp R[0] = NEW_OPCODE(WOP_subtract_lb);
R 473 core/peephole.cpp R[1] = I[0][1];
R 474 core/peephole.cpp R[2] = I[1][1];
R 482 core/peephole.cpp R[0] = NEW_OPCODE(WOP_multiply_lb);
R 483 core/peephole.cpp R[1] = I[0][1];
R 484 core/peephole.cpp R[2] = I[1][1];
R 492 core/peephole.cpp R[0] = NEW_OPCODE(WOP_divide_lb);
R 493 core/peephole.cpp R[1] = I[0][1];
R 494 core/peephole.cpp R[2] = I[1][1];
R 502 core/peephole.cpp R[0] = NEW_OPCODE(WOP_bitand_lb);
R 503 core/peephole.cpp R[1] = I[0][1];
R 504 core/peephole.cpp R[2] = I[1][1];
R 512 core/peephole.cpp R[0] = NEW_OPCODE(WOP_bitor_lb);
R 513 core/peephole.cpp R[1] = I[0][1];
R 514 core/peephole.cpp R[2] = I[1][1];
R 522 core/peephole.cpp R[0] = NEW_OPCODE(WOP_bitxor_lb);
R 523 core/peephole.cpp R[1] = I[0][1];
R 524 core/peephole.cpp R[2] = I[1][1];
R 533 core/peephole.cpp R[0] = NEW_OPCODE(WOP_iflt_lb);
R 534 core/peephole.cpp R[1] = I[2][1];
R 535 core/peephole.cpp R[2] = I[0][1];
R 536 core/peephole.cpp R[3] = I[1][1];
R 545 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifnlt_lb);
R 546 core/peephole.cpp R[1] = I[2][1];
R 547 core/peephole.cpp R[2] = I[0][1];
R 548 core/peephole.cpp R[3] = I[1][1];
R 557 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifle_lb);
R 558 core/peephole.cpp R[1] = I[2][1];
R 559 core/peephole.cpp R[2] = I[0][1];
R 560 core/peephole.cpp R[3] = I[1][1];
R 569 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifnle_lb);
R 570 core/peephole.cpp R[1] = I[2][1];
R 571 core/peephole.cpp R[2] = I[0][1];
R 572 core/peephole.cpp R[3] = I[1][1];
R 581 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifgt_lb);
R 582 core/peephole.cpp R[1] = I[2][1];
R 583 core/peephole.cpp R[2] = I[0][1];
R 584 core/peephole.cpp R[3] = I[1][1];
R 593 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifngt_lb);
R 594 core/peephole.cpp R[1] = I[2][1];
R 595 core/peephole.cpp R[2] = I[0][1];
R 596 core/peephole.cpp R[3] = I[1][1];
R 605 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifge_lb);
R 606 core/peephole.cpp R[1] = I[2][1];
R 607 core/peephole.cpp R[2] = I[0][1];
R 608 core/peephole.cpp R[3] = I[1][1];
R 617 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifnge_lb);
R 618 core/peephole.cpp R[1] = I[2][1];
R 619 core/peephole.cpp R[2] = I[0][1];
R 620 core/peephole.cpp R[3] = I[1][1];
R 629 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifeq_lb);
R 630 core/peephole.cpp R[1] = I[2][1];
R 631 core/peephole.cpp R[2] = I[0][1];
R 632 core/peephole.cpp R[3] = I[1][1];
R 641 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifne_lb);
R 642 core/peephole.cpp R[1] = I[2][1];
R 643 core/peephole.cpp R[2] = I[0][1];
R 644 core/peephole.cpp R[3] = I[1][1];
R 653 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifstricteq_lb);
R 654 core/peephole.cpp R[1] = I[2][1];
R 655 core/peephole.cpp R[2] = I[0][1];
R 656 core/peephole.cpp R[3] = I[1][1];
R 665 core/peephole.cpp R[0] = NEW_OPCODE(WOP_ifstrictne_lb);
R 666 core/peephole.cpp R[1] = I[2][1];
R 667 core/peephole.cpp R[2] = I[0][1];
R 668 core/peephole.cpp R[3] = I[1][1];
R 676 core/peephole.cpp R[0] = NEW_OPCODE(WOP_setlocal0 + I[0][1]);
R 684 core/peephole.cpp R[0] = NEW_OPCODE(WOP_storelocal);
R 685 core/peephole.cpp R[1] = I[0][1];
R 693 core/peephole.cpp R[0] = NEW_OPCODE(WOP_swap_pop);
R 422 nanojit/NativePPC.h #define SPR(spr) ((R##spr)>>5|(R##spr&31)<<5)
R 367 nanojit/NativeX64.cpp void Assembler::PUSHR(R r) { emitr(X64_pushr,r); asm_output("push %s", RQ(r)); }
R 368 nanojit/NativeX64.cpp void Assembler::POPR( R r) { emitr(X64_popr, r); asm_output("pop %s", RQ(r)); }
R 369 nanojit/NativeX64.cpp void Assembler::NOT( R r) { emitr(X64_not, r); asm_output("notl %s", RL(r)); }
R 370 nanojit/NativeX64.cpp void Assembler::NEG( R r) { emitr(X64_neg, r); asm_output("negl %s", RL(r)); }
R 371 nanojit/NativeX64.cpp void Assembler::IDIV( R r) { emitr(X64_idiv, r); asm_output("idivl edx:eax, %s",RL(r)); }
R 373 nanojit/NativeX64.cpp void Assembler::SHR( R r) { emitr(X64_shr, r); asm_output("shrl %s, ecx", RL(r)); }
R 374 nanojit/NativeX64.cpp void Assembler::SAR( R r) { emitr(X64_sar, r); asm_output("sarl %s, ecx", RL(r)); }
R 375 nanojit/NativeX64.cpp void Assembler::SHL( R r) { emitr(X64_shl, r); asm_output("shll %s, ecx", RL(r)); }
R 376 nanojit/NativeX64.cpp void Assembler::SHRQ(R r) { emitr(X64_shrq, r); asm_output("shrq %s, ecx", RQ(r)); }
R 377 nanojit/NativeX64.cpp void Assembler::SARQ(R r) { emitr(X64_sarq, r); asm_output("sarq %s, ecx", RQ(r)); }
R 378 nanojit/NativeX64.cpp void Assembler::SHLQ(R r) { emitr(X64_shlq, r); asm_output("shlq %s, ecx", RQ(r)); }
R 380 nanojit/NativeX64.cpp void Assembler::SHRI( R r, I i) { emit8(rexrb(X64_shri | U64(r&7)<<48, (R)0, r), i); asm_output("shrl %s, %d", RL(r), i); }
R 381 nanojit/NativeX64.cpp void Assembler::SARI( R r, I i) { emit8(rexrb(X64_sari | U64(r&7)<<48, (R)0, r), i); asm_output("sarl %s, %d", RL(r), i); }
R 382 nanojit/NativeX64.cpp void Assembler::SHLI( R r, I i) { emit8(rexrb(X64_shli | U64(r&7)<<48, (R)0, r), i); asm_output("shll %s, %d", RL(r), i); }
R 383 nanojit/NativeX64.cpp void Assembler::SHRQI(R r, I i) { emit8(rexrb(X64_shrqi | U64(r&7)<<48, (R)0, r), i); asm_output("shrq %s, %d", RQ(r), i); }
R 384 nanojit/NativeX64.cpp void Assembler::SARQI(R r, I i) { emit8(rexrb(X64_sarqi | U64(r&7)<<48, (R)0, r), i); asm_output("sarq %s, %d", RQ(r), i); }
R 385 nanojit/NativeX64.cpp void Assembler::SHLQI(R r, I i) { emit8(rexrb(X64_shlqi | U64(r&7)<<48, (R)0, r), i); asm_output("shlq %s, %d", RQ(r), i); }
R 387 nanojit/NativeX64.cpp void Assembler::SETE( R r) { emitr8(X64_sete, r); asm_output("sete %s", RB(r)); }
R 388 nanojit/NativeX64.cpp void Assembler::SETL( R r) { emitr8(X64_setl, r); asm_output("setl %s", RB(r)); }
R 389 nanojit/NativeX64.cpp void Assembler::SETLE(R r) { emitr8(X64_setle,r); asm_output("setle %s",RB(r)); }
R 390 nanojit/NativeX64.cpp void Assembler::SETG( R r) { emitr8(X64_setg, r); asm_output("setg %s", RB(r)); }
R 391 nanojit/NativeX64.cpp void Assembler::SETGE(R r) { emitr8(X64_setge,r); asm_output("setge %s",RB(r)); }
R 392 nanojit/NativeX64.cpp void Assembler::SETB( R r) { emitr8(X64_setb, r); asm_output("setb %s", RB(r)); }
R 393 nanojit/NativeX64.cpp void Assembler::SETBE(R r) { emitr8(X64_setbe,r); asm_output("setbe %s",RB(r)); }
R 394 nanojit/NativeX64.cpp void Assembler::SETA( R r) { emitr8(X64_seta, r); asm_output("seta %s", RB(r)); }
R 395 nanojit/NativeX64.cpp void Assembler::SETAE(R r) { emitr8(X64_setae,r); asm_output("setae %s",RB(r)); }
R 396 nanojit/NativeX64.cpp void Assembler::SETO( R r) { emitr8(X64_seto, r); asm_output("seto %s", RB(r)); }
R 398 nanojit/NativeX64.cpp void Assembler::ADDRR(R l, R r) { emitrr(X64_addrr,l,r); asm_output("addl %s, %s", RL(l),RL(r)); }
R 399 nanojit/NativeX64.cpp void Assembler::SUBRR(R l, R r) { emitrr(X64_subrr,l,r); asm_output("subl %s, %s", RL(l),RL(r)); }
R 400 nanojit/NativeX64.cpp void Assembler::ANDRR(R l, R r) { emitrr(X64_andrr,l,r); asm_output("andl %s, %s", RL(l),RL(r)); }
R 401 nanojit/NativeX64.cpp void Assembler::ORLRR(R l, R r) { emitrr(X64_orlrr,l,r); asm_output("orl %s, %s", RL(l),RL(r)); }
R 402 nanojit/NativeX64.cpp void Assembler::XORRR(R l, R r) { emitrr(X64_xorrr,l,r); asm_output("xorl %s, %s", RL(l),RL(r)); }
R 403 nanojit/NativeX64.cpp void Assembler::IMUL( R l, R r) { emitrr(X64_imul, l,r); asm_output("imull %s, %s",RL(l),RL(r)); }
R 404 nanojit/NativeX64.cpp void Assembler::CMPLR(R l, R r) { emitrr(X64_cmplr,l,r); asm_output("cmpl %s, %s", RL(l),RL(r)); }
R 405 nanojit/NativeX64.cpp void Assembler::MOVLR(R l, R r) { emitrr(X64_movlr,l,r); asm_output("movl %s, %s", RL(l),RL(r)); }
R 407 nanojit/NativeX64.cpp void Assembler::ADDQRR( R l, R r) { emitrr(X64_addqrr, l,r); asm_output("addq %s, %s", RQ(l),RQ(r)); }
R 408 nanojit/NativeX64.cpp void Assembler::SUBQRR( R l, R r) { emitrr(X64_subqrr, l,r); asm_output("subq %s, %s", RQ(l),RQ(r)); }
R 409 nanojit/NativeX64.cpp void Assembler::ANDQRR( R l, R r) { emitrr(X64_andqrr, l,r); asm_output("andq %s, %s", RQ(l),RQ(r)); }
R 410 nanojit/NativeX64.cpp void Assembler::ORQRR( R l, R r) { emitrr(X64_orqrr, l,r); asm_output("orq %s, %s", RQ(l),RQ(r)); }
R 411 nanojit/NativeX64.cpp void Assembler::XORQRR( R l, R r) { emitrr(X64_xorqrr, l,r); asm_output("xorq %s, %s", RQ(l),RQ(r)); }
R 412 nanojit/NativeX64.cpp void Assembler::CMPQR( R l, R r) { emitrr(X64_cmpqr, l,r); asm_output("cmpq %s, %s", RQ(l),RQ(r)); }
R 413 nanojit/NativeX64.cpp void Assembler::MOVQR( R l, R r) { emitrr(X64_movqr, l,r); asm_output("movq %s, %s", RQ(l),RQ(r)); }
R 414 nanojit/NativeX64.cpp void Assembler::MOVAPSR(R l, R r) { emitrr(X64_movapsr,l,r); asm_output("movaps %s, %s",RQ(l),RQ(r)); }
R 416 nanojit/NativeX64.cpp void Assembler::CMOVNO( R l, R r) { emitrr(X64_cmovno, l,r); asm_output("cmovlno %s, %s", RL(l),RL(r)); }
R 417 nanojit/NativeX64.cpp void Assembler::CMOVNE( R l, R r) { emitrr(X64_cmovne, l,r); asm_output("cmovlne %s, %s", RL(l),RL(r)); }
R 418 nanojit/NativeX64.cpp void Assembler::CMOVNL( R l, R r) { emitrr(X64_cmovnl, l,r); asm_output("cmovlnl %s, %s", RL(l),RL(r)); }
R 419 nanojit/NativeX64.cpp void Assembler::CMOVNLE(R l, R r) { emitrr(X64_cmovnle,l,r); asm_output("cmovlnle %s, %s", RL(l),RL(r)); }
R 420 nanojit/NativeX64.cpp void Assembler::CMOVNG( R l, R r) { emitrr(X64_cmovng, l,r); asm_output("cmovlng %s, %s", RL(l),RL(r)); }
R 421 nanojit/NativeX64.cpp void Assembler::CMOVNGE(R l, R r) { emitrr(X64_cmovnge,l,r); asm_output("cmovlnge %s, %s", RL(l),RL(r)); }
R 422 nanojit/NativeX64.cpp void Assembler::CMOVNB( R l, R r) { emitrr(X64_cmovnb, l,r); asm_output("cmovlnb %s, %s", RL(l),RL(r)); }
R 423 nanojit/NativeX64.cpp void Assembler::CMOVNBE(R l, R r) { emitrr(X64_cmovnbe,l,r); asm_output("cmovlnbe %s, %s", RL(l),RL(r)); }
R 424 nanojit/NativeX64.cpp void Assembler::CMOVNA( R l, R r) { emitrr(X64_cmovna, l,r); asm_output("cmovlna %s, %s", RL(l),RL(r)); }
R 425 nanojit/NativeX64.cpp void Assembler::CMOVNAE(R l, R r) { emitrr(X64_cmovnae,l,r); asm_output("cmovlnae %s, %s", RL(l),RL(r)); }
R 427 nanojit/NativeX64.cpp void Assembler::CMOVQNO( R l, R r) { emitrr(X64_cmovqno, l,r); asm_output("cmovqno %s, %s", RQ(l),RQ(r)); }
R 428 nanojit/NativeX64.cpp void Assembler::CMOVQNE( R l, R r) { emitrr(X64_cmovqne, l,r); asm_output("cmovqne %s, %s", RQ(l),RQ(r)); }
R 429 nanojit/NativeX64.cpp void Assembler::CMOVQNL( R l, R r) { emitrr(X64_cmovqnl, l,r); asm_output("cmovqnl %s, %s", RQ(l),RQ(r)); }
R 430 nanojit/NativeX64.cpp void Assembler::CMOVQNLE(R l, R r) { emitrr(X64_cmovqnle,l,r); asm_output("cmovqnle %s, %s", RQ(l),RQ(r)); }
R 431 nanojit/NativeX64.cpp void Assembler::CMOVQNG( R l, R r) { emitrr(X64_cmovqng, l,r); asm_output("cmovqng %s, %s", RQ(l),RQ(r)); }
R 432 nanojit/NativeX64.cpp void Assembler::CMOVQNGE(R l, R r) { emitrr(X64_cmovqnge,l,r); asm_output("cmovqnge %s, %s", RQ(l),RQ(r)); }
R 433 nanojit/NativeX64.cpp void Assembler::CMOVQNB( R l, R r) { emitrr(X64_cmovqnb, l,r); asm_output("cmovqnb %s, %s", RQ(l),RQ(r)); }
R 434 nanojit/NativeX64.cpp void Assembler::CMOVQNBE(R l, R r) { emitrr(X64_cmovqnbe,l,r); asm_output("cmovqnbe %s, %s", RQ(l),RQ(r)); }
R 435 nanojit/NativeX64.cpp void Assembler::CMOVQNA( R l, R r) { emitrr(X64_cmovqna, l,r); asm_output("cmovqna %s, %s", RQ(l),RQ(r)); }
R 436 nanojit/NativeX64.cpp void Assembler::CMOVQNAE(R l, R r) { emitrr(X64_cmovqnae,l,r); asm_output("cmovqnae %s, %s", RQ(l),RQ(r)); }
R 438 nanojit/NativeX64.cpp void Assembler::MOVSXDR(R l, R r) { emitrr(X64_movsxdr,l,r); asm_output("movsxd %s, %s",RQ(l),RL(r)); }
R 440 nanojit/NativeX64.cpp void Assembler::MOVZX8(R l, R r) { emitrr8(X64_movzx8,l,r); asm_output("movzx %s, %s",RQ(l),RB(r)); }
R 446 nanojit/NativeX64.cpp void Assembler::XORPS( R r) { emitrr(X64_xorps, r,r); asm_output("xorps %s, %s", RQ(r),RQ(r)); }
R 447 nanojit/NativeX64.cpp void Assembler::DIVSD( R l, R r) { emitprr(X64_divsd, l,r); asm_output("divsd %s, %s", RQ(l),RQ(r)); }
R 448 nanojit/NativeX64.cpp void Assembler::MULSD( R l, R r) { emitprr(X64_mulsd, l,r); asm_output("mulsd %s, %s", RQ(l),RQ(r)); }
R 449 nanojit/NativeX64.cpp void Assembler::ADDSD( R l, R r) { emitprr(X64_addsd, l,r); asm_output("addsd %s, %s", RQ(l),RQ(r)); }
R 450 nanojit/NativeX64.cpp void Assembler::SUBSD( R l, R r) { emitprr(X64_subsd, l,r); asm_output("subsd %s, %s", RQ(l),RQ(r)); }
R 451 nanojit/NativeX64.cpp void Assembler::CVTSQ2SD(R l, R r) { emitprr(X64_cvtsq2sd,l,r); asm_output("cvtsq2sd %s, %s",RQ(l),RQ(r)); }
R 452 nanojit/NativeX64.cpp void Assembler::CVTSI2SD(R l, R r) { emitprr(X64_cvtsi2sd,l,r); asm_output("cvtsi2sd %s, %s",RQ(l),RL(r)); }
R 453 nanojit/NativeX64.cpp void Assembler::CVTSS2SD(R l, R r) { emitprr(X64_cvtss2sd,l,r); asm_output("cvtss2sd %s, %s",RQ(l),RL(r)); }
R 454 nanojit/NativeX64.cpp void Assembler::CVTSD2SS(R l, R r) { emitprr(X64_cvtsd2ss,l,r); asm_output("cvtsd2ss %s, %s",RL(l),RQ(r)); }
R 455 nanojit/NativeX64.cpp void Assembler::UCOMISD( R l, R r) { emitprr(X64_ucomisd, l,r); asm_output("ucomisd %s, %s", RQ(l),RQ(r)); }
R 456 nanojit/NativeX64.cpp void Assembler::MOVQRX( R l, R r) { emitprr(X64_movqrx, r,l); asm_output("movq %s, %s", RQ(l),RQ(r)); } // Nb: r and l are deliberately reversed within the emitprr() call.
R 457 nanojit/NativeX64.cpp void Assembler::MOVQXR( R l, R r) { emitprr(X64_movqxr, l,r); asm_output("movq %s, %s", RQ(l),RQ(r)); }
R 460 nanojit/NativeX64.cpp void Assembler::MOVI( R r, I32 i32) { emitr_imm(X64_movi, r,i32); asm_output("movl %s, %d",RL(r),i32); }
R 461 nanojit/NativeX64.cpp void Assembler::ADDLRI(R r, I32 i32) { emitr_imm(X64_addlri,r,i32); asm_output("addl %s, %d",RL(r),i32); }
R 462 nanojit/NativeX64.cpp void Assembler::SUBLRI(R r, I32 i32) { emitr_imm(X64_sublri,r,i32); asm_output("subl %s, %d",RL(r),i32); }
R 463 nanojit/NativeX64.cpp void Assembler::ANDLRI(R r, I32 i32) { emitr_imm(X64_andlri,r,i32); asm_output("andl %s, %d",RL(r),i32); }
R 464 nanojit/NativeX64.cpp void Assembler::ORLRI( R r, I32 i32) { emitr_imm(X64_orlri, r,i32); asm_output("orl %s, %d", RL(r),i32); }
R 465 nanojit/NativeX64.cpp void Assembler::XORLRI(R r, I32 i32) { emitr_imm(X64_xorlri,r,i32); asm_output("xorl %s, %d",RL(r),i32); }
R 466 nanojit/NativeX64.cpp void Assembler::CMPLRI(R r, I32 i32) { emitr_imm(X64_cmplri,r,i32); asm_output("cmpl %s, %d",RL(r),i32); }
R 468 nanojit/NativeX64.cpp void Assembler::ADDQRI( R r, I32 i32) { emitr_imm(X64_addqri, r,i32); asm_output("addq %s, %d", RQ(r),i32); }
R 469 nanojit/NativeX64.cpp void Assembler::SUBQRI( R r, I32 i32) { emitr_imm(X64_subqri, r,i32); asm_output("subq %s, %d", RQ(r),i32); }
R 470 nanojit/NativeX64.cpp void Assembler::ANDQRI( R r, I32 i32) { emitr_imm(X64_andqri, r,i32); asm_output("andq %s, %d", RQ(r),i32); }
R 471 nanojit/NativeX64.cpp void Assembler::ORQRI( R r, I32 i32) { emitr_imm(X64_orqri, r,i32); asm_output("orq %s, %d", RQ(r),i32); }
R 472 nanojit/NativeX64.cpp void Assembler::XORQRI( R r, I32 i32) { emitr_imm(X64_xorqri, r,i32); asm_output("xorq %s, %d", RQ(r),i32); }
R 473 nanojit/NativeX64.cpp void Assembler::CMPQRI( R r, I32 i32) { emitr_imm(X64_cmpqri, r,i32); asm_output("cmpq %s, %d", RQ(r),i32); }
R 474 nanojit/NativeX64.cpp void Assembler::MOVQI32(R r, I32 i32) { emitr_imm(X64_movqi32,r,i32); asm_output("movqi32 %s, %d",RQ(r),i32); }
R 476 nanojit/NativeX64.cpp void Assembler::ADDLR8(R r, I32 i8) { emitr_imm8(X64_addlr8,r,i8); asm_output("addl %s, %d", RL(r),i8); }
R 477 nanojit/NativeX64.cpp void Assembler::SUBLR8(R r, I32 i8) { emitr_imm8(X64_sublr8,r,i8); asm_output("subl %s, %d", RL(r),i8); }
R 478 nanojit/NativeX64.cpp void Assembler::ANDLR8(R r, I32 i8) { emitr_imm8(X64_andlr8,r,i8); asm_output("andl %s, %d", RL(r),i8); }
R 479 nanojit/NativeX64.cpp void Assembler::ORLR8( R r, I32 i8) { emitr_imm8(X64_orlr8, r,i8); asm_output("orl %s, %d", RL(r),i8); }
R 480 nanojit/NativeX64.cpp void Assembler::XORLR8(R r, I32 i8) { emitr_imm8(X64_xorlr8,r,i8); asm_output("xorl %s, %d", RL(r),i8); }
R 481 nanojit/NativeX64.cpp void Assembler::CMPLR8(R r, I32 i8) { emitr_imm8(X64_cmplr8,r,i8); asm_output("cmpl %s, %d", RL(r),i8); }
R 483 nanojit/NativeX64.cpp void Assembler::ADDQR8(R r, I32 i8) { emitr_imm8(X64_addqr8,r,i8); asm_output("addq %s, %d",RQ(r),i8); }
R 484 nanojit/NativeX64.cpp void Assembler::SUBQR8(R r, I32 i8) { emitr_imm8(X64_subqr8,r,i8); asm_output("subq %s, %d",RQ(r),i8); }
R 485 nanojit/NativeX64.cpp void Assembler::ANDQR8(R r, I32 i8) { emitr_imm8(X64_andqr8,r,i8); asm_output("andq %s, %d",RQ(r),i8); }
R 486 nanojit/NativeX64.cpp void Assembler::ORQR8( R r, I32 i8) { emitr_imm8(X64_orqr8, r,i8); asm_output("orq %s, %d", RQ(r),i8); }
R 487 nanojit/NativeX64.cpp void Assembler::XORQR8(R r, I32 i8) { emitr_imm8(X64_xorqr8,r,i8); asm_output("xorq %s, %d",RQ(r),i8); }
R 488 nanojit/NativeX64.cpp void Assembler::CMPQR8(R r, I32 i8) { emitr_imm8(X64_cmpqr8,r,i8); asm_output("cmpq %s, %d",RQ(r),i8); }
R 490 nanojit/NativeX64.cpp void Assembler::IMULI(R l, R r, I32 i32) { emitrr_imm(X64_imuli,l,r,i32); asm_output("imuli %s, %s, %d",RL(l),RL(r),i32); }
R 492 nanojit/NativeX64.cpp void Assembler::MOVQI(R r, U64 u64) { emitr_imm64(X64_movqi,r,u64); asm_output("movq %s, %p",RQ(r),(void*)u64); }
R 494 nanojit/NativeX64.cpp void Assembler::LEARIP(R r, I32 d) { emitrm(X64_learip,r,d,(Register)0); asm_output("lea %s, %d(rip)",RQ(r),d); }
R 496 nanojit/NativeX64.cpp void Assembler::LEAQRM(R r1, I d, R r2) { emitrm(X64_leaqrm,r1,d,r2); asm_output("leaq %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 497 nanojit/NativeX64.cpp void Assembler::MOVLRM(R r1, I d, R r2) { emitrm(X64_movlrm,r1,d,r2); asm_output("movl %s, %d(%s)",RL(r1),d,RQ(r2)); }
R 498 nanojit/NativeX64.cpp void Assembler::MOVQRM(R r1, I d, R r2) { emitrm(X64_movqrm,r1,d,r2); asm_output("movq %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 499 nanojit/NativeX64.cpp void Assembler::MOVBMR(R r1, I d, R r2) { emitrm8(X64_movbmr,r1,d,r2); asm_output("movb %d(%s), %s",d,RQ(r1),RB(r2)); }
R 500 nanojit/NativeX64.cpp void Assembler::MOVSMR(R r1, I d, R r2) { emitprm(X64_movsmr,r1,d,r2); asm_output("movs %d(%s), %s",d,RQ(r1),RS(r2)); }
R 501 nanojit/NativeX64.cpp void Assembler::MOVLMR(R r1, I d, R r2) { emitrm(X64_movlmr,r1,d,r2); asm_output("movl %d(%s), %s",d,RQ(r1),RL(r2)); }
R 502 nanojit/NativeX64.cpp void Assembler::MOVQMR(R r1, I d, R r2) { emitrm(X64_movqmr,r1,d,r2); asm_output("movq %d(%s), %s",d,RQ(r1),RQ(r2)); }
R 504 nanojit/NativeX64.cpp void Assembler::MOVZX8M( R r1, I d, R r2) { emitrm_wide(X64_movzx8m, r1,d,r2); asm_output("movzxb %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 505 nanojit/NativeX64.cpp void Assembler::MOVZX16M(R r1, I d, R r2) { emitrm_wide(X64_movzx16m,r1,d,r2); asm_output("movzxs %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 507 nanojit/NativeX64.cpp void Assembler::MOVSX8M( R r1, I d, R r2) { emitrm_wide(X64_movsx8m, r1,d,r2); asm_output("movsxb %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 508 nanojit/NativeX64.cpp void Assembler::MOVSX16M(R r1, I d, R r2) { emitrm_wide(X64_movsx16m,r1,d,r2); asm_output("movsxs %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 510 nanojit/NativeX64.cpp void Assembler::MOVSDRM(R r1, I d, R r2) { emitprm(X64_movsdrm,r1,d,r2); asm_output("movsd %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 511 nanojit/NativeX64.cpp void Assembler::MOVSDMR(R r1, I d, R r2) { emitprm(X64_movsdmr,r1,d,r2); asm_output("movsd %d(%s), %s",d,RQ(r1),RQ(r2)); }
R 512 nanojit/NativeX64.cpp void Assembler::MOVSSRM(R r1, I d, R r2) { emitprm(X64_movssrm,r1,d,r2); asm_output("movss %s, %d(%s)",RQ(r1),d,RQ(r2)); }
R 513 nanojit/NativeX64.cpp void Assembler::MOVSSMR(R r1, I d, R r2) { emitprm(X64_movssmr,r1,d,r2); asm_output("movss %d(%s), %s",d,RQ(r1),RQ(r2)); }
R 519 nanojit/NativeX64.cpp void Assembler::JMPX(R indexreg, NIns** table) { emitrxb_imm(X64_jmpx, (R)0, indexreg, (Register)5, (int32_t)(uintptr_t)table); asm_output("jmpq [%s*8 + %p]", RQ(indexreg), (void*)table); }
R 521 nanojit/NativeX64.cpp void Assembler::JMPXB(R indexreg, R tablereg) { emitxb(X64_jmpxb, indexreg, tablereg); asm_output("jmp [%s*8 + %s]", RQ(indexreg), RQ(tablereg)); }
R 574 nanojit/NativeX64.cpp void Assembler::MOVQSPR(I d, R r) { emit(X64_movqspr | U64(d) << 56 | U64((r&7)<<3) << 40 | U64((r&8)>>1) << 24); asm_output("movq %d(rsp), %s", d, RQ(r)); } // insert r into mod/rm and rex bytes
R 576 nanojit/NativeX64.cpp void Assembler::XORPSA(R r, I32 i32) { emitxm_abs(X64_xorpsa, r, i32); asm_output("xorps %s, (0x%x)",RQ(r), i32); }
R 577 nanojit/NativeX64.cpp void Assembler::XORPSM(R r, NIns* a64) { emitxm_rel(X64_xorpsm, r, a64); asm_output("xorps %s, (%p)", RQ(r), a64); }
R 579 nanojit/NativeX64.cpp void Assembler::X86_AND8R(R r) { emit(X86_and8r | U64(r<<3|(r|4))<<56); asm_output("andb %s, %s", RB(r), RBhi(r)); }
R 580 nanojit/NativeX64.cpp void Assembler::X86_SETNP(R r) { emit(X86_setnp | U64(r|4)<<56); asm_output("setnp %s", RBhi(r)); }
R 581 nanojit/NativeX64.cpp void Assembler::X86_SETE(R r) { emit(X86_sete | U64(r)<<56); asm_output("sete %s", RB(r)); }