2022-10-14 01:59:53 +03:00
|
|
|
// This file is part of the Luau programming language and is licensed under MIT License; see LICENSE.txt for details
|
|
|
|
#include "EmitInstructionX64.h"
|
|
|
|
|
|
|
|
#include "Luau/AssemblyBuilderX64.h"
|
2023-03-31 15:21:14 +03:00
|
|
|
#include "Luau/IrRegAllocX64.h"
|
2022-10-14 01:59:53 +03:00
|
|
|
|
|
|
|
#include "CustomExecUtils.h"
|
|
|
|
#include "EmitCommonX64.h"
|
|
|
|
|
|
|
|
namespace Luau
|
|
|
|
{
|
|
|
|
namespace CodeGen
|
|
|
|
{
|
2023-03-03 15:45:38 +02:00
|
|
|
namespace X64
|
|
|
|
{
|
2022-10-14 01:59:53 +03:00
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
void emitInstCall(AssemblyBuilderX64& build, ModuleHelpers& helpers, int ra, int nparams, int nresults)
|
2022-11-04 10:02:37 -07:00
|
|
|
{
|
|
|
|
build.mov(rArg1, rState);
|
|
|
|
build.lea(rArg2, luauRegAddress(ra));
|
|
|
|
|
|
|
|
if (nparams == LUA_MULTRET)
|
|
|
|
build.mov(rArg3, qword[rState + offsetof(lua_State, top)]);
|
|
|
|
else
|
|
|
|
build.lea(rArg3, luauRegAddress(ra + 1 + nparams));
|
|
|
|
|
|
|
|
build.mov(dwordReg(rArg4), nresults);
|
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, callProlog)]);
|
|
|
|
RegisterX64 ccl = rax; // Returned from callProlog
|
|
|
|
|
|
|
|
emitUpdateBase(build);
|
|
|
|
|
|
|
|
Label cFuncCall;
|
|
|
|
|
|
|
|
build.test(byte[ccl + offsetof(Closure, isC)], 1);
|
|
|
|
build.jcc(ConditionX64::NotZero, cFuncCall);
|
|
|
|
|
|
|
|
{
|
|
|
|
RegisterX64 proto = rcx; // Sync with emitContinueCallInVm
|
|
|
|
RegisterX64 ci = rdx;
|
|
|
|
RegisterX64 argi = rsi;
|
|
|
|
RegisterX64 argend = rdi;
|
|
|
|
|
|
|
|
build.mov(proto, qword[ccl + offsetof(Closure, l.p)]);
|
|
|
|
|
|
|
|
// Switch current Closure
|
|
|
|
build.mov(sClosure, ccl); // Last use of 'ccl'
|
|
|
|
|
|
|
|
build.mov(ci, qword[rState + offsetof(lua_State, ci)]);
|
|
|
|
|
|
|
|
Label fillnil, exitfillnil;
|
|
|
|
|
|
|
|
// argi = L->top
|
|
|
|
build.mov(argi, qword[rState + offsetof(lua_State, top)]);
|
|
|
|
|
|
|
|
// argend = L->base + p->numparams
|
|
|
|
build.movzx(eax, byte[proto + offsetof(Proto, numparams)]);
|
|
|
|
build.shl(eax, kTValueSizeLog2);
|
|
|
|
build.lea(argend, addr[rBase + rax]);
|
|
|
|
|
|
|
|
// while (argi < argend) setnilvalue(argi++);
|
|
|
|
build.setLabel(fillnil);
|
|
|
|
build.cmp(argi, argend);
|
|
|
|
build.jcc(ConditionX64::NotBelow, exitfillnil);
|
|
|
|
|
|
|
|
build.mov(dword[argi + offsetof(TValue, tt)], LUA_TNIL);
|
|
|
|
build.add(argi, sizeof(TValue));
|
|
|
|
build.jmp(fillnil); // This loop rarely runs so it's not worth repeating cmp/jcc
|
|
|
|
|
|
|
|
build.setLabel(exitfillnil);
|
|
|
|
|
|
|
|
// Set L->top to ci->top as most function expect (no vararg)
|
|
|
|
build.mov(rax, qword[ci + offsetof(CallInfo, top)]);
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], rax);
|
|
|
|
|
|
|
|
build.mov(rax, qword[proto + offsetofProtoExecData]); // We'll need this value later
|
|
|
|
|
|
|
|
// But if it is vararg, update it to 'argi'
|
|
|
|
Label skipVararg;
|
|
|
|
|
|
|
|
build.test(byte[proto + offsetof(Proto, is_vararg)], 1);
|
|
|
|
build.jcc(ConditionX64::Zero, skipVararg);
|
|
|
|
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], argi);
|
|
|
|
build.setLabel(skipVararg);
|
|
|
|
|
|
|
|
// Check native function data
|
|
|
|
build.test(rax, rax);
|
|
|
|
build.jcc(ConditionX64::Zero, helpers.continueCallInVm);
|
|
|
|
|
|
|
|
// Switch current constants
|
|
|
|
build.mov(rConstants, qword[proto + offsetof(Proto, k)]);
|
|
|
|
|
|
|
|
// Switch current code
|
|
|
|
build.mov(rdx, qword[proto + offsetof(Proto, code)]);
|
|
|
|
build.mov(sCode, rdx);
|
|
|
|
|
|
|
|
build.jmp(qword[rax + offsetof(NativeProto, entryTarget)]);
|
|
|
|
}
|
|
|
|
|
|
|
|
build.setLabel(cFuncCall);
|
|
|
|
|
|
|
|
{
|
|
|
|
// results = ccl->c.f(L);
|
|
|
|
build.mov(rArg1, rState);
|
|
|
|
build.call(qword[ccl + offsetof(Closure, c.f)]); // Last use of 'ccl'
|
|
|
|
RegisterX64 results = eax;
|
|
|
|
|
|
|
|
build.test(results, results); // test here will set SF=1 for a negative number and it always sets OF to 0
|
|
|
|
build.jcc(ConditionX64::Less, helpers.exitNoContinueVm); // jl jumps if SF != OF
|
|
|
|
|
|
|
|
// We have special handling for small number of expected results below
|
|
|
|
if (nresults != 0 && nresults != 1)
|
|
|
|
{
|
|
|
|
build.mov(rArg1, rState);
|
|
|
|
build.mov(dwordReg(rArg2), nresults);
|
|
|
|
build.mov(dwordReg(rArg3), results);
|
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, callEpilogC)]);
|
|
|
|
|
|
|
|
emitUpdateBase(build);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
RegisterX64 ci = rdx;
|
|
|
|
RegisterX64 cip = rcx;
|
|
|
|
RegisterX64 vali = rsi;
|
|
|
|
|
|
|
|
build.mov(ci, qword[rState + offsetof(lua_State, ci)]);
|
|
|
|
build.lea(cip, addr[ci - sizeof(CallInfo)]);
|
|
|
|
|
|
|
|
// L->base = cip->base
|
|
|
|
build.mov(rBase, qword[cip + offsetof(CallInfo, base)]);
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, base)], rBase);
|
|
|
|
|
|
|
|
if (nresults == 1)
|
|
|
|
{
|
|
|
|
// Opportunistically copy the result we expected from (L->top - results)
|
|
|
|
build.mov(vali, qword[rState + offsetof(lua_State, top)]);
|
|
|
|
build.shl(results, kTValueSizeLog2);
|
|
|
|
build.sub(vali, qwordReg(results));
|
|
|
|
build.vmovups(xmm0, xmmword[vali]);
|
|
|
|
build.vmovups(luauReg(ra), xmm0);
|
|
|
|
|
|
|
|
Label skipnil;
|
|
|
|
|
|
|
|
// If there was no result, override the value with 'nil'
|
|
|
|
build.test(results, results);
|
|
|
|
build.jcc(ConditionX64::NotZero, skipnil);
|
|
|
|
build.mov(luauRegTag(ra), LUA_TNIL);
|
|
|
|
build.setLabel(skipnil);
|
|
|
|
}
|
|
|
|
|
|
|
|
// L->ci = cip
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, ci)], cip);
|
|
|
|
|
|
|
|
// L->top = cip->top
|
|
|
|
build.mov(rax, qword[cip + offsetof(CallInfo, top)]);
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], rax);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
void emitInstReturn(AssemblyBuilderX64& build, ModuleHelpers& helpers, int ra, int actualResults)
|
2022-10-28 01:22:49 +03:00
|
|
|
{
|
|
|
|
RegisterX64 ci = r8;
|
|
|
|
RegisterX64 cip = r9;
|
|
|
|
RegisterX64 res = rdi;
|
|
|
|
RegisterX64 nresults = esi;
|
|
|
|
|
|
|
|
build.mov(ci, qword[rState + offsetof(lua_State, ci)]);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.lea(cip, addr[ci - sizeof(CallInfo)]);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// res = ci->func; note: we assume CALL always puts func+args and expects results to start at func
|
|
|
|
build.mov(res, qword[ci + offsetof(CallInfo, func)]);
|
|
|
|
// nresults = ci->nresults
|
|
|
|
build.mov(nresults, dword[ci + offsetof(CallInfo, nresults)]);
|
|
|
|
|
|
|
|
{
|
|
|
|
Label skipResultCopy;
|
|
|
|
|
|
|
|
RegisterX64 counter = ecx;
|
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
if (actualResults == 0)
|
2022-10-28 01:22:49 +03:00
|
|
|
{
|
|
|
|
// Our instruction doesn't have any results, so just fill results expected in parent with 'nil'
|
2022-11-04 10:02:37 -07:00
|
|
|
build.test(nresults, nresults); // test here will set SF=1 for a negative number, ZF=1 for zero and OF=0
|
|
|
|
build.jcc(ConditionX64::LessEqual, skipResultCopy); // jle jumps if SF != OF or ZF == 1
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
build.mov(counter, nresults);
|
|
|
|
|
|
|
|
Label repeatNilLoop = build.setLabel();
|
|
|
|
build.mov(dword[res + offsetof(TValue, tt)], LUA_TNIL);
|
|
|
|
build.add(res, sizeof(TValue));
|
|
|
|
build.dec(counter);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotZero, repeatNilLoop);
|
2022-10-28 01:22:49 +03:00
|
|
|
}
|
2023-03-24 10:34:14 -07:00
|
|
|
else if (actualResults == 1)
|
2022-10-28 01:22:49 +03:00
|
|
|
{
|
|
|
|
// Try setting our 1 result
|
|
|
|
build.test(nresults, nresults);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::Zero, skipResultCopy);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
2022-11-04 10:02:37 -07:00
|
|
|
build.lea(counter, addr[nresults - 1]);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
build.vmovups(xmm0, luauReg(ra));
|
|
|
|
build.vmovups(xmmword[res], xmm0);
|
|
|
|
build.add(res, sizeof(TValue));
|
|
|
|
|
|
|
|
// Fill the rest of the expected results with 'nil'
|
2022-11-04 10:02:37 -07:00
|
|
|
build.test(counter, counter); // test here will set SF=1 for a negative number, ZF=1 for zero and OF=0
|
|
|
|
build.jcc(ConditionX64::LessEqual, skipResultCopy); // jle jumps if SF != OF or ZF == 1
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
Label repeatNilLoop = build.setLabel();
|
|
|
|
build.mov(dword[res + offsetof(TValue, tt)], LUA_TNIL);
|
|
|
|
build.add(res, sizeof(TValue));
|
|
|
|
build.dec(counter);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotZero, repeatNilLoop);
|
2022-10-28 01:22:49 +03:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
RegisterX64 vali = rax;
|
|
|
|
RegisterX64 valend = rdx;
|
|
|
|
|
|
|
|
// Copy return values into parent stack (but only up to nresults!)
|
|
|
|
build.test(nresults, nresults);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::Zero, skipResultCopy);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// vali = ra
|
2022-11-04 10:02:37 -07:00
|
|
|
build.lea(vali, luauRegAddress(ra));
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// Copy as much as possible for MULTRET calls, and only as much as needed otherwise
|
2023-03-24 10:34:14 -07:00
|
|
|
if (actualResults == LUA_MULTRET)
|
2022-10-28 01:22:49 +03:00
|
|
|
build.mov(valend, qword[rState + offsetof(lua_State, top)]); // valend = L->top
|
|
|
|
else
|
2023-03-24 10:34:14 -07:00
|
|
|
build.lea(valend, luauRegAddress(ra + actualResults)); // valend = ra + actualResults
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
build.mov(counter, nresults);
|
|
|
|
|
|
|
|
Label repeatValueLoop, exitValueLoop;
|
|
|
|
|
|
|
|
build.setLabel(repeatValueLoop);
|
|
|
|
build.cmp(vali, valend);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotBelow, exitValueLoop);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
build.vmovups(xmm0, xmmword[vali]);
|
|
|
|
build.vmovups(xmmword[res], xmm0);
|
|
|
|
build.add(vali, sizeof(TValue));
|
|
|
|
build.add(res, sizeof(TValue));
|
|
|
|
build.dec(counter);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotZero, repeatValueLoop);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
build.setLabel(exitValueLoop);
|
|
|
|
|
|
|
|
// Fill the rest of the expected results with 'nil'
|
2022-11-04 10:02:37 -07:00
|
|
|
build.test(counter, counter); // test here will set SF=1 for a negative number, ZF=1 for zero and OF=0
|
|
|
|
build.jcc(ConditionX64::LessEqual, skipResultCopy); // jle jumps if SF != OF or ZF == 1
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
Label repeatNilLoop = build.setLabel();
|
|
|
|
build.mov(dword[res + offsetof(TValue, tt)], LUA_TNIL);
|
|
|
|
build.add(res, sizeof(TValue));
|
|
|
|
build.dec(counter);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotZero, repeatNilLoop);
|
2022-10-28 01:22:49 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
build.setLabel(skipResultCopy);
|
|
|
|
}
|
|
|
|
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, ci)], cip); // L->ci = cip
|
|
|
|
build.mov(rBase, qword[cip + offsetof(CallInfo, base)]); // sync base = L->base while we have a chance
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, base)], rBase); // L->base = cip->base
|
|
|
|
|
|
|
|
// Start with result for LUA_MULTRET/exit value
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], res); // L->top = res
|
|
|
|
|
|
|
|
// Unlikely, but this might be the last return from VM
|
|
|
|
build.test(byte[ci + offsetof(CallInfo, flags)], LUA_CALLINFO_RETURN);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotZero, helpers.exitNoContinueVm);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
Label skipFixedRetTop;
|
2022-11-04 10:02:37 -07:00
|
|
|
build.test(nresults, nresults); // test here will set SF=1 for a negative number and it always sets OF to 0
|
|
|
|
build.jcc(ConditionX64::Less, skipFixedRetTop); // jl jumps if SF != OF
|
2022-10-28 01:22:49 +03:00
|
|
|
build.mov(rax, qword[cip + offsetof(CallInfo, top)]);
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], rax); // L->top = cip->top
|
|
|
|
build.setLabel(skipFixedRetTop);
|
|
|
|
|
|
|
|
// Returning back to the previous function is a bit tricky
|
|
|
|
// Registers alive: r9 (cip)
|
|
|
|
RegisterX64 proto = rcx;
|
|
|
|
RegisterX64 execdata = rbx;
|
|
|
|
|
|
|
|
// Change closure
|
|
|
|
build.mov(rax, qword[cip + offsetof(CallInfo, func)]);
|
|
|
|
build.mov(rax, qword[rax + offsetof(TValue, value.gc)]);
|
|
|
|
build.mov(sClosure, rax);
|
|
|
|
|
|
|
|
build.mov(proto, qword[rax + offsetof(Closure, l.p)]);
|
|
|
|
|
|
|
|
build.mov(execdata, qword[proto + offsetofProtoExecData]);
|
|
|
|
build.test(execdata, execdata);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::Zero, helpers.exitContinueVm); // Continue in interpreter if function has no native data
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// Change constants
|
|
|
|
build.mov(rConstants, qword[proto + offsetof(Proto, k)]);
|
|
|
|
|
|
|
|
// Change code
|
|
|
|
build.mov(rdx, qword[proto + offsetof(Proto, code)]);
|
|
|
|
build.mov(sCode, rdx);
|
|
|
|
|
|
|
|
build.mov(rax, qword[cip + offsetof(CallInfo, savedpc)]);
|
|
|
|
|
|
|
|
// To get instruction index from instruction pointer, we need to divide byte offset by 4
|
|
|
|
// But we will actually need to scale instruction index by 8 back to byte offset later so it cancels out
|
|
|
|
build.sub(rax, rdx);
|
|
|
|
|
|
|
|
// Get new instruction location and jump to it
|
|
|
|
build.mov(rdx, qword[execdata + offsetof(NativeProto, instTargets)]);
|
|
|
|
build.jmp(qword[rdx + rax * 2]);
|
|
|
|
}
|
|
|
|
|
2023-04-07 12:56:27 -07:00
|
|
|
void emitInstSetList(IrRegAllocX64& regs, AssemblyBuilderX64& build, int ra, int rb, int count, uint32_t index)
|
2022-10-21 10:33:43 -07:00
|
|
|
{
|
2023-03-24 10:34:14 -07:00
|
|
|
OperandX64 last = index + count - 1;
|
2022-10-21 10:33:43 -07:00
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
// Using non-volatile 'rbx' for dynamic 'count' value (for LUA_MULTRET) to skip later recomputation
|
|
|
|
// We also keep 'count' scaled by sizeof(TValue) here as it helps in the loop below
|
2022-10-21 10:33:43 -07:00
|
|
|
RegisterX64 cscaled = rbx;
|
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
if (count == LUA_MULTRET)
|
2022-10-21 10:33:43 -07:00
|
|
|
{
|
|
|
|
RegisterX64 tmp = rax;
|
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
// count = L->top - rb
|
2022-10-21 10:33:43 -07:00
|
|
|
build.mov(cscaled, qword[rState + offsetof(lua_State, top)]);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.lea(tmp, luauRegAddress(rb));
|
2022-10-21 10:33:43 -07:00
|
|
|
build.sub(cscaled, tmp); // Using byte difference
|
|
|
|
|
|
|
|
// L->top = L->ci->top
|
|
|
|
build.mov(tmp, qword[rState + offsetof(lua_State, ci)]);
|
|
|
|
build.mov(tmp, qword[tmp + offsetof(CallInfo, top)]);
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], tmp);
|
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
// last = index + count - 1;
|
2022-10-21 10:33:43 -07:00
|
|
|
last = edx;
|
|
|
|
build.mov(last, dwordReg(cscaled));
|
|
|
|
build.shr(last, kTValueSizeLog2);
|
|
|
|
build.add(last, index - 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
Label skipResize;
|
|
|
|
|
2023-04-07 12:56:27 -07:00
|
|
|
RegisterX64 table = regs.takeReg(rax, kInvalidInstIdx);
|
2022-10-21 10:33:43 -07:00
|
|
|
|
|
|
|
build.mov(table, luauRegValue(ra));
|
|
|
|
|
|
|
|
// Resize if h->sizearray < last
|
|
|
|
build.cmp(dword[table + offsetof(Table, sizearray)], last);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotBelow, skipResize);
|
2022-10-21 10:33:43 -07:00
|
|
|
|
|
|
|
// Argument setup reordered to avoid conflicts
|
|
|
|
LUAU_ASSERT(rArg3 != table);
|
|
|
|
build.mov(dwordReg(rArg3), last);
|
|
|
|
build.mov(rArg2, table);
|
|
|
|
build.mov(rArg1, rState);
|
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, luaH_resizearray)]);
|
|
|
|
build.mov(table, luauRegValue(ra)); // Reload cloberred register value
|
|
|
|
|
|
|
|
build.setLabel(skipResize);
|
|
|
|
|
|
|
|
RegisterX64 arrayDst = rdx;
|
|
|
|
RegisterX64 offset = rcx;
|
|
|
|
|
|
|
|
build.mov(arrayDst, qword[table + offsetof(Table, array)]);
|
|
|
|
|
|
|
|
const int kUnrollSetListLimit = 4;
|
|
|
|
|
2023-03-24 10:34:14 -07:00
|
|
|
if (count != LUA_MULTRET && count <= kUnrollSetListLimit)
|
2022-10-21 10:33:43 -07:00
|
|
|
{
|
2023-03-24 10:34:14 -07:00
|
|
|
for (int i = 0; i < count; ++i)
|
2022-10-21 10:33:43 -07:00
|
|
|
{
|
|
|
|
// setobj2t(L, &array[index + i - 1], rb + i);
|
|
|
|
build.vmovups(xmm0, luauRegValue(rb + i));
|
|
|
|
build.vmovups(xmmword[arrayDst + (index + i - 1) * sizeof(TValue)], xmm0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2023-03-24 10:34:14 -07:00
|
|
|
LUAU_ASSERT(count != 0);
|
2022-10-21 10:33:43 -07:00
|
|
|
|
|
|
|
build.xor_(offset, offset);
|
|
|
|
if (index != 1)
|
|
|
|
build.add(arrayDst, (index - 1) * sizeof(TValue));
|
|
|
|
|
|
|
|
Label repeatLoop, endLoop;
|
2023-03-24 10:34:14 -07:00
|
|
|
OperandX64 limit = count == LUA_MULTRET ? cscaled : OperandX64(count * sizeof(TValue));
|
2022-10-21 10:33:43 -07:00
|
|
|
|
|
|
|
// If c is static, we will always do at least one iteration
|
2023-03-24 10:34:14 -07:00
|
|
|
if (count == LUA_MULTRET)
|
2022-10-21 10:33:43 -07:00
|
|
|
{
|
|
|
|
build.cmp(offset, limit);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotBelow, endLoop);
|
2022-10-21 10:33:43 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
build.setLabel(repeatLoop);
|
|
|
|
|
|
|
|
// setobj2t(L, &array[index + i - 1], rb + i);
|
|
|
|
build.vmovups(xmm0, xmmword[offset + rBase + rb * sizeof(TValue)]); // luauReg(rb) unwrapped to add offset
|
|
|
|
build.vmovups(xmmword[offset + arrayDst], xmm0);
|
|
|
|
|
|
|
|
build.add(offset, sizeof(TValue));
|
|
|
|
build.cmp(offset, limit);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::Below, repeatLoop);
|
2022-10-21 10:33:43 -07:00
|
|
|
|
|
|
|
build.setLabel(endLoop);
|
|
|
|
}
|
|
|
|
|
2023-04-07 12:56:27 -07:00
|
|
|
callBarrierTableFast(regs, build, table, {});
|
2022-10-14 01:59:53 +03:00
|
|
|
}
|
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
void emitinstForGLoop(AssemblyBuilderX64& build, int ra, int aux, Label& loopRepeat, Label& loopExit)
|
2022-10-14 01:59:53 +03:00
|
|
|
{
|
2023-03-03 15:45:38 +02:00
|
|
|
// ipairs-style traversal is handled in IR
|
|
|
|
LUAU_ASSERT(aux >= 0);
|
2022-10-14 01:59:53 +03:00
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
// This is a fast-path for builtin table iteration, tag check for 'ra' has to be performed before emitting this instruction
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// Registers are chosen in this way to simplify fallback code for the node part
|
|
|
|
RegisterX64 table = rArg2;
|
|
|
|
RegisterX64 index = rArg3;
|
|
|
|
RegisterX64 elemPtr = rax;
|
|
|
|
|
|
|
|
build.mov(table, luauRegValue(ra + 1));
|
|
|
|
build.mov(index, luauRegValue(ra + 2));
|
|
|
|
|
|
|
|
// &array[index]
|
|
|
|
build.mov(dwordReg(elemPtr), dwordReg(index));
|
|
|
|
build.shl(dwordReg(elemPtr), kTValueSizeLog2);
|
|
|
|
build.add(elemPtr, qword[table + offsetof(Table, array)]);
|
|
|
|
|
|
|
|
// Clear extra variables since we might have more than two
|
|
|
|
for (int i = 2; i < aux; ++i)
|
|
|
|
build.mov(luauRegTag(ra + 3 + i), LUA_TNIL);
|
|
|
|
|
|
|
|
Label skipArray, skipArrayNil;
|
|
|
|
|
|
|
|
// First we advance index through the array portion
|
|
|
|
// while (unsigned(index) < unsigned(sizearray))
|
|
|
|
Label arrayLoop = build.setLabel();
|
|
|
|
build.cmp(dwordReg(index), dword[table + offsetof(Table, sizearray)]);
|
2023-03-03 15:45:38 +02:00
|
|
|
build.jcc(ConditionX64::NotBelow, skipArray);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// If element is nil, we increment the index; if it's not, we still need 'index + 1' inside
|
|
|
|
build.inc(index);
|
|
|
|
|
|
|
|
build.cmp(dword[elemPtr + offsetof(TValue, tt)], LUA_TNIL);
|
2023-03-03 15:45:38 +02:00
|
|
|
build.jcc(ConditionX64::Equal, skipArrayNil);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
|
|
|
// setpvalue(ra + 2, reinterpret_cast<void*>(uintptr_t(index + 1)));
|
|
|
|
build.mov(luauRegValue(ra + 2), index);
|
|
|
|
// Tag should already be set to lightuserdata
|
|
|
|
|
|
|
|
// setnvalue(ra + 3, double(index + 1));
|
|
|
|
build.vcvtsi2sd(xmm0, xmm0, dwordReg(index));
|
|
|
|
build.vmovsd(luauRegValue(ra + 3), xmm0);
|
|
|
|
build.mov(luauRegTag(ra + 3), LUA_TNUMBER);
|
|
|
|
|
|
|
|
// setobj2s(L, ra + 4, e);
|
|
|
|
setLuauReg(build, xmm2, ra + 4, xmmword[elemPtr]);
|
|
|
|
|
|
|
|
build.jmp(loopRepeat);
|
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
build.setLabel(skipArrayNil);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
// Index already incremented, advance to next array element
|
|
|
|
build.add(elemPtr, sizeof(TValue));
|
|
|
|
build.jmp(arrayLoop);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
build.setLabel(skipArray);
|
2022-10-28 01:22:49 +03:00
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
// Call helper to assign next node value or to signal loop exit
|
|
|
|
build.mov(rArg1, rState);
|
|
|
|
// rArg2 and rArg3 are already set
|
|
|
|
build.lea(rArg4, luauRegAddress(ra));
|
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, forgLoopNodeIter)]);
|
|
|
|
build.test(al, al);
|
|
|
|
build.jcc(ConditionX64::NotZero, loopRepeat);
|
2022-10-28 01:22:49 +03:00
|
|
|
}
|
|
|
|
|
2023-03-31 15:21:14 +03:00
|
|
|
void emitinstForGLoopFallback(AssemblyBuilderX64& build, int ra, int aux, Label& loopRepeat)
|
2022-10-28 01:22:49 +03:00
|
|
|
{
|
|
|
|
build.mov(rArg1, rState);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.mov(dwordReg(rArg2), ra);
|
|
|
|
build.mov(dwordReg(rArg3), aux);
|
2022-10-28 01:22:49 +03:00
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, forgLoopNonTableFallback)]);
|
|
|
|
emitUpdateBase(build);
|
|
|
|
build.test(al, al);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.jcc(ConditionX64::NotZero, loopRepeat);
|
2022-10-28 01:22:49 +03:00
|
|
|
}
|
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
void emitInstForGPrepXnextFallback(AssemblyBuilderX64& build, int pcpos, int ra, Label& target)
|
2022-10-28 01:22:49 +03:00
|
|
|
{
|
|
|
|
build.mov(rArg1, rState);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.lea(rArg2, luauRegAddress(ra));
|
|
|
|
build.mov(dwordReg(rArg3), pcpos + 1);
|
2022-10-28 01:22:49 +03:00
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, forgPrepXnextFallback)]);
|
|
|
|
build.jmp(target);
|
|
|
|
}
|
|
|
|
|
2022-12-09 10:07:25 -08:00
|
|
|
void emitInstGetImportFallback(AssemblyBuilderX64& build, int ra, uint32_t aux)
|
2022-10-21 10:33:43 -07:00
|
|
|
{
|
|
|
|
build.mov(rax, sClosure);
|
|
|
|
|
|
|
|
// luaV_getimport(L, cl->env, k, aux, /* propagatenil= */ false)
|
|
|
|
build.mov(rArg1, rState);
|
|
|
|
build.mov(rArg2, qword[rax + offsetof(Closure, env)]);
|
|
|
|
build.mov(rArg3, rConstants);
|
2022-11-04 10:02:37 -07:00
|
|
|
build.mov(dwordReg(rArg4), aux);
|
2022-10-21 10:33:43 -07:00
|
|
|
|
|
|
|
if (build.abi == ABIX64::Windows)
|
|
|
|
build.mov(sArg5, 0);
|
|
|
|
else
|
|
|
|
build.xor_(rArg5, rArg5);
|
|
|
|
|
|
|
|
build.call(qword[rNativeContext + offsetof(NativeContext, luaV_getimport)]);
|
|
|
|
|
|
|
|
emitUpdateBase(build);
|
|
|
|
|
|
|
|
// setobj2s(L, ra, L->top - 1)
|
|
|
|
build.mov(rax, qword[rState + offsetof(lua_State, top)]);
|
|
|
|
build.sub(rax, sizeof(TValue));
|
|
|
|
build.vmovups(xmm0, xmmword[rax]);
|
|
|
|
build.vmovups(luauReg(ra), xmm0);
|
|
|
|
|
|
|
|
// L->top--
|
|
|
|
build.mov(qword[rState + offsetof(lua_State, top)], rax);
|
|
|
|
}
|
|
|
|
|
2023-01-13 12:36:28 -08:00
|
|
|
void emitInstCoverage(AssemblyBuilderX64& build, int pcpos)
|
|
|
|
{
|
|
|
|
build.mov(rcx, sCode);
|
|
|
|
build.add(rcx, pcpos * sizeof(Instruction));
|
|
|
|
|
|
|
|
// hits = LUAU_INSN_E(*pc)
|
|
|
|
build.mov(edx, dword[rcx]);
|
|
|
|
build.sar(edx, 8);
|
|
|
|
|
|
|
|
// hits = (hits < (1 << 23) - 1) ? hits + 1 : hits;
|
|
|
|
build.xor_(eax, eax);
|
|
|
|
build.cmp(edx, (1 << 23) - 1);
|
|
|
|
build.setcc(ConditionX64::NotEqual, al);
|
|
|
|
build.add(edx, eax);
|
|
|
|
|
|
|
|
|
|
|
|
// VM_PATCH_E(pc, hits);
|
|
|
|
build.sal(edx, 8);
|
|
|
|
build.movzx(eax, byte[rcx]);
|
|
|
|
build.or_(eax, edx);
|
|
|
|
build.mov(dword[rcx], eax);
|
|
|
|
}
|
|
|
|
|
2023-03-03 15:45:38 +02:00
|
|
|
} // namespace X64
|
2022-10-14 01:59:53 +03:00
|
|
|
} // namespace CodeGen
|
|
|
|
} // namespace Luau
|