/* * Copyright (C) 2008, 2009, 2013 Apple Inc. All rights reserved. * Copyright (C) 2008 Cameron Zwarich * Copyright (C) Research In Motion Limited 2010, 2011. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of * its contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef JITStubsX86_h #define JITStubsX86_h #include "JITStubsX86Common.h" #if !CPU(X86) #error "JITStubsX86.h should only be #included if CPU(X86)" #endif #if !USE(JSVALUE32_64) #error "JITStubsX86.h only implements USE(JSVALUE32_64)" #endif namespace JSC { #if COMPILER(GCC) #if USE(MASM_PROBE) asm ( ".globl " SYMBOL_STRING(ctiMasmProbeTrampoline) "\n" HIDE_SYMBOL(ctiMasmProbeTrampoline) "\n" SYMBOL_STRING(ctiMasmProbeTrampoline) ":" "\n" "pushfd" "\n" // MacroAssembler::probe() has already generated code to store some values. // Together with the eflags pushed above, the top of stack now looks like // this: // esp[0 * ptrSize]: eflags // esp[1 * ptrSize]: return address / saved eip // esp[2 * ptrSize]: probeFunction // esp[3 * ptrSize]: arg1 // esp[4 * ptrSize]: arg2 // esp[5 * ptrSize]: saved eax // esp[6 * ptrSize]: saved esp "movl %esp, %eax" "\n" "subl $" STRINGIZE_VALUE_OF(PROBE_SIZE) ", %esp" "\n" // The X86_64 ABI specifies that the worse case stack alignment requirement // is 32 bytes. "andl $~0x1f, %esp" "\n" "movl %ebp, " STRINGIZE_VALUE_OF(PROBE_CPU_EBP_OFFSET) "(%esp)" "\n" "movl %esp, %ebp" "\n" // Save the ProbeContext*. "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_ECX_OFFSET) "(%ebp)" "\n" "movl %edx, " STRINGIZE_VALUE_OF(PROBE_CPU_EDX_OFFSET) "(%ebp)" "\n" "movl %ebx, " STRINGIZE_VALUE_OF(PROBE_CPU_EBX_OFFSET) "(%ebp)" "\n" "movl %esi, " STRINGIZE_VALUE_OF(PROBE_CPU_ESI_OFFSET) "(%ebp)" "\n" "movl %edi, " STRINGIZE_VALUE_OF(PROBE_CPU_EDI_OFFSET) "(%ebp)" "\n" "movl 0 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EFLAGS_OFFSET) "(%ebp)" "\n" "movl 1 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EIP_OFFSET) "(%ebp)" "\n" "movl 2 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_PROBE_FUNCTION_OFFSET) "(%ebp)" "\n" "movl 3 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_ARG1_OFFSET) "(%ebp)" "\n" "movl 4 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_ARG2_OFFSET) "(%ebp)" "\n" "movl 5 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EAX_OFFSET) "(%ebp)" "\n" "movl 6 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_ESP_OFFSET) "(%ebp)" "\n" "movdqa %xmm0, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM0_OFFSET) "(%ebp)" "\n" "movdqa %xmm1, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM1_OFFSET) "(%ebp)" "\n" "movdqa %xmm2, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM2_OFFSET) "(%ebp)" "\n" "movdqa %xmm3, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM3_OFFSET) "(%ebp)" "\n" "movdqa %xmm4, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM4_OFFSET) "(%ebp)" "\n" "movdqa %xmm5, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM5_OFFSET) "(%ebp)" "\n" "movdqa %xmm6, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM6_OFFSET) "(%ebp)" "\n" "movdqa %xmm7, " STRINGIZE_VALUE_OF(PROBE_CPU_XMM7_OFFSET) "(%ebp)" "\n" // Reserve stack space for the arg while maintaining the required stack // pointer 32 byte alignment: "subl $0x20, %esp" "\n" "movl %ebp, 0(%esp)" "\n" // the ProbeContext* arg. "call *" STRINGIZE_VALUE_OF(PROBE_PROBE_FUNCTION_OFFSET) "(%ebp)" "\n" // To enable probes to modify register state, we copy all registers // out of the ProbeContext before returning. "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EDX_OFFSET) "(%ebp), %edx" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EBX_OFFSET) "(%ebp), %ebx" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_ESI_OFFSET) "(%ebp), %esi" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EDI_OFFSET) "(%ebp), %edi" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM0_OFFSET) "(%ebp), %xmm0" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM1_OFFSET) "(%ebp), %xmm1" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM2_OFFSET) "(%ebp), %xmm2" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM3_OFFSET) "(%ebp), %xmm3" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM4_OFFSET) "(%ebp), %xmm4" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM5_OFFSET) "(%ebp), %xmm5" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM6_OFFSET) "(%ebp), %xmm6" "\n" "movdqa " STRINGIZE_VALUE_OF(PROBE_CPU_XMM7_OFFSET) "(%ebp), %xmm7" "\n" // There are 6 more registers left to restore: // eax, ecx, ebp, esp, eip, and eflags. // We need to handle these last few restores carefully because: // // 1. We need to push the return address on the stack for ret to use. // That means we need to write to the stack. // 2. The user probe function may have altered the restore value of esp to // point to the vicinity of one of the restore values for the remaining // registers left to be restored. // That means, for requirement 1, we may end up writing over some of the // restore values. We can check for this, and first copy the restore // values to a "safe area" on the stack before commencing with the action // for requirement 1. // 3. For requirement 2, we need to ensure that the "safe area" is // protected from interrupt handlers overwriting it. Hence, the esp needs // to be adjusted to include the "safe area" before we start copying the // the restore values. "movl %ebp, %eax" "\n" "addl $" STRINGIZE_VALUE_OF(PROBE_CPU_EFLAGS_OFFSET) ", %eax" "\n" "cmpl %eax, " STRINGIZE_VALUE_OF(PROBE_CPU_ESP_OFFSET) "(%ebp)" "\n" "jg " SYMBOL_STRING(ctiMasmProbeTrampolineEnd) "\n" // Locate the "safe area" at 2x sizeof(ProbeContext) below where the new // rsp will be. This time we don't have to 32-byte align it because we're // not using to store any xmm regs. "movl " STRINGIZE_VALUE_OF(PROBE_CPU_ESP_OFFSET) "(%ebp), %eax" "\n" "subl $2 * " STRINGIZE_VALUE_OF(PROBE_SIZE) ", %eax" "\n" "movl %eax, %esp" "\n" "subl $" STRINGIZE_VALUE_OF(PROBE_CPU_EAX_OFFSET) ", %eax" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EAX_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EAX_OFFSET) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_ECX_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_ECX_OFFSET) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EBP_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EBP_OFFSET) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_ESP_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_ESP_OFFSET) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EIP_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EIP_OFFSET) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EFLAGS_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, " STRINGIZE_VALUE_OF(PROBE_CPU_EFLAGS_OFFSET) "(%eax)" "\n" "movl %eax, %ebp" "\n" SYMBOL_STRING(ctiMasmProbeTrampolineEnd) ":" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_ESP_OFFSET) "(%ebp), %eax" "\n" "subl $5 * " STRINGIZE_VALUE_OF(PTR_SIZE) ", %eax" "\n" // At this point, %esp should be < %eax. "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EFLAGS_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, 0 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EAX_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, 1 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_ECX_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, 2 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EBP_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, 3 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax)" "\n" "movl " STRINGIZE_VALUE_OF(PROBE_CPU_EIP_OFFSET) "(%ebp), %ecx" "\n" "movl %ecx, 4 * " STRINGIZE_VALUE_OF(PTR_SIZE) "(%eax)" "\n" "movl %eax, %esp" "\n" "popfd" "\n" "popl %eax" "\n" "popl %ecx" "\n" "popl %ebp" "\n" "ret" "\n" ); #endif // USE(MASM_PROBE) #endif // COMPILER(GCC) #if COMPILER(MSVC) extern "C" { // FIXME: Since Windows doesn't use the LLInt, we have inline stubs here. // Until the LLInt is changed to support Windows, these stub needs to be updated. __declspec(naked) EncodedJSValue callToJavaScript(void* code, ExecState**, ProtoCallFrame*, Register*) { __asm { mov edx, [esp] push ebp; mov eax, ebp; mov ebp, esp; push esi; push edi; push ebx; sub esp, 0x1c; mov ecx, dword ptr[esp + 0x34]; mov esi, dword ptr[esp + 0x38]; mov ebp, dword ptr[esp + 0x3c]; sub ebp, 0x20; mov dword ptr[ebp + 0x24], 0; mov dword ptr[ebp + 0x20], 0; mov dword ptr[ebp + 0x1c], 0; mov dword ptr[ebp + 0x18], ecx; mov ebx, [ecx]; mov dword ptr[ebp + 0x14], 0; mov dword ptr[ebp + 0x10], ebx; mov dword ptr[ebp + 0xc], 0; mov dword ptr[ebp + 0x8], 1; mov dword ptr[ebp + 0x4], edx; mov dword ptr[ebp], eax; mov eax, ebp; mov edx, dword ptr[esi + 0x28]; add edx, 5; sal edx, 3; sub ebp, edx; mov dword ptr[ebp], eax; mov eax, 5; copyHeaderLoop: sub eax, 1; mov ecx, dword ptr[esi + eax * 8]; mov dword ptr 8[ebp + eax * 8], ecx; mov ecx, dword ptr 4[esi + eax * 8]; mov dword ptr 12[ebp + eax * 8], ecx; test eax, eax; jnz copyHeaderLoop; mov edx, dword ptr[esi + 0x18]; sub edx, 1; mov ecx, dword ptr[esi + 0x28]; sub ecx, 1; cmp edx, ecx; je copyArgs; xor eax, eax; mov ebx, -4; fillExtraArgsLoop: sub ecx, 1; mov dword ptr 0x30[ebp + ecx * 8], eax; mov dword ptr 0x34[ebp + ecx * 8], ebx; cmp edx, ecx; jne fillExtraArgsLoop; copyArgs: mov eax, dword ptr[esi + 0x2c]; copyArgsLoop: test edx, edx; jz copyArgsDone; sub edx, 1; mov ecx, dword ptr 0[eax + edx * 8]; mov ebx, dword ptr 4[eax + edx * 8]; mov dword ptr 0x30[ebp + edx * 8], ecx; mov dword ptr 0x34[ebp + edx * 8], ebx; jmp copyArgsLoop; copyArgsDone: mov ecx, dword ptr[esp + 0x34]; mov dword ptr[ecx], ebp; call dword ptr[esp + 0x30]; cmp dword ptr[ebp + 8], 1; je calleeFramePopped; mov ebp, dword ptr[ebp]; calleeFramePopped: mov ecx, dword ptr[ebp + 0x18]; mov ebx, dword ptr[ebp + 0x10]; mov dword ptr[ecx], ebx; add esp, 0x1c; pop ebx; pop edi; pop esi; pop ebp; ret; } } __declspec(naked) void returnFromJavaScript() { __asm { add esp, 0x1c; pop ebx; pop edi; pop esi; pop ebp; ret; } } __declspec(naked) EncodedJSValue callToNativeFunction(void* code, ExecState**, ProtoCallFrame*, Register*) { __asm { mov edx, [esp] push ebp; mov eax, ebp; mov ebp, esp; push esi; push edi; push ebx; sub esp, 0x1c; mov ecx, [esp + 0x34]; mov esi, [esp + 0x38]; mov ebp, [esp + 0x3c]; sub ebp, 0x20; mov dword ptr[ebp + 0x24], 0; mov dword ptr[ebp + 0x20], 0; mov dword ptr[ebp + 0x1c], 0; mov dword ptr[ebp + 0x18], ecx; mov ebx, [ecx]; mov dword ptr[ebp + 0x14], 0; mov dword ptr[ebp + 0x10], ebx; mov dword ptr[ebp + 0xc], 0; mov dword ptr[ebp + 0x8], 1; mov dword ptr[ebp + 0x4], edx; mov dword ptr[ebp], eax; mov eax, ebp; mov edx, dword ptr[esi + 0x28]; add edx, 5; sal edx, 3; sub ebp, edx; mov dword ptr[ebp], eax; mov eax, 5; copyHeaderLoop: sub eax, 1; mov ecx, dword ptr[esi + eax * 8]; mov dword ptr 8[ebp + eax * 8], ecx; mov ecx, dword ptr 4[esi + eax * 8]; mov dword ptr 12[ebp + eax * 8], ecx; test eax, eax; jnz copyHeaderLoop; mov edx, dword ptr[esi + 0x18]; sub edx, 1; mov ecx, dword ptr[esi + 0x28]; sub ecx, 1; cmp edx, ecx; je copyArgs; xor eax, eax; mov ebx, -4; fillExtraArgsLoop: sub ecx, 1; mov dword ptr 0x30[ebp + ecx * 8], eax; mov dword ptr 0x34[ebp + ecx * 8], ebx; cmp edx, ecx; jne fillExtraArgsLoop; copyArgs: mov eax, dword ptr[esi + 0x2c]; copyArgsLoop: test edx, edx; jz copyArgsDone; sub edx, 1; mov ecx, dword ptr 0[eax + edx * 8]; mov ebx, dword ptr 4[eax + edx * 8]; mov dword ptr 0x30[ebp + edx * 8], ecx; mov dword ptr 0x34[ebp + edx * 8], ebx; jmp copyArgsLoop; copyArgsDone: mov ecx, dword ptr[esp + 0x34]; mov dword ptr[ecx], ebp; mov edi, dword ptr[esp + 0x30]; mov dword ptr[esp + 0x30], ebp; mov ecx, ebp; call edi; cmp dword ptr[ebp + 8], 1; je calleeFramePopped; mov ebp, dword ptr[ebp]; calleeFramePopped: mov ecx, dword ptr[ebp + 0x18]; mov ebx, dword ptr[ebp + 0x10]; mov dword ptr[ecx], ebx; add esp, 0x1c; pop ebx; pop edi; pop esi; pop ebp; ret; } } } #endif // COMPILER(MSVC) } // namespace JSC #endif // JITStubsX86_h