Home | History | Annotate | Download | only in X64
      1 #
      2 # ConvertAsm.py: Automatically generated from CopyMem.asm
      3 #
      4 #------------------------------------------------------------------------------
      5 #
      6 # Copyright (c) 2006 - 2009, Intel Corporation. All rights reserved.<BR>
      7 # This program and the accompanying materials
      8 # are licensed and made available under the terms and conditions of the BSD License
      9 # which accompanies this distribution.  The full text of the license may be found at
     10 # http://opensource.org/licenses/bsd-license.php.
     11 #
     12 # THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS,
     13 # WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED.
     14 #
     15 # Module Name:
     16 #
     17 #   CopyMem.S
     18 #
     19 # Abstract:
     20 #
     21 #   CopyMem function
     22 #
     23 # Notes:
     24 #
     25 #------------------------------------------------------------------------------
     26 
     27 
     28 #------------------------------------------------------------------------------
     29 #  VOID *
     30 #  EFIAPI
     31 #  InternalMemCopyMem (
     32 #    IN VOID   *Destination,
     33 #    IN VOID   *Source,
     34 #    IN UINTN  Count
     35 #    )
     36 #------------------------------------------------------------------------------
     37 ASM_GLOBAL ASM_PFX(InternalMemCopyMem)
     38 ASM_PFX(InternalMemCopyMem):
     39     pushq   %rsi
     40     pushq   %rdi
     41     movq    %rdx, %rsi                  # rsi <- Source
     42     movq    %rcx, %rdi                  # rdi <- Destination
     43     leaq    -1(%rsi, %r8,), %r9         # r9 <- Last byte of Source
     44     cmpq    %rdi, %rsi
     45     movq    %rdi, %rax                  # rax <- Destination as return value
     46     jae     L0                          # Copy forward if Source > Destination
     47     cmpq    %rdi, %r9                   # Overlapped?
     48     jae     L_CopyBackward              # Copy backward if overlapped
     49 L0:
     50     xorq    %rcx, %rcx
     51     subq    %rdi, %rcx                  # rcx <- -rdi
     52     andq    $15, %rcx                   # rcx + rsi should be 16 bytes aligned
     53     jz      L1                          # skip if rcx == 0
     54     cmpq    %r8, %rcx
     55     cmova   %r8, %rcx
     56     subq    %rcx, %r8
     57     rep     movsb
     58 L1:
     59     movq    %r8, %rcx
     60     andq    $15, %r8
     61     shrq    $4, %rcx                    # rcx <- # of DQwords to copy
     62     jz      L_CopyBytes
     63     movdqa  %xmm0, 0x18(%rsp)           # save xmm0 on stack
     64 L2:
     65     movdqu  (%rsi), %xmm0               # rsi may not be 16-byte aligned
     66     movntdq  %xmm0, (%rdi)              # rdi should be 16-byte aligned
     67     addq    $16, %rsi
     68     addq    $16, %rdi
     69     loop    L2
     70     mfence
     71     movdqa  0x18(%rsp), %xmm0           # restore xmm0
     72     jmp     L_CopyBytes                 # copy remaining bytes
     73 L_CopyBackward:
     74     movq    %r9, %rsi                   # rsi <- Last byte of Source
     75     leaq    -1(%rdi, %r8,), %rdi        # rdi <- Last byte of Destination
     76     std
     77 L_CopyBytes:
     78     movq    %r8, %rcx
     79     rep     movsb
     80     cld
     81     popq    %rdi
     82     popq    %rsi
     83     ret
     84