• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1#
2# ConvertAsm.py: Automatically generated from CopyMem.asm
3#
4#------------------------------------------------------------------------------
5#
6# Copyright (c) 2006 - 2009, Intel Corporation. All rights reserved.<BR>
7# This program and the accompanying materials
8# are licensed and made available under the terms and conditions of the BSD License
9# which accompanies this distribution.  The full text of the license may be found at
10# http://opensource.org/licenses/bsd-license.php.
11#
12# THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS,
13# WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED.
14#
15# Module Name:
16#
17#   CopyMem.S
18#
19# Abstract:
20#
21#   CopyMem function
22#
23# Notes:
24#
25#------------------------------------------------------------------------------
26
27
28#------------------------------------------------------------------------------
29#  VOID *
30#  EFIAPI
31#  InternalMemCopyMem (
32#    IN VOID   *Destination,
33#    IN VOID   *Source,
34#    IN UINTN  Count
35#    )
36#------------------------------------------------------------------------------
37ASM_GLOBAL ASM_PFX(InternalMemCopyMem)
38ASM_PFX(InternalMemCopyMem):
39    pushq   %rsi
40    pushq   %rdi
41    movq    %rdx, %rsi                  # rsi <- Source
42    movq    %rcx, %rdi                  # rdi <- Destination
43    leaq    -1(%rsi, %r8,), %r9         # r9 <- Last byte of Source
44    cmpq    %rdi, %rsi
45    movq    %rdi, %rax                  # rax <- Destination as return value
46    jae     L0                          # Copy forward if Source > Destination
47    cmpq    %rdi, %r9                   # Overlapped?
48    jae     L_CopyBackward              # Copy backward if overlapped
49L0:
50    xorq    %rcx, %rcx
51    subq    %rdi, %rcx                  # rcx <- -rdi
52    andq    $15, %rcx                   # rcx + rsi should be 16 bytes aligned
53    jz      L1                          # skip if rcx == 0
54    cmpq    %r8, %rcx
55    cmova   %r8, %rcx
56    subq    %rcx, %r8
57    rep     movsb
58L1:
59    movq    %r8, %rcx
60    andq    $15, %r8
61    shrq    $4, %rcx                    # rcx <- # of DQwords to copy
62    jz      L_CopyBytes
63    movdqa  %xmm0, 0x18(%rsp)           # save xmm0 on stack
64L2:
65    movdqu  (%rsi), %xmm0               # rsi may not be 16-byte aligned
66    movntdq  %xmm0, (%rdi)              # rdi should be 16-byte aligned
67    addq    $16, %rsi
68    addq    $16, %rdi
69    loop    L2
70    mfence
71    movdqa  0x18(%rsp), %xmm0           # restore xmm0
72    jmp     L_CopyBytes                 # copy remaining bytes
73L_CopyBackward:
74    movq    %r9, %rsi                   # rsi <- Last byte of Source
75    leaq    -1(%rdi, %r8,), %rdi        # rdi <- Last byte of Destination
76    std
77L_CopyBytes:
78    movq    %r8, %rcx
79    rep     movsb
80    cld
81    popq    %rdi
82    popq    %rsi
83    ret
84