.text
.type   _x86_AES_encrypt_compact,@function
.align  16
_x86_AES_encrypt_compact:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        movl    %edi,20(%esp)
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,24(%esp)
        movl    -128(%ebp),%edi
        movl    -96(%ebp),%esi
        movl    -64(%ebp),%edi
        movl    -32(%ebp),%esi
        movl    (%ebp),%edi
        movl    32(%ebp),%esi
        movl    64(%ebp),%edi
        movl    96(%ebp),%esi
.align  16
.L000loop:
        movl    %eax,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %bh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,4(%esp)

        movl    %ebx,%esi
        andl    $255,%esi
        shrl    $16,%ebx
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %ch,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,8(%esp)

        movl    %ecx,%esi
        andl    $255,%esi
        shrl    $24,%ecx
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %dh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edx
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movzbl  %bh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi

        andl    $255,%edx
        movzbl  -128(%ebp,%edx,1),%edx
        movzbl  %ah,%eax
        movzbl  -128(%ebp,%eax,1),%eax
        shll    $8,%eax
        xorl    %eax,%edx
        movl    4(%esp),%eax
        andl    $255,%ebx
        movzbl  -128(%ebp,%ebx,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%edx
        movl    8(%esp),%ebx
        movzbl  -128(%ebp,%ecx,1),%ecx
        shll    $24,%ecx
        xorl    %ecx,%edx
        movl    %esi,%ecx

        movl    $2155905152,%ebp
        andl    %ecx,%ebp
        leal    (%ecx,%ecx,1),%edi
        movl    %ebp,%esi
        shrl    $7,%ebp
        andl    $4278124286,%edi
        subl    %ebp,%esi
        movl    %ecx,%ebp
        andl    $454761243,%esi
        rorl    $16,%ebp
        xorl    %edi,%esi
        movl    %ecx,%edi
        xorl    %esi,%ecx
        rorl    $24,%edi
        xorl    %ebp,%esi
        roll    $24,%ecx
        xorl    %edi,%esi
        movl    $2155905152,%ebp
        xorl    %esi,%ecx
        andl    %edx,%ebp
        leal    (%edx,%edx,1),%edi
        movl    %ebp,%esi
        shrl    $7,%ebp
        andl    $4278124286,%edi
        subl    %ebp,%esi
        movl    %edx,%ebp
        andl    $454761243,%esi
        rorl    $16,%ebp
        xorl    %edi,%esi
        movl    %edx,%edi
        xorl    %esi,%edx
        rorl    $24,%edi
        xorl    %ebp,%esi
        roll    $24,%edx
        xorl    %edi,%esi
        movl    $2155905152,%ebp
        xorl    %esi,%edx
        andl    %eax,%ebp
        leal    (%eax,%eax,1),%edi
        movl    %ebp,%esi
        shrl    $7,%ebp
        andl    $4278124286,%edi
        subl    %ebp,%esi
        movl    %eax,%ebp
        andl    $454761243,%esi
        rorl    $16,%ebp
        xorl    %edi,%esi
        movl    %eax,%edi
        xorl    %esi,%eax
        rorl    $24,%edi
        xorl    %ebp,%esi
        roll    $24,%eax
        xorl    %edi,%esi
        movl    $2155905152,%ebp
        xorl    %esi,%eax
        andl    %ebx,%ebp
        leal    (%ebx,%ebx,1),%edi
        movl    %ebp,%esi
        shrl    $7,%ebp
        andl    $4278124286,%edi
        subl    %ebp,%esi
        movl    %ebx,%ebp
        andl    $454761243,%esi
        rorl    $16,%ebp
        xorl    %edi,%esi
        movl    %ebx,%edi
        xorl    %esi,%ebx
        rorl    $24,%edi
        xorl    %ebp,%esi
        roll    $24,%ebx
        xorl    %edi,%esi
        xorl    %esi,%ebx
        movl    20(%esp),%edi
        movl    28(%esp),%ebp
        addl    $16,%edi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        cmpl    24(%esp),%edi
        movl    %edi,20(%esp)
        jb      .L000loop
        movl    %eax,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %bh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,4(%esp)

        movl    %ebx,%esi
        andl    $255,%esi
        shrl    $16,%ebx
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %ch,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,8(%esp)

        movl    %ecx,%esi
        andl    $255,%esi
        shrl    $24,%ecx
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %dh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edx
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movzbl  %bh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi

        movl    20(%esp),%edi
        andl    $255,%edx
        movzbl  -128(%ebp,%edx,1),%edx
        movzbl  %ah,%eax
        movzbl  -128(%ebp,%eax,1),%eax
        shll    $8,%eax
        xorl    %eax,%edx
        movl    4(%esp),%eax
        andl    $255,%ebx
        movzbl  -128(%ebp,%ebx,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%edx
        movl    8(%esp),%ebx
        movzbl  -128(%ebp,%ecx,1),%ecx
        shll    $24,%ecx
        xorl    %ecx,%edx
        movl    %esi,%ecx

        xorl    16(%edi),%eax
        xorl    20(%edi),%ebx
        xorl    24(%edi),%ecx
        xorl    28(%edi),%edx
        ret
.size   _x86_AES_encrypt_compact,.-_x86_AES_encrypt_compact
.type   _sse_AES_encrypt_compact,@function
.align  16
_sse_AES_encrypt_compact:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        pxor    (%edi),%mm0
        pxor    8(%edi),%mm4
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,24(%esp)
        movl    $454761243,%eax
        movl    %eax,8(%esp)
        movl    %eax,12(%esp)
        movl    -128(%ebp),%eax
        movl    -96(%ebp),%ebx
        movl    -64(%ebp),%ecx
        movl    -32(%ebp),%edx
        movl    (%ebp),%eax
        movl    32(%ebp),%ebx
        movl    64(%ebp),%ecx
        movl    96(%ebp),%edx
.align  16
.L001loop:
        pshufw  $8,%mm0,%mm1
        pshufw  $13,%mm4,%mm5
        movd    %mm1,%eax
        movd    %mm5,%ebx
        movl    %edi,20(%esp)
        movzbl  %al,%esi
        movzbl  %ah,%edx
        pshufw  $13,%mm0,%mm2
        movzbl  -128(%ebp,%esi,1),%ecx
        movzbl  %bl,%edi
        movzbl  -128(%ebp,%edx,1),%edx
        shrl    $16,%eax
        shll    $8,%edx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shll    $16,%esi
        pshufw  $8,%mm4,%mm6
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %ah,%edi
        shll    $24,%esi
        shrl    $16,%ebx
        orl     %esi,%edx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shll    $8,%esi
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %al,%edi
        shll    $24,%esi
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bl,%edi
        movd    %mm2,%eax
        movd    %ecx,%mm0
        movzbl  -128(%ebp,%edi,1),%ecx
        movzbl  %ah,%edi
        shll    $16,%ecx
        movd    %mm6,%ebx
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shll    $24,%esi
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bl,%edi
        shll    $8,%esi
        shrl    $16,%ebx
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %al,%edi
        shrl    $16,%eax
        movd    %ecx,%mm1
        movzbl  -128(%ebp,%edi,1),%ecx
        movzbl  %ah,%edi
        shll    $16,%ecx
        andl    $255,%eax
        orl     %esi,%ecx
        punpckldq       %mm1,%mm0
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shll    $24,%esi
        andl    $255,%ebx
        movzbl  -128(%ebp,%eax,1),%eax
        orl     %esi,%ecx
        shll    $16,%eax
        movzbl  -128(%ebp,%edi,1),%esi
        orl     %eax,%edx
        shll    $8,%esi
        movzbl  -128(%ebp,%ebx,1),%ebx
        orl     %esi,%ecx
        orl     %ebx,%edx
        movl    20(%esp),%edi
        movd    %ecx,%mm4
        movd    %edx,%mm5
        punpckldq       %mm5,%mm4
        addl    $16,%edi
        cmpl    24(%esp),%edi
        ja      .L002out
        movq    8(%esp),%mm2
        pxor    %mm3,%mm3
        pxor    %mm7,%mm7
        movq    %mm0,%mm1
        movq    %mm4,%mm5
        pcmpgtb %mm0,%mm3
        pcmpgtb %mm4,%mm7
        pand    %mm2,%mm3
        pand    %mm2,%mm7
        pshufw  $177,%mm0,%mm2
        pshufw  $177,%mm4,%mm6
        paddb   %mm0,%mm0
        paddb   %mm4,%mm4
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        pshufw  $177,%mm2,%mm3
        pshufw  $177,%mm6,%mm7
        pxor    %mm0,%mm1
        pxor    %mm4,%mm5
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        movq    %mm3,%mm2
        movq    %mm7,%mm6
        pslld   $8,%mm3
        pslld   $8,%mm7
        psrld   $24,%mm2
        psrld   $24,%mm6
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        movq    %mm1,%mm3
        movq    %mm5,%mm7
        movq    (%edi),%mm2
        movq    8(%edi),%mm6
        psrld   $8,%mm1
        psrld   $8,%mm5
        movl    -128(%ebp),%eax
        pslld   $24,%mm3
        pslld   $24,%mm7
        movl    -64(%ebp),%ebx
        pxor    %mm1,%mm0
        pxor    %mm5,%mm4
        movl    (%ebp),%ecx
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        movl    64(%ebp),%edx
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        jmp     .L001loop
.align  16
.L002out:
        pxor    (%edi),%mm0
        pxor    8(%edi),%mm4
        ret
.size   _sse_AES_encrypt_compact,.-_sse_AES_encrypt_compact
.type   _x86_AES_encrypt,@function
.align  16
_x86_AES_encrypt:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        movl    %edi,20(%esp)
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,24(%esp)
.align  16
.L003loop:
        movl    %eax,%esi
        andl    $255,%esi
        movl    (%ebp,%esi,8),%esi
        movzbl  %bh,%edi
        xorl    3(%ebp,%edi,8),%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        xorl    2(%ebp,%edi,8),%esi
        movl    %edx,%edi
        shrl    $24,%edi
        xorl    1(%ebp,%edi,8),%esi
        movl    %esi,4(%esp)

        movl    %ebx,%esi
        andl    $255,%esi
        shrl    $16,%ebx
        movl    (%ebp,%esi,8),%esi
        movzbl  %ch,%edi
        xorl    3(%ebp,%edi,8),%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        xorl    2(%ebp,%edi,8),%esi
        movl    %eax,%edi
        shrl    $24,%edi
        xorl    1(%ebp,%edi,8),%esi
        movl    %esi,8(%esp)

        movl    %ecx,%esi
        andl    $255,%esi
        shrl    $24,%ecx
        movl    (%ebp,%esi,8),%esi
        movzbl  %dh,%edi
        xorl    3(%ebp,%edi,8),%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edx
        andl    $255,%edi
        xorl    2(%ebp,%edi,8),%esi
        movzbl  %bh,%edi
        xorl    1(%ebp,%edi,8),%esi

        movl    20(%esp),%edi
        movl    (%ebp,%edx,8),%edx
        movzbl  %ah,%eax
        xorl    3(%ebp,%eax,8),%edx
        movl    4(%esp),%eax
        andl    $255,%ebx
        xorl    2(%ebp,%ebx,8),%edx
        movl    8(%esp),%ebx
        xorl    1(%ebp,%ecx,8),%edx
        movl    %esi,%ecx

        addl    $16,%edi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        cmpl    24(%esp),%edi
        movl    %edi,20(%esp)
        jb      .L003loop
        movl    %eax,%esi
        andl    $255,%esi
        movl    2(%ebp,%esi,8),%esi
        andl    $255,%esi
        movzbl  %bh,%edi
        movl    (%ebp,%edi,8),%edi
        andl    $65280,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movl    (%ebp,%edi,8),%edi
        andl    $16711680,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $24,%edi
        movl    2(%ebp,%edi,8),%edi
        andl    $4278190080,%edi
        xorl    %edi,%esi
        movl    %esi,4(%esp)
        movl    %ebx,%esi
        andl    $255,%esi
        shrl    $16,%ebx
        movl    2(%ebp,%esi,8),%esi
        andl    $255,%esi
        movzbl  %ch,%edi
        movl    (%ebp,%edi,8),%edi
        andl    $65280,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movl    (%ebp,%edi,8),%edi
        andl    $16711680,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $24,%edi
        movl    2(%ebp,%edi,8),%edi
        andl    $4278190080,%edi
        xorl    %edi,%esi
        movl    %esi,8(%esp)
        movl    %ecx,%esi
        andl    $255,%esi
        shrl    $24,%ecx
        movl    2(%ebp,%esi,8),%esi
        andl    $255,%esi
        movzbl  %dh,%edi
        movl    (%ebp,%edi,8),%edi
        andl    $65280,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edx
        andl    $255,%edi
        movl    (%ebp,%edi,8),%edi
        andl    $16711680,%edi
        xorl    %edi,%esi
        movzbl  %bh,%edi
        movl    2(%ebp,%edi,8),%edi
        andl    $4278190080,%edi
        xorl    %edi,%esi
        movl    20(%esp),%edi
        andl    $255,%edx
        movl    2(%ebp,%edx,8),%edx
        andl    $255,%edx
        movzbl  %ah,%eax
        movl    (%ebp,%eax,8),%eax
        andl    $65280,%eax
        xorl    %eax,%edx
        movl    4(%esp),%eax
        andl    $255,%ebx
        movl    (%ebp,%ebx,8),%ebx
        andl    $16711680,%ebx
        xorl    %ebx,%edx
        movl    8(%esp),%ebx
        movl    2(%ebp,%ecx,8),%ecx
        andl    $4278190080,%ecx
        xorl    %ecx,%edx
        movl    %esi,%ecx
        addl    $16,%edi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        ret
.align  64
.LAES_Te:
.long   2774754246,2774754246
.long   2222750968,2222750968
.long   2574743534,2574743534
.long   2373680118,2373680118
.long   234025727,234025727
.long   3177933782,3177933782
.long   2976870366,2976870366
.long   1422247313,1422247313
.long   1345335392,1345335392
.long   50397442,50397442
.long   2842126286,2842126286
.long   2099981142,2099981142
.long   436141799,436141799
.long   1658312629,1658312629
.long   3870010189,3870010189
.long   2591454956,2591454956
.long   1170918031,1170918031
.long   2642575903,2642575903
.long   1086966153,1086966153
.long   2273148410,2273148410
.long   368769775,368769775
.long   3948501426,3948501426
.long   3376891790,3376891790
.long   200339707,200339707
.long   3970805057,3970805057
.long   1742001331,1742001331
.long   4255294047,4255294047
.long   3937382213,3937382213
.long   3214711843,3214711843
.long   4154762323,4154762323
.long   2524082916,2524082916
.long   1539358875,1539358875
.long   3266819957,3266819957
.long   486407649,486407649
.long   2928907069,2928907069
.long   1780885068,1780885068
.long   1513502316,1513502316
.long   1094664062,1094664062
.long   49805301,49805301
.long   1338821763,1338821763
.long   1546925160,1546925160
.long   4104496465,4104496465
.long   887481809,887481809
.long   150073849,150073849
.long   2473685474,2473685474
.long   1943591083,1943591083
.long   1395732834,1395732834
.long   1058346282,1058346282
.long   201589768,201589768
.long   1388824469,1388824469
.long   1696801606,1696801606
.long   1589887901,1589887901
.long   672667696,672667696
.long   2711000631,2711000631
.long   251987210,251987210
.long   3046808111,3046808111
.long   151455502,151455502
.long   907153956,907153956
.long   2608889883,2608889883
.long   1038279391,1038279391
.long   652995533,652995533
.long   1764173646,1764173646
.long   3451040383,3451040383
.long   2675275242,2675275242
.long   453576978,453576978
.long   2659418909,2659418909
.long   1949051992,1949051992
.long   773462580,773462580
.long   756751158,756751158
.long   2993581788,2993581788
.long   3998898868,3998898868
.long   4221608027,4221608027
.long   4132590244,4132590244
.long   1295727478,1295727478
.long   1641469623,1641469623
.long   3467883389,3467883389
.long   2066295122,2066295122
.long   1055122397,1055122397
.long   1898917726,1898917726
.long   2542044179,2542044179
.long   4115878822,4115878822
.long   1758581177,1758581177
.long   0,0
.long   753790401,753790401
.long   1612718144,1612718144
.long   536673507,536673507
.long   3367088505,3367088505
.long   3982187446,3982187446
.long   3194645204,3194645204
.long   1187761037,1187761037
.long   3653156455,3653156455
.long   1262041458,1262041458
.long   3729410708,3729410708
.long   3561770136,3561770136
.long   3898103984,3898103984
.long   1255133061,1255133061
.long   1808847035,1808847035
.long   720367557,720367557
.long   3853167183,3853167183
.long   385612781,385612781
.long   3309519750,3309519750
.long   3612167578,3612167578
.long   1429418854,1429418854
.long   2491778321,2491778321
.long   3477423498,3477423498
.long   284817897,284817897
.long   100794884,100794884
.long   2172616702,2172616702
.long   4031795360,4031795360
.long   1144798328,1144798328
.long   3131023141,3131023141
.long   3819481163,3819481163
.long   4082192802,4082192802
.long   4272137053,4272137053
.long   3225436288,3225436288
.long   2324664069,2324664069
.long   2912064063,2912064063
.long   3164445985,3164445985
.long   1211644016,1211644016
.long   83228145,83228145
.long   3753688163,3753688163
.long   3249976951,3249976951
.long   1977277103,1977277103
.long   1663115586,1663115586
.long   806359072,806359072
.long   452984805,452984805
.long   250868733,250868733
.long   1842533055,1842533055
.long   1288555905,1288555905
.long   336333848,336333848
.long   890442534,890442534
.long   804056259,804056259
.long   3781124030,3781124030
.long   2727843637,2727843637
.long   3427026056,3427026056
.long   957814574,957814574
.long   1472513171,1472513171
.long   4071073621,4071073621
.long   2189328124,2189328124
.long   1195195770,1195195770
.long   2892260552,2892260552
.long   3881655738,3881655738
.long   723065138,723065138
.long   2507371494,2507371494
.long   2690670784,2690670784
.long   2558624025,2558624025
.long   3511635870,3511635870
.long   2145180835,2145180835
.long   1713513028,1713513028
.long   2116692564,2116692564
.long   2878378043,2878378043
.long   2206763019,2206763019
.long   3393603212,3393603212
.long   703524551,703524551
.long   3552098411,3552098411
.long   1007948840,1007948840
.long   2044649127,2044649127
.long   3797835452,3797835452
.long   487262998,487262998
.long   1994120109,1994120109
.long   1004593371,1004593371
.long   1446130276,1446130276
.long   1312438900,1312438900
.long   503974420,503974420
.long   3679013266,3679013266
.long   168166924,168166924
.long   1814307912,1814307912
.long   3831258296,3831258296
.long   1573044895,1573044895
.long   1859376061,1859376061
.long   4021070915,4021070915
.long   2791465668,2791465668
.long   2828112185,2828112185
.long   2761266481,2761266481
.long   937747667,937747667
.long   2339994098,2339994098
.long   854058965,854058965
.long   1137232011,1137232011
.long   1496790894,1496790894
.long   3077402074,3077402074
.long   2358086913,2358086913
.long   1691735473,1691735473
.long   3528347292,3528347292
.long   3769215305,3769215305
.long   3027004632,3027004632
.long   4199962284,4199962284
.long   133494003,133494003
.long   636152527,636152527
.long   2942657994,2942657994
.long   2390391540,2390391540
.long   3920539207,3920539207
.long   403179536,403179536
.long   3585784431,3585784431
.long   2289596656,2289596656
.long   1864705354,1864705354
.long   1915629148,1915629148
.long   605822008,605822008
.long   4054230615,4054230615
.long   3350508659,3350508659
.long   1371981463,1371981463
.long   602466507,602466507
.long   2094914977,2094914977
.long   2624877800,2624877800
.long   555687742,555687742
.long   3712699286,3712699286
.long   3703422305,3703422305
.long   2257292045,2257292045
.long   2240449039,2240449039
.long   2423288032,2423288032
.long   1111375484,1111375484
.long   3300242801,3300242801
.long   2858837708,2858837708
.long   3628615824,3628615824
.long   84083462,84083462
.long   32962295,32962295
.long   302911004,302911004
.long   2741068226,2741068226
.long   1597322602,1597322602
.long   4183250862,4183250862
.long   3501832553,3501832553
.long   2441512471,2441512471
.long   1489093017,1489093017
.long   656219450,656219450
.long   3114180135,3114180135
.long   954327513,954327513
.long   335083755,335083755
.long   3013122091,3013122091
.long   856756514,856756514
.long   3144247762,3144247762
.long   1893325225,1893325225
.long   2307821063,2307821063
.long   2811532339,2811532339
.long   3063651117,3063651117
.long   572399164,572399164
.long   2458355477,2458355477
.long   552200649,552200649
.long   1238290055,1238290055
.long   4283782570,4283782570
.long   2015897680,2015897680
.long   2061492133,2061492133
.long   2408352771,2408352771
.long   4171342169,4171342169
.long   2156497161,2156497161
.long   386731290,386731290
.long   3669999461,3669999461
.long   837215959,837215959
.long   3326231172,3326231172
.long   3093850320,3093850320
.long   3275833730,3275833730
.long   2962856233,2962856233
.long   1999449434,1999449434
.long   286199582,286199582
.long   3417354363,3417354363
.long   4233385128,4233385128
.long   3602627437,3602627437
.long   974525996,974525996
.byte   99,124,119,123,242,107,111,197
.byte   48,1,103,43,254,215,171,118
.byte   202,130,201,125,250,89,71,240
.byte   173,212,162,175,156,164,114,192
.byte   183,253,147,38,54,63,247,204
.byte   52,165,229,241,113,216,49,21
.byte   4,199,35,195,24,150,5,154
.byte   7,18,128,226,235,39,178,117
.byte   9,131,44,26,27,110,90,160
.byte   82,59,214,179,41,227,47,132
.byte   83,209,0,237,32,252,177,91
.byte   106,203,190,57,74,76,88,207
.byte   208,239,170,251,67,77,51,133
.byte   69,249,2,127,80,60,159,168
.byte   81,163,64,143,146,157,56,245
.byte   188,182,218,33,16,255,243,210
.byte   205,12,19,236,95,151,68,23
.byte   196,167,126,61,100,93,25,115
.byte   96,129,79,220,34,42,144,136
.byte   70,238,184,20,222,94,11,219
.byte   224,50,58,10,73,6,36,92
.byte   194,211,172,98,145,149,228,121
.byte   231,200,55,109,141,213,78,169
.byte   108,86,244,234,101,122,174,8
.byte   186,120,37,46,28,166,180,198
.byte   232,221,116,31,75,189,139,138
.byte   112,62,181,102,72,3,246,14
.byte   97,53,87,185,134,193,29,158
.byte   225,248,152,17,105,217,142,148
.byte   155,30,135,233,206,85,40,223
.byte   140,161,137,13,191,230,66,104
.byte   65,153,45,15,176,84,187,22
.byte   99,124,119,123,242,107,111,197
.byte   48,1,103,43,254,215,171,118
.byte   202,130,201,125,250,89,71,240
.byte   173,212,162,175,156,164,114,192
.byte   183,253,147,38,54,63,247,204
.byte   52,165,229,241,113,216,49,21
.byte   4,199,35,195,24,150,5,154
.byte   7,18,128,226,235,39,178,117
.byte   9,131,44,26,27,110,90,160
.byte   82,59,214,179,41,227,47,132
.byte   83,209,0,237,32,252,177,91
.byte   106,203,190,57,74,76,88,207
.byte   208,239,170,251,67,77,51,133
.byte   69,249,2,127,80,60,159,168
.byte   81,163,64,143,146,157,56,245
.byte   188,182,218,33,16,255,243,210
.byte   205,12,19,236,95,151,68,23
.byte   196,167,126,61,100,93,25,115
.byte   96,129,79,220,34,42,144,136
.byte   70,238,184,20,222,94,11,219
.byte   224,50,58,10,73,6,36,92
.byte   194,211,172,98,145,149,228,121
.byte   231,200,55,109,141,213,78,169
.byte   108,86,244,234,101,122,174,8
.byte   186,120,37,46,28,166,180,198
.byte   232,221,116,31,75,189,139,138
.byte   112,62,181,102,72,3,246,14
.byte   97,53,87,185,134,193,29,158
.byte   225,248,152,17,105,217,142,148
.byte   155,30,135,233,206,85,40,223
.byte   140,161,137,13,191,230,66,104
.byte   65,153,45,15,176,84,187,22
.byte   99,124,119,123,242,107,111,197
.byte   48,1,103,43,254,215,171,118
.byte   202,130,201,125,250,89,71,240
.byte   173,212,162,175,156,164,114,192
.byte   183,253,147,38,54,63,247,204
.byte   52,165,229,241,113,216,49,21
.byte   4,199,35,195,24,150,5,154
.byte   7,18,128,226,235,39,178,117
.byte   9,131,44,26,27,110,90,160
.byte   82,59,214,179,41,227,47,132
.byte   83,209,0,237,32,252,177,91
.byte   106,203,190,57,74,76,88,207
.byte   208,239,170,251,67,77,51,133
.byte   69,249,2,127,80,60,159,168
.byte   81,163,64,143,146,157,56,245
.byte   188,182,218,33,16,255,243,210
.byte   205,12,19,236,95,151,68,23
.byte   196,167,126,61,100,93,25,115
.byte   96,129,79,220,34,42,144,136
.byte   70,238,184,20,222,94,11,219
.byte   224,50,58,10,73,6,36,92
.byte   194,211,172,98,145,149,228,121
.byte   231,200,55,109,141,213,78,169
.byte   108,86,244,234,101,122,174,8
.byte   186,120,37,46,28,166,180,198
.byte   232,221,116,31,75,189,139,138
.byte   112,62,181,102,72,3,246,14
.byte   97,53,87,185,134,193,29,158
.byte   225,248,152,17,105,217,142,148
.byte   155,30,135,233,206,85,40,223
.byte   140,161,137,13,191,230,66,104
.byte   65,153,45,15,176,84,187,22
.byte   99,124,119,123,242,107,111,197
.byte   48,1,103,43,254,215,171,118
.byte   202,130,201,125,250,89,71,240
.byte   173,212,162,175,156,164,114,192
.byte   183,253,147,38,54,63,247,204
.byte   52,165,229,241,113,216,49,21
.byte   4,199,35,195,24,150,5,154
.byte   7,18,128,226,235,39,178,117
.byte   9,131,44,26,27,110,90,160
.byte   82,59,214,179,41,227,47,132
.byte   83,209,0,237,32,252,177,91
.byte   106,203,190,57,74,76,88,207
.byte   208,239,170,251,67,77,51,133
.byte   69,249,2,127,80,60,159,168
.byte   81,163,64,143,146,157,56,245
.byte   188,182,218,33,16,255,243,210
.byte   205,12,19,236,95,151,68,23
.byte   196,167,126,61,100,93,25,115
.byte   96,129,79,220,34,42,144,136
.byte   70,238,184,20,222,94,11,219
.byte   224,50,58,10,73,6,36,92
.byte   194,211,172,98,145,149,228,121
.byte   231,200,55,109,141,213,78,169
.byte   108,86,244,234,101,122,174,8
.byte   186,120,37,46,28,166,180,198
.byte   232,221,116,31,75,189,139,138
.byte   112,62,181,102,72,3,246,14
.byte   97,53,87,185,134,193,29,158
.byte   225,248,152,17,105,217,142,148
.byte   155,30,135,233,206,85,40,223
.byte   140,161,137,13,191,230,66,104
.byte   65,153,45,15,176,84,187,22
.long   1,2,4,8
.long   16,32,64,128
.long   27,54,0,0
.long   0,0,0,0
.size   _x86_AES_encrypt,.-_x86_AES_encrypt
.globl  AES_encrypt
.type   AES_encrypt,@function
.align  16
AES_encrypt:
.L_AES_encrypt_begin:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        pushl   %ebp
        pushl   %ebx
        pushl   %esi
        pushl   %edi
        movl    20(%esp),%esi
        movl    28(%esp),%edi
        movl    %esp,%eax
        subl    $36,%esp
        andl    $-64,%esp
        leal    -127(%edi),%ebx
        subl    %esp,%ebx
        negl    %ebx
        andl    $960,%ebx
        subl    %ebx,%esp
        addl    $4,%esp
        movl    %eax,28(%esp)
        call    .L004pic_point
.L004pic_point:
        popl    %ebp
        leal    OPENSSL_ia32cap_P,%eax
        leal    .LAES_Te-.L004pic_point(%ebp),%ebp
        leal    764(%esp),%ebx
        subl    %ebp,%ebx
        andl    $768,%ebx
        leal    2176(%ebp,%ebx,1),%ebp
        btl     $25,(%eax)
        jnc     .L005x86
        movq    (%esi),%mm0
        movq    8(%esi),%mm4
        call    _sse_AES_encrypt_compact
        movl    28(%esp),%esp
        movl    24(%esp),%esi
        movq    %mm0,(%esi)
        movq    %mm4,8(%esi)
        emms
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.align  16
.L005x86:
        movl    %ebp,24(%esp)
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        call    _x86_AES_encrypt_compact
        movl    28(%esp),%esp
        movl    24(%esp),%esi
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.size   AES_encrypt,.-.L_AES_encrypt_begin
.type   _x86_AES_decrypt_compact,@function
.align  16
_x86_AES_decrypt_compact:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        movl    %edi,20(%esp)
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,24(%esp)
        movl    -128(%ebp),%edi
        movl    -96(%ebp),%esi
        movl    -64(%ebp),%edi
        movl    -32(%ebp),%esi
        movl    (%ebp),%edi
        movl    32(%ebp),%esi
        movl    64(%ebp),%edi
        movl    96(%ebp),%esi
.align  16
.L006loop:
        movl    %eax,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %dh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %ebx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,4(%esp)
        movl    %ebx,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %ah,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,8(%esp)
        movl    %ecx,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %bh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        andl    $255,%edx
        movzbl  -128(%ebp,%edx,1),%edx
        movzbl  %ch,%ecx
        movzbl  -128(%ebp,%ecx,1),%ecx
        shll    $8,%ecx
        xorl    %ecx,%edx
        movl    %esi,%ecx
        shrl    $16,%ebx
        andl    $255,%ebx
        movzbl  -128(%ebp,%ebx,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%edx
        shrl    $24,%eax
        movzbl  -128(%ebp,%eax,1),%eax
        shll    $24,%eax
        xorl    %eax,%edx
        movl    $2155905152,%edi
        andl    %ecx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ecx,%ecx,1),%eax
        subl    %edi,%esi
        andl    $4278124286,%eax
        andl    $454761243,%esi
        xorl    %esi,%eax
        movl    $2155905152,%edi
        andl    %eax,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%eax,%eax,1),%ebx
        subl    %edi,%esi
        andl    $4278124286,%ebx
        andl    $454761243,%esi
        xorl    %ecx,%eax
        xorl    %esi,%ebx
        movl    $2155905152,%edi
        andl    %ebx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ebx,%ebx,1),%ebp
        subl    %edi,%esi
        andl    $4278124286,%ebp
        andl    $454761243,%esi
        xorl    %ecx,%ebx
        roll    $8,%ecx
        xorl    %esi,%ebp
        xorl    %eax,%ecx
        xorl    %ebp,%eax
        xorl    %ebx,%ecx
        xorl    %ebp,%ebx
        roll    $24,%eax
        xorl    %ebp,%ecx
        roll    $16,%ebx
        xorl    %eax,%ecx
        roll    $8,%ebp
        xorl    %ebx,%ecx
        movl    4(%esp),%eax
        xorl    %ebp,%ecx
        movl    %ecx,12(%esp)
        movl    $2155905152,%edi
        andl    %edx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%edx,%edx,1),%ebx
        subl    %edi,%esi
        andl    $4278124286,%ebx
        andl    $454761243,%esi
        xorl    %esi,%ebx
        movl    $2155905152,%edi
        andl    %ebx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ebx,%ebx,1),%ecx
        subl    %edi,%esi
        andl    $4278124286,%ecx
        andl    $454761243,%esi
        xorl    %edx,%ebx
        xorl    %esi,%ecx
        movl    $2155905152,%edi
        andl    %ecx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ecx,%ecx,1),%ebp
        subl    %edi,%esi
        andl    $4278124286,%ebp
        andl    $454761243,%esi
        xorl    %edx,%ecx
        roll    $8,%edx
        xorl    %esi,%ebp
        xorl    %ebx,%edx
        xorl    %ebp,%ebx
        xorl    %ecx,%edx
        xorl    %ebp,%ecx
        roll    $24,%ebx
        xorl    %ebp,%edx
        roll    $16,%ecx
        xorl    %ebx,%edx
        roll    $8,%ebp
        xorl    %ecx,%edx
        movl    8(%esp),%ebx
        xorl    %ebp,%edx
        movl    %edx,16(%esp)
        movl    $2155905152,%edi
        andl    %eax,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%eax,%eax,1),%ecx
        subl    %edi,%esi
        andl    $4278124286,%ecx
        andl    $454761243,%esi
        xorl    %esi,%ecx
        movl    $2155905152,%edi
        andl    %ecx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ecx,%ecx,1),%edx
        subl    %edi,%esi
        andl    $4278124286,%edx
        andl    $454761243,%esi
        xorl    %eax,%ecx
        xorl    %esi,%edx
        movl    $2155905152,%edi
        andl    %edx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%edx,%edx,1),%ebp
        subl    %edi,%esi
        andl    $4278124286,%ebp
        andl    $454761243,%esi
        xorl    %eax,%edx
        roll    $8,%eax
        xorl    %esi,%ebp
        xorl    %ecx,%eax
        xorl    %ebp,%ecx
        xorl    %edx,%eax
        xorl    %ebp,%edx
        roll    $24,%ecx
        xorl    %ebp,%eax
        roll    $16,%edx
        xorl    %ecx,%eax
        roll    $8,%ebp
        xorl    %edx,%eax
        xorl    %ebp,%eax
        movl    $2155905152,%edi
        andl    %ebx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ebx,%ebx,1),%ecx
        subl    %edi,%esi
        andl    $4278124286,%ecx
        andl    $454761243,%esi
        xorl    %esi,%ecx
        movl    $2155905152,%edi
        andl    %ecx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%ecx,%ecx,1),%edx
        subl    %edi,%esi
        andl    $4278124286,%edx
        andl    $454761243,%esi
        xorl    %ebx,%ecx
        xorl    %esi,%edx
        movl    $2155905152,%edi
        andl    %edx,%edi
        movl    %edi,%esi
        shrl    $7,%edi
        leal    (%edx,%edx,1),%ebp
        subl    %edi,%esi
        andl    $4278124286,%ebp
        andl    $454761243,%esi
        xorl    %ebx,%edx
        roll    $8,%ebx
        xorl    %esi,%ebp
        xorl    %ecx,%ebx
        xorl    %ebp,%ecx
        xorl    %edx,%ebx
        xorl    %ebp,%edx
        roll    $24,%ecx
        xorl    %ebp,%ebx
        roll    $16,%edx
        xorl    %ecx,%ebx
        roll    $8,%ebp
        xorl    %edx,%ebx
        movl    12(%esp),%ecx
        xorl    %ebp,%ebx
        movl    16(%esp),%edx
        movl    20(%esp),%edi
        movl    28(%esp),%ebp
        addl    $16,%edi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        cmpl    24(%esp),%edi
        movl    %edi,20(%esp)
        jb      .L006loop
        movl    %eax,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %dh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %ebx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,4(%esp)
        movl    %ebx,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %ah,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,8(%esp)
        movl    %ecx,%esi
        andl    $255,%esi
        movzbl  -128(%ebp,%esi,1),%esi
        movzbl  %bh,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $24,%edi
        movzbl  -128(%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    20(%esp),%edi
        andl    $255,%edx
        movzbl  -128(%ebp,%edx,1),%edx
        movzbl  %ch,%ecx
        movzbl  -128(%ebp,%ecx,1),%ecx
        shll    $8,%ecx
        xorl    %ecx,%edx
        movl    %esi,%ecx
        shrl    $16,%ebx
        andl    $255,%ebx
        movzbl  -128(%ebp,%ebx,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%edx
        movl    8(%esp),%ebx
        shrl    $24,%eax
        movzbl  -128(%ebp,%eax,1),%eax
        shll    $24,%eax
        xorl    %eax,%edx
        movl    4(%esp),%eax
        xorl    16(%edi),%eax
        xorl    20(%edi),%ebx
        xorl    24(%edi),%ecx
        xorl    28(%edi),%edx
        ret
.size   _x86_AES_decrypt_compact,.-_x86_AES_decrypt_compact
.type   _sse_AES_decrypt_compact,@function
.align  16
_sse_AES_decrypt_compact:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        pxor    (%edi),%mm0
        pxor    8(%edi),%mm4
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,24(%esp)
        movl    $454761243,%eax
        movl    %eax,8(%esp)
        movl    %eax,12(%esp)
        movl    -128(%ebp),%eax
        movl    -96(%ebp),%ebx
        movl    -64(%ebp),%ecx
        movl    -32(%ebp),%edx
        movl    (%ebp),%eax
        movl    32(%ebp),%ebx
        movl    64(%ebp),%ecx
        movl    96(%ebp),%edx
.align  16
.L007loop:
        pshufw  $12,%mm0,%mm1
        pshufw  $9,%mm4,%mm5
        movd    %mm1,%eax
        movd    %mm5,%ebx
        movl    %edi,20(%esp)
        movzbl  %al,%esi
        movzbl  %ah,%edx
        pshufw  $6,%mm0,%mm2
        movzbl  -128(%ebp,%esi,1),%ecx
        movzbl  %bl,%edi
        movzbl  -128(%ebp,%edx,1),%edx
        shrl    $16,%eax
        shll    $8,%edx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shll    $16,%esi
        pshufw  $3,%mm4,%mm6
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %ah,%edi
        shll    $24,%esi
        shrl    $16,%ebx
        orl     %esi,%edx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shll    $24,%esi
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %al,%edi
        shll    $8,%esi
        movd    %mm2,%eax
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bl,%edi
        shll    $16,%esi
        movd    %mm6,%ebx
        movd    %ecx,%mm0
        movzbl  -128(%ebp,%edi,1),%ecx
        movzbl  %al,%edi
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bl,%edi
        orl     %esi,%edx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %ah,%edi
        shll    $16,%esi
        shrl    $16,%eax
        orl     %esi,%edx
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %bh,%edi
        shrl    $16,%ebx
        shll    $8,%esi
        movd    %edx,%mm1
        movzbl  -128(%ebp,%edi,1),%edx
        movzbl  %bh,%edi
        shll    $24,%edx
        andl    $255,%ebx
        orl     %esi,%edx
        punpckldq       %mm1,%mm0
        movzbl  -128(%ebp,%edi,1),%esi
        movzbl  %al,%edi
        shll    $8,%esi
        movzbl  %ah,%eax
        movzbl  -128(%ebp,%ebx,1),%ebx
        orl     %esi,%ecx
        movzbl  -128(%ebp,%edi,1),%esi
        orl     %ebx,%edx
        shll    $16,%esi
        movzbl  -128(%ebp,%eax,1),%eax
        orl     %esi,%edx
        shll    $24,%eax
        orl     %eax,%ecx
        movl    20(%esp),%edi
        movd    %edx,%mm4
        movd    %ecx,%mm5
        punpckldq       %mm5,%mm4
        addl    $16,%edi
        cmpl    24(%esp),%edi
        ja      .L008out
        movq    %mm0,%mm3
        movq    %mm4,%mm7
        pshufw  $228,%mm0,%mm2
        pshufw  $228,%mm4,%mm6
        movq    %mm0,%mm1
        movq    %mm4,%mm5
        pshufw  $177,%mm0,%mm0
        pshufw  $177,%mm4,%mm4
        pslld   $8,%mm2
        pslld   $8,%mm6
        psrld   $8,%mm3
        psrld   $8,%mm7
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        pslld   $16,%mm2
        pslld   $16,%mm6
        psrld   $16,%mm3
        psrld   $16,%mm7
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        movq    8(%esp),%mm3
        pxor    %mm2,%mm2
        pxor    %mm6,%mm6
        pcmpgtb %mm1,%mm2
        pcmpgtb %mm5,%mm6
        pand    %mm3,%mm2
        pand    %mm3,%mm6
        paddb   %mm1,%mm1
        paddb   %mm5,%mm5
        pxor    %mm2,%mm1
        pxor    %mm6,%mm5
        movq    %mm1,%mm3
        movq    %mm5,%mm7
        movq    %mm1,%mm2
        movq    %mm5,%mm6
        pxor    %mm1,%mm0
        pxor    %mm5,%mm4
        pslld   $24,%mm3
        pslld   $24,%mm7
        psrld   $8,%mm2
        psrld   $8,%mm6
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        movq    8(%esp),%mm2
        pxor    %mm3,%mm3
        pxor    %mm7,%mm7
        pcmpgtb %mm1,%mm3
        pcmpgtb %mm5,%mm7
        pand    %mm2,%mm3
        pand    %mm2,%mm7
        paddb   %mm1,%mm1
        paddb   %mm5,%mm5
        pxor    %mm3,%mm1
        pxor    %mm7,%mm5
        pshufw  $177,%mm1,%mm3
        pshufw  $177,%mm5,%mm7
        pxor    %mm1,%mm0
        pxor    %mm5,%mm4
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        pxor    %mm3,%mm3
        pxor    %mm7,%mm7
        pcmpgtb %mm1,%mm3
        pcmpgtb %mm5,%mm7
        pand    %mm2,%mm3
        pand    %mm2,%mm7
        paddb   %mm1,%mm1
        paddb   %mm5,%mm5
        pxor    %mm3,%mm1
        pxor    %mm7,%mm5
        pxor    %mm1,%mm0
        pxor    %mm5,%mm4
        movq    %mm1,%mm3
        movq    %mm5,%mm7
        pshufw  $177,%mm1,%mm2
        pshufw  $177,%mm5,%mm6
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        pslld   $8,%mm1
        pslld   $8,%mm5
        psrld   $8,%mm3
        psrld   $8,%mm7
        movq    (%edi),%mm2
        movq    8(%edi),%mm6
        pxor    %mm1,%mm0
        pxor    %mm5,%mm4
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        movl    -128(%ebp),%eax
        pslld   $16,%mm1
        pslld   $16,%mm5
        movl    -64(%ebp),%ebx
        psrld   $16,%mm3
        psrld   $16,%mm7
        movl    (%ebp),%ecx
        pxor    %mm1,%mm0
        pxor    %mm5,%mm4
        movl    64(%ebp),%edx
        pxor    %mm3,%mm0
        pxor    %mm7,%mm4
        pxor    %mm2,%mm0
        pxor    %mm6,%mm4
        jmp     .L007loop
.align  16
.L008out:
        pxor    (%edi),%mm0
        pxor    8(%edi),%mm4
        ret
.size   _sse_AES_decrypt_compact,.-_sse_AES_decrypt_compact
.type   _x86_AES_decrypt,@function
.align  16
_x86_AES_decrypt:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        movl    %edi,20(%esp)
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,24(%esp)
.align  16
.L009loop:
        movl    %eax,%esi
        andl    $255,%esi
        movl    (%ebp,%esi,8),%esi
        movzbl  %dh,%edi
        xorl    3(%ebp,%edi,8),%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        xorl    2(%ebp,%edi,8),%esi
        movl    %ebx,%edi
        shrl    $24,%edi
        xorl    1(%ebp,%edi,8),%esi
        movl    %esi,4(%esp)

        movl    %ebx,%esi
        andl    $255,%esi
        movl    (%ebp,%esi,8),%esi
        movzbl  %ah,%edi
        xorl    3(%ebp,%edi,8),%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        xorl    2(%ebp,%edi,8),%esi
        movl    %ecx,%edi
        shrl    $24,%edi
        xorl    1(%ebp,%edi,8),%esi
        movl    %esi,8(%esp)

        movl    %ecx,%esi
        andl    $255,%esi
        movl    (%ebp,%esi,8),%esi
        movzbl  %bh,%edi
        xorl    3(%ebp,%edi,8),%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edi
        xorl    2(%ebp,%edi,8),%esi
        movl    %edx,%edi
        shrl    $24,%edi
        xorl    1(%ebp,%edi,8),%esi

        movl    20(%esp),%edi
        andl    $255,%edx
        movl    (%ebp,%edx,8),%edx
        movzbl  %ch,%ecx
        xorl    3(%ebp,%ecx,8),%edx
        movl    %esi,%ecx
        shrl    $16,%ebx
        andl    $255,%ebx
        xorl    2(%ebp,%ebx,8),%edx
        movl    8(%esp),%ebx
        shrl    $24,%eax
        xorl    1(%ebp,%eax,8),%edx
        movl    4(%esp),%eax

        addl    $16,%edi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        cmpl    24(%esp),%edi
        movl    %edi,20(%esp)
        jb      .L009loop
        leal    2176(%ebp),%ebp
        movl    -128(%ebp),%edi
        movl    -96(%ebp),%esi
        movl    -64(%ebp),%edi
        movl    -32(%ebp),%esi
        movl    (%ebp),%edi
        movl    32(%ebp),%esi
        movl    64(%ebp),%edi
        movl    96(%ebp),%esi
        leal    -128(%ebp),%ebp
        movl    %eax,%esi
        andl    $255,%esi
        movzbl  (%ebp,%esi,1),%esi
        movzbl  %dh,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %ebx,%edi
        shrl    $24,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,4(%esp)
        movl    %ebx,%esi
        andl    $255,%esi
        movzbl  (%ebp,%esi,1),%esi
        movzbl  %ah,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %ecx,%edi
        shrl    $24,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    %esi,8(%esp)
        movl    %ecx,%esi
        andl    $255,%esi
        movzbl  (%ebp,%esi,1),%esi
        movzbl  %bh,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $8,%edi
        xorl    %edi,%esi
        movl    %eax,%edi
        shrl    $16,%edi
        andl    $255,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $16,%edi
        xorl    %edi,%esi
        movl    %edx,%edi
        shrl    $24,%edi
        movzbl  (%ebp,%edi,1),%edi
        shll    $24,%edi
        xorl    %edi,%esi
        movl    20(%esp),%edi
        andl    $255,%edx
        movzbl  (%ebp,%edx,1),%edx
        movzbl  %ch,%ecx
        movzbl  (%ebp,%ecx,1),%ecx
        shll    $8,%ecx
        xorl    %ecx,%edx
        movl    %esi,%ecx
        shrl    $16,%ebx
        andl    $255,%ebx
        movzbl  (%ebp,%ebx,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%edx
        movl    8(%esp),%ebx
        shrl    $24,%eax
        movzbl  (%ebp,%eax,1),%eax
        shll    $24,%eax
        xorl    %eax,%edx
        movl    4(%esp),%eax
        leal    -2048(%ebp),%ebp
        addl    $16,%edi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        ret
.align  64
.LAES_Td:
.long   1353184337,1353184337
.long   1399144830,1399144830
.long   3282310938,3282310938
.long   2522752826,2522752826
.long   3412831035,3412831035
.long   4047871263,4047871263
.long   2874735276,2874735276
.long   2466505547,2466505547
.long   1442459680,1442459680
.long   4134368941,4134368941
.long   2440481928,2440481928
.long   625738485,625738485
.long   4242007375,4242007375
.long   3620416197,3620416197
.long   2151953702,2151953702
.long   2409849525,2409849525
.long   1230680542,1230680542
.long   1729870373,1729870373
.long   2551114309,2551114309
.long   3787521629,3787521629
.long   41234371,41234371
.long   317738113,317738113
.long   2744600205,2744600205
.long   3338261355,3338261355
.long   3881799427,3881799427
.long   2510066197,2510066197
.long   3950669247,3950669247
.long   3663286933,3663286933
.long   763608788,763608788
.long   3542185048,3542185048
.long   694804553,694804553
.long   1154009486,1154009486
.long   1787413109,1787413109
.long   2021232372,2021232372
.long   1799248025,1799248025
.long   3715217703,3715217703
.long   3058688446,3058688446
.long   397248752,397248752
.long   1722556617,1722556617
.long   3023752829,3023752829
.long   407560035,407560035
.long   2184256229,2184256229
.long   1613975959,1613975959
.long   1165972322,1165972322
.long   3765920945,3765920945
.long   2226023355,2226023355
.long   480281086,480281086
.long   2485848313,2485848313
.long   1483229296,1483229296
.long   436028815,436028815
.long   2272059028,2272059028
.long   3086515026,3086515026
.long   601060267,601060267
.long   3791801202,3791801202
.long   1468997603,1468997603
.long   715871590,715871590
.long   120122290,120122290
.long   63092015,63092015
.long   2591802758,2591802758
.long   2768779219,2768779219
.long   4068943920,4068943920
.long   2997206819,2997206819
.long   3127509762,3127509762
.long   1552029421,1552029421
.long   723308426,723308426
.long   2461301159,2461301159
.long   4042393587,4042393587
.long   2715969870,2715969870
.long   3455375973,3455375973
.long   3586000134,3586000134
.long   526529745,526529745
.long   2331944644,2331944644
.long   2639474228,2639474228
.long   2689987490,2689987490
.long   853641733,853641733
.long   1978398372,1978398372
.long   971801355,971801355
.long   2867814464,2867814464
.long   111112542,111112542
.long   1360031421,1360031421
.long   4186579262,4186579262
.long   1023860118,1023860118
.long   2919579357,2919579357
.long   1186850381,1186850381
.long   3045938321,3045938321
.long   90031217,90031217
.long   1876166148,1876166148
.long   4279586912,4279586912
.long   620468249,620468249
.long   2548678102,2548678102
.long   3426959497,3426959497
.long   2006899047,2006899047
.long   3175278768,3175278768
.long   2290845959,2290845959
.long   945494503,945494503
.long   3689859193,3689859193
.long   1191869601,1191869601
.long   3910091388,3910091388
.long   3374220536,3374220536
.long   0,0
.long   2206629897,2206629897
.long   1223502642,1223502642
.long   2893025566,2893025566
.long   1316117100,1316117100
.long   4227796733,4227796733
.long   1446544655,1446544655
.long   517320253,517320253
.long   658058550,658058550
.long   1691946762,1691946762
.long   564550760,564550760
.long   3511966619,3511966619
.long   976107044,976107044
.long   2976320012,2976320012
.long   266819475,266819475
.long   3533106868,3533106868
.long   2660342555,2660342555
.long   1338359936,1338359936
.long   2720062561,2720062561
.long   1766553434,1766553434
.long   370807324,370807324
.long   179999714,179999714
.long   3844776128,3844776128
.long   1138762300,1138762300
.long   488053522,488053522
.long   185403662,185403662
.long   2915535858,2915535858
.long   3114841645,3114841645
.long   3366526484,3366526484
.long   2233069911,2233069911
.long   1275557295,1275557295
.long   3151862254,3151862254
.long   4250959779,4250959779
.long   2670068215,2670068215
.long   3170202204,3170202204
.long   3309004356,3309004356
.long   880737115,880737115
.long   1982415755,1982415755
.long   3703972811,3703972811
.long   1761406390,1761406390
.long   1676797112,1676797112
.long   3403428311,3403428311
.long   277177154,277177154
.long   1076008723,1076008723
.long   538035844,538035844
.long   2099530373,2099530373
.long   4164795346,4164795346
.long   288553390,288553390
.long   1839278535,1839278535
.long   1261411869,1261411869
.long   4080055004,4080055004
.long   3964831245,3964831245
.long   3504587127,3504587127
.long   1813426987,1813426987
.long   2579067049,2579067049
.long   4199060497,4199060497
.long   577038663,577038663
.long   3297574056,3297574056
.long   440397984,440397984
.long   3626794326,3626794326
.long   4019204898,4019204898
.long   3343796615,3343796615
.long   3251714265,3251714265
.long   4272081548,4272081548
.long   906744984,906744984
.long   3481400742,3481400742
.long   685669029,685669029
.long   646887386,646887386
.long   2764025151,2764025151
.long   3835509292,3835509292
.long   227702864,227702864
.long   2613862250,2613862250
.long   1648787028,1648787028
.long   3256061430,3256061430
.long   3904428176,3904428176
.long   1593260334,1593260334
.long   4121936770,4121936770
.long   3196083615,3196083615
.long   2090061929,2090061929
.long   2838353263,2838353263
.long   3004310991,3004310991
.long   999926984,999926984
.long   2809993232,2809993232
.long   1852021992,1852021992
.long   2075868123,2075868123
.long   158869197,158869197
.long   4095236462,4095236462
.long   28809964,28809964
.long   2828685187,2828685187
.long   1701746150,1701746150
.long   2129067946,2129067946
.long   147831841,147831841
.long   3873969647,3873969647
.long   3650873274,3650873274
.long   3459673930,3459673930
.long   3557400554,3557400554
.long   3598495785,3598495785
.long   2947720241,2947720241
.long   824393514,824393514
.long   815048134,815048134
.long   3227951669,3227951669
.long   935087732,935087732
.long   2798289660,2798289660
.long   2966458592,2966458592
.long   366520115,366520115
.long   1251476721,1251476721
.long   4158319681,4158319681
.long   240176511,240176511
.long   804688151,804688151
.long   2379631990,2379631990
.long   1303441219,1303441219
.long   1414376140,1414376140
.long   3741619940,3741619940
.long   3820343710,3820343710
.long   461924940,461924940
.long   3089050817,3089050817
.long   2136040774,2136040774
.long   82468509,82468509
.long   1563790337,1563790337
.long   1937016826,1937016826
.long   776014843,776014843
.long   1511876531,1511876531
.long   1389550482,1389550482
.long   861278441,861278441
.long   323475053,323475053
.long   2355222426,2355222426
.long   2047648055,2047648055
.long   2383738969,2383738969
.long   2302415851,2302415851
.long   3995576782,3995576782
.long   902390199,902390199
.long   3991215329,3991215329
.long   1018251130,1018251130
.long   1507840668,1507840668
.long   1064563285,1064563285
.long   2043548696,2043548696
.long   3208103795,3208103795
.long   3939366739,3939366739
.long   1537932639,1537932639
.long   342834655,342834655
.long   2262516856,2262516856
.long   2180231114,2180231114
.long   1053059257,1053059257
.long   741614648,741614648
.long   1598071746,1598071746
.long   1925389590,1925389590
.long   203809468,203809468
.long   2336832552,2336832552
.long   1100287487,1100287487
.long   1895934009,1895934009
.long   3736275976,3736275976
.long   2632234200,2632234200
.long   2428589668,2428589668
.long   1636092795,1636092795
.long   1890988757,1890988757
.long   1952214088,1952214088
.long   1113045200,1113045200
.byte   82,9,106,213,48,54,165,56
.byte   191,64,163,158,129,243,215,251
.byte   124,227,57,130,155,47,255,135
.byte   52,142,67,68,196,222,233,203
.byte   84,123,148,50,166,194,35,61
.byte   238,76,149,11,66,250,195,78
.byte   8,46,161,102,40,217,36,178
.byte   118,91,162,73,109,139,209,37
.byte   114,248,246,100,134,104,152,22
.byte   212,164,92,204,93,101,182,146
.byte   108,112,72,80,253,237,185,218
.byte   94,21,70,87,167,141,157,132
.byte   144,216,171,0,140,188,211,10
.byte   247,228,88,5,184,179,69,6
.byte   208,44,30,143,202,63,15,2
.byte   193,175,189,3,1,19,138,107
.byte   58,145,17,65,79,103,220,234
.byte   151,242,207,206,240,180,230,115
.byte   150,172,116,34,231,173,53,133
.byte   226,249,55,232,28,117,223,110
.byte   71,241,26,113,29,41,197,137
.byte   111,183,98,14,170,24,190,27
.byte   252,86,62,75,198,210,121,32
.byte   154,219,192,254,120,205,90,244
.byte   31,221,168,51,136,7,199,49
.byte   177,18,16,89,39,128,236,95
.byte   96,81,127,169,25,181,74,13
.byte   45,229,122,159,147,201,156,239
.byte   160,224,59,77,174,42,245,176
.byte   200,235,187,60,131,83,153,97
.byte   23,43,4,126,186,119,214,38
.byte   225,105,20,99,85,33,12,125
.byte   82,9,106,213,48,54,165,56
.byte   191,64,163,158,129,243,215,251
.byte   124,227,57,130,155,47,255,135
.byte   52,142,67,68,196,222,233,203
.byte   84,123,148,50,166,194,35,61
.byte   238,76,149,11,66,250,195,78
.byte   8,46,161,102,40,217,36,178
.byte   118,91,162,73,109,139,209,37
.byte   114,248,246,100,134,104,152,22
.byte   212,164,92,204,93,101,182,146
.byte   108,112,72,80,253,237,185,218
.byte   94,21,70,87,167,141,157,132
.byte   144,216,171,0,140,188,211,10
.byte   247,228,88,5,184,179,69,6
.byte   208,44,30,143,202,63,15,2
.byte   193,175,189,3,1,19,138,107
.byte   58,145,17,65,79,103,220,234
.byte   151,242,207,206,240,180,230,115
.byte   150,172,116,34,231,173,53,133
.byte   226,249,55,232,28,117,223,110
.byte   71,241,26,113,29,41,197,137
.byte   111,183,98,14,170,24,190,27
.byte   252,86,62,75,198,210,121,32
.byte   154,219,192,254,120,205,90,244
.byte   31,221,168,51,136,7,199,49
.byte   177,18,16,89,39,128,236,95
.byte   96,81,127,169,25,181,74,13
.byte   45,229,122,159,147,201,156,239
.byte   160,224,59,77,174,42,245,176
.byte   200,235,187,60,131,83,153,97
.byte   23,43,4,126,186,119,214,38
.byte   225,105,20,99,85,33,12,125
.byte   82,9,106,213,48,54,165,56
.byte   191,64,163,158,129,243,215,251
.byte   124,227,57,130,155,47,255,135
.byte   52,142,67,68,196,222,233,203
.byte   84,123,148,50,166,194,35,61
.byte   238,76,149,11,66,250,195,78
.byte   8,46,161,102,40,217,36,178
.byte   118,91,162,73,109,139,209,37
.byte   114,248,246,100,134,104,152,22
.byte   212,164,92,204,93,101,182,146
.byte   108,112,72,80,253,237,185,218
.byte   94,21,70,87,167,141,157,132
.byte   144,216,171,0,140,188,211,10
.byte   247,228,88,5,184,179,69,6
.byte   208,44,30,143,202,63,15,2
.byte   193,175,189,3,1,19,138,107
.byte   58,145,17,65,79,103,220,234
.byte   151,242,207,206,240,180,230,115
.byte   150,172,116,34,231,173,53,133
.byte   226,249,55,232,28,117,223,110
.byte   71,241,26,113,29,41,197,137
.byte   111,183,98,14,170,24,190,27
.byte   252,86,62,75,198,210,121,32
.byte   154,219,192,254,120,205,90,244
.byte   31,221,168,51,136,7,199,49
.byte   177,18,16,89,39,128,236,95
.byte   96,81,127,169,25,181,74,13
.byte   45,229,122,159,147,201,156,239
.byte   160,224,59,77,174,42,245,176
.byte   200,235,187,60,131,83,153,97
.byte   23,43,4,126,186,119,214,38
.byte   225,105,20,99,85,33,12,125
.byte   82,9,106,213,48,54,165,56
.byte   191,64,163,158,129,243,215,251
.byte   124,227,57,130,155,47,255,135
.byte   52,142,67,68,196,222,233,203
.byte   84,123,148,50,166,194,35,61
.byte   238,76,149,11,66,250,195,78
.byte   8,46,161,102,40,217,36,178
.byte   118,91,162,73,109,139,209,37
.byte   114,248,246,100,134,104,152,22
.byte   212,164,92,204,93,101,182,146
.byte   108,112,72,80,253,237,185,218
.byte   94,21,70,87,167,141,157,132
.byte   144,216,171,0,140,188,211,10
.byte   247,228,88,5,184,179,69,6
.byte   208,44,30,143,202,63,15,2
.byte   193,175,189,3,1,19,138,107
.byte   58,145,17,65,79,103,220,234
.byte   151,242,207,206,240,180,230,115
.byte   150,172,116,34,231,173,53,133
.byte   226,249,55,232,28,117,223,110
.byte   71,241,26,113,29,41,197,137
.byte   111,183,98,14,170,24,190,27
.byte   252,86,62,75,198,210,121,32
.byte   154,219,192,254,120,205,90,244
.byte   31,221,168,51,136,7,199,49
.byte   177,18,16,89,39,128,236,95
.byte   96,81,127,169,25,181,74,13
.byte   45,229,122,159,147,201,156,239
.byte   160,224,59,77,174,42,245,176
.byte   200,235,187,60,131,83,153,97
.byte   23,43,4,126,186,119,214,38
.byte   225,105,20,99,85,33,12,125
.size   _x86_AES_decrypt,.-_x86_AES_decrypt
.globl  AES_decrypt
.type   AES_decrypt,@function
.align  16
AES_decrypt:
.L_AES_decrypt_begin:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        pushl   %ebp
        pushl   %ebx
        pushl   %esi
        pushl   %edi
        movl    20(%esp),%esi
        movl    28(%esp),%edi
        movl    %esp,%eax
        subl    $36,%esp
        andl    $-64,%esp
        leal    -127(%edi),%ebx
        subl    %esp,%ebx
        negl    %ebx
        andl    $960,%ebx
        subl    %ebx,%esp
        addl    $4,%esp
        movl    %eax,28(%esp)
        call    .L010pic_point
.L010pic_point:
        popl    %ebp
        leal    OPENSSL_ia32cap_P,%eax
        leal    .LAES_Td-.L010pic_point(%ebp),%ebp
        leal    764(%esp),%ebx
        subl    %ebp,%ebx
        andl    $768,%ebx
        leal    2176(%ebp,%ebx,1),%ebp
        btl     $25,(%eax)
        jnc     .L011x86
        movq    (%esi),%mm0
        movq    8(%esi),%mm4
        call    _sse_AES_decrypt_compact
        movl    28(%esp),%esp
        movl    24(%esp),%esi
        movq    %mm0,(%esi)
        movq    %mm4,8(%esi)
        emms
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.align  16
.L011x86:
        movl    %ebp,24(%esp)
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        call    _x86_AES_decrypt_compact
        movl    28(%esp),%esp
        movl    24(%esp),%esi
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.size   AES_decrypt,.-.L_AES_decrypt_begin
.globl  AES_cbc_encrypt
.type   AES_cbc_encrypt,@function
.align  16
AES_cbc_encrypt:
.L_AES_cbc_encrypt_begin:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        pushl   %ebp
        pushl   %ebx
        pushl   %esi
        pushl   %edi
        movl    28(%esp),%ecx
        cmpl    $0,%ecx
        je      .L012drop_out
        call    .L013pic_point
.L013pic_point:
        popl    %ebp
        leal    OPENSSL_ia32cap_P,%eax
        cmpl    $0,40(%esp)
        leal    .LAES_Te-.L013pic_point(%ebp),%ebp
        jne     .L014picked_te
        leal    .LAES_Td-.LAES_Te(%ebp),%ebp
.L014picked_te:
        pushfl
        cld
        cmpl    $512,%ecx
        jb      .L015slow_way
        testl   $15,%ecx
        jnz     .L015slow_way
        btl     $28,(%eax)
        jc      .L015slow_way
        leal    -324(%esp),%esi
        andl    $-64,%esi
        movl    %ebp,%eax
        leal    2304(%ebp),%ebx
        movl    %esi,%edx
        andl    $4095,%eax
        andl    $4095,%ebx
        andl    $4095,%edx
        cmpl    %ebx,%edx
        jb      .L016tbl_break_out
        subl    %ebx,%edx
        subl    %edx,%esi
        jmp     .L017tbl_ok
.align  4
.L016tbl_break_out:
        subl    %eax,%edx
        andl    $4095,%edx
        addl    $384,%edx
        subl    %edx,%esi
.align  4
.L017tbl_ok:
        leal    24(%esp),%edx
        xchgl   %esi,%esp
        addl    $4,%esp
        movl    %ebp,24(%esp)
        movl    %esi,28(%esp)
        movl    (%edx),%eax
        movl    4(%edx),%ebx
        movl    12(%edx),%edi
        movl    16(%edx),%esi
        movl    20(%edx),%edx
        movl    %eax,32(%esp)
        movl    %ebx,36(%esp)
        movl    %ecx,40(%esp)
        movl    %edi,44(%esp)
        movl    %esi,48(%esp)
        movl    $0,316(%esp)
        movl    %edi,%ebx
        movl    $61,%ecx
        subl    %ebp,%ebx
        movl    %edi,%esi
        andl    $4095,%ebx
        leal    76(%esp),%edi
        cmpl    $2304,%ebx
        jb      .L018do_copy
        cmpl    $3852,%ebx
        jb      .L019skip_copy
.align  4
.L018do_copy:
        movl    %edi,44(%esp)
.long   2784229001
.L019skip_copy:
        movl    $16,%edi
.align  4
.L020prefetch_tbl:
        movl    (%ebp),%eax
        movl    32(%ebp),%ebx
        movl    64(%ebp),%ecx
        movl    96(%ebp),%esi
        leal    128(%ebp),%ebp
        subl    $1,%edi
        jnz     .L020prefetch_tbl
        subl    $2048,%ebp
        movl    32(%esp),%esi
        movl    48(%esp),%edi
        cmpl    $0,%edx
        je      .L021fast_decrypt
        movl    (%edi),%eax
        movl    4(%edi),%ebx
.align  16
.L022fast_enc_loop:
        movl    8(%edi),%ecx
        movl    12(%edi),%edx
        xorl    (%esi),%eax
        xorl    4(%esi),%ebx
        xorl    8(%esi),%ecx
        xorl    12(%esi),%edx
        movl    44(%esp),%edi
        call    _x86_AES_encrypt
        movl    32(%esp),%esi
        movl    36(%esp),%edi
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        leal    16(%esi),%esi
        movl    40(%esp),%ecx
        movl    %esi,32(%esp)
        leal    16(%edi),%edx
        movl    %edx,36(%esp)
        subl    $16,%ecx
        movl    %ecx,40(%esp)
        jnz     .L022fast_enc_loop
        movl    48(%esp),%esi
        movl    8(%edi),%ecx
        movl    12(%edi),%edx
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        cmpl    $0,316(%esp)
        movl    44(%esp),%edi
        je      .L023skip_ezero
        movl    $60,%ecx
        xorl    %eax,%eax
.align  4
.long   2884892297
.L023skip_ezero:
        movl    28(%esp),%esp
        popfl
.L012drop_out:
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L021fast_decrypt:
        cmpl    36(%esp),%esi
        je      .L024fast_dec_in_place
        movl    %edi,52(%esp)
.align  4
.align  16
.L025fast_dec_loop:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    44(%esp),%edi
        call    _x86_AES_decrypt
        movl    52(%esp),%edi
        movl    40(%esp),%esi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        movl    36(%esp),%edi
        movl    32(%esp),%esi
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    40(%esp),%ecx
        movl    %esi,52(%esp)
        leal    16(%esi),%esi
        movl    %esi,32(%esp)
        leal    16(%edi),%edi
        movl    %edi,36(%esp)
        subl    $16,%ecx
        movl    %ecx,40(%esp)
        jnz     .L025fast_dec_loop
        movl    52(%esp),%edi
        movl    48(%esp),%esi
        movl    (%edi),%eax
        movl    4(%edi),%ebx
        movl    8(%edi),%ecx
        movl    12(%edi),%edx
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        jmp     .L026fast_dec_out
.align  16
.L024fast_dec_in_place:
.L027fast_dec_in_place_loop:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        leal    60(%esp),%edi
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    44(%esp),%edi
        call    _x86_AES_decrypt
        movl    48(%esp),%edi
        movl    36(%esp),%esi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        leal    16(%esi),%esi
        movl    %esi,36(%esp)
        leal    60(%esp),%esi
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    32(%esp),%esi
        movl    40(%esp),%ecx
        leal    16(%esi),%esi
        movl    %esi,32(%esp)
        subl    $16,%ecx
        movl    %ecx,40(%esp)
        jnz     .L027fast_dec_in_place_loop
.align  4
.L026fast_dec_out:
        cmpl    $0,316(%esp)
        movl    44(%esp),%edi
        je      .L028skip_dzero
        movl    $60,%ecx
        xorl    %eax,%eax
.align  4
.long   2884892297
.L028skip_dzero:
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L015slow_way:
        movl    (%eax),%eax
        movl    36(%esp),%edi
        leal    -80(%esp),%esi
        andl    $-64,%esi
        leal    -143(%edi),%ebx
        subl    %esi,%ebx
        negl    %ebx
        andl    $960,%ebx
        subl    %ebx,%esi
        leal    768(%esi),%ebx
        subl    %ebp,%ebx
        andl    $768,%ebx
        leal    2176(%ebp,%ebx,1),%ebp
        leal    24(%esp),%edx
        xchgl   %esi,%esp
        addl    $4,%esp
        movl    %ebp,24(%esp)
        movl    %esi,28(%esp)
        movl    %eax,52(%esp)
        movl    (%edx),%eax
        movl    4(%edx),%ebx
        movl    16(%edx),%esi
        movl    20(%edx),%edx
        movl    %eax,32(%esp)
        movl    %ebx,36(%esp)
        movl    %ecx,40(%esp)
        movl    %edi,44(%esp)
        movl    %esi,48(%esp)
        movl    %esi,%edi
        movl    %eax,%esi
        cmpl    $0,%edx
        je      .L029slow_decrypt
        cmpl    $16,%ecx
        movl    %ebx,%edx
        jb      .L030slow_enc_tail
        btl     $25,52(%esp)
        jnc     .L031slow_enc_x86
        movq    (%edi),%mm0
        movq    8(%edi),%mm4
.align  16
.L032slow_enc_loop_sse:
        pxor    (%esi),%mm0
        pxor    8(%esi),%mm4
        movl    44(%esp),%edi
        call    _sse_AES_encrypt_compact
        movl    32(%esp),%esi
        movl    36(%esp),%edi
        movl    40(%esp),%ecx
        movq    %mm0,(%edi)
        movq    %mm4,8(%edi)
        leal    16(%esi),%esi
        movl    %esi,32(%esp)
        leal    16(%edi),%edx
        movl    %edx,36(%esp)
        subl    $16,%ecx
        cmpl    $16,%ecx
        movl    %ecx,40(%esp)
        jae     .L032slow_enc_loop_sse
        testl   $15,%ecx
        jnz     .L030slow_enc_tail
        movl    48(%esp),%esi
        movq    %mm0,(%esi)
        movq    %mm4,8(%esi)
        emms
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L031slow_enc_x86:
        movl    (%edi),%eax
        movl    4(%edi),%ebx
.align  4
.L033slow_enc_loop_x86:
        movl    8(%edi),%ecx
        movl    12(%edi),%edx
        xorl    (%esi),%eax
        xorl    4(%esi),%ebx
        xorl    8(%esi),%ecx
        xorl    12(%esi),%edx
        movl    44(%esp),%edi
        call    _x86_AES_encrypt_compact
        movl    32(%esp),%esi
        movl    36(%esp),%edi
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    40(%esp),%ecx
        leal    16(%esi),%esi
        movl    %esi,32(%esp)
        leal    16(%edi),%edx
        movl    %edx,36(%esp)
        subl    $16,%ecx
        cmpl    $16,%ecx
        movl    %ecx,40(%esp)
        jae     .L033slow_enc_loop_x86
        testl   $15,%ecx
        jnz     .L030slow_enc_tail
        movl    48(%esp),%esi
        movl    8(%edi),%ecx
        movl    12(%edi),%edx
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L030slow_enc_tail:
        emms
        movl    %edx,%edi
        movl    $16,%ebx
        subl    %ecx,%ebx
        cmpl    %esi,%edi
        je      .L034enc_in_place
.align  4
.long   2767451785
        jmp     .L035enc_skip_in_place
.L034enc_in_place:
        leal    (%edi,%ecx,1),%edi
.L035enc_skip_in_place:
        movl    %ebx,%ecx
        xorl    %eax,%eax
.align  4
.long   2868115081
        movl    48(%esp),%edi
        movl    %edx,%esi
        movl    (%edi),%eax
        movl    4(%edi),%ebx
        movl    $16,40(%esp)
        jmp     .L033slow_enc_loop_x86
.align  16
.L029slow_decrypt:
        btl     $25,52(%esp)
        jnc     .L036slow_dec_loop_x86
.align  4
.L037slow_dec_loop_sse:
        movq    (%esi),%mm0
        movq    8(%esi),%mm4
        movl    44(%esp),%edi
        call    _sse_AES_decrypt_compact
        movl    32(%esp),%esi
        leal    60(%esp),%eax
        movl    36(%esp),%ebx
        movl    40(%esp),%ecx
        movl    48(%esp),%edi
        movq    (%esi),%mm1
        movq    8(%esi),%mm5
        pxor    (%edi),%mm0
        pxor    8(%edi),%mm4
        movq    %mm1,(%edi)
        movq    %mm5,8(%edi)
        subl    $16,%ecx
        jc      .L038slow_dec_partial_sse
        movq    %mm0,(%ebx)
        movq    %mm4,8(%ebx)
        leal    16(%ebx),%ebx
        movl    %ebx,36(%esp)
        leal    16(%esi),%esi
        movl    %esi,32(%esp)
        movl    %ecx,40(%esp)
        jnz     .L037slow_dec_loop_sse
        emms
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L038slow_dec_partial_sse:
        movq    %mm0,(%eax)
        movq    %mm4,8(%eax)
        emms
        addl    $16,%ecx
        movl    %ebx,%edi
        movl    %eax,%esi
.align  4
.long   2767451785
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L036slow_dec_loop_x86:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        leal    60(%esp),%edi
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    44(%esp),%edi
        call    _x86_AES_decrypt_compact
        movl    48(%esp),%edi
        movl    40(%esp),%esi
        xorl    (%edi),%eax
        xorl    4(%edi),%ebx
        xorl    8(%edi),%ecx
        xorl    12(%edi),%edx
        subl    $16,%esi
        jc      .L039slow_dec_partial_x86
        movl    %esi,40(%esp)
        movl    36(%esp),%esi
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        leal    16(%esi),%esi
        movl    %esi,36(%esp)
        leal    60(%esp),%esi
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    32(%esp),%esi
        leal    16(%esi),%esi
        movl    %esi,32(%esp)
        jnz     .L036slow_dec_loop_x86
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
        pushfl
.align  16
.L039slow_dec_partial_x86:
        leal    60(%esp),%esi
        movl    %eax,(%esi)
        movl    %ebx,4(%esi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        movl    32(%esp),%esi
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    40(%esp),%ecx
        movl    36(%esp),%edi
        leal    60(%esp),%esi
.align  4
.long   2767451785
        movl    28(%esp),%esp
        popfl
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.size   AES_cbc_encrypt,.-.L_AES_cbc_encrypt_begin
.type   _x86_AES_set_encrypt_key,@function
.align  16
_x86_AES_set_encrypt_key:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        pushl   %ebp
        pushl   %ebx
        pushl   %esi
        pushl   %edi
        movl    24(%esp),%esi
        movl    32(%esp),%edi
        testl   $-1,%esi
        jz      .L040badpointer
        testl   $-1,%edi
        jz      .L040badpointer
        call    .L041pic_point
.L041pic_point:
        popl    %ebp
        leal    .LAES_Te-.L041pic_point(%ebp),%ebp
        leal    2176(%ebp),%ebp
        movl    -128(%ebp),%eax
        movl    -96(%ebp),%ebx
        movl    -64(%ebp),%ecx
        movl    -32(%ebp),%edx
        movl    (%ebp),%eax
        movl    32(%ebp),%ebx
        movl    64(%ebp),%ecx
        movl    96(%ebp),%edx
        movl    28(%esp),%ecx
        cmpl    $128,%ecx
        je      .L04210rounds
        cmpl    $192,%ecx
        je      .L04312rounds
        cmpl    $256,%ecx
        je      .L04414rounds
        movl    $-2,%eax
        jmp     .L045exit
.L04210rounds:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        xorl    %ecx,%ecx
        jmp     .L04610shortcut
.align  4
.L04710loop:
        movl    (%edi),%eax
        movl    12(%edi),%edx
.L04610shortcut:
        movzbl  %dl,%esi
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $24,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shrl    $16,%edx
        movzbl  %dl,%esi
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $8,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%eax
        xorl    896(%ebp,%ecx,4),%eax
        movl    %eax,16(%edi)
        xorl    4(%edi),%eax
        movl    %eax,20(%edi)
        xorl    8(%edi),%eax
        movl    %eax,24(%edi)
        xorl    12(%edi),%eax
        movl    %eax,28(%edi)
        incl    %ecx
        addl    $16,%edi
        cmpl    $10,%ecx
        jl      .L04710loop
        movl    $10,80(%edi)
        xorl    %eax,%eax
        jmp     .L045exit
.L04312rounds:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    16(%esi),%ecx
        movl    20(%esi),%edx
        movl    %ecx,16(%edi)
        movl    %edx,20(%edi)
        xorl    %ecx,%ecx
        jmp     .L04812shortcut
.align  4
.L04912loop:
        movl    (%edi),%eax
        movl    20(%edi),%edx
.L04812shortcut:
        movzbl  %dl,%esi
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $24,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shrl    $16,%edx
        movzbl  %dl,%esi
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $8,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%eax
        xorl    896(%ebp,%ecx,4),%eax
        movl    %eax,24(%edi)
        xorl    4(%edi),%eax
        movl    %eax,28(%edi)
        xorl    8(%edi),%eax
        movl    %eax,32(%edi)
        xorl    12(%edi),%eax
        movl    %eax,36(%edi)
        cmpl    $7,%ecx
        je      .L05012break
        incl    %ecx
        xorl    16(%edi),%eax
        movl    %eax,40(%edi)
        xorl    20(%edi),%eax
        movl    %eax,44(%edi)
        addl    $24,%edi
        jmp     .L04912loop
.L05012break:
        movl    $12,72(%edi)
        xorl    %eax,%eax
        jmp     .L045exit
.L04414rounds:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    8(%esi),%ecx
        movl    12(%esi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,8(%edi)
        movl    %edx,12(%edi)
        movl    16(%esi),%eax
        movl    20(%esi),%ebx
        movl    24(%esi),%ecx
        movl    28(%esi),%edx
        movl    %eax,16(%edi)
        movl    %ebx,20(%edi)
        movl    %ecx,24(%edi)
        movl    %edx,28(%edi)
        xorl    %ecx,%ecx
        jmp     .L05114shortcut
.align  4
.L05214loop:
        movl    28(%edi),%edx
.L05114shortcut:
        movl    (%edi),%eax
        movzbl  %dl,%esi
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $24,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shrl    $16,%edx
        movzbl  %dl,%esi
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $8,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shll    $16,%ebx
        xorl    %ebx,%eax
        xorl    896(%ebp,%ecx,4),%eax
        movl    %eax,32(%edi)
        xorl    4(%edi),%eax
        movl    %eax,36(%edi)
        xorl    8(%edi),%eax
        movl    %eax,40(%edi)
        xorl    12(%edi),%eax
        movl    %eax,44(%edi)
        cmpl    $6,%ecx
        je      .L05314break
        incl    %ecx
        movl    %eax,%edx
        movl    16(%edi),%eax
        movzbl  %dl,%esi
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shrl    $16,%edx
        shll    $8,%ebx
        movzbl  %dl,%esi
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        movzbl  %dh,%esi
        shll    $16,%ebx
        xorl    %ebx,%eax
        movzbl  -128(%ebp,%esi,1),%ebx
        shll    $24,%ebx
        xorl    %ebx,%eax
        movl    %eax,48(%edi)
        xorl    20(%edi),%eax
        movl    %eax,52(%edi)
        xorl    24(%edi),%eax
        movl    %eax,56(%edi)
        xorl    28(%edi),%eax
        movl    %eax,60(%edi)
        addl    $32,%edi
        jmp     .L05214loop
.L05314break:
        movl    $14,48(%edi)
        xorl    %eax,%eax
        jmp     .L045exit
.L040badpointer:
        movl    $-1,%eax
.L045exit:
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.size   _x86_AES_set_encrypt_key,.-_x86_AES_set_encrypt_key
.globl  AES_set_encrypt_key
.type   AES_set_encrypt_key,@function
.align  16
AES_set_encrypt_key:
.L_AES_set_encrypt_key_begin:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        call    _x86_AES_set_encrypt_key
        ret
.size   AES_set_encrypt_key,.-.L_AES_set_encrypt_key_begin
.globl  AES_set_decrypt_key
.type   AES_set_decrypt_key,@function
.align  16
AES_set_decrypt_key:
.L_AES_set_decrypt_key_begin:
        #ifdef __CET__

.byte   243,15,30,251
        #endif

        call    _x86_AES_set_encrypt_key
        cmpl    $0,%eax
        je      .L054proceed
        ret
.L054proceed:
        pushl   %ebp
        pushl   %ebx
        pushl   %esi
        pushl   %edi
        movl    28(%esp),%esi
        movl    240(%esi),%ecx
        leal    (,%ecx,4),%ecx
        leal    (%esi,%ecx,4),%edi
.align  4
.L055invert:
        movl    (%esi),%eax
        movl    4(%esi),%ebx
        movl    (%edi),%ecx
        movl    4(%edi),%edx
        movl    %eax,(%edi)
        movl    %ebx,4(%edi)
        movl    %ecx,(%esi)
        movl    %edx,4(%esi)
        movl    8(%esi),%eax
        movl    12(%esi),%ebx
        movl    8(%edi),%ecx
        movl    12(%edi),%edx
        movl    %eax,8(%edi)
        movl    %ebx,12(%edi)
        movl    %ecx,8(%esi)
        movl    %edx,12(%esi)
        addl    $16,%esi
        subl    $16,%edi
        cmpl    %edi,%esi
        jne     .L055invert
        movl    28(%esp),%edi
        movl    240(%edi),%esi
        leal    -2(%esi,%esi,1),%esi
        leal    (%edi,%esi,8),%esi
        movl    %esi,28(%esp)
        movl    16(%edi),%eax
.align  4
.L056permute:
        addl    $16,%edi
        movl    $2155905152,%ebp
        andl    %eax,%ebp
        leal    (%eax,%eax,1),%ebx
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%ebx
        andl    $454761243,%esi
        xorl    %esi,%ebx
        movl    $2155905152,%ebp
        andl    %ebx,%ebp
        leal    (%ebx,%ebx,1),%ecx
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%ecx
        andl    $454761243,%esi
        xorl    %eax,%ebx
        xorl    %esi,%ecx
        movl    $2155905152,%ebp
        andl    %ecx,%ebp
        leal    (%ecx,%ecx,1),%edx
        movl    %ebp,%esi
        shrl    $7,%ebp
        xorl    %eax,%ecx
        subl    %ebp,%esi
        andl    $4278124286,%edx
        andl    $454761243,%esi
        roll    $8,%eax
        xorl    %esi,%edx
        movl    4(%edi),%ebp
        xorl    %ebx,%eax
        xorl    %edx,%ebx
        xorl    %ecx,%eax
        roll    $24,%ebx
        xorl    %edx,%ecx
        xorl    %edx,%eax
        roll    $16,%ecx
        xorl    %ebx,%eax
        roll    $8,%edx
        xorl    %ecx,%eax
        movl    %ebp,%ebx
        xorl    %edx,%eax
        movl    %eax,(%edi)
        movl    $2155905152,%ebp
        andl    %ebx,%ebp
        leal    (%ebx,%ebx,1),%ecx
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%ecx
        andl    $454761243,%esi
        xorl    %esi,%ecx
        movl    $2155905152,%ebp
        andl    %ecx,%ebp
        leal    (%ecx,%ecx,1),%edx
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%edx
        andl    $454761243,%esi
        xorl    %ebx,%ecx
        xorl    %esi,%edx
        movl    $2155905152,%ebp
        andl    %edx,%ebp
        leal    (%edx,%edx,1),%eax
        movl    %ebp,%esi
        shrl    $7,%ebp
        xorl    %ebx,%edx
        subl    %ebp,%esi
        andl    $4278124286,%eax
        andl    $454761243,%esi
        roll    $8,%ebx
        xorl    %esi,%eax
        movl    8(%edi),%ebp
        xorl    %ecx,%ebx
        xorl    %eax,%ecx
        xorl    %edx,%ebx
        roll    $24,%ecx
        xorl    %eax,%edx
        xorl    %eax,%ebx
        roll    $16,%edx
        xorl    %ecx,%ebx
        roll    $8,%eax
        xorl    %edx,%ebx
        movl    %ebp,%ecx
        xorl    %eax,%ebx
        movl    %ebx,4(%edi)
        movl    $2155905152,%ebp
        andl    %ecx,%ebp
        leal    (%ecx,%ecx,1),%edx
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%edx
        andl    $454761243,%esi
        xorl    %esi,%edx
        movl    $2155905152,%ebp
        andl    %edx,%ebp
        leal    (%edx,%edx,1),%eax
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%eax
        andl    $454761243,%esi
        xorl    %ecx,%edx
        xorl    %esi,%eax
        movl    $2155905152,%ebp
        andl    %eax,%ebp
        leal    (%eax,%eax,1),%ebx
        movl    %ebp,%esi
        shrl    $7,%ebp
        xorl    %ecx,%eax
        subl    %ebp,%esi
        andl    $4278124286,%ebx
        andl    $454761243,%esi
        roll    $8,%ecx
        xorl    %esi,%ebx
        movl    12(%edi),%ebp
        xorl    %edx,%ecx
        xorl    %ebx,%edx
        xorl    %eax,%ecx
        roll    $24,%edx
        xorl    %ebx,%eax
        xorl    %ebx,%ecx
        roll    $16,%eax
        xorl    %edx,%ecx
        roll    $8,%ebx
        xorl    %eax,%ecx
        movl    %ebp,%edx
        xorl    %ebx,%ecx
        movl    %ecx,8(%edi)
        movl    $2155905152,%ebp
        andl    %edx,%ebp
        leal    (%edx,%edx,1),%eax
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%eax
        andl    $454761243,%esi
        xorl    %esi,%eax
        movl    $2155905152,%ebp
        andl    %eax,%ebp
        leal    (%eax,%eax,1),%ebx
        movl    %ebp,%esi
        shrl    $7,%ebp
        subl    %ebp,%esi
        andl    $4278124286,%ebx
        andl    $454761243,%esi
        xorl    %edx,%eax
        xorl    %esi,%ebx
        movl    $2155905152,%ebp
        andl    %ebx,%ebp
        leal    (%ebx,%ebx,1),%ecx
        movl    %ebp,%esi
        shrl    $7,%ebp
        xorl    %edx,%ebx
        subl    %ebp,%esi
        andl    $4278124286,%ecx
        andl    $454761243,%esi
        roll    $8,%edx
        xorl    %esi,%ecx
        movl    16(%edi),%ebp
        xorl    %eax,%edx
        xorl    %ecx,%eax
        xorl    %ebx,%edx
        roll    $24,%eax
        xorl    %ecx,%ebx
        xorl    %ecx,%edx
        roll    $16,%ebx
        xorl    %eax,%edx
        roll    $8,%ecx
        xorl    %ebx,%edx
        movl    %ebp,%eax
        xorl    %ecx,%edx
        movl    %edx,12(%edi)
        cmpl    28(%esp),%edi
        jb      .L056permute
        xorl    %eax,%eax
        popl    %edi
        popl    %esi
        popl    %ebx
        popl    %ebp
        ret
.size   AES_set_decrypt_key,.-.L_AES_set_decrypt_key_begin
.byte   65,69,83,32,102,111,114,32,120,56,54,44,32,67,82,89
.byte   80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114
.byte   111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
.comm   OPENSSL_ia32cap_P,16,4

        .section ".note.gnu.property", "a"
        .p2align 2
        .long 1f - 0f
        .long 4f - 1f
        .long 5
0:
        .asciz "GNU"
1:
        .p2align 2
        .long 0xc0000002
        .long 3f - 2f
2:
        .long 3
3:
        .p2align 2
4:
