yuezonghe | 824eb0c | 2024-06-27 02:32:26 -0700 | [diff] [blame] | 1 | /*
|
| 2 | ;uInt longest_match_x64(
|
| 3 | ; deflate_state *s,
|
| 4 | ; IPos cur_match); // current match
|
| 5 |
|
| 6 | ; gvmat64.S -- Asm portion of the optimized longest_match for 32 bits x86_64
|
| 7 | ; (AMD64 on Athlon 64, Opteron, Phenom
|
| 8 | ; and Intel EM64T on Pentium 4 with EM64T, Pentium D, Core 2 Duo, Core I5/I7)
|
| 9 | ; this file is translation from gvmat64.asm to GCC 4.x (for Linux, Mac XCode)
|
| 10 | ; Copyright (C) 1995-2010 Jean-loup Gailly, Brian Raiter and Gilles Vollant.
|
| 11 | ;
|
| 12 | ; File written by Gilles Vollant, by converting to assembly the longest_match
|
| 13 | ; from Jean-loup Gailly in deflate.c of zLib and infoZip zip.
|
| 14 | ; and by taking inspiration on asm686 with masm, optimised assembly code
|
| 15 | ; from Brian Raiter, written 1998
|
| 16 | ;
|
| 17 | ; This software is provided 'as-is', without any express or implied
|
| 18 | ; warranty. In no event will the authors be held liable for any damages
|
| 19 | ; arising from the use of this software.
|
| 20 | ;
|
| 21 | ; Permission is granted to anyone to use this software for any purpose,
|
| 22 | ; including commercial applications, and to alter it and redistribute it
|
| 23 | ; freely, subject to the following restrictions:
|
| 24 | ;
|
| 25 | ; 1. The origin of this software must not be misrepresented; you must not
|
| 26 | ; claim that you wrote the original software. If you use this software
|
| 27 | ; in a product, an acknowledgment in the product documentation would be
|
| 28 | ; appreciated but is not required.
|
| 29 | ; 2. Altered source versions must be plainly marked as such, and must not be
|
| 30 | ; misrepresented as being the original software
|
| 31 | ; 3. This notice may not be removed or altered from any source distribution.
|
| 32 | ;
|
| 33 | ; http://www.zlib.net
|
| 34 | ; http://www.winimage.com/zLibDll
|
| 35 | ; http://www.muppetlabs.com/~breadbox/software/assembly.html
|
| 36 | ;
|
| 37 | ; to compile this file for zLib, I use option:
|
| 38 | ; gcc -c -arch x86_64 gvmat64.S
|
| 39 |
|
| 40 |
|
| 41 | ;uInt longest_match(s, cur_match)
|
| 42 | ; deflate_state *s;
|
| 43 | ; IPos cur_match; // current match /
|
| 44 | ;
|
| 45 | ; with XCode for Mac, I had strange error with some jump on intel syntax
|
| 46 | ; this is why BEFORE_JMP and AFTER_JMP are used
|
| 47 | */
|
| 48 |
|
| 49 |
|
| 50 | #define BEFORE_JMP .att_syntax
|
| 51 | #define AFTER_JMP .intel_syntax noprefix
|
| 52 |
|
| 53 | #ifndef NO_UNDERLINE
|
| 54 | # define match_init _match_init
|
| 55 | # define longest_match _longest_match
|
| 56 | #endif
|
| 57 |
|
| 58 | .intel_syntax noprefix
|
| 59 |
|
| 60 | .globl match_init, longest_match
|
| 61 | .text
|
| 62 | longest_match:
|
| 63 |
|
| 64 |
|
| 65 |
|
| 66 | #define LocalVarsSize 96
|
| 67 | /*
|
| 68 | ; register used : rax,rbx,rcx,rdx,rsi,rdi,r8,r9,r10,r11,r12
|
| 69 | ; free register : r14,r15
|
| 70 | ; register can be saved : rsp
|
| 71 | */
|
| 72 |
|
| 73 | #define chainlenwmask (rsp + 8 - LocalVarsSize)
|
| 74 | #define nicematch (rsp + 16 - LocalVarsSize)
|
| 75 |
|
| 76 | #define save_rdi (rsp + 24 - LocalVarsSize)
|
| 77 | #define save_rsi (rsp + 32 - LocalVarsSize)
|
| 78 | #define save_rbx (rsp + 40 - LocalVarsSize)
|
| 79 | #define save_rbp (rsp + 48 - LocalVarsSize)
|
| 80 | #define save_r12 (rsp + 56 - LocalVarsSize)
|
| 81 | #define save_r13 (rsp + 64 - LocalVarsSize)
|
| 82 | #define save_r14 (rsp + 72 - LocalVarsSize)
|
| 83 | #define save_r15 (rsp + 80 - LocalVarsSize)
|
| 84 |
|
| 85 |
|
| 86 | /*
|
| 87 | ; all the +4 offsets are due to the addition of pending_buf_size (in zlib
|
| 88 | ; in the deflate_state structure since the asm code was first written
|
| 89 | ; (if you compile with zlib 1.0.4 or older, remove the +4).
|
| 90 | ; Note : these value are good with a 8 bytes boundary pack structure
|
| 91 | */
|
| 92 |
|
| 93 | #define MAX_MATCH 258
|
| 94 | #define MIN_MATCH 3
|
| 95 | #define MIN_LOOKAHEAD (MAX_MATCH+MIN_MATCH+1)
|
| 96 |
|
| 97 | /*
|
| 98 | ;;; Offsets for fields in the deflate_state structure. These numbers
|
| 99 | ;;; are calculated from the definition of deflate_state, with the
|
| 100 | ;;; assumption that the compiler will dword-align the fields. (Thus,
|
| 101 | ;;; changing the definition of deflate_state could easily cause this
|
| 102 | ;;; program to crash horribly, without so much as a warning at
|
| 103 | ;;; compile time. Sigh.)
|
| 104 |
|
| 105 | ; all the +zlib1222add offsets are due to the addition of fields
|
| 106 | ; in zlib in the deflate_state structure since the asm code was first written
|
| 107 | ; (if you compile with zlib 1.0.4 or older, use "zlib1222add equ (-4)").
|
| 108 | ; (if you compile with zlib between 1.0.5 and 1.2.2.1, use "zlib1222add equ 0").
|
| 109 | ; if you compile with zlib 1.2.2.2 or later , use "zlib1222add equ 8").
|
| 110 | */
|
| 111 |
|
| 112 |
|
| 113 |
|
| 114 | /* you can check the structure offset by running
|
| 115 |
|
| 116 | #include <stdlib.h>
|
| 117 | #include <stdio.h>
|
| 118 | #include "deflate.h"
|
| 119 |
|
| 120 | void print_depl()
|
| 121 | {
|
| 122 | deflate_state ds;
|
| 123 | deflate_state *s=&ds;
|
| 124 | printf("size pointer=%u\n",(int)sizeof(void*));
|
| 125 |
|
| 126 | printf("#define dsWSize %u\n",(int)(((char*)&(s->w_size))-((char*)s)));
|
| 127 | printf("#define dsWMask %u\n",(int)(((char*)&(s->w_mask))-((char*)s)));
|
| 128 | printf("#define dsWindow %u\n",(int)(((char*)&(s->window))-((char*)s)));
|
| 129 | printf("#define dsPrev %u\n",(int)(((char*)&(s->prev))-((char*)s)));
|
| 130 | printf("#define dsMatchLen %u\n",(int)(((char*)&(s->match_length))-((char*)s)));
|
| 131 | printf("#define dsPrevMatch %u\n",(int)(((char*)&(s->prev_match))-((char*)s)));
|
| 132 | printf("#define dsStrStart %u\n",(int)(((char*)&(s->strstart))-((char*)s)));
|
| 133 | printf("#define dsMatchStart %u\n",(int)(((char*)&(s->match_start))-((char*)s)));
|
| 134 | printf("#define dsLookahead %u\n",(int)(((char*)&(s->lookahead))-((char*)s)));
|
| 135 | printf("#define dsPrevLen %u\n",(int)(((char*)&(s->prev_length))-((char*)s)));
|
| 136 | printf("#define dsMaxChainLen %u\n",(int)(((char*)&(s->max_chain_length))-((char*)s)));
|
| 137 | printf("#define dsGoodMatch %u\n",(int)(((char*)&(s->good_match))-((char*)s)));
|
| 138 | printf("#define dsNiceMatch %u\n",(int)(((char*)&(s->nice_match))-((char*)s)));
|
| 139 | }
|
| 140 | */
|
| 141 |
|
| 142 | #define dsWSize 68
|
| 143 | #define dsWMask 76
|
| 144 | #define dsWindow 80
|
| 145 | #define dsPrev 96
|
| 146 | #define dsMatchLen 144
|
| 147 | #define dsPrevMatch 148
|
| 148 | #define dsStrStart 156
|
| 149 | #define dsMatchStart 160
|
| 150 | #define dsLookahead 164
|
| 151 | #define dsPrevLen 168
|
| 152 | #define dsMaxChainLen 172
|
| 153 | #define dsGoodMatch 188
|
| 154 | #define dsNiceMatch 192
|
| 155 |
|
| 156 | #define window_size [ rcx + dsWSize]
|
| 157 | #define WMask [ rcx + dsWMask]
|
| 158 | #define window_ad [ rcx + dsWindow]
|
| 159 | #define prev_ad [ rcx + dsPrev]
|
| 160 | #define strstart [ rcx + dsStrStart]
|
| 161 | #define match_start [ rcx + dsMatchStart]
|
| 162 | #define Lookahead [ rcx + dsLookahead] //; 0ffffffffh on infozip
|
| 163 | #define prev_length [ rcx + dsPrevLen]
|
| 164 | #define max_chain_length [ rcx + dsMaxChainLen]
|
| 165 | #define good_match [ rcx + dsGoodMatch]
|
| 166 | #define nice_match [ rcx + dsNiceMatch]
|
| 167 |
|
| 168 | /*
|
| 169 | ; windows:
|
| 170 | ; parameter 1 in rcx(deflate state s), param 2 in rdx (cur match)
|
| 171 |
|
| 172 | ; see http://weblogs.asp.net/oldnewthing/archive/2004/01/14/58579.aspx and
|
| 173 | ; http://msdn.microsoft.com/library/en-us/kmarch/hh/kmarch/64bitAMD_8e951dd2-ee77-4728-8702-55ce4b5dd24a.xml.asp
|
| 174 | ;
|
| 175 | ; All registers must be preserved across the call, except for
|
| 176 | ; rax, rcx, rdx, r8, r9, r10, and r11, which are scratch.
|
| 177 |
|
| 178 | ;
|
| 179 | ; gcc on macosx-linux:
|
| 180 | ; see http://www.x86-64.org/documentation/abi-0.99.pdf
|
| 181 | ; param 1 in rdi, param 2 in rsi
|
| 182 | ; rbx, rsp, rbp, r12 to r15 must be preserved
|
| 183 |
|
| 184 | ;;; Save registers that the compiler may be using, and adjust esp to
|
| 185 | ;;; make room for our stack frame.
|
| 186 |
|
| 187 |
|
| 188 | ;;; Retrieve the function arguments. r8d will hold cur_match
|
| 189 | ;;; throughout the entire function. edx will hold the pointer to the
|
| 190 | ;;; deflate_state structure during the function's setup (before
|
| 191 | ;;; entering the main loop.
|
| 192 |
|
| 193 | ; ms: parameter 1 in rcx (deflate_state* s), param 2 in edx -> r8 (cur match)
|
| 194 | ; mac: param 1 in rdi, param 2 rsi
|
| 195 | ; this clear high 32 bits of r8, which can be garbage in both r8 and rdx
|
| 196 | */
|
| 197 | mov [save_rbx],rbx
|
| 198 | mov [save_rbp],rbp
|
| 199 |
|
| 200 |
|
| 201 | mov rcx,rdi
|
| 202 |
|
| 203 | mov r8d,esi
|
| 204 |
|
| 205 |
|
| 206 | mov [save_r12],r12
|
| 207 | mov [save_r13],r13
|
| 208 | mov [save_r14],r14
|
| 209 | mov [save_r15],r15
|
| 210 |
|
| 211 |
|
| 212 | //;;; uInt wmask = s->w_mask;
|
| 213 | //;;; unsigned chain_length = s->max_chain_length;
|
| 214 | //;;; if (s->prev_length >= s->good_match) {
|
| 215 | //;;; chain_length >>= 2;
|
| 216 | //;;; }
|
| 217 |
|
| 218 |
|
| 219 | mov edi, prev_length
|
| 220 | mov esi, good_match
|
| 221 | mov eax, WMask
|
| 222 | mov ebx, max_chain_length
|
| 223 | cmp edi, esi
|
| 224 | jl LastMatchGood
|
| 225 | shr ebx, 2
|
| 226 | LastMatchGood:
|
| 227 |
|
| 228 | //;;; chainlen is decremented once beforehand so that the function can
|
| 229 | //;;; use the sign flag instead of the zero flag for the exit test.
|
| 230 | //;;; It is then shifted into the high word, to make room for the wmask
|
| 231 | //;;; value, which it will always accompany.
|
| 232 |
|
| 233 | dec ebx
|
| 234 | shl ebx, 16
|
| 235 | or ebx, eax
|
| 236 |
|
| 237 | //;;; on zlib only
|
| 238 | //;;; if ((uInt)nice_match > s->lookahead) nice_match = s->lookahead;
|
| 239 |
|
| 240 |
|
| 241 |
|
| 242 | mov eax, nice_match
|
| 243 | mov [chainlenwmask], ebx
|
| 244 | mov r10d, Lookahead
|
| 245 | cmp r10d, eax
|
| 246 | cmovnl r10d, eax
|
| 247 | mov [nicematch],r10d
|
| 248 |
|
| 249 |
|
| 250 |
|
| 251 | //;;; register Bytef *scan = s->window + s->strstart;
|
| 252 | mov r10, window_ad
|
| 253 | mov ebp, strstart
|
| 254 | lea r13, [r10 + rbp]
|
| 255 |
|
| 256 | //;;; Determine how many bytes the scan ptr is off from being
|
| 257 | //;;; dword-aligned.
|
| 258 |
|
| 259 | mov r9,r13
|
| 260 | neg r13
|
| 261 | and r13,3
|
| 262 |
|
| 263 | //;;; IPos limit = s->strstart > (IPos)MAX_DIST(s) ?
|
| 264 | //;;; s->strstart - (IPos)MAX_DIST(s) : NIL;
|
| 265 |
|
| 266 |
|
| 267 | mov eax, window_size
|
| 268 | sub eax, MIN_LOOKAHEAD
|
| 269 |
|
| 270 |
|
| 271 | xor edi,edi
|
| 272 | sub ebp, eax
|
| 273 |
|
| 274 | mov r11d, prev_length
|
| 275 |
|
| 276 | cmovng ebp,edi
|
| 277 |
|
| 278 | //;;; int best_len = s->prev_length;
|
| 279 |
|
| 280 |
|
| 281 | //;;; Store the sum of s->window + best_len in esi locally, and in esi.
|
| 282 |
|
| 283 | lea rsi,[r10+r11]
|
| 284 |
|
| 285 | //;;; register ush scan_start = *(ushf*)scan;
|
| 286 | //;;; register ush scan_end = *(ushf*)(scan+best_len-1);
|
| 287 | //;;; Posf *prev = s->prev;
|
| 288 |
|
| 289 | movzx r12d,word ptr [r9]
|
| 290 | movzx ebx, word ptr [r9 + r11 - 1]
|
| 291 |
|
| 292 | mov rdi, prev_ad
|
| 293 |
|
| 294 | //;;; Jump into the main loop.
|
| 295 |
|
| 296 | mov edx, [chainlenwmask]
|
| 297 |
|
| 298 | cmp bx,word ptr [rsi + r8 - 1]
|
| 299 | jz LookupLoopIsZero
|
| 300 |
|
| 301 |
|
| 302 |
|
| 303 | LookupLoop1:
|
| 304 | and r8d, edx
|
| 305 |
|
| 306 | movzx r8d, word ptr [rdi + r8*2]
|
| 307 | cmp r8d, ebp
|
| 308 | jbe LeaveNow
|
| 309 |
|
| 310 |
|
| 311 |
|
| 312 | sub edx, 0x00010000
|
| 313 | BEFORE_JMP
|
| 314 | js LeaveNow
|
| 315 | AFTER_JMP
|
| 316 |
|
| 317 | LoopEntry1:
|
| 318 | cmp bx,word ptr [rsi + r8 - 1]
|
| 319 | BEFORE_JMP
|
| 320 | jz LookupLoopIsZero
|
| 321 | AFTER_JMP
|
| 322 |
|
| 323 | LookupLoop2:
|
| 324 | and r8d, edx
|
| 325 |
|
| 326 | movzx r8d, word ptr [rdi + r8*2]
|
| 327 | cmp r8d, ebp
|
| 328 | BEFORE_JMP
|
| 329 | jbe LeaveNow
|
| 330 | AFTER_JMP
|
| 331 | sub edx, 0x00010000
|
| 332 | BEFORE_JMP
|
| 333 | js LeaveNow
|
| 334 | AFTER_JMP
|
| 335 |
|
| 336 | LoopEntry2:
|
| 337 | cmp bx,word ptr [rsi + r8 - 1]
|
| 338 | BEFORE_JMP
|
| 339 | jz LookupLoopIsZero
|
| 340 | AFTER_JMP
|
| 341 |
|
| 342 | LookupLoop4:
|
| 343 | and r8d, edx
|
| 344 |
|
| 345 | movzx r8d, word ptr [rdi + r8*2]
|
| 346 | cmp r8d, ebp
|
| 347 | BEFORE_JMP
|
| 348 | jbe LeaveNow
|
| 349 | AFTER_JMP
|
| 350 | sub edx, 0x00010000
|
| 351 | BEFORE_JMP
|
| 352 | js LeaveNow
|
| 353 | AFTER_JMP
|
| 354 |
|
| 355 | LoopEntry4:
|
| 356 |
|
| 357 | cmp bx,word ptr [rsi + r8 - 1]
|
| 358 | BEFORE_JMP
|
| 359 | jnz LookupLoop1
|
| 360 | jmp LookupLoopIsZero
|
| 361 | AFTER_JMP
|
| 362 | /*
|
| 363 | ;;; do {
|
| 364 | ;;; match = s->window + cur_match;
|
| 365 | ;;; if (*(ushf*)(match+best_len-1) != scan_end ||
|
| 366 | ;;; *(ushf*)match != scan_start) continue;
|
| 367 | ;;; [...]
|
| 368 | ;;; } while ((cur_match = prev[cur_match & wmask]) > limit
|
| 369 | ;;; && --chain_length != 0);
|
| 370 | ;;;
|
| 371 | ;;; Here is the inner loop of the function. The function will spend the
|
| 372 | ;;; majority of its time in this loop, and majority of that time will
|
| 373 | ;;; be spent in the first ten instructions.
|
| 374 | ;;;
|
| 375 | ;;; Within this loop:
|
| 376 | ;;; ebx = scanend
|
| 377 | ;;; r8d = curmatch
|
| 378 | ;;; edx = chainlenwmask - i.e., ((chainlen << 16) | wmask)
|
| 379 | ;;; esi = windowbestlen - i.e., (window + bestlen)
|
| 380 | ;;; edi = prev
|
| 381 | ;;; ebp = limit
|
| 382 | */
|
| 383 | .balign 16
|
| 384 | LookupLoop:
|
| 385 | and r8d, edx
|
| 386 |
|
| 387 | movzx r8d, word ptr [rdi + r8*2]
|
| 388 | cmp r8d, ebp
|
| 389 | BEFORE_JMP
|
| 390 | jbe LeaveNow
|
| 391 | AFTER_JMP
|
| 392 | sub edx, 0x00010000
|
| 393 | BEFORE_JMP
|
| 394 | js LeaveNow
|
| 395 | AFTER_JMP
|
| 396 |
|
| 397 | LoopEntry:
|
| 398 |
|
| 399 | cmp bx,word ptr [rsi + r8 - 1]
|
| 400 | BEFORE_JMP
|
| 401 | jnz LookupLoop1
|
| 402 | AFTER_JMP
|
| 403 | LookupLoopIsZero:
|
| 404 | cmp r12w, word ptr [r10 + r8]
|
| 405 | BEFORE_JMP
|
| 406 | jnz LookupLoop1
|
| 407 | AFTER_JMP
|
| 408 |
|
| 409 |
|
| 410 | //;;; Store the current value of chainlen.
|
| 411 | mov [chainlenwmask], edx
|
| 412 | /*
|
| 413 | ;;; Point edi to the string under scrutiny, and esi to the string we
|
| 414 | ;;; are hoping to match it up with. In actuality, esi and edi are
|
| 415 | ;;; both pointed (MAX_MATCH_8 - scanalign) bytes ahead, and edx is
|
| 416 | ;;; initialized to -(MAX_MATCH_8 - scanalign).
|
| 417 | */
|
| 418 | lea rsi,[r8+r10]
|
| 419 | mov rdx, 0xfffffffffffffef8 //; -(MAX_MATCH_8)
|
| 420 | lea rsi, [rsi + r13 + 0x0108] //;MAX_MATCH_8]
|
| 421 | lea rdi, [r9 + r13 + 0x0108] //;MAX_MATCH_8]
|
| 422 |
|
| 423 | prefetcht1 [rsi+rdx]
|
| 424 | prefetcht1 [rdi+rdx]
|
| 425 |
|
| 426 | /*
|
| 427 | ;;; Test the strings for equality, 8 bytes at a time. At the end,
|
| 428 | ;;; adjust rdx so that it is offset to the exact byte that mismatched.
|
| 429 | ;;;
|
| 430 | ;;; We already know at this point that the first three bytes of the
|
| 431 | ;;; strings match each other, and they can be safely passed over before
|
| 432 | ;;; starting the compare loop. So what this code does is skip over 0-3
|
| 433 | ;;; bytes, as much as necessary in order to dword-align the edi
|
| 434 | ;;; pointer. (rsi will still be misaligned three times out of four.)
|
| 435 | ;;;
|
| 436 | ;;; It should be confessed that this loop usually does not represent
|
| 437 | ;;; much of the total running time. Replacing it with a more
|
| 438 | ;;; straightforward "rep cmpsb" would not drastically degrade
|
| 439 | ;;; performance.
|
| 440 | */
|
| 441 |
|
| 442 | LoopCmps:
|
| 443 | mov rax, [rsi + rdx]
|
| 444 | xor rax, [rdi + rdx]
|
| 445 | jnz LeaveLoopCmps
|
| 446 |
|
| 447 | mov rax, [rsi + rdx + 8]
|
| 448 | xor rax, [rdi + rdx + 8]
|
| 449 | jnz LeaveLoopCmps8
|
| 450 |
|
| 451 |
|
| 452 | mov rax, [rsi + rdx + 8+8]
|
| 453 | xor rax, [rdi + rdx + 8+8]
|
| 454 | jnz LeaveLoopCmps16
|
| 455 |
|
| 456 | add rdx,8+8+8
|
| 457 |
|
| 458 | BEFORE_JMP
|
| 459 | jnz LoopCmps
|
| 460 | jmp LenMaximum
|
| 461 | AFTER_JMP
|
| 462 |
|
| 463 | LeaveLoopCmps16: add rdx,8
|
| 464 | LeaveLoopCmps8: add rdx,8
|
| 465 | LeaveLoopCmps:
|
| 466 |
|
| 467 | test eax, 0x0000FFFF
|
| 468 | jnz LenLower
|
| 469 |
|
| 470 | test eax,0xffffffff
|
| 471 |
|
| 472 | jnz LenLower32
|
| 473 |
|
| 474 | add rdx,4
|
| 475 | shr rax,32
|
| 476 | or ax,ax
|
| 477 | BEFORE_JMP
|
| 478 | jnz LenLower
|
| 479 | AFTER_JMP
|
| 480 |
|
| 481 | LenLower32:
|
| 482 | shr eax,16
|
| 483 | add rdx,2
|
| 484 |
|
| 485 | LenLower:
|
| 486 | sub al, 1
|
| 487 | adc rdx, 0
|
| 488 | //;;; Calculate the length of the match. If it is longer than MAX_MATCH,
|
| 489 | //;;; then automatically accept it as the best possible match and leave.
|
| 490 |
|
| 491 | lea rax, [rdi + rdx]
|
| 492 | sub rax, r9
|
| 493 | cmp eax, MAX_MATCH
|
| 494 | BEFORE_JMP
|
| 495 | jge LenMaximum
|
| 496 | AFTER_JMP
|
| 497 | /*
|
| 498 | ;;; If the length of the match is not longer than the best match we
|
| 499 | ;;; have so far, then forget it and return to the lookup loop.
|
| 500 | ;///////////////////////////////////
|
| 501 | */
|
| 502 | cmp eax, r11d
|
| 503 | jg LongerMatch
|
| 504 |
|
| 505 | lea rsi,[r10+r11]
|
| 506 |
|
| 507 | mov rdi, prev_ad
|
| 508 | mov edx, [chainlenwmask]
|
| 509 | BEFORE_JMP
|
| 510 | jmp LookupLoop
|
| 511 | AFTER_JMP
|
| 512 | /*
|
| 513 | ;;; s->match_start = cur_match;
|
| 514 | ;;; best_len = len;
|
| 515 | ;;; if (len >= nice_match) break;
|
| 516 | ;;; scan_end = *(ushf*)(scan+best_len-1);
|
| 517 | */
|
| 518 | LongerMatch:
|
| 519 | mov r11d, eax
|
| 520 | mov match_start, r8d
|
| 521 | cmp eax, [nicematch]
|
| 522 | BEFORE_JMP
|
| 523 | jge LeaveNow
|
| 524 | AFTER_JMP
|
| 525 |
|
| 526 | lea rsi,[r10+rax]
|
| 527 |
|
| 528 | movzx ebx, word ptr [r9 + rax - 1]
|
| 529 | mov rdi, prev_ad
|
| 530 | mov edx, [chainlenwmask]
|
| 531 | BEFORE_JMP
|
| 532 | jmp LookupLoop
|
| 533 | AFTER_JMP
|
| 534 |
|
| 535 | //;;; Accept the current string, with the maximum possible length.
|
| 536 |
|
| 537 | LenMaximum:
|
| 538 | mov r11d,MAX_MATCH
|
| 539 | mov match_start, r8d
|
| 540 |
|
| 541 | //;;; if ((uInt)best_len <= s->lookahead) return (uInt)best_len;
|
| 542 | //;;; return s->lookahead;
|
| 543 |
|
| 544 | LeaveNow:
|
| 545 | mov eax, Lookahead
|
| 546 | cmp r11d, eax
|
| 547 | cmovng eax, r11d
|
| 548 |
|
| 549 |
|
| 550 |
|
| 551 | //;;; Restore the stack and return from whence we came.
|
| 552 |
|
| 553 |
|
| 554 | // mov rsi,[save_rsi]
|
| 555 | // mov rdi,[save_rdi]
|
| 556 | mov rbx,[save_rbx]
|
| 557 | mov rbp,[save_rbp]
|
| 558 | mov r12,[save_r12]
|
| 559 | mov r13,[save_r13]
|
| 560 | mov r14,[save_r14]
|
| 561 | mov r15,[save_r15]
|
| 562 |
|
| 563 |
|
| 564 | ret 0
|
| 565 | //; please don't remove this string !
|
| 566 | //; Your can freely use gvmat64 in any free or commercial app
|
| 567 | //; but it is far better don't remove the string in the binary!
|
| 568 | // db 0dh,0ah,"asm686 with masm, optimised assembly code from Brian Raiter, written 1998, converted to amd 64 by Gilles Vollant 2005",0dh,0ah,0
|
| 569 |
|
| 570 |
|
| 571 | match_init:
|
| 572 | ret 0
|
| 573 |
|
| 574 |
|