summaryrefslogtreecommitdiffstats
path: root/common/jpeg/simd/jiss2fst-64.asm
blob: 8b664a67bc010833bd713869d96f11910f960101 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
;
; jiss2fst.asm - fast integer IDCT (64-bit SSE2)
;
; Copyright 2009 Pierre Ossman <ossman@cendio.se> for Cendio AB
; Copyright 2009 D. R. Commander
;
; Based on
; x86 SIMD extension for IJG JPEG library
; Copyright (C) 1999-2006, MIYASAKA Masaru.
; For conditions of distribution and use, see copyright notice in jsimdext.inc
;
; This file should be assembled with NASM (Netwide Assembler),
; can *not* be assembled with Microsoft's MASM or any compatible
; assembler (including Borland's Turbo Assembler).
; NASM is available from http://nasm.sourceforge.net/ or
; http://sourceforge.net/projecpt/showfiles.php?group_id=6208
;
; This file contains a fast, not so accurate integer implementation of
; the inverse DCT (Discrete Cosine Transform). The following code is
; based directly on the IJG's original jidctfst.c; see the jidctfst.c
; for more details.
;
; [TAB8]

%include "jsimdext.inc"
%include "jdct.inc"

; --------------------------------------------------------------------------

%define CONST_BITS	8	; 14 is also OK.
%define PASS1_BITS	2

%if IFAST_SCALE_BITS != PASS1_BITS
%error "'IFAST_SCALE_BITS' must be equal to 'PASS1_BITS'."
%endif

%if CONST_BITS == 8
F_1_082	equ	277		; FIX(1.082392200)
F_1_414	equ	362		; FIX(1.414213562)
F_1_847	equ	473		; FIX(1.847759065)
F_2_613	equ	669		; FIX(2.613125930)
F_1_613	equ	(F_2_613 - 256)	; FIX(2.613125930) - FIX(1)
%else
; NASM cannot do compile-time arithmetic on floating-point constants.
%define	DESCALE(x,n)  (((x)+(1<<((n)-1)))>>(n))
F_1_082	equ	DESCALE(1162209775,30-CONST_BITS)	; FIX(1.082392200)
F_1_414	equ	DESCALE(1518500249,30-CONST_BITS)	; FIX(1.414213562)
F_1_847	equ	DESCALE(1984016188,30-CONST_BITS)	; FIX(1.847759065)
F_2_613	equ	DESCALE(2805822602,30-CONST_BITS)	; FIX(2.613125930)
F_1_613	equ	(F_2_613 - (1 << CONST_BITS))	; FIX(2.613125930) - FIX(1)
%endif

; --------------------------------------------------------------------------
	SECTION	SEG_CONST

; PRE_MULTIPLY_SCALE_BITS <= 2 (to avoid overflow)
; CONST_BITS + CONST_SHIFT + PRE_MULTIPLY_SCALE_BITS == 16 (for pmulhw)

%define PRE_MULTIPLY_SCALE_BITS   2
%define CONST_SHIFT     (16 - PRE_MULTIPLY_SCALE_BITS - CONST_BITS)

	alignz	16
	global	EXTN(jconst_idct_ifast_sse2)

EXTN(jconst_idct_ifast_sse2):

PW_F1414	times 8 dw  F_1_414 << CONST_SHIFT
PW_F1847	times 8 dw  F_1_847 << CONST_SHIFT
PW_MF1613	times 8 dw -F_1_613 << CONST_SHIFT
PW_F1082	times 8 dw  F_1_082 << CONST_SHIFT
PB_CENTERJSAMP	times 16 db CENTERJSAMPLE

	alignz	16

; --------------------------------------------------------------------------
	SECTION	SEG_TEXT
	BITS	64
;
; Perform dequantization and inverse DCT on one block of coefficients.
;
; GLOBAL(void)
; jsimd_idct_ifast_sse2 (void * dct_table, JCOEFPTR coef_block,
;                       JSAMPARRAY output_buf, JDIMENSION output_col)
;

; r10 = jpeg_component_info * compptr
; r11 = JCOEFPTR coef_block
; r12 = JSAMPARRAY output_buf
; r13 = JDIMENSION output_col

%define original_rbp	rbp+0
%define wk(i)		rbp-(WK_NUM-(i))*SIZEOF_XMMWORD	; xmmword wk[WK_NUM]
%define WK_NUM		2

	align	16
	global	EXTN(jsimd_idct_ifast_sse2)

EXTN(jsimd_idct_ifast_sse2):
	push	rbp
	mov	rax,rsp				; rax = original rbp
	sub	rsp, byte 4
	and	rsp, byte (-SIZEOF_XMMWORD)	; align to 128 bits
	mov	[rsp],eax
	mov	rbp,rsp				; rbp = aligned rbp
	lea	rsp, [wk(0)]
	collect_args

	; ---- Pass 1: process columns from input.

	mov	rdx, r10	; quantptr
	mov	rsi, r11		; inptr

%ifndef NO_ZERO_COLUMN_TEST_IFAST_SSE2
	mov	eax, DWORD [DWBLOCK(1,0,rsi,SIZEOF_JCOEF)]
	or	eax, DWORD [DWBLOCK(2,0,rsi,SIZEOF_JCOEF)]
	jnz	near .columnDCT

	movdqa	xmm0, XMMWORD [XMMBLOCK(1,0,rsi,SIZEOF_JCOEF)]
	movdqa	xmm1, XMMWORD [XMMBLOCK(2,0,rsi,SIZEOF_JCOEF)]
	por	xmm0, XMMWORD [XMMBLOCK(3,0,rsi,SIZEOF_JCOEF)]
	por	xmm1, XMMWORD [XMMBLOCK(4,0,rsi,SIZEOF_JCOEF)]
	por	xmm0, XMMWORD [XMMBLOCK(5,0,rsi,SIZEOF_JCOEF)]
	por	xmm1, XMMWORD [XMMBLOCK(6,0,rsi,SIZEOF_JCOEF)]
	por	xmm0, XMMWORD [XMMBLOCK(7,0,rsi,SIZEOF_JCOEF)]
	por	xmm1,xmm0
	packsswb xmm1,xmm1
	packsswb xmm1,xmm1
	movd	eax,xmm1
	test	rax,rax
	jnz	short .columnDCT

	; -- AC terms all zero

	movdqa	xmm0, XMMWORD [XMMBLOCK(0,0,rsi,SIZEOF_JCOEF)]
	pmullw	xmm0, XMMWORD [XMMBLOCK(0,0,rdx,SIZEOF_ISLOW_MULT_TYPE)]

	movdqa    xmm7,xmm0		; xmm0=in0=(00 01 02 03 04 05 06 07)
	punpcklwd xmm0,xmm0		; xmm0=(00 00 01 01 02 02 03 03)
	punpckhwd xmm7,xmm7		; xmm7=(04 04 05 05 06 06 07 07)

	pshufd	xmm6,xmm0,0x00		; xmm6=col0=(00 00 00 00 00 00 00 00)
	pshufd	xmm2,xmm0,0x55		; xmm2=col1=(01 01 01 01 01 01 01 01)
	pshufd	xmm5,xmm0,0xAA		; xmm5=col2=(02 02 02 02 02 02 02 02)
	pshufd	xmm0,xmm0,0xFF		; xmm0=col3=(03 03 03 03 03 03 03 03)
	pshufd	xmm1,xmm7,0x00		; xmm1=col4=(04 04 04 04 04 04 04 04)
	pshufd	xmm4,xmm7,0x55		; xmm4=col5=(05 05 05 05 05 05 05 05)
	pshufd	xmm3,xmm7,0xAA		; xmm3=col6=(06 06 06 06 06 06 06 06)
	pshufd	xmm7,xmm7,0xFF		; xmm7=col7=(07 07 07 07 07 07 07 07)

	movdqa	XMMWORD [wk(0)], xmm2	; wk(0)=col1
	movdqa	XMMWORD [wk(1)], xmm0	; wk(1)=col3
	jmp	near .column_end
%endif
.columnDCT:

	; -- Even part

	movdqa	xmm0, XMMWORD [XMMBLOCK(0,0,rsi,SIZEOF_JCOEF)]
	movdqa	xmm1, XMMWORD [XMMBLOCK(2,0,rsi,SIZEOF_JCOEF)]
	pmullw	xmm0, XMMWORD [XMMBLOCK(0,0,rdx,SIZEOF_IFAST_MULT_TYPE)]
	pmullw	xmm1, XMMWORD [XMMBLOCK(2,0,rdx,SIZEOF_IFAST_MULT_TYPE)]
	movdqa	xmm2, XMMWORD [XMMBLOCK(4,0,rsi,SIZEOF_JCOEF)]
	movdqa	xmm3, XMMWORD [XMMBLOCK(6,0,rsi,SIZEOF_JCOEF)]
	pmullw	xmm2, XMMWORD [XMMBLOCK(4,0,rdx,SIZEOF_IFAST_MULT_TYPE)]
	pmullw	xmm3, XMMWORD [XMMBLOCK(6,0,rdx,SIZEOF_IFAST_MULT_TYPE)]

	movdqa	xmm4,xmm0
	movdqa	xmm5,xmm1
	psubw	xmm0,xmm2		; xmm0=tmp11
	psubw	xmm1,xmm3
	paddw	xmm4,xmm2		; xmm4=tmp10
	paddw	xmm5,xmm3		; xmm5=tmp13

	psllw	xmm1,PRE_MULTIPLY_SCALE_BITS
	pmulhw	xmm1,[rel PW_F1414]
	psubw	xmm1,xmm5		; xmm1=tmp12

	movdqa	xmm6,xmm4
	movdqa	xmm7,xmm0
	psubw	xmm4,xmm5		; xmm4=tmp3
	psubw	xmm0,xmm1		; xmm0=tmp2
	paddw	xmm6,xmm5		; xmm6=tmp0
	paddw	xmm7,xmm1		; xmm7=tmp1

	movdqa	XMMWORD [wk(1)], xmm4	; wk(1)=tmp3
	movdqa	XMMWORD [wk(0)], xmm0	; wk(0)=tmp2

	; -- Odd part

	movdqa	xmm2, XMMWORD [XMMBLOCK(1,0,rsi,SIZEOF_JCOEF)]
	movdqa	xmm3, XMMWORD [XMMBLOCK(3,0,rsi,SIZEOF_JCOEF)]
	pmullw	xmm2, XMMWORD [XMMBLOCK(1,0,rdx,SIZEOF_IFAST_MULT_TYPE)]
	pmullw	xmm3, XMMWORD [XMMBLOCK(3,0,rdx,SIZEOF_IFAST_MULT_TYPE)]
	movdqa	xmm5, XMMWORD [XMMBLOCK(5,0,rsi,SIZEOF_JCOEF)]
	movdqa	xmm1, XMMWORD [XMMBLOCK(7,0,rsi,SIZEOF_JCOEF)]
	pmullw	xmm5, XMMWORD [XMMBLOCK(5,0,rdx,SIZEOF_IFAST_MULT_TYPE)]
	pmullw	xmm1, XMMWORD [XMMBLOCK(7,0,rdx,SIZEOF_IFAST_MULT_TYPE)]

	movdqa	xmm4,xmm2
	movdqa	xmm0,xmm5
	psubw	xmm2,xmm1		; xmm2=z12
	psubw	xmm5,xmm3		; xmm5=z10
	paddw	xmm4,xmm1		; xmm4=z11
	paddw	xmm0,xmm3		; xmm0=z13

	movdqa	xmm1,xmm5		; xmm1=z10(unscaled)
	psllw	xmm2,PRE_MULTIPLY_SCALE_BITS
	psllw	xmm5,PRE_MULTIPLY_SCALE_BITS

	movdqa	xmm3,xmm4
	psubw	xmm4,xmm0
	paddw	xmm3,xmm0		; xmm3=tmp7

	psllw	xmm4,PRE_MULTIPLY_SCALE_BITS
	pmulhw	xmm4,[rel PW_F1414]	; xmm4=tmp11

	; To avoid overflow...
	;
	; (Original)
	; tmp12 = -2.613125930 * z10 + z5;
	;
	; (This implementation)
	; tmp12 = (-1.613125930 - 1) * z10 + z5;
	;       = -1.613125930 * z10 - z10 + z5;

	movdqa	xmm0,xmm5
	paddw	xmm5,xmm2
	pmulhw	xmm5,[rel PW_F1847]	; xmm5=z5
	pmulhw	xmm0,[rel PW_MF1613]
	pmulhw	xmm2,[rel PW_F1082]
	psubw	xmm0,xmm1
	psubw	xmm2,xmm5		; xmm2=tmp10
	paddw	xmm0,xmm5		; xmm0=tmp12

	; -- Final output stage

	psubw	xmm0,xmm3		; xmm0=tmp6
	movdqa	xmm1,xmm6
	movdqa	xmm5,xmm7
	paddw	xmm6,xmm3		; xmm6=data0=(00 01 02 03 04 05 06 07)
	paddw	xmm7,xmm0		; xmm7=data1=(10 11 12 13 14 15 16 17)
	psubw	xmm1,xmm3		; xmm1=data7=(70 71 72 73 74 75 76 77)
	psubw	xmm5,xmm0		; xmm5=data6=(60 61 62 63 64 65 66 67)
	psubw	xmm4,xmm0		; xmm4=tmp5

	movdqa    xmm3,xmm6		; transpose coefficients(phase 1)
	punpcklwd xmm6,xmm7		; xmm6=(00 10 01 11 02 12 03 13)
	punpckhwd xmm3,xmm7		; xmm3=(04 14 05 15 06 16 07 17)
	movdqa    xmm0,xmm5		; transpose coefficients(phase 1)
	punpcklwd xmm5,xmm1		; xmm5=(60 70 61 71 62 72 63 73)
	punpckhwd xmm0,xmm1		; xmm0=(64 74 65 75 66 76 67 77)

	movdqa	xmm7, XMMWORD [wk(0)]	; xmm7=tmp2
	movdqa	xmm1, XMMWORD [wk(1)]	; xmm1=tmp3

	movdqa	XMMWORD [wk(0)], xmm5	; wk(0)=(60 70 61 71 62 72 63 73)
	movdqa	XMMWORD [wk(1)], xmm0	; wk(1)=(64 74 65 75 66 76 67 77)

	paddw	xmm2,xmm4		; xmm2=tmp4
	movdqa	xmm5,xmm7
	movdqa	xmm0,xmm1
	paddw	xmm7,xmm4		; xmm7=data2=(20 21 22 23 24 25 26 27)
	paddw	xmm1,xmm2		; xmm1=data4=(40 41 42 43 44 45 46 47)
	psubw	xmm5,xmm4		; xmm5=data5=(50 51 52 53 54 55 56 57)
	psubw	xmm0,xmm2		; xmm0=data3=(30 31 32 33 34 35 36 37)

	movdqa    xmm4,xmm7		; transpose coefficients(phase 1)
	punpcklwd xmm7,xmm0		; xmm7=(20 30 21 31 22 32 23 33)
	punpckhwd xmm4,xmm0		; xmm4=(24 34 25 35 26 36 27 37)
	movdqa    xmm2,xmm1		; transpose coefficients(phase 1)
	punpcklwd xmm1,xmm5		; xmm1=(40 50 41 51 42 52 43 53)
	punpckhwd xmm2,xmm5		; xmm2=(44 54 45 55 46 56 47 57)

	movdqa    xmm0,xmm3		; transpose coefficients(phase 2)
	punpckldq xmm3,xmm4		; xmm3=(04 14 24 34 05 15 25 35)
	punpckhdq xmm0,xmm4		; xmm0=(06 16 26 36 07 17 27 37)
	movdqa    xmm5,xmm6		; transpose coefficients(phase 2)
	punpckldq xmm6,xmm7		; xmm6=(00 10 20 30 01 11 21 31)
	punpckhdq xmm5,xmm7		; xmm5=(02 12 22 32 03 13 23 33)

	movdqa	xmm4, XMMWORD [wk(0)]	; xmm4=(60 70 61 71 62 72 63 73)
	movdqa	xmm7, XMMWORD [wk(1)]	; xmm7=(64 74 65 75 66 76 67 77)

	movdqa	XMMWORD [wk(0)], xmm3	; wk(0)=(04 14 24 34 05 15 25 35)
	movdqa	XMMWORD [wk(1)], xmm0	; wk(1)=(06 16 26 36 07 17 27 37)

	movdqa    xmm3,xmm1		; transpose coefficients(phase 2)
	punpckldq xmm1,xmm4		; xmm1=(40 50 60 70 41 51 61 71)
	punpckhdq xmm3,xmm4		; xmm3=(42 52 62 72 43 53 63 73)
	movdqa    xmm0,xmm2		; transpose coefficients(phase 2)
	punpckldq xmm2,xmm7		; xmm2=(44 54 64 74 45 55 65 75)
	punpckhdq xmm0,xmm7		; xmm0=(46 56 66 76 47 57 67 77)

	movdqa     xmm4,xmm6		; transpose coefficients(phase 3)
	punpcklqdq xmm6,xmm1		; xmm6=col0=(00 10 20 30 40 50 60 70)
	punpckhqdq xmm4,xmm1		; xmm4=col1=(01 11 21 31 41 51 61 71)
	movdqa     xmm7,xmm5		; transpose coefficients(phase 3)
	punpcklqdq xmm5,xmm3		; xmm5=col2=(02 12 22 32 42 52 62 72)
	punpckhqdq xmm7,xmm3		; xmm7=col3=(03 13 23 33 43 53 63 73)

	movdqa	xmm1, XMMWORD [wk(0)]	; xmm1=(04 14 24 34 05 15 25 35)
	movdqa	xmm3, XMMWORD [wk(1)]	; xmm3=(06 16 26 36 07 17 27 37)

	movdqa	XMMWORD [wk(0)], xmm4	; wk(0)=col1
	movdqa	XMMWORD [wk(1)], xmm7	; wk(1)=col3

	movdqa     xmm4,xmm1		; transpose coefficients(phase 3)
	punpcklqdq xmm1,xmm2		; xmm1=col4=(04 14 24 34 44 54 64 74)
	punpckhqdq xmm4,xmm2		; xmm4=col5=(05 15 25 35 45 55 65 75)
	movdqa     xmm7,xmm3		; transpose coefficients(phase 3)
	punpcklqdq xmm3,xmm0		; xmm3=col6=(06 16 26 36 46 56 66 76)
	punpckhqdq xmm7,xmm0		; xmm7=col7=(07 17 27 37 47 57 67 77)
.column_end:

	; -- Prefetch the next coefficient block

	prefetchnta [rsi + DCTSIZE2*SIZEOF_JCOEF + 0*32]
	prefetchnta [rsi + DCTSIZE2*SIZEOF_JCOEF + 1*32]
	prefetchnta [rsi + DCTSIZE2*SIZEOF_JCOEF + 2*32]
	prefetchnta [rsi + DCTSIZE2*SIZEOF_JCOEF + 3*32]

	; ---- Pass 2: process rows from work array, store into output array.

	mov	rax, [original_rbp]
	mov	rdi, r12	; (JSAMPROW *)
	mov	rax, r13

	; -- Even part

	; xmm6=col0, xmm5=col2, xmm1=col4, xmm3=col6

	movdqa	xmm2,xmm6
	movdqa	xmm0,xmm5
	psubw	xmm6,xmm1		; xmm6=tmp11
	psubw	xmm5,xmm3
	paddw	xmm2,xmm1		; xmm2=tmp10
	paddw	xmm0,xmm3		; xmm0=tmp13

	psllw	xmm5,PRE_MULTIPLY_SCALE_BITS
	pmulhw	xmm5,[rel PW_F1414]
	psubw	xmm5,xmm0		; xmm5=tmp12

	movdqa	xmm1,xmm2
	movdqa	xmm3,xmm6
	psubw	xmm2,xmm0		; xmm2=tmp3
	psubw	xmm6,xmm5		; xmm6=tmp2
	paddw	xmm1,xmm0		; xmm1=tmp0
	paddw	xmm3,xmm5		; xmm3=tmp1

	movdqa	xmm0, XMMWORD [wk(0)]	; xmm0=col1
	movdqa	xmm5, XMMWORD [wk(1)]	; xmm5=col3

	movdqa	XMMWORD [wk(0)], xmm2	; wk(0)=tmp3
	movdqa	XMMWORD [wk(1)], xmm6	; wk(1)=tmp2

	; -- Odd part

	; xmm0=col1, xmm5=col3, xmm4=col5, xmm7=col7

	movdqa	xmm2,xmm0
	movdqa	xmm6,xmm4
	psubw	xmm0,xmm7		; xmm0=z12
	psubw	xmm4,xmm5		; xmm4=z10
	paddw	xmm2,xmm7		; xmm2=z11
	paddw	xmm6,xmm5		; xmm6=z13

	movdqa	xmm7,xmm4		; xmm7=z10(unscaled)
	psllw	xmm0,PRE_MULTIPLY_SCALE_BITS
	psllw	xmm4,PRE_MULTIPLY_SCALE_BITS

	movdqa	xmm5,xmm2
	psubw	xmm2,xmm6
	paddw	xmm5,xmm6		; xmm5=tmp7

	psllw	xmm2,PRE_MULTIPLY_SCALE_BITS
	pmulhw	xmm2,[rel PW_F1414]	; xmm2=tmp11

	; To avoid overflow...
	;
	; (Original)
	; tmp12 = -2.613125930 * z10 + z5;
	;
	; (This implementation)
	; tmp12 = (-1.613125930 - 1) * z10 + z5;
	;       = -1.613125930 * z10 - z10 + z5;

	movdqa	xmm6,xmm4
	paddw	xmm4,xmm0
	pmulhw	xmm4,[rel PW_F1847]	; xmm4=z5
	pmulhw	xmm6,[rel PW_MF1613]
	pmulhw	xmm0,[rel PW_F1082]
	psubw	xmm6,xmm7
	psubw	xmm0,xmm4		; xmm0=tmp10
	paddw	xmm6,xmm4		; xmm6=tmp12

	; -- Final output stage

	psubw	xmm6,xmm5		; xmm6=tmp6
	movdqa	xmm7,xmm1
	movdqa	xmm4,xmm3
	paddw	xmm1,xmm5		; xmm1=data0=(00 10 20 30 40 50 60 70)
	paddw	xmm3,xmm6		; xmm3=data1=(01 11 21 31 41 51 61 71)
	psraw	xmm1,(PASS1_BITS+3)	; descale
	psraw	xmm3,(PASS1_BITS+3)	; descale
	psubw	xmm7,xmm5		; xmm7=data7=(07 17 27 37 47 57 67 77)
	psubw	xmm4,xmm6		; xmm4=data6=(06 16 26 36 46 56 66 76)
	psraw	xmm7,(PASS1_BITS+3)	; descale
	psraw	xmm4,(PASS1_BITS+3)	; descale
	psubw	xmm2,xmm6		; xmm2=tmp5

	packsswb  xmm1,xmm4	; xmm1=(00 10 20 30 40 50 60 70 06 16 26 36 46 56 66 76)
	packsswb  xmm3,xmm7	; xmm3=(01 11 21 31 41 51 61 71 07 17 27 37 47 57 67 77)

	movdqa	xmm5, XMMWORD [wk(1)]	; xmm5=tmp2
	movdqa	xmm6, XMMWORD [wk(0)]	; xmm6=tmp3

	paddw	xmm0,xmm2		; xmm0=tmp4
	movdqa	xmm4,xmm5
	movdqa	xmm7,xmm6
	paddw	xmm5,xmm2		; xmm5=data2=(02 12 22 32 42 52 62 72)
	paddw	xmm6,xmm0		; xmm6=data4=(04 14 24 34 44 54 64 74)
	psraw	xmm5,(PASS1_BITS+3)	; descale
	psraw	xmm6,(PASS1_BITS+3)	; descale
	psubw	xmm4,xmm2		; xmm4=data5=(05 15 25 35 45 55 65 75)
	psubw	xmm7,xmm0		; xmm7=data3=(03 13 23 33 43 53 63 73)
	psraw	xmm4,(PASS1_BITS+3)	; descale
	psraw	xmm7,(PASS1_BITS+3)	; descale

	movdqa    xmm2,[rel PB_CENTERJSAMP]	; xmm2=[rel PB_CENTERJSAMP]

	packsswb  xmm5,xmm6	; xmm5=(02 12 22 32 42 52 62 72 04 14 24 34 44 54 64 74)
	packsswb  xmm7,xmm4	; xmm7=(03 13 23 33 43 53 63 73 05 15 25 35 45 55 65 75)

	paddb     xmm1,xmm2
	paddb     xmm3,xmm2
	paddb     xmm5,xmm2
	paddb     xmm7,xmm2

	movdqa    xmm0,xmm1	; transpose coefficients(phase 1)
	punpcklbw xmm1,xmm3	; xmm1=(00 01 10 11 20 21 30 31 40 41 50 51 60 61 70 71)
	punpckhbw xmm0,xmm3	; xmm0=(06 07 16 17 26 27 36 37 46 47 56 57 66 67 76 77)
	movdqa    xmm6,xmm5	; transpose coefficients(phase 1)
	punpcklbw xmm5,xmm7	; xmm5=(02 03 12 13 22 23 32 33 42 43 52 53 62 63 72 73)
	punpckhbw xmm6,xmm7	; xmm6=(04 05 14 15 24 25 34 35 44 45 54 55 64 65 74 75)

	movdqa    xmm4,xmm1	; transpose coefficients(phase 2)
	punpcklwd xmm1,xmm5	; xmm1=(00 01 02 03 10 11 12 13 20 21 22 23 30 31 32 33)
	punpckhwd xmm4,xmm5	; xmm4=(40 41 42 43 50 51 52 53 60 61 62 63 70 71 72 73)
	movdqa    xmm2,xmm6	; transpose coefficients(phase 2)
	punpcklwd xmm6,xmm0	; xmm6=(04 05 06 07 14 15 16 17 24 25 26 27 34 35 36 37)
	punpckhwd xmm2,xmm0	; xmm2=(44 45 46 47 54 55 56 57 64 65 66 67 74 75 76 77)

	movdqa    xmm3,xmm1	; transpose coefficients(phase 3)
	punpckldq xmm1,xmm6	; xmm1=(00 01 02 03 04 05 06 07 10 11 12 13 14 15 16 17)
	punpckhdq xmm3,xmm6	; xmm3=(20 21 22 23 24 25 26 27 30 31 32 33 34 35 36 37)
	movdqa    xmm7,xmm4	; transpose coefficients(phase 3)
	punpckldq xmm4,xmm2	; xmm4=(40 41 42 43 44 45 46 47 50 51 52 53 54 55 56 57)
	punpckhdq xmm7,xmm2	; xmm7=(60 61 62 63 64 65 66 67 70 71 72 73 74 75 76 77)

	pshufd	xmm5,xmm1,0x4E	; xmm5=(10 11 12 13 14 15 16 17 00 01 02 03 04 05 06 07)
	pshufd	xmm0,xmm3,0x4E	; xmm0=(30 31 32 33 34 35 36 37 20 21 22 23 24 25 26 27)
	pshufd	xmm6,xmm4,0x4E	; xmm6=(50 51 52 53 54 55 56 57 40 41 42 43 44 45 46 47)
	pshufd	xmm2,xmm7,0x4E	; xmm2=(70 71 72 73 74 75 76 77 60 61 62 63 64 65 66 67)

	mov	rdx, JSAMPROW [rdi+0*SIZEOF_JSAMPROW]
	mov	rsi, JSAMPROW [rdi+2*SIZEOF_JSAMPROW]
	movq	XMM_MMWORD [rdx+rax*SIZEOF_JSAMPLE], xmm1
	movq	XMM_MMWORD [rsi+rax*SIZEOF_JSAMPLE], xmm3
	mov	rdx, JSAMPROW [rdi+4*SIZEOF_JSAMPROW]
	mov	rsi, JSAMPROW [rdi+6*SIZEOF_JSAMPROW]
	movq	XMM_MMWORD [rdx+rax*SIZEOF_JSAMPLE], xmm4
	movq	XMM_MMWORD [rsi+rax*SIZEOF_JSAMPLE], xmm7

	mov	rdx, JSAMPROW [rdi+1*SIZEOF_JSAMPROW]
	mov	rsi, JSAMPROW [rdi+3*SIZEOF_JSAMPROW]
	movq	XMM_MMWORD [rdx+rax*SIZEOF_JSAMPLE], xmm5
	movq	XMM_MMWORD [rsi+rax*SIZEOF_JSAMPLE], xmm0
	mov	rdx, JSAMPROW [rdi+5*SIZEOF_JSAMPROW]
	mov	rsi, JSAMPROW [rdi+7*SIZEOF_JSAMPROW]
	movq	XMM_MMWORD [rdx+rax*SIZEOF_JSAMPLE], xmm6
	movq	XMM_MMWORD [rsi+rax*SIZEOF_JSAMPLE], xmm2

	uncollect_args
	mov	rsp,rbp		; rsp <- aligned rbp
	pop	rsp		; rsp <- original rbp
	pop	rbp
	ret
	ret

; For some reason, the OS X linker does not honor the request to align the
; segment unless we do this.
	align	16