Coverage Report

Created: 2026-04-01 07:17

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/src/ghostpdl/base/gsroprun1.h
Line
Count
Source
1
/* Copyright (C) 2001-2023 Artifex Software, Inc.
2
   All Rights Reserved.
3
4
   This software is provided AS-IS with no warranty, either express or
5
   implied.
6
7
   This software is distributed under license and may not be copied,
8
   modified or distributed except as expressly authorized under the terms
9
   of the license contained in the file LICENSE in this distribution.
10
11
   Refer to licensing information at http://www.artifex.com or contact
12
   Artifex Software, Inc.,  39 Mesa Street, Suite 108A, San Francisco,
13
   CA 94129, USA, for further information.
14
*/
15
16
17
/* This file is repeatedly included by gsroprun.c to 'autogenerate' many
18
 * different versions of roprun code. DO NOT USE THIS FILE EXCEPT FROM
19
 * gsroprun.c.
20
 */
21
22
/* Set the following defines as appropriate on entry:
23
 *   TEMPLATE_NAME (Compulsory)  The name of the function to generate
24
 *   SPECIFIC_ROP  (Optional)    If set, the function will base its decision
25
 *                               about whether to provide S and T upon
26
 *                               this value.
27
 *   SPECIFIC_CODE (Optional)    If set, this should expand out to code to
28
 *                               perform the rop. Will be invoked as:
29
 *                               SPECIFIC_ROP(OUT,D,S,T)
30
 *   S_CONST       (Optional)    If set, S will be taken to be constant, else
31
 *                               S will be read from a pointer.
32
 *   T_CONST       (Optional)    If set, T will be taken to be constant, else
33
 *                               T will be read from a pointer.
34
 */
35
36
#if defined(TEMPLATE_NAME)
37
38
#ifdef SPECIFIC_ROP
39
#if rop3_uses_S(SPECIFIC_ROP)
40
#define S_USED
41
#endif
42
#if rop3_uses_T(SPECIFIC_ROP)
43
#define T_USED
44
#endif
45
#else /* !SPECIFIC_ROP */
46
#define S_USED
47
#define T_USED
48
#endif /* SPECIFIC_ROP */
49
50
/* We work in 'chunks' here; for bigendian machines, we can safely use
51
 * chunks of 'int' size. For little endian machines where we have a cheap
52
 * endian swap, we can do likewise. For others, we'll work at the byte
53
 * level. */
54
#if !ARCH_IS_BIG_ENDIAN && !defined(ENDIAN_SWAP_INT)
55
#define CHUNKSIZE 8
56
#define CHUNK byte
57
#define CHUNKONES 255
58
59
#define ADJUST_TO_CHUNK(d,dpos) do {} while (0)
60
61
#else /* ARCH_IS_BIG_ENDIAN || defined(ENDIAN_SWAP_INT) */
62
#if ARCH_LOG2_SIZEOF_INT == 2
63
1.45G
#define CHUNKSIZE 32
64
223M
#define CHUNK unsigned int
65
109M
#define CHUNKONES 0xFFFFFFFFU
66
67
#if ARCH_SIZEOF_PTR == (1<<ARCH_LOG2_SIZEOF_INT)
68
#define ROP_PTRDIFF_T int
69
#else
70
#define ROP_PTRDIFF_T int64_t
71
#endif
72
#define ADJUST_TO_CHUNK(d, dpos)                      \
73
109M
    do { int offset = ((ROP_PTRDIFF_T)d) & ((CHUNKSIZE>>3)-1);  \
74
109M
         d = (CHUNK *)(void *)(((byte *)(void *)d)-offset);   \
75
109M
         dpos += offset<<3;                           \
76
109M
     } while (0)
77
#else
78
/* FIXME: Write more code in here when we find an example. */
79
#endif
80
#endif /* ARCH_IS_BIG_ENDIAN || defined(ENDIAN_SWAP_INT) */
81
82
/* We define an 'RE' macro that reverses the endianness of a chunk, if we
83
 * need it, and does nothing otherwise. */
84
#if !ARCH_IS_BIG_ENDIAN && defined(ENDIAN_SWAP_INT) && (CHUNKSIZE != 8)
85
694M
#define RE(I) ((CHUNK)ENDIAN_SWAP_INT(I))
86
#else /* ARCH_IS_BIG_ENDIAN || !defined(ENDIAN_SWAP_INT) || (CHUNKSIZE == 8) */
87
#define RE(I) (I)
88
#endif /* ARCH_IS_BIG_ENDIAN || !defined(ENDIAN_SWAP_INT) || (CHUNKSIZE == 8) */
89
90
/* In some cases we will need to fetch values from a pointer, and 'skew'
91
 * them. We need 2 variants of this macro. One that is 'SAFE' to use when
92
 * SKEW might be 0, and one that can be faster, because we know that SKEW
93
 * is non zero. */
94
#define SKEW_FETCH(S,s,SKEW) \
95
485M
    do { S = RE((RE(s[0])<<SKEW) | (RE(s[1])>>(CHUNKSIZE-SKEW))); s++; } while (0)
96
#define SAFE_SKEW_FETCH(S,s,SKEW,L,R)                                    \
97
99.3M
    do { S = RE(((L) ? 0 : (RE(s[0])<<SKEW)) | ((R) ? 0 : (RE(s[1])>>(CHUNKSIZE-SKEW)))); s++; } while (0)
98
99
#if defined(S_USED) && !defined(S_CONST)
100
#define S_SKEW
101
485M
#define FETCH_S           SKEW_FETCH(S,s,s_skew)
102
99.3M
#define SAFE_FETCH_S(L,R) SAFE_SKEW_FETCH(S,s,s_skew,L,R)
103
#else /* !defined(S_USED) || defined(S_CONST) */
104
#define FETCH_S
105
#define SAFE_FETCH_S(L,R)
106
#endif /* !defined(S_USED) || defined(S_CONST) */
107
108
#if defined(T_USED) && !defined(T_CONST)
109
#define T_SKEW
110
0
#define FETCH_T           SKEW_FETCH(T,t,t_skew)
111
0
#define SAFE_FETCH_T(L,R) SAFE_SKEW_FETCH(T,t,t_skew,L,R)
112
#else /* !defined(T_USED) || defined(T_CONST) */
113
#define FETCH_T
114
#define SAFE_FETCH_T(L,R)
115
#endif /* !defined(T_USED) || defined(T_CONST) */
116
117
static void TEMPLATE_NAME(rop_run_op *op, byte *d_, int len)
118
54.6M
{
119
#ifndef SPECIFIC_CODE
120
    rop_proc     proc = rop_proc_table[op->rop];
121
20.0M
#define SPECIFIC_CODE(OUT_, D_,S_,T_) OUT_ = proc(D_,S_,T_)
122
#endif /* !defined(SPECIFIC_CODE) */
123
54.6M
    CHUNK        lmask, rmask;
124
#ifdef S_USED
125
#ifdef S_CONST
126
0
    CHUNK        S = (CHUNK)op->s.c;
127
#else /* !defined(S_CONST) */
128
    const CHUNK *s = (CHUNK *)(void *)op->s.b.ptr;
129
54.6M
    CHUNK        S;
130
    int          s_skew;
131
#endif /* !defined(S_CONST) */
132
#else /* !defined(S_USED) */
133
#define S 0
134
#undef S_CONST
135
#endif /* !defined(S_USED) */
136
#ifdef T_USED
137
#ifdef T_CONST
138
4.44M
    CHUNK        T = (CHUNK)op->t.c;
139
#else /* !defined(T_CONST) */
140
    const CHUNK *t = (CHUNK *)(void *)op->t.b.ptr;
141
0
    CHUNK        T;
142
    int          t_skew;
143
#endif /* !defined(T_CONST) */
144
#else /* !defined(T_USED) */
145
#define T 0
146
#undef T_CONST
147
#endif /* !defined(T_USED) */
148
#if defined(S_SKEW) || defined(T_SKEW)
149
    int skewflags = 0;
150
#endif
151
54.6M
    CHUNK        D;
152
54.6M
    int          dpos = op->dpos;
153
54.6M
    CHUNK       *d = (CHUNK *)(void *)d_;
154
155
    /* Align d to CHUNKSIZE */
156
54.6M
    ADJUST_TO_CHUNK(d,dpos);
157
158
    /* On entry len = length in 'depth' chunks. Change it to be the length
159
     * in bits, and add on the number of bits we skip at the start of the
160
     * run. */
161
54.6M
    len    = len * op->depth + dpos;
162
163
    /* lmask = the set of bits to alter in the output bitmap on the left
164
     * hand edge of the run. rmask = the set of bits NOT to alter in the
165
     * output bitmap on the right hand edge of the run. */
166
54.6M
    lmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & dpos)));
167
54.6M
    rmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & len)));
168
54.6M
    if (rmask == CHUNKONES) rmask = 0;
169
170
#if defined(S_CONST) || defined(T_CONST)
171
    /* S and T should be supplied as 'depth' bits. Duplicate them up to be
172
     * byte size (if they are supplied byte sized, that's fine too). */
173
4.44M
    if (op->depth & 1) {
174
#ifdef S_CONST
175
        S |= S<<1;
176
#endif /* !defined(S_CONST) */
177
4.44M
#ifdef T_CONST
178
4.44M
        T |= T<<1;
179
4.44M
#endif /* !defined(T_CONST) */
180
4.44M
    }
181
4.44M
    if (op->depth & 3) {
182
#ifdef S_CONST
183
        S |= S<<2;
184
#endif /* !defined(S_CONST) */
185
4.44M
#ifdef T_CONST
186
4.44M
        T |= T<<2;
187
4.44M
#endif /* !defined(T_CONST) */
188
4.44M
    }
189
4.44M
    if (op->depth & 7) {
190
#ifdef S_CONST
191
        S |= S<<4;
192
#endif /* !defined(S_CONST) */
193
4.44M
#ifdef T_CONST
194
4.44M
        T |= T<<4;
195
4.44M
#endif /* !defined(T_CONST) */
196
4.44M
    }
197
#if CHUNKSIZE > 8
198
4.44M
    if (op->depth & 15) {
199
#ifdef S_CONST
200
        S |= S<<8;
201
#endif /* !defined(S_CONST) */
202
4.44M
#ifdef T_CONST
203
4.44M
        T |= T<<8;
204
4.44M
#endif /* !defined(T_CONST) */
205
4.44M
    }
206
#endif /* CHUNKSIZE > 8 */
207
#if CHUNKSIZE > 16
208
4.44M
    if (op->depth & 31) {
209
#ifdef S_CONST
210
        S |= S<<16;
211
#endif /* !defined(S_CONST) */
212
4.44M
#ifdef T_CONST
213
4.44M
        T |= T<<16;
214
4.44M
#endif /* !defined(T_CONST) */
215
4.44M
    }
216
#endif /* CHUNKSIZE > 16 */
217
#endif /* defined(S_CONST) || defined(T_CONST) */
218
219
    /* Note #1: This mirrors what the original code did, but I think it has
220
     * the risk of moving s and t back beyond officially allocated space. We
221
     * may be saved by the fact that all blocks have a word or two in front
222
     * of them due to the allocator. If we ever get valgrind properly marking
223
     * allocated blocks as readable etc, then this may throw some spurious
224
     * errors. RJW. */
225
#ifdef S_SKEW
226
    {
227
        int slen, slen2;
228
        int spos = op->s.b.pos;
229
54.6M
        ADJUST_TO_CHUNK(s, spos);
230
        s_skew = spos - dpos;
231
54.6M
        if (s_skew < 0) {
232
39.9M
            s_skew += CHUNKSIZE;
233
39.9M
            s--;
234
39.9M
            skewflags |= 1; /* Suppress reading off left edge */
235
39.9M
        }
236
        /* We are allowed to read all the data bits, so: len - dpos + tpos
237
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
238
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
239
         * This is larger, then suppress. */
240
54.6M
        slen  = (len + s_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
241
54.6M
        slen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
242
54.6M
        if ((s_skew == 0) || (slen < slen2)) {
243
30.5M
            skewflags |= 4; /* Suppress reading off the right edge */
244
30.5M
        }
245
    }
246
#endif /* !defined(S_SKEW) */
247
#ifdef T_SKEW
248
    {
249
        int tlen, tlen2;
250
        int tpos = op->t.b.pos;
251
0
        ADJUST_TO_CHUNK(t, tpos);
252
        t_skew = tpos - dpos;
253
0
        if (t_skew < 0) {
254
0
            t_skew += CHUNKSIZE;
255
0
            t--;
256
0
            skewflags |= 2; /* Suppress reading off left edge */
257
0
        }
258
        /* We are allowed to read all the data bits, so: len - dpos + tpos
259
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
260
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
261
         * This is larger, then suppress. */
262
0
        tlen  = (len + t_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
263
0
        tlen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
264
0
        if ((t_skew == 0) || (tlen < tlen2)) {
265
0
            skewflags |= 8; /* Suppress reading off the right edge */
266
0
        }
267
    }
268
#endif /* !defined(T_SKEW) */
269
270
54.6M
    len -= CHUNKSIZE; /* len = bytes to do - CHUNKSIZE */
271
    /* len <= 0 means 1 word or less to do */
272
54.6M
    if (len <= 0) {
273
        /* Short case - starts and ends in the same chunk */
274
241k
        lmask &= ~rmask; /* Combined mask = bits to alter */
275
241k
        SAFE_FETCH_S(skewflags & 1,skewflags & 4);
276
241k
        SAFE_FETCH_T(skewflags & 2,skewflags & 8);
277
241k
        SPECIFIC_CODE(D, *d, S, T);
278
241k
        *d = (*d & ~lmask) | (D & lmask);
279
241k
        return;
280
241k
    }
281
54.4M
    if ((lmask != CHUNKONES)
282
#if defined(S_SKEW) || defined(T_SKEW)
283
9.78M
        || (skewflags & 3)
284
#endif
285
54.4M
        ) {
286
        /* Unaligned left hand case */
287
44.6M
        SAFE_FETCH_S(skewflags & 1,s_skew == 0);
288
44.6M
        SAFE_FETCH_T(skewflags & 2,t_skew == 0);
289
44.6M
        SPECIFIC_CODE(D, *d, S, T);
290
44.6M
        *d = (*d & ~lmask) | (D & lmask);
291
44.6M
        d++;
292
44.6M
        len -= CHUNKSIZE;
293
44.6M
    }
294
54.4M
    if (len > 0) {
295
        /* Simple middle case (complete destination chunks). */
296
#ifdef S_SKEW
297
40.4M
        if (s_skew == 0) {
298
#ifdef T_SKEW
299
0
            if (t_skew == 0) {
300
0
                do {
301
0
                    SPECIFIC_CODE(*d, *d, *s++, *t++);
302
0
                    d++;
303
0
                    len -= CHUNKSIZE;
304
0
                } while (len > 0);
305
0
            } else
306
0
#endif /* !defined(T_SKEW) */
307
0
            {
308
443M
                do {
309
443M
                    FETCH_T;
310
443M
                    SPECIFIC_CODE(*d, *d, *s++, T);
311
443M
                    d++;
312
443M
                    len -= CHUNKSIZE;
313
443M
                } while (len > 0);
314
0
            }
315
10.0M
        } else
316
30.3M
#endif /* !defined(S_SKEW) */
317
30.3M
        {
318
#ifdef T_SKEW
319
0
            if (t_skew == 0) {
320
0
                do {
321
0
                    FETCH_S;
322
0
                    SPECIFIC_CODE(*d, *d, S, *t++);
323
0
                    d++;
324
0
                    len -= CHUNKSIZE;
325
0
                } while (len > 0);
326
0
            } else
327
0
#endif /* !defined(T_SKEW) */
328
0
            {
329
485M
                do {
330
485M
                    FETCH_S;
331
485M
                    FETCH_T;
332
485M
                    SPECIFIC_CODE(*d, *d, S, T);
333
485M
                    d++;
334
485M
                    len -= CHUNKSIZE;
335
485M
                } while (len > 0);
336
0
            }
337
30.3M
        }
338
40.4M
    }
339
    /* Unaligned right hand case */
340
54.4M
    SAFE_FETCH_S(0,skewflags & 4);
341
54.4M
    SAFE_FETCH_T(0,skewflags & 8);
342
54.4M
    SPECIFIC_CODE(D, *d, S, T);
343
54.4M
    *d = (*d & rmask) | (D & ~rmask);
344
54.4M
}
gsroprun.c:notS_rop_run1_const_t
Line
Count
Source
118
138k
{
119
#ifndef SPECIFIC_CODE
120
    rop_proc     proc = rop_proc_table[op->rop];
121
#define SPECIFIC_CODE(OUT_, D_,S_,T_) OUT_ = proc(D_,S_,T_)
122
#endif /* !defined(SPECIFIC_CODE) */
123
138k
    CHUNK        lmask, rmask;
124
138k
#ifdef S_USED
125
#ifdef S_CONST
126
    CHUNK        S = (CHUNK)op->s.c;
127
#else /* !defined(S_CONST) */
128
138k
    const CHUNK *s = (CHUNK *)(void *)op->s.b.ptr;
129
138k
    CHUNK        S;
130
138k
    int          s_skew;
131
138k
#endif /* !defined(S_CONST) */
132
#else /* !defined(S_USED) */
133
#define S 0
134
#undef S_CONST
135
#endif /* !defined(S_USED) */
136
#ifdef T_USED
137
#ifdef T_CONST
138
    CHUNK        T = (CHUNK)op->t.c;
139
#else /* !defined(T_CONST) */
140
    const CHUNK *t = (CHUNK *)(void *)op->t.b.ptr;
141
    CHUNK        T;
142
    int          t_skew;
143
#endif /* !defined(T_CONST) */
144
#else /* !defined(T_USED) */
145
138k
#define T 0
146
138k
#undef T_CONST
147
138k
#endif /* !defined(T_USED) */
148
138k
#if defined(S_SKEW) || defined(T_SKEW)
149
138k
    int skewflags = 0;
150
138k
#endif
151
138k
    CHUNK        D;
152
138k
    int          dpos = op->dpos;
153
138k
    CHUNK       *d = (CHUNK *)(void *)d_;
154
155
    /* Align d to CHUNKSIZE */
156
138k
    ADJUST_TO_CHUNK(d,dpos);
157
158
    /* On entry len = length in 'depth' chunks. Change it to be the length
159
     * in bits, and add on the number of bits we skip at the start of the
160
     * run. */
161
138k
    len    = len * op->depth + dpos;
162
163
    /* lmask = the set of bits to alter in the output bitmap on the left
164
     * hand edge of the run. rmask = the set of bits NOT to alter in the
165
     * output bitmap on the right hand edge of the run. */
166
138k
    lmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & dpos)));
167
138k
    rmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & len)));
168
138k
    if (rmask == CHUNKONES) rmask = 0;
169
170
#if defined(S_CONST) || defined(T_CONST)
171
    /* S and T should be supplied as 'depth' bits. Duplicate them up to be
172
     * byte size (if they are supplied byte sized, that's fine too). */
173
    if (op->depth & 1) {
174
#ifdef S_CONST
175
        S |= S<<1;
176
#endif /* !defined(S_CONST) */
177
#ifdef T_CONST
178
        T |= T<<1;
179
#endif /* !defined(T_CONST) */
180
    }
181
    if (op->depth & 3) {
182
#ifdef S_CONST
183
        S |= S<<2;
184
#endif /* !defined(S_CONST) */
185
#ifdef T_CONST
186
        T |= T<<2;
187
#endif /* !defined(T_CONST) */
188
    }
189
    if (op->depth & 7) {
190
#ifdef S_CONST
191
        S |= S<<4;
192
#endif /* !defined(S_CONST) */
193
#ifdef T_CONST
194
        T |= T<<4;
195
#endif /* !defined(T_CONST) */
196
    }
197
#if CHUNKSIZE > 8
198
    if (op->depth & 15) {
199
#ifdef S_CONST
200
        S |= S<<8;
201
#endif /* !defined(S_CONST) */
202
#ifdef T_CONST
203
        T |= T<<8;
204
#endif /* !defined(T_CONST) */
205
    }
206
#endif /* CHUNKSIZE > 8 */
207
#if CHUNKSIZE > 16
208
    if (op->depth & 31) {
209
#ifdef S_CONST
210
        S |= S<<16;
211
#endif /* !defined(S_CONST) */
212
#ifdef T_CONST
213
        T |= T<<16;
214
#endif /* !defined(T_CONST) */
215
    }
216
#endif /* CHUNKSIZE > 16 */
217
#endif /* defined(S_CONST) || defined(T_CONST) */
218
219
    /* Note #1: This mirrors what the original code did, but I think it has
220
     * the risk of moving s and t back beyond officially allocated space. We
221
     * may be saved by the fact that all blocks have a word or two in front
222
     * of them due to the allocator. If we ever get valgrind properly marking
223
     * allocated blocks as readable etc, then this may throw some spurious
224
     * errors. RJW. */
225
138k
#ifdef S_SKEW
226
138k
    {
227
138k
        int slen, slen2;
228
138k
        int spos = op->s.b.pos;
229
138k
        ADJUST_TO_CHUNK(s, spos);
230
138k
        s_skew = spos - dpos;
231
138k
        if (s_skew < 0) {
232
91.1k
            s_skew += CHUNKSIZE;
233
91.1k
            s--;
234
91.1k
            skewflags |= 1; /* Suppress reading off left edge */
235
91.1k
        }
236
        /* We are allowed to read all the data bits, so: len - dpos + tpos
237
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
238
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
239
         * This is larger, then suppress. */
240
138k
        slen  = (len + s_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
241
138k
        slen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
242
138k
        if ((s_skew == 0) || (slen < slen2)) {
243
119k
            skewflags |= 4; /* Suppress reading off the right edge */
244
119k
        }
245
138k
    }
246
138k
#endif /* !defined(S_SKEW) */
247
#ifdef T_SKEW
248
    {
249
        int tlen, tlen2;
250
        int tpos = op->t.b.pos;
251
        ADJUST_TO_CHUNK(t, tpos);
252
        t_skew = tpos - dpos;
253
        if (t_skew < 0) {
254
            t_skew += CHUNKSIZE;
255
            t--;
256
            skewflags |= 2; /* Suppress reading off left edge */
257
        }
258
        /* We are allowed to read all the data bits, so: len - dpos + tpos
259
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
260
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
261
         * This is larger, then suppress. */
262
        tlen  = (len + t_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
263
        tlen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
264
        if ((t_skew == 0) || (tlen < tlen2)) {
265
            skewflags |= 8; /* Suppress reading off the right edge */
266
        }
267
    }
268
#endif /* !defined(T_SKEW) */
269
270
138k
    len -= CHUNKSIZE; /* len = bytes to do - CHUNKSIZE */
271
    /* len <= 0 means 1 word or less to do */
272
138k
    if (len <= 0) {
273
        /* Short case - starts and ends in the same chunk */
274
3
        lmask &= ~rmask; /* Combined mask = bits to alter */
275
3
        SAFE_FETCH_S(skewflags & 1,skewflags & 4);
276
3
        SAFE_FETCH_T(skewflags & 2,skewflags & 8);
277
3
        SPECIFIC_CODE(D, *d, S, T);
278
3
        *d = (*d & ~lmask) | (D & lmask);
279
3
        return;
280
3
    }
281
138k
    if ((lmask != CHUNKONES)
282
19.2k
#if defined(S_SKEW) || defined(T_SKEW)
283
19.2k
        || (skewflags & 3)
284
138k
#endif
285
138k
        ) {
286
        /* Unaligned left hand case */
287
119k
        SAFE_FETCH_S(skewflags & 1,s_skew == 0);
288
119k
        SAFE_FETCH_T(skewflags & 2,t_skew == 0);
289
119k
        SPECIFIC_CODE(D, *d, S, T);
290
119k
        *d = (*d & ~lmask) | (D & lmask);
291
119k
        d++;
292
119k
        len -= CHUNKSIZE;
293
119k
    }
294
138k
    if (len > 0) {
295
        /* Simple middle case (complete destination chunks). */
296
137k
#ifdef S_SKEW
297
137k
        if (s_skew == 0) {
298
#ifdef T_SKEW
299
            if (t_skew == 0) {
300
                do {
301
                    SPECIFIC_CODE(*d, *d, *s++, *t++);
302
                    d++;
303
                    len -= CHUNKSIZE;
304
                } while (len > 0);
305
            } else
306
#endif /* !defined(T_SKEW) */
307
40.6k
            {
308
1.21M
                do {
309
1.21M
                    FETCH_T;
310
1.21M
                    SPECIFIC_CODE(*d, *d, *s++, T);
311
1.21M
                    d++;
312
1.21M
                    len -= CHUNKSIZE;
313
1.21M
                } while (len > 0);
314
40.6k
            }
315
40.6k
        } else
316
97.1k
#endif /* !defined(S_SKEW) */
317
97.1k
        {
318
#ifdef T_SKEW
319
            if (t_skew == 0) {
320
                do {
321
                    FETCH_S;
322
                    SPECIFIC_CODE(*d, *d, S, *t++);
323
                    d++;
324
                    len -= CHUNKSIZE;
325
                } while (len > 0);
326
            } else
327
#endif /* !defined(T_SKEW) */
328
97.1k
            {
329
987k
                do {
330
987k
                    FETCH_S;
331
987k
                    FETCH_T;
332
987k
                    SPECIFIC_CODE(*d, *d, S, T);
333
987k
                    d++;
334
987k
                    len -= CHUNKSIZE;
335
987k
                } while (len > 0);
336
97.1k
            }
337
97.1k
        }
338
137k
    }
339
    /* Unaligned right hand case */
340
138k
    SAFE_FETCH_S(0,skewflags & 4);
341
138k
    SAFE_FETCH_T(0,skewflags & 8);
342
138k
    SPECIFIC_CODE(D, *d, S, T);
343
138k
    *d = (*d & rmask) | (D & ~rmask);
344
138k
}
Unexecuted instantiation: gsroprun.c:invert_rop_run1
Unexecuted instantiation: gsroprun.c:xor_rop_run1_const_t
gsroprun.c:sets_rop_run1
Line
Count
Source
118
45.5M
{
119
#ifndef SPECIFIC_CODE
120
    rop_proc     proc = rop_proc_table[op->rop];
121
#define SPECIFIC_CODE(OUT_, D_,S_,T_) OUT_ = proc(D_,S_,T_)
122
#endif /* !defined(SPECIFIC_CODE) */
123
45.5M
    CHUNK        lmask, rmask;
124
45.5M
#ifdef S_USED
125
#ifdef S_CONST
126
    CHUNK        S = (CHUNK)op->s.c;
127
#else /* !defined(S_CONST) */
128
45.5M
    const CHUNK *s = (CHUNK *)(void *)op->s.b.ptr;
129
45.5M
    CHUNK        S;
130
45.5M
    int          s_skew;
131
45.5M
#endif /* !defined(S_CONST) */
132
#else /* !defined(S_USED) */
133
#define S 0
134
#undef S_CONST
135
#endif /* !defined(S_USED) */
136
#ifdef T_USED
137
#ifdef T_CONST
138
    CHUNK        T = (CHUNK)op->t.c;
139
#else /* !defined(T_CONST) */
140
    const CHUNK *t = (CHUNK *)(void *)op->t.b.ptr;
141
    CHUNK        T;
142
    int          t_skew;
143
#endif /* !defined(T_CONST) */
144
#else /* !defined(T_USED) */
145
45.5M
#define T 0
146
45.5M
#undef T_CONST
147
45.5M
#endif /* !defined(T_USED) */
148
45.5M
#if defined(S_SKEW) || defined(T_SKEW)
149
45.5M
    int skewflags = 0;
150
45.5M
#endif
151
45.5M
    CHUNK        D;
152
45.5M
    int          dpos = op->dpos;
153
45.5M
    CHUNK       *d = (CHUNK *)(void *)d_;
154
155
    /* Align d to CHUNKSIZE */
156
45.5M
    ADJUST_TO_CHUNK(d,dpos);
157
158
    /* On entry len = length in 'depth' chunks. Change it to be the length
159
     * in bits, and add on the number of bits we skip at the start of the
160
     * run. */
161
45.5M
    len    = len * op->depth + dpos;
162
163
    /* lmask = the set of bits to alter in the output bitmap on the left
164
     * hand edge of the run. rmask = the set of bits NOT to alter in the
165
     * output bitmap on the right hand edge of the run. */
166
45.5M
    lmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & dpos)));
167
45.5M
    rmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & len)));
168
45.5M
    if (rmask == CHUNKONES) rmask = 0;
169
170
#if defined(S_CONST) || defined(T_CONST)
171
    /* S and T should be supplied as 'depth' bits. Duplicate them up to be
172
     * byte size (if they are supplied byte sized, that's fine too). */
173
    if (op->depth & 1) {
174
#ifdef S_CONST
175
        S |= S<<1;
176
#endif /* !defined(S_CONST) */
177
#ifdef T_CONST
178
        T |= T<<1;
179
#endif /* !defined(T_CONST) */
180
    }
181
    if (op->depth & 3) {
182
#ifdef S_CONST
183
        S |= S<<2;
184
#endif /* !defined(S_CONST) */
185
#ifdef T_CONST
186
        T |= T<<2;
187
#endif /* !defined(T_CONST) */
188
    }
189
    if (op->depth & 7) {
190
#ifdef S_CONST
191
        S |= S<<4;
192
#endif /* !defined(S_CONST) */
193
#ifdef T_CONST
194
        T |= T<<4;
195
#endif /* !defined(T_CONST) */
196
    }
197
#if CHUNKSIZE > 8
198
    if (op->depth & 15) {
199
#ifdef S_CONST
200
        S |= S<<8;
201
#endif /* !defined(S_CONST) */
202
#ifdef T_CONST
203
        T |= T<<8;
204
#endif /* !defined(T_CONST) */
205
    }
206
#endif /* CHUNKSIZE > 8 */
207
#if CHUNKSIZE > 16
208
    if (op->depth & 31) {
209
#ifdef S_CONST
210
        S |= S<<16;
211
#endif /* !defined(S_CONST) */
212
#ifdef T_CONST
213
        T |= T<<16;
214
#endif /* !defined(T_CONST) */
215
    }
216
#endif /* CHUNKSIZE > 16 */
217
#endif /* defined(S_CONST) || defined(T_CONST) */
218
219
    /* Note #1: This mirrors what the original code did, but I think it has
220
     * the risk of moving s and t back beyond officially allocated space. We
221
     * may be saved by the fact that all blocks have a word or two in front
222
     * of them due to the allocator. If we ever get valgrind properly marking
223
     * allocated blocks as readable etc, then this may throw some spurious
224
     * errors. RJW. */
225
45.5M
#ifdef S_SKEW
226
45.5M
    {
227
45.5M
        int slen, slen2;
228
45.5M
        int spos = op->s.b.pos;
229
45.5M
        ADJUST_TO_CHUNK(s, spos);
230
45.5M
        s_skew = spos - dpos;
231
45.5M
        if (s_skew < 0) {
232
31.4M
            s_skew += CHUNKSIZE;
233
31.4M
            s--;
234
31.4M
            skewflags |= 1; /* Suppress reading off left edge */
235
31.4M
        }
236
        /* We are allowed to read all the data bits, so: len - dpos + tpos
237
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
238
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
239
         * This is larger, then suppress. */
240
45.5M
        slen  = (len + s_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
241
45.5M
        slen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
242
45.5M
        if ((s_skew == 0) || (slen < slen2)) {
243
25.9M
            skewflags |= 4; /* Suppress reading off the right edge */
244
25.9M
        }
245
45.5M
    }
246
45.5M
#endif /* !defined(S_SKEW) */
247
#ifdef T_SKEW
248
    {
249
        int tlen, tlen2;
250
        int tpos = op->t.b.pos;
251
        ADJUST_TO_CHUNK(t, tpos);
252
        t_skew = tpos - dpos;
253
        if (t_skew < 0) {
254
            t_skew += CHUNKSIZE;
255
            t--;
256
            skewflags |= 2; /* Suppress reading off left edge */
257
        }
258
        /* We are allowed to read all the data bits, so: len - dpos + tpos
259
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
260
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
261
         * This is larger, then suppress. */
262
        tlen  = (len + t_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
263
        tlen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
264
        if ((t_skew == 0) || (tlen < tlen2)) {
265
            skewflags |= 8; /* Suppress reading off the right edge */
266
        }
267
    }
268
#endif /* !defined(T_SKEW) */
269
270
45.5M
    len -= CHUNKSIZE; /* len = bytes to do - CHUNKSIZE */
271
    /* len <= 0 means 1 word or less to do */
272
45.5M
    if (len <= 0) {
273
        /* Short case - starts and ends in the same chunk */
274
226k
        lmask &= ~rmask; /* Combined mask = bits to alter */
275
226k
        SAFE_FETCH_S(skewflags & 1,skewflags & 4);
276
226k
        SAFE_FETCH_T(skewflags & 2,skewflags & 8);
277
226k
        SPECIFIC_CODE(D, *d, S, T);
278
226k
        *d = (*d & ~lmask) | (D & lmask);
279
226k
        return;
280
226k
    }
281
45.2M
    if ((lmask != CHUNKONES)
282
9.29M
#if defined(S_SKEW) || defined(T_SKEW)
283
9.29M
        || (skewflags & 3)
284
45.2M
#endif
285
45.2M
        ) {
286
        /* Unaligned left hand case */
287
35.9M
        SAFE_FETCH_S(skewflags & 1,s_skew == 0);
288
35.9M
        SAFE_FETCH_T(skewflags & 2,t_skew == 0);
289
35.9M
        SPECIFIC_CODE(D, *d, S, T);
290
35.9M
        *d = (*d & ~lmask) | (D & lmask);
291
35.9M
        d++;
292
35.9M
        len -= CHUNKSIZE;
293
35.9M
    }
294
45.2M
    if (len > 0) {
295
        /* Simple middle case (complete destination chunks). */
296
34.6M
#ifdef S_SKEW
297
34.6M
        if (s_skew == 0) {
298
#ifdef T_SKEW
299
            if (t_skew == 0) {
300
                do {
301
                    SPECIFIC_CODE(*d, *d, *s++, *t++);
302
                    d++;
303
                    len -= CHUNKSIZE;
304
                } while (len > 0);
305
            } else
306
#endif /* !defined(T_SKEW) */
307
9.60M
            {
308
437M
                do {
309
437M
                    FETCH_T;
310
437M
                    SPECIFIC_CODE(*d, *d, *s++, T);
311
437M
                    d++;
312
437M
                    len -= CHUNKSIZE;
313
437M
                } while (len > 0);
314
9.60M
            }
315
9.60M
        } else
316
25.0M
#endif /* !defined(S_SKEW) */
317
25.0M
        {
318
#ifdef T_SKEW
319
            if (t_skew == 0) {
320
                do {
321
                    FETCH_S;
322
                    SPECIFIC_CODE(*d, *d, S, *t++);
323
                    d++;
324
                    len -= CHUNKSIZE;
325
                } while (len > 0);
326
            } else
327
#endif /* !defined(T_SKEW) */
328
25.0M
            {
329
469M
                do {
330
469M
                    FETCH_S;
331
469M
                    FETCH_T;
332
469M
                    SPECIFIC_CODE(*d, *d, S, T);
333
469M
                    d++;
334
469M
                    len -= CHUNKSIZE;
335
469M
                } while (len > 0);
336
25.0M
            }
337
25.0M
        }
338
34.6M
    }
339
    /* Unaligned right hand case */
340
45.2M
    SAFE_FETCH_S(0,skewflags & 4);
341
45.2M
    SAFE_FETCH_T(0,skewflags & 8);
342
45.2M
    SPECIFIC_CODE(D, *d, S, T);
343
45.2M
    *d = (*d & rmask) | (D & ~rmask);
344
45.2M
}
gsroprun.c:dors_rop_run1_const_t
Line
Count
Source
118
4.57M
{
119
#ifndef SPECIFIC_CODE
120
    rop_proc     proc = rop_proc_table[op->rop];
121
#define SPECIFIC_CODE(OUT_, D_,S_,T_) OUT_ = proc(D_,S_,T_)
122
#endif /* !defined(SPECIFIC_CODE) */
123
4.57M
    CHUNK        lmask, rmask;
124
4.57M
#ifdef S_USED
125
#ifdef S_CONST
126
    CHUNK        S = (CHUNK)op->s.c;
127
#else /* !defined(S_CONST) */
128
4.57M
    const CHUNK *s = (CHUNK *)(void *)op->s.b.ptr;
129
4.57M
    CHUNK        S;
130
4.57M
    int          s_skew;
131
4.57M
#endif /* !defined(S_CONST) */
132
#else /* !defined(S_USED) */
133
#define S 0
134
#undef S_CONST
135
#endif /* !defined(S_USED) */
136
#ifdef T_USED
137
#ifdef T_CONST
138
    CHUNK        T = (CHUNK)op->t.c;
139
#else /* !defined(T_CONST) */
140
    const CHUNK *t = (CHUNK *)(void *)op->t.b.ptr;
141
    CHUNK        T;
142
    int          t_skew;
143
#endif /* !defined(T_CONST) */
144
#else /* !defined(T_USED) */
145
4.57M
#define T 0
146
4.57M
#undef T_CONST
147
4.57M
#endif /* !defined(T_USED) */
148
4.57M
#if defined(S_SKEW) || defined(T_SKEW)
149
4.57M
    int skewflags = 0;
150
4.57M
#endif
151
4.57M
    CHUNK        D;
152
4.57M
    int          dpos = op->dpos;
153
4.57M
    CHUNK       *d = (CHUNK *)(void *)d_;
154
155
    /* Align d to CHUNKSIZE */
156
4.57M
    ADJUST_TO_CHUNK(d,dpos);
157
158
    /* On entry len = length in 'depth' chunks. Change it to be the length
159
     * in bits, and add on the number of bits we skip at the start of the
160
     * run. */
161
4.57M
    len    = len * op->depth + dpos;
162
163
    /* lmask = the set of bits to alter in the output bitmap on the left
164
     * hand edge of the run. rmask = the set of bits NOT to alter in the
165
     * output bitmap on the right hand edge of the run. */
166
4.57M
    lmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & dpos)));
167
4.57M
    rmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & len)));
168
4.57M
    if (rmask == CHUNKONES) rmask = 0;
169
170
#if defined(S_CONST) || defined(T_CONST)
171
    /* S and T should be supplied as 'depth' bits. Duplicate them up to be
172
     * byte size (if they are supplied byte sized, that's fine too). */
173
    if (op->depth & 1) {
174
#ifdef S_CONST
175
        S |= S<<1;
176
#endif /* !defined(S_CONST) */
177
#ifdef T_CONST
178
        T |= T<<1;
179
#endif /* !defined(T_CONST) */
180
    }
181
    if (op->depth & 3) {
182
#ifdef S_CONST
183
        S |= S<<2;
184
#endif /* !defined(S_CONST) */
185
#ifdef T_CONST
186
        T |= T<<2;
187
#endif /* !defined(T_CONST) */
188
    }
189
    if (op->depth & 7) {
190
#ifdef S_CONST
191
        S |= S<<4;
192
#endif /* !defined(S_CONST) */
193
#ifdef T_CONST
194
        T |= T<<4;
195
#endif /* !defined(T_CONST) */
196
    }
197
#if CHUNKSIZE > 8
198
    if (op->depth & 15) {
199
#ifdef S_CONST
200
        S |= S<<8;
201
#endif /* !defined(S_CONST) */
202
#ifdef T_CONST
203
        T |= T<<8;
204
#endif /* !defined(T_CONST) */
205
    }
206
#endif /* CHUNKSIZE > 8 */
207
#if CHUNKSIZE > 16
208
    if (op->depth & 31) {
209
#ifdef S_CONST
210
        S |= S<<16;
211
#endif /* !defined(S_CONST) */
212
#ifdef T_CONST
213
        T |= T<<16;
214
#endif /* !defined(T_CONST) */
215
    }
216
#endif /* CHUNKSIZE > 16 */
217
#endif /* defined(S_CONST) || defined(T_CONST) */
218
219
    /* Note #1: This mirrors what the original code did, but I think it has
220
     * the risk of moving s and t back beyond officially allocated space. We
221
     * may be saved by the fact that all blocks have a word or two in front
222
     * of them due to the allocator. If we ever get valgrind properly marking
223
     * allocated blocks as readable etc, then this may throw some spurious
224
     * errors. RJW. */
225
4.57M
#ifdef S_SKEW
226
4.57M
    {
227
4.57M
        int slen, slen2;
228
4.57M
        int spos = op->s.b.pos;
229
4.57M
        ADJUST_TO_CHUNK(s, spos);
230
4.57M
        s_skew = spos - dpos;
231
4.57M
        if (s_skew < 0) {
232
4.31M
            s_skew += CHUNKSIZE;
233
4.31M
            s--;
234
4.31M
            skewflags |= 1; /* Suppress reading off left edge */
235
4.31M
        }
236
        /* We are allowed to read all the data bits, so: len - dpos + tpos
237
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
238
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
239
         * This is larger, then suppress. */
240
4.57M
        slen  = (len + s_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
241
4.57M
        slen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
242
4.57M
        if ((s_skew == 0) || (slen < slen2)) {
243
2.13M
            skewflags |= 4; /* Suppress reading off the right edge */
244
2.13M
        }
245
4.57M
    }
246
4.57M
#endif /* !defined(S_SKEW) */
247
#ifdef T_SKEW
248
    {
249
        int tlen, tlen2;
250
        int tpos = op->t.b.pos;
251
        ADJUST_TO_CHUNK(t, tpos);
252
        t_skew = tpos - dpos;
253
        if (t_skew < 0) {
254
            t_skew += CHUNKSIZE;
255
            t--;
256
            skewflags |= 2; /* Suppress reading off left edge */
257
        }
258
        /* We are allowed to read all the data bits, so: len - dpos + tpos
259
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
260
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
261
         * This is larger, then suppress. */
262
        tlen  = (len + t_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
263
        tlen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
264
        if ((t_skew == 0) || (tlen < tlen2)) {
265
            skewflags |= 8; /* Suppress reading off the right edge */
266
        }
267
    }
268
#endif /* !defined(T_SKEW) */
269
270
4.57M
    len -= CHUNKSIZE; /* len = bytes to do - CHUNKSIZE */
271
    /* len <= 0 means 1 word or less to do */
272
4.57M
    if (len <= 0) {
273
        /* Short case - starts and ends in the same chunk */
274
9.32k
        lmask &= ~rmask; /* Combined mask = bits to alter */
275
9.32k
        SAFE_FETCH_S(skewflags & 1,skewflags & 4);
276
9.32k
        SAFE_FETCH_T(skewflags & 2,skewflags & 8);
277
9.32k
        SPECIFIC_CODE(D, *d, S, T);
278
9.32k
        *d = (*d & ~lmask) | (D & lmask);
279
9.32k
        return;
280
9.32k
    }
281
4.56M
    if ((lmask != CHUNKONES)
282
214k
#if defined(S_SKEW) || defined(T_SKEW)
283
214k
        || (skewflags & 3)
284
4.56M
#endif
285
4.56M
        ) {
286
        /* Unaligned left hand case */
287
4.35M
        SAFE_FETCH_S(skewflags & 1,s_skew == 0);
288
4.35M
        SAFE_FETCH_T(skewflags & 2,t_skew == 0);
289
4.35M
        SPECIFIC_CODE(D, *d, S, T);
290
4.35M
        *d = (*d & ~lmask) | (D & lmask);
291
4.35M
        d++;
292
4.35M
        len -= CHUNKSIZE;
293
4.35M
    }
294
4.56M
    if (len > 0) {
295
        /* Simple middle case (complete destination chunks). */
296
2.73M
#ifdef S_SKEW
297
2.73M
        if (s_skew == 0) {
298
#ifdef T_SKEW
299
            if (t_skew == 0) {
300
                do {
301
                    SPECIFIC_CODE(*d, *d, *s++, *t++);
302
                    d++;
303
                    len -= CHUNKSIZE;
304
                } while (len > 0);
305
            } else
306
#endif /* !defined(T_SKEW) */
307
144k
            {
308
1.10M
                do {
309
1.10M
                    FETCH_T;
310
1.10M
                    SPECIFIC_CODE(*d, *d, *s++, T);
311
1.10M
                    d++;
312
1.10M
                    len -= CHUNKSIZE;
313
1.10M
                } while (len > 0);
314
144k
            }
315
144k
        } else
316
2.59M
#endif /* !defined(S_SKEW) */
317
2.59M
        {
318
#ifdef T_SKEW
319
            if (t_skew == 0) {
320
                do {
321
                    FETCH_S;
322
                    SPECIFIC_CODE(*d, *d, S, *t++);
323
                    d++;
324
                    len -= CHUNKSIZE;
325
                } while (len > 0);
326
            } else
327
#endif /* !defined(T_SKEW) */
328
2.59M
            {
329
7.06M
                do {
330
7.06M
                    FETCH_S;
331
7.06M
                    FETCH_T;
332
7.06M
                    SPECIFIC_CODE(*d, *d, S, T);
333
7.06M
                    d++;
334
7.06M
                    len -= CHUNKSIZE;
335
7.06M
                } while (len > 0);
336
2.59M
            }
337
2.59M
        }
338
2.73M
    }
339
    /* Unaligned right hand case */
340
4.56M
    SAFE_FETCH_S(0,skewflags & 4);
341
4.56M
    SAFE_FETCH_T(0,skewflags & 8);
342
4.56M
    SPECIFIC_CODE(D, *d, S, T);
343
4.56M
    *d = (*d & rmask) | (D & ~rmask);
344
4.56M
}
Unexecuted instantiation: gsroprun.c:generic_rop_run1
gsroprun.c:generic_rop_run1_const_t
Line
Count
Source
118
4.44M
{
119
4.44M
#ifndef SPECIFIC_CODE
120
4.44M
    rop_proc     proc = rop_proc_table[op->rop];
121
4.44M
#define SPECIFIC_CODE(OUT_, D_,S_,T_) OUT_ = proc(D_,S_,T_)
122
4.44M
#endif /* !defined(SPECIFIC_CODE) */
123
4.44M
    CHUNK        lmask, rmask;
124
4.44M
#ifdef S_USED
125
#ifdef S_CONST
126
    CHUNK        S = (CHUNK)op->s.c;
127
#else /* !defined(S_CONST) */
128
4.44M
    const CHUNK *s = (CHUNK *)(void *)op->s.b.ptr;
129
4.44M
    CHUNK        S;
130
4.44M
    int          s_skew;
131
4.44M
#endif /* !defined(S_CONST) */
132
#else /* !defined(S_USED) */
133
#define S 0
134
#undef S_CONST
135
#endif /* !defined(S_USED) */
136
4.44M
#ifdef T_USED
137
4.44M
#ifdef T_CONST
138
4.44M
    CHUNK        T = (CHUNK)op->t.c;
139
#else /* !defined(T_CONST) */
140
    const CHUNK *t = (CHUNK *)(void *)op->t.b.ptr;
141
    CHUNK        T;
142
    int          t_skew;
143
#endif /* !defined(T_CONST) */
144
#else /* !defined(T_USED) */
145
#define T 0
146
#undef T_CONST
147
#endif /* !defined(T_USED) */
148
4.44M
#if defined(S_SKEW) || defined(T_SKEW)
149
4.44M
    int skewflags = 0;
150
4.44M
#endif
151
4.44M
    CHUNK        D;
152
4.44M
    int          dpos = op->dpos;
153
4.44M
    CHUNK       *d = (CHUNK *)(void *)d_;
154
155
    /* Align d to CHUNKSIZE */
156
4.44M
    ADJUST_TO_CHUNK(d,dpos);
157
158
    /* On entry len = length in 'depth' chunks. Change it to be the length
159
     * in bits, and add on the number of bits we skip at the start of the
160
     * run. */
161
4.44M
    len    = len * op->depth + dpos;
162
163
    /* lmask = the set of bits to alter in the output bitmap on the left
164
     * hand edge of the run. rmask = the set of bits NOT to alter in the
165
     * output bitmap on the right hand edge of the run. */
166
4.44M
    lmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & dpos)));
167
4.44M
    rmask  = RE((CHUNKONES>>((CHUNKSIZE-1) & len)));
168
4.44M
    if (rmask == CHUNKONES) rmask = 0;
169
170
4.44M
#if defined(S_CONST) || defined(T_CONST)
171
    /* S and T should be supplied as 'depth' bits. Duplicate them up to be
172
     * byte size (if they are supplied byte sized, that's fine too). */
173
4.44M
    if (op->depth & 1) {
174
#ifdef S_CONST
175
        S |= S<<1;
176
#endif /* !defined(S_CONST) */
177
4.44M
#ifdef T_CONST
178
4.44M
        T |= T<<1;
179
4.44M
#endif /* !defined(T_CONST) */
180
4.44M
    }
181
4.44M
    if (op->depth & 3) {
182
#ifdef S_CONST
183
        S |= S<<2;
184
#endif /* !defined(S_CONST) */
185
4.44M
#ifdef T_CONST
186
4.44M
        T |= T<<2;
187
4.44M
#endif /* !defined(T_CONST) */
188
4.44M
    }
189
4.44M
    if (op->depth & 7) {
190
#ifdef S_CONST
191
        S |= S<<4;
192
#endif /* !defined(S_CONST) */
193
4.44M
#ifdef T_CONST
194
4.44M
        T |= T<<4;
195
4.44M
#endif /* !defined(T_CONST) */
196
4.44M
    }
197
4.44M
#if CHUNKSIZE > 8
198
4.44M
    if (op->depth & 15) {
199
#ifdef S_CONST
200
        S |= S<<8;
201
#endif /* !defined(S_CONST) */
202
4.44M
#ifdef T_CONST
203
4.44M
        T |= T<<8;
204
4.44M
#endif /* !defined(T_CONST) */
205
4.44M
    }
206
4.44M
#endif /* CHUNKSIZE > 8 */
207
4.44M
#if CHUNKSIZE > 16
208
4.44M
    if (op->depth & 31) {
209
#ifdef S_CONST
210
        S |= S<<16;
211
#endif /* !defined(S_CONST) */
212
4.44M
#ifdef T_CONST
213
4.44M
        T |= T<<16;
214
4.44M
#endif /* !defined(T_CONST) */
215
4.44M
    }
216
4.44M
#endif /* CHUNKSIZE > 16 */
217
4.44M
#endif /* defined(S_CONST) || defined(T_CONST) */
218
219
    /* Note #1: This mirrors what the original code did, but I think it has
220
     * the risk of moving s and t back beyond officially allocated space. We
221
     * may be saved by the fact that all blocks have a word or two in front
222
     * of them due to the allocator. If we ever get valgrind properly marking
223
     * allocated blocks as readable etc, then this may throw some spurious
224
     * errors. RJW. */
225
4.44M
#ifdef S_SKEW
226
4.44M
    {
227
4.44M
        int slen, slen2;
228
4.44M
        int spos = op->s.b.pos;
229
4.44M
        ADJUST_TO_CHUNK(s, spos);
230
4.44M
        s_skew = spos - dpos;
231
4.44M
        if (s_skew < 0) {
232
4.04M
            s_skew += CHUNKSIZE;
233
4.04M
            s--;
234
4.04M
            skewflags |= 1; /* Suppress reading off left edge */
235
4.04M
        }
236
        /* We are allowed to read all the data bits, so: len - dpos + tpos
237
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
238
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
239
         * This is larger, then suppress. */
240
4.44M
        slen  = (len + s_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
241
4.44M
        slen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
242
4.44M
        if ((s_skew == 0) || (slen < slen2)) {
243
2.30M
            skewflags |= 4; /* Suppress reading off the right edge */
244
2.30M
        }
245
4.44M
    }
246
4.44M
#endif /* !defined(S_SKEW) */
247
#ifdef T_SKEW
248
    {
249
        int tlen, tlen2;
250
        int tpos = op->t.b.pos;
251
        ADJUST_TO_CHUNK(t, tpos);
252
        t_skew = tpos - dpos;
253
        if (t_skew < 0) {
254
            t_skew += CHUNKSIZE;
255
            t--;
256
            skewflags |= 2; /* Suppress reading off left edge */
257
        }
258
        /* We are allowed to read all the data bits, so: len - dpos + tpos
259
         * We're allowed to read in CHUNKS, so: CHUNKUP(len-dpos+tpos).
260
         * This code will actually read CHUNKUP(len)+CHUNKSIZE bits. If
261
         * This is larger, then suppress. */
262
        tlen  = (len + t_skew    + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
263
        tlen2 = (len + CHUNKSIZE + CHUNKSIZE-1) & ~(CHUNKSIZE-1);
264
        if ((t_skew == 0) || (tlen < tlen2)) {
265
            skewflags |= 8; /* Suppress reading off the right edge */
266
        }
267
    }
268
#endif /* !defined(T_SKEW) */
269
270
4.44M
    len -= CHUNKSIZE; /* len = bytes to do - CHUNKSIZE */
271
    /* len <= 0 means 1 word or less to do */
272
4.44M
    if (len <= 0) {
273
        /* Short case - starts and ends in the same chunk */
274
5.13k
        lmask &= ~rmask; /* Combined mask = bits to alter */
275
5.13k
        SAFE_FETCH_S(skewflags & 1,skewflags & 4);
276
5.13k
        SAFE_FETCH_T(skewflags & 2,skewflags & 8);
277
5.13k
        SPECIFIC_CODE(D, *d, S, T);
278
5.13k
        *d = (*d & ~lmask) | (D & lmask);
279
5.13k
        return;
280
5.13k
    }
281
4.43M
    if ((lmask != CHUNKONES)
282
258k
#if defined(S_SKEW) || defined(T_SKEW)
283
258k
        || (skewflags & 3)
284
4.43M
#endif
285
4.43M
        ) {
286
        /* Unaligned left hand case */
287
4.17M
        SAFE_FETCH_S(skewflags & 1,s_skew == 0);
288
4.17M
        SAFE_FETCH_T(skewflags & 2,t_skew == 0);
289
4.17M
        SPECIFIC_CODE(D, *d, S, T);
290
4.17M
        *d = (*d & ~lmask) | (D & lmask);
291
4.17M
        d++;
292
4.17M
        len -= CHUNKSIZE;
293
4.17M
    }
294
4.43M
    if (len > 0) {
295
        /* Simple middle case (complete destination chunks). */
296
2.83M
#ifdef S_SKEW
297
2.83M
        if (s_skew == 0) {
298
#ifdef T_SKEW
299
            if (t_skew == 0) {
300
                do {
301
                    SPECIFIC_CODE(*d, *d, *s++, *t++);
302
                    d++;
303
                    len -= CHUNKSIZE;
304
                } while (len > 0);
305
            } else
306
#endif /* !defined(T_SKEW) */
307
282k
            {
308
3.45M
                do {
309
3.45M
                    FETCH_T;
310
3.45M
                    SPECIFIC_CODE(*d, *d, *s++, T);
311
3.45M
                    d++;
312
3.45M
                    len -= CHUNKSIZE;
313
3.45M
                } while (len > 0);
314
282k
            }
315
282k
        } else
316
2.55M
#endif /* !defined(S_SKEW) */
317
2.55M
        {
318
#ifdef T_SKEW
319
            if (t_skew == 0) {
320
                do {
321
                    FETCH_S;
322
                    SPECIFIC_CODE(*d, *d, S, *t++);
323
                    d++;
324
                    len -= CHUNKSIZE;
325
                } while (len > 0);
326
            } else
327
#endif /* !defined(T_SKEW) */
328
2.55M
            {
329
8.01M
                do {
330
8.01M
                    FETCH_S;
331
8.01M
                    FETCH_T;
332
8.01M
                    SPECIFIC_CODE(*d, *d, S, T);
333
8.01M
                    d++;
334
8.01M
                    len -= CHUNKSIZE;
335
8.01M
                } while (len > 0);
336
2.55M
            }
337
2.55M
        }
338
2.83M
    }
339
    /* Unaligned right hand case */
340
4.43M
    SAFE_FETCH_S(0,skewflags & 4);
341
4.43M
    SAFE_FETCH_T(0,skewflags & 8);
342
4.43M
    SPECIFIC_CODE(D, *d, S, T);
343
4.43M
    *d = (*d & rmask) | (D & ~rmask);
344
4.43M
}
Unexecuted instantiation: gsroprun.c:generic_rop_run1_const_st
345
346
#undef ADJUST_TO_CHUNK
347
#undef CHUNKSIZE
348
#undef CHUNK
349
#undef CHUNKONES
350
#undef FETCH_S
351
#undef FETCH_T
352
#undef SAFE_FETCH_S
353
#undef SAFE_FETCH_T
354
#undef RE
355
#undef S
356
#undef S_USED
357
#undef S_CONST
358
#undef S_SKEW
359
#undef SKEW_FETCH
360
#undef SAFE_SKEW_FETCH
361
#undef SPECIFIC_CODE
362
#undef SPECIFIC_ROP
363
#undef T
364
#undef T_USED
365
#undef T_CONST
366
#undef T_SKEW
367
#undef TEMPLATE_NAME
368
#undef ROP_PTRDIFF_T
369
370
#else
371
int dummy;
372
#endif