diff options
-rw-r--r-- | draw/draw_device.c | 37 | ||||
-rw-r--r-- | draw/draw_scale.c | 1394 | ||||
-rw-r--r-- | draw/draw_simple_scale.c | 68 | ||||
-rw-r--r-- | fitz/fitz.h | 2 |
4 files changed, 74 insertions, 1427 deletions
diff --git a/draw/draw_device.c b/draw/draw_device.c index 9a238b73..1cb3e4b9 100644 --- a/draw/draw_device.c +++ b/draw/draw_device.c @@ -821,7 +821,7 @@ fz_draw_fill_shade(fz_device *devp, fz_shade *shade, fz_matrix ctm, float alpha) } static fz_pixmap * -fz_transform_pixmap(fz_context *ctx, fz_pixmap *image, fz_matrix *ctm, int x, int y, int dx, int dy, int gridfit) +fz_transform_pixmap(fz_context *ctx, fz_pixmap *image, fz_matrix *ctm, int x, int y, int dx, int dy, int gridfit, fz_bbox *clip) { fz_pixmap *scaled; @@ -831,7 +831,7 @@ fz_transform_pixmap(fz_context *ctx, fz_pixmap *image, fz_matrix *ctm, int x, in fz_matrix m = *ctm; if (gridfit) fz_gridfit_matrix(&m); - scaled = fz_scale_pixmap(ctx, image, m.e, m.f, m.a, m.d); + scaled = fz_scale_pixmap(ctx, image, m.e, m.f, m.a, m.d, clip); if (!scaled) return NULL; ctm->a = scaled->w; @@ -845,9 +845,17 @@ fz_transform_pixmap(fz_context *ctx, fz_pixmap *image, fz_matrix *ctm, int x, in { /* Other orthogonal flip/rotation cases */ fz_matrix m = *ctm; + fz_bbox rclip; if (gridfit) fz_gridfit_matrix(&m); - scaled = fz_scale_pixmap(ctx, image, m.f, m.e, m.b, m.c); + if (clip) + { + rclip.x0 = clip->y0; + rclip.y0 = clip->x0; + rclip.x1 = clip->y1; + rclip.y1 = clip->x1; + } + scaled = fz_scale_pixmap(ctx, image, m.f, m.e, m.b, m.c, (clip ? &rclip : 0)); if (!scaled) return NULL; ctm->b = scaled->w; @@ -860,7 +868,7 @@ fz_transform_pixmap(fz_context *ctx, fz_pixmap *image, fz_matrix *ctm, int x, in /* Downscale, non rectilinear case */ if (dx > 0 && dy > 0) { - scaled = fz_scale_pixmap(ctx, image, 0, 0, (float)dx, (float)dy); + scaled = fz_scale_pixmap(ctx, image, 0, 0, (float)dx, (float)dy, NULL); return scaled; } @@ -878,6 +886,9 @@ fz_draw_fill_image(fz_device *devp, fz_pixmap *image, fz_matrix ctm, float alpha fz_context *ctx = dev->ctx; fz_draw_state *state = &dev->stack[dev->top]; fz_colorspace *model = state->dest->colorspace; + fz_bbox clip = fz_bound_pixmap(state->dest); + + clip = fz_intersect_bbox(clip, state->scissor); fz_var(scaled); @@ -915,14 +926,14 @@ fz_draw_fill_image(fz_device *devp, fz_pixmap *image, fz_matrix ctm, float alpha if (dx < image->w && dy < image->h) { int gridfit = alpha == 1.0f && !(dev->flags & FZ_DRAWDEV_FLAGS_TYPE3); - scaled = fz_transform_pixmap(ctx, image, &ctm, state->dest->x, state->dest->y, dx, dy, gridfit); + scaled = fz_transform_pixmap(ctx, image, &ctm, state->dest->x, state->dest->y, dx, dy, gridfit, &clip); if (!scaled) { if (dx < 1) dx = 1; if (dy < 1) dy = 1; - scaled = fz_scale_pixmap(ctx, image, image->x, image->y, dx, dy); + scaled = fz_scale_pixmap(ctx, image, image->x, image->y, dx, dy, NULL); } if (scaled) image = scaled; @@ -973,6 +984,9 @@ fz_draw_fill_image_mask(fz_device *devp, fz_pixmap *image, fz_matrix ctm, int i; fz_draw_state *state = &dev->stack[dev->top]; fz_colorspace *model = state->dest->colorspace; + fz_bbox clip = fz_bound_pixmap(state->dest); + + clip = fz_intersect_bbox(clip, state->scissor); if (image->w == 0 || image->h == 0) return; @@ -985,14 +999,14 @@ fz_draw_fill_image_mask(fz_device *devp, fz_pixmap *image, fz_matrix ctm, if (dx < image->w && dy < image->h) { int gridfit = alpha == 1.0f && !(dev->flags & FZ_DRAWDEV_FLAGS_TYPE3); - scaled = fz_transform_pixmap(dev->ctx, image, &ctm, state->dest->x, state->dest->y, dx, dy, gridfit); + scaled = fz_transform_pixmap(dev->ctx, image, &ctm, state->dest->x, state->dest->y, dx, dy, gridfit, &clip); if (!scaled) { if (dx < 1) dx = 1; if (dy < 1) dy = 1; - scaled = fz_scale_pixmap(dev->ctx, image, image->x, image->y, dx, dy); + scaled = fz_scale_pixmap(dev->ctx, image, image->x, image->y, dx, dy, NULL); } if (scaled) image = scaled; @@ -1025,6 +1039,9 @@ fz_draw_clip_image_mask(fz_device *devp, fz_pixmap *image, fz_rect *rect, fz_mat int dx, dy; fz_draw_state *state = push_stack(dev); fz_colorspace *model = state->dest->colorspace; + fz_bbox clip = fz_bound_pixmap(state->dest); + + clip = fz_intersect_bbox(clip, state->scissor); fz_var(mask); fz_var(dest); @@ -1066,14 +1083,14 @@ fz_draw_clip_image_mask(fz_device *devp, fz_pixmap *image, fz_rect *rect, fz_mat if (dx < image->w && dy < image->h) { int gridfit = !(dev->flags & FZ_DRAWDEV_FLAGS_TYPE3); - scaled = fz_transform_pixmap(dev->ctx, image, &ctm, state->dest->x, state->dest->y, dx, dy, gridfit); + scaled = fz_transform_pixmap(dev->ctx, image, &ctm, state->dest->x, state->dest->y, dx, dy, gridfit, &clip); if (!scaled) { if (dx < 1) dx = 1; if (dy < 1) dy = 1; - scaled = fz_scale_pixmap(dev->ctx, image, image->x, image->y, dx, dy); + scaled = fz_scale_pixmap(dev->ctx, image, image->x, image->y, dx, dy, NULL); } if (scaled) image = scaled; diff --git a/draw/draw_scale.c b/draw/draw_scale.c deleted file mode 100644 index e478633c..00000000 --- a/draw/draw_scale.c +++ /dev/null @@ -1,1394 +0,0 @@ -/* -This code does smooth scaling of a pixmap. - -This function returns a new pixmap representing the area starting at (0,0) -given by taking the source pixmap src, scaling it to width w, and height h, -and then positioning it at (frac(x),frac(y)). -*/ - -#include "fitz.h" - -/* Do we special case handling of single pixel high/wide images? The - * 'purest' handling is given by not special casing them, but certain - * files that use such images 'stack' them to give full images. Not - * special casing them results in then being fainter and giving noticeable - * rounding errors. - */ -#define SINGLE_PIXEL_SPECIALS - -/* If we're compiling as thumb code, then we need to tell the compiler - * to enter and exit ARM mode around our assembly sections. If we move - * the ARM functions to a separate file and arrange for it to be compiled - * without thumb mode, we can save some time on entry. - */ -#ifdef ARCH_ARM -#ifdef ARCH_THUMB -#define ENTER_ARM ".balign 4\nmov r12,pc\nbx r12\n0:.arm\n" -#define ENTER_THUMB "9:.thumb\n" -#else -#define ENTER_ARM -#define ENTER_THUMB -#endif -#endif - -#ifdef DEBUG_SCALING -#ifdef WIN32 -#include <windows.h> -static void debug_print(const char *fmt, ...) -{ - va_list args; - char text[256]; - va_start(args, fmt); - vsprintf(text, fmt, args); - va_end(args); - OutputDebugStringA(text); - printf(text); -} -#else -static void debug_print(const char *fmt, ...) -{ - va_list args; - va_start(args, fmt); - vfprintf(stderr, fmt, args); - va_end(args); -} -#endif -#endif -#ifdef DEBUG_SCALING -#define DBUG(A) debug_print A -#else -#define DBUG(A) do {} while(0==1) -#endif - -/* -Consider a row of source samples, src, of width src_w, positioned at x, -scaled to width dst_w. - -src[i] is centred at: x + (i + 0.5)*dst_w/src_w - -Therefore the distance between the centre of the jth output pixel and -the centre of the ith source sample is: - -dist[j,i] = j + 0.5 - (x + (i + 0.5)*dst_w/src_w) - -When scaling up, therefore: - -dst[j] = SUM(filter(dist[j,i]) * src[i]) - (for all ints i) - -This can be simplified by noticing that filters are only non zero within -a given filter width (henceforth called W). So: - -dst[j] = SUM(filter(dist[j,i]) * src[i]) - (for ints i, s.t. (j*src_w/dst_w)-W < i < (j*src_w/dst_w)+W) - -When scaling down, each filtered source sample is stretched to be wider -to avoid aliasing issues. This effectively reduces the distance between -centres. - -dst[j] = SUM(filter(dist[j,i] * F) * F * src[i]) - (where F = dst_w/src_w) - (for ints i, s.t. (j-W)/F < i < (j+W)/F) - -*/ - -typedef struct fz_scale_filter_s fz_scale_filter; - -struct fz_scale_filter_s -{ - int width; - float (*fn)(fz_scale_filter *, float); -}; - -/* Image scale filters */ - -static float -triangle(fz_scale_filter *filter, float f) -{ - if (f >= 1) - return 0; - return 1-f; -} - -static float -box(fz_scale_filter *filter, float f) -{ - if (f >= 0.5f) - return 0; - return 1; -} - -static float -simple(fz_scale_filter *filter, float x) -{ - if (x >= 1) - return 0; - return 1 + (2*x - 3)*x*x; -} - -static float -lanczos2(fz_scale_filter *filter, float x) -{ - if (x >= 2) - return 0; - return sinf(M_PI*x) * sinf(M_PI*x/2) / (M_PI*x) / (M_PI*x/2); -} - -static float -lanczos3(fz_scale_filter *filter, float f) -{ - if (f >= 3) - return 0; - return sinf(M_PI*f) * sinf(M_PI*f/3) / (M_PI*f) / (M_PI*f/3); -} - -/* -The Mitchell family of filters is defined: - - f(x) = 1 { (12-9B-6C)x^3 + (-18+12B+6C)x^2 + (6-2B) for x < 1 - - { - 6 { (-B-6C)x^3+(6B+30C)x^2+(-12B-48C)x+(8B+24C) for 1<=x<=2 - -The 'best' ones lie along the line B+2C = 1. -The literature suggests that B=1/3, C=1/3 is best. - - f(x) = 1 { (12-3-2)x^3 - (-18+4+2)x^2 + (16/3) for x < 1 - - { - 6 { (-7/3)x^3 + 12x^2 - 20x + (32/3) for 1<=x<=2 - - f(x) = 1 { 21x^3 - 36x^2 + 16 for x < 1 - - { - 18{ -7x^3 + 36x^2 - 60x + 32 for 1<=x<=2 -*/ - -static float -mitchell(fz_scale_filter *filter, float x) -{ - if (x >= 2) - return 0; - if (x >= 1) - return (32 + x*(-60 + x*(36 - 7*x)))/18; - return (16 + x*x*(-36 + 21*x))/18; -} - -fz_scale_filter fz_scale_filter_box = { 1, box }; -fz_scale_filter fz_scale_filter_triangle = { 1, triangle }; -fz_scale_filter fz_scale_filter_simple = { 1, simple }; -fz_scale_filter fz_scale_filter_lanczos2 = { 2, lanczos2 }; -fz_scale_filter fz_scale_filter_lanczos3 = { 3, lanczos3 }; -fz_scale_filter fz_scale_filter_mitchell = { 2, mitchell }; - -/* -We build ourselves a set of tables to contain the precalculated weights -for a given set of scale settings. - -The first dst_w entries in index are the index into index of the -sets of weight for each destination pixel. - -Each of the sets of weights is a set of values consisting of: - the minimum source pixel index used for this destination pixel - the number of weights used for this destination pixel - the weights themselves - -So to calculate dst[i] we do the following: - - weights = &index[index[i]]; - min = *weights++; - len = *weights++; - dst[i] = 0; - while (--len > 0) - dst[i] += src[min++] * *weights++ - -in addition, we guarantee that at the end of this process weights will now -point to the weights value for dst pixel i+1. - -In the simplest version of this algorithm, we would scale the whole image -horizontally first into a temporary buffer, then scale that temporary -buffer again vertically to give us our result. Using such a simple -algorithm would mean that could use the same style of weights for both -horizontal and vertical scaling. - -Unfortunately, this would also require a large temporary buffer, -particularly in the case where we are scaling up. - -We therefore modify the algorithm as follows; we scale scanlines from the -source image horizontally into a temporary buffer, until we have all the -contributors for a given output scanline. We then produce that output -scanline from the temporary buffer. In this way we restrict the height -of the temporary buffer to a small fraction of the final size. - -Unfortunately, this means that the pseudo code for recombining a -scanline of fully scaled pixels is as follows: - - weights = &index[index[y]]; - min = *weights++; - len = *weights++; - for (x=0 to dst_w) - min2 = min - len2 = len - weights2 = weights - dst[x] = 0; - while (--len2 > 0) - dst[x] += temp[x][(min2++) % tmp_buf_height] * *weights2++ - -i.e. it requires a % operation for every source pixel - this is typically -expensive. - -To avoid this, we alter the order in which vertical weights are stored, -so that they are ordered in the same order as the temporary buffer lines -would appear. This simplifies the algorithm to: - - weights = &index[index[y]]; - min = *weights++; - len = *weights++; - for (x=0 to dst_w) - min2 = 0 - len2 = len - weights2 = weights - dst[x] = 0; - while (--len2 > 0) - dst[x] += temp[i][min2++] * *weights2++ - -This means that len may be larger than it needs to be (due to the -possible inclusion of a zero weight row or two), but in practise this -is only an increase of 1 or 2 at worst. - -We implement this by generating the weights as normal (but ensuring we -leave enough space) and then reordering afterwards. - -*/ - -typedef struct fz_weights_s fz_weights; - -struct fz_weights_s -{ - int flip; - int count; - int max_len; - int n; - int new_line; - int index[1]; -}; - -static fz_weights * -new_weights(fz_context *ctx, fz_scale_filter *filter, int src_w, float dst_w, int dst_w_i, int n, int flip) -{ - int max_len; - fz_weights *weights; - - if (src_w > dst_w) - { - /* Scaling down, so there will be a maximum of - * 2*filterwidth*src_w/dst_w src pixels - * contributing to each dst pixel. */ - max_len = (int)ceilf((2 * filter->width * src_w)/dst_w); - if (max_len > src_w) - max_len = src_w; - } - else - { - /* Scaling up, so there will be a maximum of - * 2*filterwidth src pixels contributing to each dst pixel. - */ - max_len = 2 * filter->width; - } - /* We need the size of the struct, - * plus dst_w*sizeof(int) for the index - * plus (2+max_len)*sizeof(int) for the weights - * plus room for an extra set of weights for reordering. - */ - weights = fz_malloc(ctx, sizeof(*weights)+(max_len+3)*(dst_w_i+1)*sizeof(int)); - if (!weights) - return NULL; - weights->count = -1; - weights->max_len = max_len; - weights->index[0] = dst_w_i; - weights->n = n; - weights->flip = flip; - return weights; -} - -static void -init_weights(fz_weights *weights, int j) -{ - int index; - - assert(weights->count == j-1); - weights->count++; - weights->new_line = 1; - if (j == 0) - index = weights->index[0]; - else - { - index = weights->index[j-1]; - index += 2 + weights->index[index+1]; - } - weights->index[j] = index; /* row pointer */ - weights->index[index] = 0; /* min */ - weights->index[index+1] = 0; /* len */ -} - -static void -add_weight(fz_weights *weights, int j, int i, fz_scale_filter *filter, - float x, float F, float G, int src_w, float dst_w) -{ - float dist = j - x + 0.5f - ((i + 0.5f)*dst_w/src_w); - float f; - int min, len, index, weight; - - dist *= G; - if (dist < 0) - dist = -dist; - f = filter->fn(filter, dist)*F; - weight = (int)(256*f+0.5f); - if (weight == 0) - return; - - /* wrap i back into range */ -#ifdef MIRROR_WRAP - do - { - if (i < 0) - i = -1-i; - else if (i >= src_w) - i = 2*src_w-1-i; - else - break; - } - while (1); -#elif defined(WRAP) - if (i < 0) - i = 0; - else if (i >= src_w) - i = src_w-1; -#else - if (i < 0) - { - i = 0; - weight = 0; - } - else if (i >= src_w) - { - i = src_w-1; - weight = 0; - } - if (weight == 0) - return; -#endif - - DBUG(("add_weight[%d][%d] = %d(%g) dist=%g\n",j,i,weight,f,dist)); - - if (weights->new_line) - { - /* New line */ - weights->new_line = 0; - index = weights->index[j]; /* row pointer */ - weights->index[index] = i; /* min */ - weights->index[index+1] = 0; /* len */ - } - index = weights->index[j]; - min = weights->index[index++]; - len = weights->index[index++]; - while (i < min) - { - /* This only happens in rare cases, but we need to insert - * one earlier. In exceedingly rare cases we may need to - * insert more than one earlier. */ - int k; - - for (k = len; k > 0; k--) - { - weights->index[index+k] = weights->index[index+k-1]; - } - weights->index[index] = 0; - min--; - len++; - weights->index[index-2] = min; - weights->index[index-1] = len; - } - if (i-min >= len) - { - /* The usual case */ - while (i-min >= ++len) - { - weights->index[index+len-1] = 0; - } - assert(len-1 == i-min); - weights->index[index+i-min] = weight; - weights->index[index-1] = len; - assert(len <= weights->max_len); - } - else - { - /* Infrequent case */ - weights->index[index+i-min] += weight; - } -} - -static void -reorder_weights(fz_weights *weights, int j, int src_w) -{ - int idx = weights->index[j]; - int min = weights->index[idx++]; - int len = weights->index[idx++]; - int max = weights->max_len; - int tmp = idx+max; - int i, off; - - /* Copy into the temporary area */ - memcpy(&weights->index[tmp], &weights->index[idx], sizeof(int)*len); - - /* Pad out if required */ - assert(len <= max); - assert(min+len <= src_w); - off = 0; - if (len < max) - { - memset(&weights->index[tmp+len], 0, sizeof(int)*(max-len)); - len = max; - if (min + len > src_w) - { - off = min + len - src_w; - min = src_w - len; - weights->index[idx-2] = min; - } - weights->index[idx-1] = len; - } - - /* Copy back into the proper places */ - for (i = 0; i < len; i++) - { - weights->index[idx+((min+i+off) % max)] = weights->index[tmp+i]; - } -} - -/* Due to rounding and edge effects, the sums for the weights sometimes don't - * add up to 256. This causes visible rendering effects. Therefore, we take - * pains to ensure that they 1) never exceed 256, and 2) add up to exactly - * 256 for all pixels that are completely covered. See bug #691629. */ -static void -check_weights(fz_weights *weights, int j, int w, float x, float wf) -{ - int idx, len; - int sum = 0; - int max = -256; - int maxidx = 0; - int i; - - idx = weights->index[j]; - idx++; /* min */ - len = weights->index[idx++]; - - for(i=0; i < len; i++) - { - int v = weights->index[idx++]; - sum += v; - if (v > max) - { - max = v; - maxidx = idx; - } - } - /* If we aren't the first or last pixel, OR if the sum is too big - * then adjust it. */ - if (((j != 0) && (j != w-1)) || (sum > 256)) - weights->index[maxidx-1] += 256-sum; - /* Otherwise, if we are the first pixel, and it's fully covered, then - * adjust it. */ - else if ((j == 0) && (x < 0.0001F) && (sum != 256)) - weights->index[maxidx-1] += 256-sum; - /* Finally, if we are the last pixel, and it's fully covered, then - * adjust it. */ - else if ((j == w-1) && ((float)w-wf < 0.0001F) && (sum != 256)) - weights->index[maxidx-1] += 256-sum; - DBUG(("total weight %d = %d\n", j, sum)); -} - -static fz_weights * -make_weights(fz_context *ctx, int src_w, float x, float dst_w, fz_scale_filter *filter, int vertical, int dst_w_int, int n, int flip) -{ - fz_weights *weights; - float F, G; - float window; - int j; - - if (dst_w < src_w) - { - /* Scaling down */ - F = dst_w / src_w; - G = 1; - } - else - { - /* Scaling up */ - F = 1; - G = src_w / dst_w; - } - window = filter->width / F; - DBUG(("make_weights src_w=%d x=%g dst_w=%g dst_w_int=%d F=%g window=%g\n", src_w, x, dst_w, dst_w_int, F, window)); - weights = new_weights(ctx, filter, src_w, dst_w, dst_w_int, n, flip); - if (!weights) - return NULL; - for (j = 0; j < dst_w_int; j++) - { - /* find the position of the centre of dst[j] in src space */ - float centre = (j - x + 0.5f)*src_w/dst_w - 0.5f; - int l, r; - l = ceilf(centre - window); - r = floorf(centre + window); - DBUG(("%d: centre=%g l=%d r=%d\n", j, centre, l, r)); - init_weights(weights, j); - for (; l <= r; l++) - { - add_weight(weights, j, l, filter, x, F, G, src_w, dst_w); - } - check_weights(weights, j, dst_w_int, x, dst_w); - if (vertical) - { - reorder_weights(weights, j, src_w); - } - } - weights->count++; /* weights->count = dst_w_int now */ - return weights; -} - -static void -scale_row_to_temp(int *dst, unsigned char *src, fz_weights *weights) -{ - int *contrib = &weights->index[weights->index[0]]; - int len, i, j, n; - unsigned char *min; - - n = weights->n; - if (weights->flip) - { - dst += (weights->count-1)*n; - for (i=weights->count; i > 0; i--) - { - min = &src[n * *contrib++]; - len = *contrib++; - for (j = 0; j < n; j++) - dst[j] = 0; - while (len-- > 0) - { - for (j = n; j > 0; j--) - *dst++ += *min++ * *contrib; - dst -= n; - contrib++; - } - dst -= n; - } - } - else - { - for (i=weights->count; i > 0; i--) - { - min = &src[n * *contrib++]; - len = *contrib++; - for (j = 0; j < n; j++) - dst[j] = 0; - while (len-- > 0) - { - for (j = n; j > 0; j--) - *dst++ += *min++ * *contrib; - dst -= n; - contrib++; - } - dst += n; - } - } -} - -#ifdef ARCH_ARM - -static void -scale_row_to_temp1(int *dst, unsigned char *src, fz_weights *weights) -__attribute__((naked)); - -static void -scale_row_to_temp2(int *dst, unsigned char *src, fz_weights *weights) -__attribute__((naked)); - -static void -scale_row_to_temp4(int *dst, unsigned char *src, fz_weights *weights) -__attribute__((naked)); - -static void -scale_row_from_temp(unsigned char *dst, int *src, fz_weights *weights, int width, int row) -__attribute__((naked)); - -static void -scale_row_to_temp1(int *dst, unsigned char *src, fz_weights *weights) -{ - /* possible optimisation in here; unroll inner loops to avoid stall. */ - asm volatile( - ENTER_ARM - "stmfd r13!,{r4-r5,r9,r14} \n" - "@ r0 = dst \n" - "@ r1 = src \n" - "@ r2 = weights \n" - "ldr r12,[r2],#4 @ r12= flip \n" - "ldr r3, [r2],#16 @ r3 = count r2 = &index\n" - "ldr r4, [r2] @ r4 = index[0] \n" - "cmp r12,#0 @ if (flip) \n" - "beq 4f @ { \n" - "add r2, r2, r4, LSL #2 @ r2 = &index[index[0]] \n" - "add r0, r0, r3, LSL #2 @ dst += count \n" - "1: \n" - "ldr r4, [r2], #4 @ r4 = *contrib++ \n" - "ldr r9, [r2], #4 @ r9 = len = *contrib++ \n" - "mov r5, #0 @ r5 = a = 0 \n" - "add r4, r1, r4 @ r4 = min = &src[r4] \n" - "cmp r9, #0 @ while (len-- > 0) \n" - "beq 3f @ { \n" - "2: \n" - "ldr r12,[r2], #4 @ r12 = *contrib++ \n" - "ldrb r14,[r4], #1 @ r14 = *min++ \n" - "subs r9, r9, #1 @ r9 = len-- \n" - "@stall on r14 \n" - "mla r5, r12,r14,r5 @ g += r14 * r12 \n" - "bgt 2b @ } \n" - "3: \n" - "str r5,[r0, #-4]! @ *--dst=a \n" - "subs r3, r3, #1 @ i-- \n" - "bgt 1b @ \n" - "ldmfd r13!,{r4-r5,r9,PC} @ pop, return to thumb \n" - "4:" - "add r2, r2, r4, LSL #2 @ r2 = &index[index[0]] \n" - "5:" - "ldr r4, [r2], #4 @ r4 = *contrib++ \n" - "ldr r9, [r2], #4 @ r9 = len = *contrib++ \n" - "mov r5, #0 @ r5 = a = 0 \n" - "add r4, r1, r4 @ r4 = min = &src[r4] \n" - "cmp r9, #0 @ while (len-- > 0) \n" - "beq 7f @ { \n" - "6: \n" - "ldr r12,[r2], #4 @ r12 = *contrib++ \n" - "ldrb r14,[r4], #1 @ r14 = *min++ \n" - "subs r9, r9, #1 @ r9 = len-- \n" - "@stall on r14 \n" - "mla r5, r12,r14,r5 @ a += r14 * r12 \n" - "bgt 6b @ } \n" - "7: \n" - "str r5, [r0], #4 @ *dst++=a \n" - "subs r3, r3, #1 @ i-- \n" - "bgt 5b @ \n" - "ldmfd r13!,{r4-r5,r9,PC} @ pop, return to thumb \n" - ENTER_THUMB - ); -} - -static void -scale_row_to_temp2(int *dst, unsigned char *src, fz_weights *weights) -{ - asm volatile( - ENTER_ARM - "stmfd r13!,{r4-r6,r9-r11,r14} \n" - "@ r0 = dst \n" - "@ r1 = src \n" - "@ r2 = weights \n" - "ldr r12,[r2],#4 @ r12= flip \n" - "ldr r3, [r2],#16 @ r3 = count r2 = &index\n" - "ldr r4, [r2] @ r4 = index[0] \n" - "cmp r12,#0 @ if (flip) \n" - "beq 4f @ { \n" - "add r2, r2, r4, LSL #2 @ r2 = &index[index[0]] \n" - "add r0, r0, r3, LSL #3 @ dst += 2*count \n" - "1: \n" - "ldr r4, [r2], #4 @ r4 = *contrib++ \n" - "ldr r9, [r2], #4 @ r9 = len = *contrib++ \n" - "mov r5, #0 @ r5 = g = 0 \n" - "mov r6, #0 @ r6 = a = 0 \n" - "add r4, r1, r4, LSL #1 @ r4 = min = &src[2*r4] \n" - "cmp r9, #0 @ while (len-- > 0) \n" - "beq 3f @ { \n" - "2: \n" - "ldr r14,[r2], #4 @ r14 = *contrib++ \n" - "ldrb r11,[r4], #1 @ r11 = *min++ \n" - "ldrb r12,[r4], #1 @ r12 = *min++ \n" - "subs r9, r9, #1 @ r9 = len-- \n" - "mla r5, r14,r11,r5 @ g += r11 * r14 \n" - "mla r6, r14,r12,r6 @ a += r12 * r14 \n" - "bgt 2b @ } \n" - "3: \n" - "stmdb r0!,{r5,r6} @ *--dst=a;*--dst=g; \n" - "subs r3, r3, #1 @ i-- \n" - "bgt 1b @ \n" - "ldmfd r13!,{r4-r6,r9-r11,PC} @ pop, return to thumb \n" - "4:" - "add r2, r2, r4, LSL #2 @ r2 = &index[index[0]] \n" - "5:" - "ldr r4, [r2], #4 @ r4 = *contrib++ \n" - "ldr r9, [r2], #4 @ r9 = len = *contrib++ \n" - "mov r5, #0 @ r5 = g = 0 \n" - "mov r6, #0 @ r6 = a = 0 \n" - "add r4, r1, r4, LSL #1 @ r4 = min = &src[2*r4] \n" - "cmp r9, #0 @ while (len-- > 0) \n" - "beq 7f @ { \n" - "6: \n" - "ldr r14,[r2], #4 @ r10 = *contrib++ \n" - "ldrb r11,[r4], #1 @ r11 = *min++ \n" - "ldrb r12,[r4], #1 @ r12 = *min++ \n" - "subs r9, r9, #1 @ r9 = len-- \n" - "mla r5, r14,r11,r5 @ g += r11 * r14 \n" - "mla r6, r14,r12,r6 @ a += r12 * r14 \n" - "bgt 6b @ } \n" - "7: \n" - "stmia r0!,{r5,r6} @ *dst++=r;*dst++=g; \n" - "subs r3, r3, #1 @ i-- \n" - "bgt 5b @ \n" - "ldmfd r13!,{r4-r6,r9-r11,PC} @ pop, return to thumb \n" - ENTER_THUMB - ); -} - -static void -scale_row_to_temp4(int *dst, unsigned char *src, fz_weights *weights) -{ - asm volatile( - ENTER_ARM - "stmfd r13!,{r4-r11,r14} \n" - "@ r0 = dst \n" - "@ r1 = src \n" - "@ r2 = weights \n" - "ldr r12,[r2],#4 @ r12= flip \n" - "ldr r3, [r2],#16 @ r3 = count r2 = &index\n" - "ldr r4, [r2] @ r4 = index[0] \n" - "cmp r12,#0 @ if (flip) \n" - "beq 4f @ { \n" - "add r2, r2, r4, LSL #2 @ r2 = &index[index[0]] \n" - "add r0, r0, r3, LSL #4 @ dst += 4*count \n" - "1: \n" - "ldr r4, [r2], #4 @ r4 = *contrib++ \n" - "ldr r9, [r2], #4 @ r9 = len = *contrib++ \n" - "mov r5, #0 @ r5 = r = 0 \n" - "mov r6, #0 @ r6 = g = 0 \n" - "mov r7, #0 @ r7 = b = 0 \n" - "mov r8, #0 @ r8 = a = 0 \n" - "add r4, r1, r4, LSL #2 @ r4 = min = &src[4*r4] \n" - "cmp r9, #0 @ while (len-- > 0) \n" - "beq 3f @ { \n" - "2: \n" - "ldr r10,[r2], #4 @ r10 = *contrib++ \n" - "ldrb r11,[r4], #1 @ r11 = *min++ \n" - "ldrb r12,[r4], #1 @ r12 = *min++ \n" - "ldrb r14,[r4], #1 @ r14 = *min++ \n" - "mla r5, r10,r11,r5 @ r += r11 * r10 \n" - "ldrb r11,[r4], #1 @ r11 = *min++ \n" - "mla r6, r10,r12,r6 @ g += r12 * r10 \n" - "mla r7, r10,r14,r7 @ b += r14 * r10 \n" - "mla r8, r10,r11,r8 @ a += r11 * r10 \n" - "subs r9, r9, #1 @ r9 = len-- \n" - "bgt 2b @ } \n" - "3: \n" - "stmdb r0!,{r5,r6,r7,r8} @ *--dst=a;*--dst=b; \n" - " @ *--dst=g;*--dst=r; \n" - "subs r3, r3, #1 @ i-- \n" - "bgt 1b @ \n" - "ldmfd r13!,{r4-r11,PC} @ pop, return to thumb \n" - "4:" - "add r2, r2, r4, LSL #2 @ r2 = &index[index[0]] \n" - "5:" - "ldr r4, [r2], #4 @ r4 = *contrib++ \n" - "ldr r9, [r2], #4 @ r9 = len = *contrib++ \n" - "mov r5, #0 @ r5 = r = 0 \n" - "mov r6, #0 @ r6 = g = 0 \n" - "mov r7, #0 @ r7 = b = 0 \n" - "mov r8, #0 @ r8 = a = 0 \n" - "add r4, r1, r4, LSL #2 @ r4 = min = &src[4*r4] \n" - "cmp r9, #0 @ while (len-- > 0) \n" - "beq 7f @ { \n" - "6: \n" - "ldr r10,[r2], #4 @ r10 = *contrib++ \n" - "ldrb r11,[r4], #1 @ r11 = *min++ \n" - "ldrb r12,[r4], #1 @ r12 = *min++ \n" - "ldrb r14,[r4], #1 @ r14 = *min++ \n" - "mla r5, r10,r11,r5 @ r += r11 * r10 \n" - "ldrb r11,[r4], #1 @ r11 = *min++ \n" - "mla r6, r10,r12,r6 @ g += r12 * r10 \n" - "mla r7, r10,r14,r7 @ b += r14 * r10 \n" - "mla r8, r10,r11,r8 @ a += r11 * r10 \n" - "subs r9, r9, #1 @ r9 = len-- \n" - "bgt 6b @ } \n" - "7: \n" - "stmia r0!,{r5,r6,r7,r8} @ *dst++=r;*dst++=g; \n" - " @ *dst++=b;*dst++=a; \n" - "subs r3, r3, #1 @ i-- \n" - "bgt 5b @ \n" - "ldmfd r13!,{r4-r11,PC} @ pop, return to thumb \n" - ENTER_THUMB - ); -} - -static void -scale_row_from_temp(unsigned char *dst, int *src, fz_weights *weights, int width, int row) -{ - asm volatile( - ENTER_ARM - "ldr r12,[r13] @ r12= row \n" - "add r2, r2, #20 @ r2 = weights->index \n" - "stmfd r13!,{r4-r11,r14} \n" - "@ r0 = dst \n" - "@ r1 = src \n" - "@ r2 = &weights->index[0] \n" - "@ r3 = width \n" - "@ r12= row \n" - "ldr r4, [r2, r12, LSL #2] @ r4 = index[row] \n" - "add r2, r2, #4 @ r2 = &index[1] \n" - "mov r6, r3 @ r6 = x = width \n" - "ldr r14,[r2, r4, LSL #2]! @ r2 = contrib = index[index[row]+1]\n" - " @ r14= len = *contrib \n" - "1: \n" - "mov r5, r1 @ r5 = min = src \n" - "mov r7, #1<<15 @ r7 = val = 1<<15 \n" - "movs r8, r14 @ r8 = len2 = len \n" - "add r9, r2, #4 @ r9 = contrib2 \n" - "ble 3f @ while (len2-- > 0) { \n" - "2: \n" - "ldr r10,[r9], #4 @ r10 = *contrib2++ \n" - "ldr r12,[r5], r3, LSL #2 @ r12 = *min r5 = min += width\n" - "subs r8, r8, #1 @ len2-- \n" - "@ stall r12 \n" - "mla r7, r10,r12,r7 @ val += r12 * r10 \n" - "bgt 2b @ } \n" - "3: \n" - "movs r7, r7, asr #16 @ r7 = val >>= 16 \n" - "movlt r7, #0 @ if (r7 < 0) r7 = 0 \n" - "cmp r7, #255 @ if (r7 > 255) \n" - "add r1, r1, #4 @ src++ \n" - "movgt r7, #255 @ r7 = 255 \n" - "subs r6, r6, #1 @ x-- \n" - "strb r7, [r0], #1 @ *dst++ = val \n" - "bgt 1b @ \n" - "ldmfd r13!,{r4-r11,PC} @ pop, return to thumb \n" - ENTER_THUMB - ); -} - -#else - -static void -scale_row_to_temp1(int *dst, unsigned char *src, fz_weights *weights) -{ - int *contrib = &weights->index[weights->index[0]]; - int len, i; - unsigned char *min; - - assert(weights->n == 1); - if (weights->flip) - { - dst += weights->count; - for (i=weights->count; i > 0; i--) - { - int val = 0; - min = &src[*contrib++]; - len = *contrib++; - while (len-- > 0) - { - val += *min++ * *contrib++; - } - *--dst = val; - } - } - else - { - for (i=weights->count; i > 0; i--) - { - int val = 0; - min = &src[*contrib++]; - len = *contrib++; - while (len-- > 0) - { - val += *min++ * *contrib++; - } - *dst++ = val; - } - } -} - -static void -scale_row_to_temp2(int *dst, unsigned char *src, fz_weights *weights) -{ - int *contrib = &weights->index[weights->index[0]]; - int len, i; - unsigned char *min; - - assert(weights->n == 2); - if (weights->flip) - { - dst += 2*weights->count; - for (i=weights->count; i > 0; i--) - { - int c1 = 0; - int c2 = 0; - min = &src[2 * *contrib++]; - len = *contrib++; - while (len-- > 0) - { - c1 += *min++ * *contrib; - c2 += *min++ * *contrib++; - } - *--dst = c2; - *--dst = c1; - } - } - else - { - for (i=weights->count; i > 0; i--) - { - int c1 = 0; - int c2 = 0; - min = &src[2 * *contrib++]; - len = *contrib++; - while (len-- > 0) - { - c1 += *min++ * *contrib; - c2 += *min++ * *contrib++; - } - *dst++ = c1; - *dst++ = c2; - } - } -} - -static void -scale_row_to_temp4(int *dst, unsigned char *src, fz_weights *weights) -{ - int *contrib = &weights->index[weights->index[0]]; - int len, i; - unsigned char *min; - - assert(weights->n == 4); - if (weights->flip) - { - dst += 4*weights->count; - for (i=weights->count; i > 0; i--) - { - int r = 0; - int g = 0; - int b = 0; - int a = 0; - min = &src[4 * *contrib++]; - len = *contrib++; - while (len-- > 0) - { - r += *min++ * *contrib; - g += *min++ * *contrib; - b += *min++ * *contrib; - a += *min++ * *contrib++; - } - *--dst = a; - *--dst = b; - *--dst = g; - *--dst = r; - } - } - else - { - for (i=weights->count; i > 0; i--) - { - int r = 0; - int g = 0; - int b = 0; - int a = 0; - min = &src[4 * *contrib++]; - len = *contrib++; - while (len-- > 0) - { - r += *min++ * *contrib; - g += *min++ * *contrib; - b += *min++ * *contrib; - a += *min++ * *contrib++; - } - *dst++ = r; - *dst++ = g; - *dst++ = b; - *dst++ = a; - } - } -} - -static void -scale_row_from_temp(unsigned char *dst, int *src, fz_weights *weights, int width, int row) -{ - int *contrib = &weights->index[weights->index[row]]; - int len, x; - - contrib++; /* Skip min */ - len = *contrib++; - for (x=width; x > 0; x--) - { - int *min = src; - int val = 0; - int len2 = len; - int *contrib2 = contrib; - - while (len2-- > 0) - { - val += *min * *contrib2++; - min += width; - } - val = (val+(1<<15))>>16; - if (val < 0) - val = 0; - else if (val > 255) - val = 255; - *dst++ = val; - src++; - } -} -#endif - -#ifdef SINGLE_PIXEL_SPECIALS -static void -duplicate_single_pixel(unsigned char *dst, unsigned char *src, int n, int w, int h) -{ - int i; - - for (i = n; i > 0; i--) - *dst++ = *src++; - for (i = (w*h-1)*n; i > 0; i--) - { - *dst = dst[-n]; - dst++; - } -} - -static void -scale_single_row(unsigned char *dst, unsigned char *src, fz_weights *weights, int src_w, int h) -{ - int *contrib = &weights->index[weights->index[0]]; - int min, len, i, j, val, n; - int tmp[FZ_MAX_COLORS]; - - n = weights->n; - /* Scale a single row */ - if (weights->flip) - { - dst += (weights->count-1)*n; - for (i=weights->count; i > 0; i--) - { - min = *contrib++; - len = *contrib++; - min *= n; - for (j = 0; j < n; j++) - tmp[j] = 0; - while (len-- > 0) - { - for (j = 0; j < n; j++) - tmp[j] += src[min++] * *contrib; - contrib++; - } - for (j = 0; j < n; j++) - { - val = (tmp[j]+(1<<7))>>8; - if (val < 0) - val = 0; - else if (val > 255) - val = 255; - *dst++ = val; - } - dst -= 2*n; - } - dst += n * (weights->count+1); - } - else - { - for (i=weights->count; i > 0; i--) - { - min = *contrib++; - len = *contrib++; - min *= n; - for (j = 0; j < n; j++) - tmp[j] = 0; - while (len-- > 0) - { - for (j = 0; j < n; j++) - tmp[j] += src[min++] * *contrib; - contrib++; - } - for (j = 0; j < n; j++) - { - val = (tmp[j]+(1<<7))>>8; - if (val < 0) - val = 0; - else if (val > 255) - val = 255; - *dst++ = val; - } - } - } - /* And then duplicate it h times */ - n *= weights->count; - while (--h > 0) - { - memcpy(dst, dst-n, n); - dst += n; - } -} - -static void -scale_single_col(unsigned char *dst, unsigned char *src, fz_weights *weights, int src_w, int n, int w, int flip_y) -{ - int *contrib = &weights->index[weights->index[0]]; - int min, len, i, j, val; - int tmp[FZ_MAX_COLORS]; - - if (flip_y) - { - src_w = (src_w-1)*n; - w = (w-1)*n; - for (i=weights->count; i > 0; i--) - { - /* Scale the next pixel in the column */ - min = *contrib++; - len = *contrib++; - min = src_w-min*n; - for (j = 0; j < n; j++) - tmp[j] = 0; - while (len-- > 0) - { - for (j = 0; j < n; j++) - tmp[j] += src[src_w-min+j] * *contrib; - contrib++; - } - for (j = 0; j < n; j++) - { - val = (tmp[j]+(1<<7))>>8; - if (val < 0) - val = 0; - else if (val > 255) - val = 255; - *dst++ = val; - } - /* And then duplicate it across the row */ - for (j = w; j > 0; j--) - { - *dst = dst[-n]; - dst++; - } - } - } - else - { - w = (w-1)*n; - for (i=weights->count; i > 0; i--) - { - /* Scale the next pixel in the column */ - min = *contrib++; - len = *contrib++; - min *= n; - for (j = 0; j < n; j++) - tmp[j] = 0; - while (len-- > 0) - { - for (j = 0; j < n; j++) - tmp[j] += src[min++] * *contrib; - contrib++; - } - for (j = 0; j < n; j++) - { - val = (tmp[j]+(1<<7))>>8; - if (val < 0) - val = 0; - else if (val > 255) - val = 255; - *dst++ = val; - } - /* And then duplicate it across the row */ - for (j = w; j > 0; j--) - { - *dst = dst[-n]; - dst++; - } - } - } -} -#endif /* SINGLE_PIXEL_SPECIALS */ - - -fz_pixmap * -fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, float h) -{ - fz_scale_filter *filter = &fz_scale_filter_simple; - fz_weights *contrib_rows = NULL; - fz_weights *contrib_cols = NULL; - fz_pixmap *output = NULL; - int *temp = NULL; - int max_row, temp_span, temp_rows, row; - int dst_w_int, dst_h_int, dst_x_int, dst_y_int; - int flip_x, flip_y; - - DBUG(("Scale: (%d,%d) to (%g,%g) at (%g,%g)\n",src->w,src->h,w,h,x,y)); - - /* Find the destination bbox, width/height, and sub pixel offset, - * allowing for whether we're flipping or not. */ - /* Note that the x and y sub pixel offsets here are different. - * The (x,y) position given describes where the bottom left corner - * of the source image should be mapped to (i.e. where (0,h) in image - * space ends up, not the more logical and sane (0,0)). Also there - * are differences in the way we scale horizontally and vertically. - * When scaling rows horizontally, we always read forwards through - * the source, and store either forwards or in reverse as required. - * When scaling vertically, we always store out forwards, but may - * feed source rows in in a different order. - * - * Consider the image rectangle 'r' to which the image is mapped, - * and the (possibly) larger rectangle 'R', given by expanding 'r' to - * complete pixels. - * - * x can either be r.xmin-R.xmin or R.xmax-r.xmax depending on whether - * the image is x flipped or not. Whatever happens 0 <= x < 1. - * y is always R.ymax - r.ymax. - */ - /* dst_x_int is calculated to be the left of the scaled image, and - * x (the sub_pixel_offset) is the distance in from either the left - * or right pixel expanded edge. */ - flip_x = (w < 0); - if (flip_x) - { - float tmp; - w = -w; - dst_x_int = floor(x-w); - tmp = ceilf(x); - dst_w_int = (int)tmp; - x = tmp - x; - dst_w_int -= dst_x_int; - } - else - { - dst_x_int = floor(x); - x -= (float)dst_x_int; - dst_w_int = (int)ceilf(x + w); - } - flip_y = (h < 0); - /* dst_y_int is calculated to be the bottom of the scaled image, but - * y (the sub pixel offset) has to end up being the value at the top. - */ - if (flip_y) - { - h = -h; - dst_y_int = floor(y-h); - dst_h_int = (int)ceilf(y) - dst_y_int; - } else { - dst_y_int = floor(y); - y += h; - dst_h_int = (int)ceilf(y) - dst_y_int; - } - /* y is the top edge position in floats. We want it to be the - * distance down from the next pixel boundary. */ - y = ceilf(y) - y; - - DBUG(("Result image: (%d,%d) at (%d,%d) (subpix=%g,%g)\n", dst_w_int, dst_h_int, dst_x_int, dst_y_int, x, y)); - - /* Step 1: Calculate the weights for columns and rows */ -#ifdef SINGLE_PIXEL_SPECIALS - if (src->w == 1) - { - contrib_cols = NULL; - } - else -#endif /* SINGLE_PIXEL_SPECIALS */ - { - contrib_cols = make_weights(ctx, src->w, x, w, filter, 0, dst_w_int, src->n, flip_x); - if (!contrib_cols) - goto cleanup; - } -#ifdef SINGLE_PIXEL_SPECIALS - if (src->h == 1) - { - contrib_rows = NULL; - } - else -#endif /* SINGLE_PIXEL_SPECIALS */ - { - contrib_rows = make_weights(ctx, src->h, y, h, filter, 1, dst_h_int, src->n, flip_y); - if (!contrib_rows) - goto cleanup; - } - - assert(!contrib_cols || contrib_cols->count == dst_w_int); - assert(!contrib_rows || contrib_rows->count == dst_h_int); - output = fz_new_pixmap(ctx, src->colorspace, dst_w_int, dst_h_int); - output->x = dst_x_int; - output->y = dst_y_int; - - /* Step 2: Apply the weights */ -#ifdef SINGLE_PIXEL_SPECIALS - if (!contrib_rows) - { - /* Only 1 source pixel high. */ - if (!contrib_cols) - { - /* Only 1 pixel in the entire image! */ - duplicate_single_pixel(output->samples, src->samples, src->n, dst_w_int, dst_h_int); - } - else - { - /* Scale the row once, then copy it. */ - scale_single_row(output->samples, src->samples, contrib_cols, src->w, dst_h_int); - } - } - else if (!contrib_cols) - { - /* Only 1 source pixel wide. Scale the col and duplicate. */ - scale_single_col(output->samples, src->samples, contrib_rows, src->h, src->n, dst_w_int, flip_y); - } - else -#endif /* SINGLE_PIXEL_SPECIALS */ - { - void (*row_scale)(int *dst, unsigned char *src, fz_weights *weights); - - temp_span = contrib_cols->count * src->n; - temp_rows = contrib_rows->max_len; - if (temp_span <= 0 || temp_rows > INT_MAX / temp_span) - goto cleanup; - temp = fz_malloc_array(ctx, temp_span*temp_rows, sizeof(int)); - if (!temp) - goto cleanup; - switch (src->n) - { - default: - row_scale = scale_row_to_temp; - break; - case 1: /* Image mask case */ - row_scale = scale_row_to_temp1; - break; - case 2: /* Greyscale with alpha case */ - row_scale = scale_row_to_temp2; - break; - case 4: /* RGBA */ - row_scale = scale_row_to_temp4; - break; - } - max_row = 0; - for (row = 0; row < contrib_rows->count; row++) - { - /* - Which source rows do we need to have scaled into the - temporary buffer in order to be able to do the final - scale? - */ - int row_index = contrib_rows->index[row]; - int row_min = contrib_rows->index[row_index++]; - int row_len = contrib_rows->index[row_index++]; - while (max_row < row_min+row_len) - { - /* Scale another row */ - assert(max_row < src->h); - DBUG(("scaling row %d to temp\n", max_row)); - (*row_scale)(&temp[temp_span*(max_row % temp_rows)], &src->samples[(flip_y ? (src->h-1-max_row): max_row)*src->w*src->n], contrib_cols); - max_row++; - } - - DBUG(("scaling row %d from temp\n", row)); - scale_row_from_temp(&output->samples[row*output->w*output->n], temp, contrib_rows, temp_span, row); - } - fz_free(ctx, temp); - } - -cleanup: - fz_free(ctx, contrib_rows); - fz_free(ctx, contrib_cols); - return output; -} diff --git a/draw/draw_simple_scale.c b/draw/draw_simple_scale.c index 178b182f..cb419126 100644 --- a/draw/draw_simple_scale.c +++ b/draw/draw_simple_scale.c @@ -218,16 +218,17 @@ typedef struct fz_weights_s fz_weights; struct fz_weights_s { - int flip; - int count; - int max_len; - int n; - int new_line; + int flip; /* true if outputting reversed */ + int count; /* number of output pixels we have records for in this table */ + int max_len; /* Maximum number of weights for any one output pixel */ + int n; /* number of components (src->n) */ + int new_line; /* True if no weights for the current output pixel */ + int patch_l; /* How many output pixels we skip over */ int index[1]; }; static fz_weights * -new_weights(fz_context *ctx, fz_scale_filter *filter, int src_w, float dst_w, int dst_w_i, int n, int flip) +new_weights(fz_context *ctx, fz_scale_filter *filter, int src_w, float dst_w, int patch_w, int n, int flip, int patch_l) { int max_len; fz_weights *weights; @@ -249,26 +250,29 @@ new_weights(fz_context *ctx, fz_scale_filter *filter, int src_w, float dst_w, in max_len = 2 * filter->width; } /* We need the size of the struct, - * plus dst_w*sizeof(int) for the index + * plus patch_w*sizeof(int) for the index * plus (2+max_len)*sizeof(int) for the weights * plus room for an extra set of weights for reordering. */ - weights = fz_malloc(ctx, sizeof(*weights)+(max_len+3)*(dst_w_i+1)*sizeof(int)); + weights = fz_malloc(ctx, sizeof(*weights)+(max_len+3)*(patch_w+1)*sizeof(int)); if (!weights) return NULL; weights->count = -1; weights->max_len = max_len; - weights->index[0] = dst_w_i; + weights->index[0] = patch_w; weights->n = n; + weights->patch_l = patch_l; weights->flip = flip; return weights; } +/* j is destination pixel in the patch_l..patch_l+patch_w range */ static void init_weights(fz_weights *weights, int j) { int index; + j -= weights->patch_l; assert(weights->count == j-1); weights->count++; weights->new_line = 1; @@ -316,6 +320,8 @@ add_weight(fz_weights *weights, int j, int i, fz_scale_filter *filter, DBUG(("add_weight[%d][%d] = %d(%g) dist=%g\n",j,i,weight,f,dist)); + /* Move j from patch_l...patch_l+patch_w range to 0..patch_w range */ + j -= weights->patch_l; if (weights->new_line) { /* New line */ @@ -366,7 +372,7 @@ add_weight(fz_weights *weights, int j, int i, fz_scale_filter *filter, static void reorder_weights(fz_weights *weights, int j, int src_w) { - int idx = weights->index[j]; + int idx = weights->index[j - weights->patch_l]; int min = weights->index[idx++]; int len = weights->index[idx++]; int max = weights->max_len; @@ -413,7 +419,7 @@ check_weights(fz_weights *weights, int j, int w, float x, float wf) int maxidx = 0; int i; - idx = weights->index[j]; + idx = weights->index[j - weights->patch_l]; idx++; /* min */ len = weights->index[idx++]; @@ -443,7 +449,7 @@ check_weights(fz_weights *weights, int j, int w, float x, float wf) } static fz_weights * -make_weights(fz_context *ctx, int src_w, float x, float dst_w, fz_scale_filter *filter, int vertical, int dst_w_int, int n, int flip) +make_weights(fz_context *ctx, int src_w, float x, float dst_w, fz_scale_filter *filter, int vertical, int patch_l, int patch_r, int n, int flip) { fz_weights *weights; float F, G; @@ -463,11 +469,11 @@ make_weights(fz_context *ctx, int src_w, float x, float dst_w, fz_scale_filter * G = src_w / dst_w; } window = filter->width / F; - DBUG(("make_weights src_w=%d x=%g dst_w=%g dst_w_int=%d F=%g window=%g\n", src_w, x, dst_w, dst_w_int, F, window)); - weights = new_weights(ctx, filter, src_w, dst_w, dst_w_int, n, flip); + DBUG(("make_weights src_w=%d x=%g dst_w=%g patch_l=%d patch_r=%d F=%g window=%g\n", src_w, x, dst_w, patch_l, patch_r, F, window)); + weights = new_weights(ctx, filter, src_w, dst_w, patch_r-patch_l, n, flip, patch_l); if (!weights) return NULL; - for (j = 0; j < dst_w_int; j++) + for (j = patch_l; j < patch_r; j++) { /* find the position of the centre of dst[j] in src space */ float centre = (j - x + 0.5f)*src_w/dst_w - 0.5f; @@ -480,7 +486,7 @@ make_weights(fz_context *ctx, int src_w, float x, float dst_w, fz_scale_filter * { add_weight(weights, j, l, filter, x, F, G, src_w, dst_w); } - check_weights(weights, j, dst_w_int, x, dst_w); + check_weights(weights, j, patch_r-patch_l, x, dst_w); if (vertical) { reorder_weights(weights, j, src_w); @@ -1167,7 +1173,7 @@ scale_single_col(unsigned char *dst, unsigned char *src, fz_weights *weights, in #endif /* SINGLE_PIXEL_SPECIALS */ fz_pixmap * -fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, float h) +fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, float h, fz_bbox *clip) { fz_scale_filter *filter = &fz_scale_filter_simple; fz_weights *contrib_rows = NULL; @@ -1177,6 +1183,7 @@ fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, floa int max_row, temp_span, temp_rows, row; int dst_w_int, dst_h_int, dst_x_int, dst_y_int; int flip_x, flip_y; + fz_bbox patch; fz_var(contrib_cols); fz_var(contrib_rows); @@ -1245,21 +1252,38 @@ fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, floa fz_try(ctx) { + /* Step 0: Calculate the patch */ + patch.x0 = 0; + patch.y0 = 0; + patch.x1 = dst_w_int; + patch.y1 = dst_h_int; + if (clip) + { + if (patch.x0 < clip->x0 - dst_x_int) + patch.x0 = clip->x0 - dst_x_int; + if (patch.x1 > clip->x1 - dst_x_int) + patch.x1 = clip->x1 - dst_x_int; + if (patch.y0 < clip->y0 - dst_y_int) + patch.y0 = clip->y0 - dst_y_int; + if (patch.y1 > clip->y1 - dst_y_int) + patch.y1 = clip->y1 - dst_y_int; + } + /* Step 1: Calculate the weights for columns and rows */ #ifdef SINGLE_PIXEL_SPECIALS if (src->w == 1) contrib_cols = NULL; else #endif /* SINGLE_PIXEL_SPECIALS */ - contrib_cols = make_weights(ctx, src->w, x, w, filter, 0, dst_w_int, src->n, flip_x); + contrib_cols = make_weights(ctx, src->w, x, w, filter, 0, patch.x0, patch.x1, src->n, flip_x); #ifdef SINGLE_PIXEL_SPECIALS if (src->h == 1) contrib_rows = NULL; else #endif /* SINGLE_PIXEL_SPECIALS */ - contrib_rows = make_weights(ctx, src->h, y, h, filter, 1, dst_h_int, src->n, flip_y); + contrib_rows = make_weights(ctx, src->h, y, h, filter, 1, patch.y0, patch.y1, src->n, flip_y); - output = fz_new_pixmap(ctx, src->colorspace, dst_w_int, dst_h_int); + output = fz_new_pixmap(ctx, src->colorspace, patch.x1-patch.x0, patch.y1-patch.y0); } fz_catch(ctx) { @@ -1267,8 +1291,8 @@ fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, floa fz_free(ctx, contrib_rows); fz_rethrow(ctx); } - output->x = dst_x_int; - output->y = dst_y_int; + output->x = dst_x_int + patch.x0; + output->y = dst_y_int + patch.y0; /* Step 2: Apply the weights */ #ifdef SINGLE_PIXEL_SPECIALS diff --git a/fitz/fitz.h b/fitz/fitz.h index 3971f24c..49db7dc2 100644 --- a/fitz/fitz.h +++ b/fitz/fitz.h @@ -953,7 +953,7 @@ void fz_invert_pixmap(fz_pixmap *pix); void fz_gamma_pixmap(fz_pixmap *pix, float gamma); unsigned int fz_pixmap_size(fz_pixmap *pix); -fz_pixmap *fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, float h); +fz_pixmap *fz_scale_pixmap(fz_context *ctx, fz_pixmap *src, float x, float y, float w, float h, fz_bbox *clip); void fz_write_pnm(fz_context *ctx, fz_pixmap *pixmap, char *filename); void fz_write_pam(fz_context *ctx, fz_pixmap *pixmap, char *filename, int savealpha); |