131 |
image_swap(IMAGE * image1, |
image_swap(IMAGE * image1, |
132 |
IMAGE * image2) |
IMAGE * image2) |
133 |
{ |
{ |
134 |
uint8_t *tmp; |
SWAP(uint8_t*, image1->y, image2->y); |
135 |
|
SWAP(uint8_t*, image1->u, image2->u); |
136 |
tmp = image1->y; |
SWAP(uint8_t*, image1->v, image2->v); |
|
image1->y = image2->y; |
|
|
image2->y = tmp; |
|
|
|
|
|
tmp = image1->u; |
|
|
image1->u = image2->u; |
|
|
image2->u = tmp; |
|
|
|
|
|
tmp = image1->v; |
|
|
image1->v = image2->v; |
|
|
image2->v = tmp; |
|
137 |
} |
} |
138 |
|
|
139 |
|
|
501 |
#undef IMG_V |
#undef IMG_V |
502 |
} |
} |
503 |
|
|
504 |
DPRINTF(DPRINTF_DEBUG,"chroma_optimized_pixels = %i/%i", pixels, width*height/4); |
DPRINTF(XVID_DEBUG_DEBUG,"chroma_optimized_pixels = %i/%i", pixels, width*height/4); |
505 |
} |
} |
506 |
|
|
507 |
|
|
1019 |
|
|
1020 |
for (y = 0; y < height; y++) |
for (y = 0; y < height; y++) |
1021 |
for (x = 0; x < width; x++) |
for (x = 0; x < width; x++) |
1022 |
sum += ABS(img1->y[x + y * stride] - img2->y[x + y * stride]); |
sum += abs(img1->y[x + y * stride] - img2->y[x + y * stride]); |
1023 |
|
|
1024 |
for (y = 0; y < height2; y++) |
for (y = 0; y < height2; y++) |
1025 |
for (x = 0; x < width2; x++) |
for (x = 0; x < width2; x++) |
1026 |
sum += ABS(img1->u[x + y * stride2] - img2->u[x + y * stride2]); |
sum += abs(img1->u[x + y * stride2] - img2->u[x + y * stride2]); |
1027 |
|
|
1028 |
for (y = 0; y < height2; y++) |
for (y = 0; y < height2; y++) |
1029 |
for (x = 0; x < width2; x++) |
for (x = 0; x < width2; x++) |
1030 |
sum += ABS(img1->v[x + y * stride2] - img2->v[x + y * stride2]); |
sum += abs(img1->v[x + y * stride2] - img2->v[x + y * stride2]); |
1031 |
|
|
1032 |
return (float) sum / (width * height * 3 / 2); |
return (float) sum / (width * height * 3 / 2); |
1033 |
} |
} |