... | ... |
@@ -49,8 +49,7 @@ static int aura_decode_frame(AVCodecContext *avctx, |
49 | 49 |
void *data, int *got_frame, |
50 | 50 |
AVPacket *pkt) |
51 | 51 |
{ |
52 |
- AuraDecodeContext *s=avctx->priv_data; |
|
53 |
- |
|
52 |
+ AuraDecodeContext *s = avctx->priv_data; |
|
54 | 53 |
uint8_t *Y, *U, *V; |
55 | 54 |
uint8_t val; |
56 | 55 |
int x, y; |
... | ... |
@@ -68,12 +67,12 @@ static int aura_decode_frame(AVCodecContext *avctx, |
68 | 68 |
/* pixel data starts 48 bytes in, after 3x16-byte tables */ |
69 | 69 |
buf += 48; |
70 | 70 |
|
71 |
- if(s->frame.data[0]) |
|
71 |
+ if (s->frame.data[0]) |
|
72 | 72 |
avctx->release_buffer(avctx, &s->frame); |
73 | 73 |
|
74 | 74 |
s->frame.buffer_hints = FF_BUFFER_HINTS_VALID; |
75 | 75 |
s->frame.reference = 0; |
76 |
- if(ff_get_buffer(avctx, &s->frame) < 0) { |
|
76 |
+ if (ff_get_buffer(avctx, &s->frame) < 0) { |
|
77 | 77 |
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); |
78 | 78 |
return -1; |
79 | 79 |
} |
... | ... |
@@ -85,23 +84,23 @@ static int aura_decode_frame(AVCodecContext *avctx, |
85 | 85 |
/* iterate through each line in the height */ |
86 | 86 |
for (y = 0; y < avctx->height; y++) { |
87 | 87 |
/* reset predictors */ |
88 |
- val = *buf++; |
|
88 |
+ val = *buf++; |
|
89 | 89 |
U[0] = val & 0xF0; |
90 | 90 |
Y[0] = val << 4; |
91 |
- val = *buf++; |
|
91 |
+ val = *buf++; |
|
92 | 92 |
V[0] = val & 0xF0; |
93 | 93 |
Y[1] = Y[0] + delta_table[val & 0xF]; |
94 |
- Y += 2; U++; V++; |
|
94 |
+ Y += 2; U++; V++; |
|
95 | 95 |
|
96 | 96 |
/* iterate through the remaining pixel groups (4 pixels/group) */ |
97 | 97 |
for (x = 1; x < (avctx->width >> 1); x++) { |
98 |
- val = *buf++; |
|
98 |
+ val = *buf++; |
|
99 | 99 |
U[0] = U[-1] + delta_table[val >> 4]; |
100 | 100 |
Y[0] = Y[-1] + delta_table[val & 0xF]; |
101 |
- val = *buf++; |
|
101 |
+ val = *buf++; |
|
102 | 102 |
V[0] = V[-1] + delta_table[val >> 4]; |
103 | 103 |
Y[1] = Y[ 0] + delta_table[val & 0xF]; |
104 |
- Y += 2; U++; V++; |
|
104 |
+ Y += 2; U++; V++; |
|
105 | 105 |
} |
106 | 106 |
Y += s->frame.linesize[0] - avctx->width; |
107 | 107 |
U += s->frame.linesize[1] - (avctx->width >> 1); |
... | ... |
@@ -109,7 +108,7 @@ static int aura_decode_frame(AVCodecContext *avctx, |
109 | 109 |
} |
110 | 110 |
|
111 | 111 |
*got_frame = 1; |
112 |
- *(AVFrame*)data= s->frame; |
|
112 |
+ *(AVFrame*)data = s->frame; |
|
113 | 113 |
|
114 | 114 |
return pkt->size; |
115 | 115 |
} |