Originally committed as revision 14372 to svn://svn.ffmpeg.org/ffmpeg/trunk
Diego Biurrun authored on 2008/07/25 02:09:28... | ... |
@@ -508,7 +508,7 @@ retry: |
508 | 508 |
s->padding_bug_score= 256*256*256*64; |
509 | 509 |
|
510 | 510 |
/* very ugly XVID padding bug detection FIXME/XXX solve this differently |
511 |
- * lets hope this at least works |
|
511 |
+ * Let us hope this at least works. |
|
512 | 512 |
*/ |
513 | 513 |
if( s->resync_marker==0 && s->data_partitioning==0 && s->divx_version==0 |
514 | 514 |
&& s->codec_id==CODEC_ID_MPEG4 && s->vo_type==0) |
... | ... |
@@ -24,7 +24,7 @@ |
24 | 24 |
* Motion estimation template. |
25 | 25 |
*/ |
26 | 26 |
|
27 |
-//lets hope gcc will remove the unused vars ...(gcc 3.2.2 seems to do it ...) |
|
27 |
+//Let us hope gcc will remove the unused vars ...(gcc 3.2.2 seems to do it ...) |
|
28 | 28 |
#define LOAD_COMMON\ |
29 | 29 |
uint32_t av_unused * const score_map= c->score_map;\ |
30 | 30 |
const int av_unused xmin= c->xmin;\ |
... | ... |
@@ -1396,7 +1396,7 @@ static av_always_inline void mpeg_motion_lowres(MpegEncContext *s, |
1396 | 1396 |
linesize = s->current_picture.linesize[0] << field_based; |
1397 | 1397 |
uvlinesize = s->current_picture.linesize[1] << field_based; |
1398 | 1398 |
|
1399 |
- if(s->quarter_sample){ //FIXME obviously not perfect but qpel wont work in lowres anyway |
|
1399 |
+ if(s->quarter_sample){ //FIXME obviously not perfect but qpel will not work in lowres anyway |
|
1400 | 1400 |
motion_x/=2; |
1401 | 1401 |
motion_y/=2; |
1402 | 1402 |
} |
... | ... |
@@ -230,7 +230,7 @@ void av_resample_compensate(AVResampleContext *c, int sample_delta, int compensa |
230 | 230 |
* @param consumed the number of samples of src which have been consumed are returned here |
231 | 231 |
* @param src_size the number of unconsumed samples available |
232 | 232 |
* @param dst_size the amount of space in samples available in dst |
233 |
- * @param update_ctx if this is 0 then the context wont be modified, that way several channels can be resampled with the same context |
|
233 |
+ * @param update_ctx If this is 0 then the context will not be modified, that way several channels can be resampled with the same context. |
|
234 | 234 |
* @return the number of samples written in dst or -1 if an error occurred |
235 | 235 |
*/ |
236 | 236 |
int av_resample(AVResampleContext *c, short *dst, short *src, int *consumed, int src_size, int dst_size, int update_ctx){ |
... | ... |
@@ -155,7 +155,7 @@ static int dc1394_v1_read_header(AVFormatContext *c, AVFormatParameters * ap) |
155 | 155 |
if (dc1394_read_common(c,ap,&fmt,&fps) != 0) |
156 | 156 |
return -1; |
157 | 157 |
|
158 |
- /* Now lets prep the hardware */ |
|
158 |
+ /* Now let us prep the hardware. */ |
|
159 | 159 |
dc1394->handle = dc1394_create_handle(0); /* FIXME: gotta have ap->port */ |
160 | 160 |
if (!dc1394->handle) { |
161 | 161 |
av_log(c, AV_LOG_ERROR, "Can't acquire dc1394 handle on port %d\n", 0 /* ap->port */); |
... | ... |
@@ -248,7 +248,7 @@ static int dc1394_v2_read_header(AVFormatContext *c, AVFormatParameters * ap) |
248 | 248 |
if (dc1394_read_common(c,ap,&fmt,&fps) != 0) |
249 | 249 |
return -1; |
250 | 250 |
|
251 |
- /* Now lets prep the hardware */ |
|
251 |
+ /* Now let us prep the hardware. */ |
|
252 | 252 |
dc1394->d = dc1394_new(); |
253 | 253 |
dc1394_camera_enumerate (dc1394->d, &list); |
254 | 254 |
if ( !list || list->num == 0) { |
... | ... |
@@ -255,7 +255,7 @@ int dv_assemble_frame(DVMuxContext *c, AVStream* st, |
255 | 255 |
av_log(st->codec, AV_LOG_ERROR, "Can't process DV frame #%d. Insufficient video data or severe sync problem.\n", c->frames); |
256 | 256 |
av_fifo_generic_write(&c->audio_data[i], data, data_size, NULL); |
257 | 257 |
|
258 |
- /* Lets see if we've got enough audio for one DV frame */ |
|
258 |
+ /* Let us see if we've got enough audio for one DV frame. */ |
|
259 | 259 |
c->has_audio |= ((reqasize <= av_fifo_size(&c->audio_data[i])) << i); |
260 | 260 |
|
261 | 261 |
break; |
... | ... |
@@ -263,7 +263,7 @@ int dv_assemble_frame(DVMuxContext *c, AVStream* st, |
263 | 263 |
break; |
264 | 264 |
} |
265 | 265 |
|
266 |
- /* Lets see if we have enough data to construct one DV frame */ |
|
266 |
+ /* Let us see if we have enough data to construct one DV frame. */ |
|
267 | 267 |
if (c->has_video == 1 && c->has_audio + 1 == 1<<c->n_ast) { |
268 | 268 |
dv_inject_metadata(c, *frame); |
269 | 269 |
c->has_audio = 0; |
... | ... |
@@ -74,7 +74,7 @@ static inline av_const SoftFloat av_normalize1_sf(SoftFloat a){ |
74 | 74 |
/** |
75 | 75 |
* |
76 | 76 |
* @return will not be more denormalized then a+b, so if either input is |
77 |
- * normalized then the output wont be worse then the other input |
|
77 |
+ * normalized then the output will not be worse then the other input |
|
78 | 78 |
* if both are normalized then the output will be normalized |
79 | 79 |
*/ |
80 | 80 |
static inline av_const SoftFloat av_mul_sf(SoftFloat a, SoftFloat b){ |