84f26f52 |
/*
* Interface to the Android Stagefright library for
* H/W accelerated H.264 decoding
*
* Copyright (C) 2011 Mohamed Naufal
* Copyright (C) 2011 Martin Storsjö
* |
9cff6d9e |
* This file is part of FFmpeg. |
84f26f52 |
* |
9cff6d9e |
* FFmpeg is free software; you can redistribute it and/or |
84f26f52 |
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
* |
9cff6d9e |
* FFmpeg is distributed in the hope that it will be useful, |
84f26f52 |
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public |
9cff6d9e |
* License along with FFmpeg; if not, write to the Free Software |
84f26f52 |
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <binder/ProcessState.h>
#include <media/stagefright/MetaData.h>
#include <media/stagefright/MediaBufferGroup.h>
#include <media/stagefright/MediaDebug.h>
#include <media/stagefright/MediaDefs.h>
#include <media/stagefright/OMXClient.h>
#include <media/stagefright/OMXCodec.h>
#include <utils/List.h>
#include <new> |
7b915a40 |
#include <map> |
84f26f52 |
extern "C" {
#include "avcodec.h"
#include "libavutil/imgutils.h" |
dda8afc3 |
#include "internal.h" |
84f26f52 |
}
#define OMX_QCOM_COLOR_FormatYVU420SemiPlanar 0x7FA30C00
using namespace android;
struct Frame {
status_t status;
size_t size;
int64_t time;
int key;
uint8_t *buffer; |
1d48e88d |
AVFrame *vframe; |
84f26f52 |
};
|
7b915a40 |
struct TimeStamp {
int64_t pts;
int64_t reordered_opaque; |
84f26f52 |
};
class CustomSource;
struct StagefrightContext {
AVCodecContext *avctx;
AVBitStreamFilterContext *bsfc;
uint8_t* orig_extradata;
int orig_extradata_size; |
e7f9aa6f |
sp<MediaSource> *source; |
84f26f52 |
List<Frame*> *in_queue, *out_queue;
pthread_mutex_t in_mutex, out_mutex;
pthread_cond_t condition;
pthread_t decode_thread_id;
Frame *end_frame;
bool source_done; |
a85996d8 |
volatile sig_atomic_t thread_started, thread_exited, stop_decode; |
84f26f52 |
|
1d48e88d |
AVFrame *prev_frame; |
7b915a40 |
std::map<int64_t, TimeStamp> *ts_map;
int64_t frame_index; |
84f26f52 |
uint8_t *dummy_buf;
int dummy_bufsize;
OMXClient *client; |
e7f9aa6f |
sp<MediaSource> *decoder; |
84f26f52 |
const char *decoder_component;
};
class CustomSource : public MediaSource {
public:
CustomSource(AVCodecContext *avctx, sp<MetaData> meta) {
s = (StagefrightContext*)avctx->priv_data;
source_meta = meta;
frame_size = (avctx->width * avctx->height * 3) / 2;
buf_group.add_buffer(new MediaBuffer(frame_size));
}
virtual sp<MetaData> getFormat() {
return source_meta;
}
virtual status_t start(MetaData *params) {
return OK;
}
virtual status_t stop() {
return OK;
}
virtual status_t read(MediaBuffer **buffer,
const MediaSource::ReadOptions *options) {
Frame *frame;
status_t ret;
|
23ea48f2 |
if (s->thread_exited)
return ERROR_END_OF_STREAM; |
84f26f52 |
pthread_mutex_lock(&s->in_mutex);
while (s->in_queue->empty())
pthread_cond_wait(&s->condition, &s->in_mutex);
frame = *s->in_queue->begin();
ret = frame->status;
if (ret == OK) {
ret = buf_group.acquire_buffer(buffer);
if (ret == OK) {
memcpy((*buffer)->data(), frame->buffer, frame->size);
(*buffer)->set_range(0, frame->size);
(*buffer)->meta_data()->clear();
(*buffer)->meta_data()->setInt32(kKeyIsSyncFrame,frame->key);
(*buffer)->meta_data()->setInt64(kKeyTime, frame->time);
} else {
av_log(s->avctx, AV_LOG_ERROR, "Failed to acquire MediaBuffer\n");
}
av_freep(&frame->buffer);
}
s->in_queue->erase(s->in_queue->begin());
pthread_mutex_unlock(&s->in_mutex);
av_freep(&frame);
return ret;
}
private:
MediaBufferGroup buf_group;
sp<MetaData> source_meta;
StagefrightContext *s;
int frame_size;
};
void* decode_thread(void *arg)
{
AVCodecContext *avctx = (AVCodecContext*)arg;
StagefrightContext *s = (StagefrightContext*)avctx->priv_data; |
79393a83 |
const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(avctx->pix_fmt); |
84f26f52 |
Frame* frame;
MediaBuffer *buffer; |
1d48e88d |
int32_t w, h; |
84f26f52 |
int decode_done = 0; |
1d48e88d |
int ret;
int src_linesize[3];
const uint8_t *src_data[3];
int64_t out_frame_index = 0;
|
84f26f52 |
do {
buffer = NULL;
frame = (Frame*)av_mallocz(sizeof(Frame));
if (!frame) {
frame = s->end_frame;
frame->status = AVERROR(ENOMEM);
decode_done = 1;
s->end_frame = NULL; |
1d48e88d |
goto push_frame;
}
frame->status = (*s->decoder)->read(&buffer);
if (frame->status == OK) {
sp<MetaData> outFormat = (*s->decoder)->getFormat();
outFormat->findInt32(kKeyWidth , &w);
outFormat->findInt32(kKeyHeight, &h); |
48ad1f39 |
frame->vframe = av_frame_alloc(); |
1d48e88d |
if (!frame->vframe) {
frame->status = AVERROR(ENOMEM);
decode_done = 1;
buffer->release();
goto push_frame;
} |
d1c96b28 |
ret = ff_get_buffer(avctx, frame->vframe, AV_GET_BUFFER_FLAG_REF); |
1d48e88d |
if (ret < 0) {
frame->status = ret;
decode_done = 1;
buffer->release();
goto push_frame;
}
// The OMX.SEC decoder doesn't signal the modified width/height
if (s->decoder_component && !strncmp(s->decoder_component, "OMX.SEC", 7) &&
(w & 15 || h & 15)) {
if (((w + 15)&~15) * ((h + 15)&~15) * 3/2 == buffer->range_length()) {
w = (w + 15)&~15;
h = (h + 15)&~15;
}
}
if (!avctx->width || !avctx->height || avctx->width > w || avctx->height > h) {
avctx->width = w;
avctx->height = h;
}
|
2343a99c |
src_linesize[0] = av_image_get_linesize(avctx->pix_fmt, w, 0);
src_linesize[1] = av_image_get_linesize(avctx->pix_fmt, w, 1);
src_linesize[2] = av_image_get_linesize(avctx->pix_fmt, w, 2); |
1d48e88d |
src_data[0] = (uint8_t*)buffer->data();
src_data[1] = src_data[0] + src_linesize[0] * h; |
2343a99c |
src_data[2] = src_data[1] + src_linesize[1] * -(-h>>pix_desc->log2_chroma_h); |
1d48e88d |
av_image_copy(frame->vframe->data, frame->vframe->linesize,
src_data, src_linesize,
avctx->pix_fmt, avctx->width, avctx->height);
buffer->meta_data()->findInt64(kKeyTime, &out_frame_index);
if (out_frame_index && s->ts_map->count(out_frame_index) > 0) {
frame->vframe->pts = (*s->ts_map)[out_frame_index].pts;
frame->vframe->reordered_opaque = (*s->ts_map)[out_frame_index].reordered_opaque;
s->ts_map->erase(out_frame_index);
}
buffer->release(); |
84f26f52 |
} else if (frame->status == INFO_FORMAT_CHANGED) {
if (buffer)
buffer->release();
av_free(frame);
continue;
} else {
decode_done = 1;
} |
1d48e88d |
push_frame: |
295f1395 |
while (true) {
pthread_mutex_lock(&s->out_mutex);
if (s->out_queue->size() >= 10) {
pthread_mutex_unlock(&s->out_mutex);
usleep(10000);
continue;
}
break;
} |
84f26f52 |
s->out_queue->push_back(frame);
pthread_mutex_unlock(&s->out_mutex);
} while (!decode_done && !s->stop_decode);
s->thread_exited = true;
return 0;
}
static av_cold int Stagefright_init(AVCodecContext *avctx)
{
StagefrightContext *s = (StagefrightContext*)avctx->priv_data;
sp<MetaData> meta, outFormat;
int32_t colorFormat = 0;
int ret;
if (!avctx->extradata || !avctx->extradata_size || avctx->extradata[0] != 1)
return -1;
s->avctx = avctx;
s->bsfc = av_bitstream_filter_init("h264_mp4toannexb");
if (!s->bsfc) {
av_log(avctx, AV_LOG_ERROR, "Cannot open the h264_mp4toannexb BSF!\n");
return -1;
}
s->orig_extradata_size = avctx->extradata_size;
s->orig_extradata = (uint8_t*) av_mallocz(avctx->extradata_size +
FF_INPUT_BUFFER_PADDING_SIZE);
if (!s->orig_extradata) {
ret = AVERROR(ENOMEM);
goto fail;
}
memcpy(s->orig_extradata, avctx->extradata, avctx->extradata_size);
meta = new MetaData;
if (meta == NULL) {
ret = AVERROR(ENOMEM);
goto fail;
}
meta->setCString(kKeyMIMEType, MEDIA_MIMETYPE_VIDEO_AVC);
meta->setInt32(kKeyWidth, avctx->width);
meta->setInt32(kKeyHeight, avctx->height);
meta->setData(kKeyAVCC, kTypeAVCC, avctx->extradata, avctx->extradata_size);
android::ProcessState::self()->startThreadPool();
|
e7f9aa6f |
s->source = new sp<MediaSource>();
*s->source = new CustomSource(avctx, meta); |
84f26f52 |
s->in_queue = new List<Frame*>;
s->out_queue = new List<Frame*>; |
7b915a40 |
s->ts_map = new std::map<int64_t, TimeStamp>; |
84f26f52 |
s->client = new OMXClient;
s->end_frame = (Frame*)av_mallocz(sizeof(Frame));
if (s->source == NULL || !s->in_queue || !s->out_queue || !s->client || |
7b915a40 |
!s->ts_map || !s->end_frame) { |
84f26f52 |
ret = AVERROR(ENOMEM);
goto fail;
}
if (s->client->connect() != OK) {
av_log(avctx, AV_LOG_ERROR, "Cannot connect OMX client\n");
ret = -1;
goto fail;
}
|
e7f9aa6f |
s->decoder = new sp<MediaSource>();
*s->decoder = OMXCodec::Create(s->client->interface(), meta,
false, *s->source, NULL, |
84f26f52 |
OMXCodec::kClientNeedsFramebuffer); |
e7f9aa6f |
if ((*s->decoder)->start() != OK) { |
84f26f52 |
av_log(avctx, AV_LOG_ERROR, "Cannot start decoder\n");
ret = -1;
s->client->disconnect();
goto fail;
}
|
e7f9aa6f |
outFormat = (*s->decoder)->getFormat(); |
84f26f52 |
outFormat->findInt32(kKeyColorFormat, &colorFormat);
if (colorFormat == OMX_QCOM_COLOR_FormatYVU420SemiPlanar ||
colorFormat == OMX_COLOR_FormatYUV420SemiPlanar) |
ac627b3d |
avctx->pix_fmt = AV_PIX_FMT_NV21; |
2343a99c |
else if (colorFormat == OMX_COLOR_FormatYCbYCr) |
ac627b3d |
avctx->pix_fmt = AV_PIX_FMT_YUYV422; |
2343a99c |
else if (colorFormat == OMX_COLOR_FormatCbYCrY) |
ac627b3d |
avctx->pix_fmt = AV_PIX_FMT_UYVY422; |
84f26f52 |
else |
ac627b3d |
avctx->pix_fmt = AV_PIX_FMT_YUV420P; |
84f26f52 |
outFormat->findCString(kKeyDecoderComponent, &s->decoder_component);
if (s->decoder_component)
s->decoder_component = av_strdup(s->decoder_component);
pthread_mutex_init(&s->in_mutex, NULL);
pthread_mutex_init(&s->out_mutex, NULL);
pthread_cond_init(&s->condition, NULL);
return 0;
fail:
av_bitstream_filter_close(s->bsfc);
av_freep(&s->orig_extradata);
av_freep(&s->end_frame);
delete s->in_queue;
delete s->out_queue; |
7b915a40 |
delete s->ts_map; |
84f26f52 |
delete s->client;
return ret;
}
static int Stagefright_decode_frame(AVCodecContext *avctx, void *data, |
4012cd6c |
int *got_frame, AVPacket *avpkt) |
84f26f52 |
{
StagefrightContext *s = (StagefrightContext*)avctx->priv_data;
Frame *frame;
status_t status;
int orig_size = avpkt->size;
AVPacket pkt = *avpkt; |
1d48e88d |
AVFrame *ret_frame; |
84f26f52 |
|
a85996d8 |
if (!s->thread_started) {
pthread_create(&s->decode_thread_id, NULL, &decode_thread, avctx);
s->thread_started = true;
}
|
84f26f52 |
if (avpkt && avpkt->data) {
av_bitstream_filter_filter(s->bsfc, avctx, NULL, &pkt.data, &pkt.size,
avpkt->data, avpkt->size, avpkt->flags & AV_PKT_FLAG_KEY);
avpkt = &pkt;
}
if (!s->source_done) {
if(!s->dummy_buf) {
s->dummy_buf = (uint8_t*)av_malloc(avpkt->size);
if (!s->dummy_buf)
return AVERROR(ENOMEM);
s->dummy_bufsize = avpkt->size;
memcpy(s->dummy_buf, avpkt->data, avpkt->size);
}
frame = (Frame*)av_mallocz(sizeof(Frame));
if (avpkt->data) {
frame->status = OK; |
09a49e89 |
frame->size = avpkt->size; |
84f26f52 |
frame->key = avpkt->flags & AV_PKT_FLAG_KEY ? 1 : 0;
frame->buffer = (uint8_t*)av_malloc(avpkt->size);
if (!frame->buffer) {
av_freep(&frame);
return AVERROR(ENOMEM);
}
uint8_t *ptr = avpkt->data;
// The OMX.SEC decoder fails without this. |
09a49e89 |
if (avpkt->size == orig_size + avctx->extradata_size) { |
84f26f52 |
ptr += avctx->extradata_size; |
09a49e89 |
frame->size = orig_size;
} |
84f26f52 |
memcpy(frame->buffer, ptr, orig_size); |
f51b7e52 |
if (avpkt == &pkt)
av_free(avpkt->data); |
7b915a40 |
frame->time = ++s->frame_index;
(*s->ts_map)[s->frame_index].pts = avpkt->pts;
(*s->ts_map)[s->frame_index].reordered_opaque = avctx->reordered_opaque; |
84f26f52 |
} else {
frame->status = ERROR_END_OF_STREAM;
s->source_done = true;
}
while (true) {
if (s->thread_exited) {
s->source_done = true;
break;
}
pthread_mutex_lock(&s->in_mutex);
if (s->in_queue->size() >= 10) {
pthread_mutex_unlock(&s->in_mutex);
usleep(10000);
continue;
}
s->in_queue->push_back(frame);
pthread_cond_signal(&s->condition);
pthread_mutex_unlock(&s->in_mutex);
break;
}
}
while (true) {
pthread_mutex_lock(&s->out_mutex);
if (!s->out_queue->empty()) break;
pthread_mutex_unlock(&s->out_mutex);
if (s->source_done) {
usleep(10000);
continue;
} else {
return orig_size;
}
}
frame = *s->out_queue->begin();
s->out_queue->erase(s->out_queue->begin());
pthread_mutex_unlock(&s->out_mutex);
|
1d48e88d |
ret_frame = frame->vframe; |
84f26f52 |
status = frame->status;
av_freep(&frame);
if (status == ERROR_END_OF_STREAM)
return 0;
if (status != OK) {
if (status == AVERROR(ENOMEM))
return status;
av_log(avctx, AV_LOG_ERROR, "Decode failed: %x\n", status);
return -1;
}
|
d1c96b28 |
if (s->prev_frame)
av_frame_free(&s->prev_frame); |
1d48e88d |
s->prev_frame = ret_frame; |
84f26f52 |
|
4012cd6c |
*got_frame = 1; |
1d48e88d |
*(AVFrame*)data = *ret_frame;
return orig_size; |
84f26f52 |
}
static av_cold int Stagefright_close(AVCodecContext *avctx)
{
StagefrightContext *s = (StagefrightContext*)avctx->priv_data;
Frame *frame;
|
a85996d8 |
if (s->thread_started) { |
0ef2e660 |
if (!s->thread_exited) {
s->stop_decode = 1; |
84f26f52 |
|
0ef2e660 |
// Make sure decode_thread() doesn't get stuck
pthread_mutex_lock(&s->out_mutex);
while (!s->out_queue->empty()) {
frame = *s->out_queue->begin();
s->out_queue->erase(s->out_queue->begin()); |
d1c96b28 |
if (frame->vframe)
av_frame_free(&frame->vframe); |
0ef2e660 |
av_freep(&frame);
}
pthread_mutex_unlock(&s->out_mutex);
// Feed a dummy frame prior to signalling EOF.
// This is required to terminate the decoder(OMX.SEC)
// when only one frame is read during stream info detection.
if (s->dummy_buf && (frame = (Frame*)av_mallocz(sizeof(Frame)))) {
frame->status = OK;
frame->size = s->dummy_bufsize;
frame->key = 1;
frame->buffer = s->dummy_buf;
pthread_mutex_lock(&s->in_mutex);
s->in_queue->push_back(frame);
pthread_cond_signal(&s->condition);
pthread_mutex_unlock(&s->in_mutex);
s->dummy_buf = NULL;
} |
295f1395 |
|
84f26f52 |
pthread_mutex_lock(&s->in_mutex); |
0ef2e660 |
s->end_frame->status = ERROR_END_OF_STREAM;
s->in_queue->push_back(s->end_frame); |
84f26f52 |
pthread_cond_signal(&s->condition);
pthread_mutex_unlock(&s->in_mutex); |
0ef2e660 |
s->end_frame = NULL; |
84f26f52 |
}
|
0ef2e660 |
pthread_join(s->decode_thread_id, NULL); |
84f26f52 |
|
d1c96b28 |
if (s->prev_frame)
av_frame_free(&s->prev_frame); |
84f26f52 |
|
0ef2e660 |
s->thread_started = false; |
a85996d8 |
}
|
84f26f52 |
while (!s->in_queue->empty()) {
frame = *s->in_queue->begin();
s->in_queue->erase(s->in_queue->begin());
if (frame->size)
av_freep(&frame->buffer);
av_freep(&frame);
}
while (!s->out_queue->empty()) {
frame = *s->out_queue->begin();
s->out_queue->erase(s->out_queue->begin()); |
d1c96b28 |
if (frame->vframe)
av_frame_free(&frame->vframe); |
84f26f52 |
av_freep(&frame);
}
|
e7f9aa6f |
(*s->decoder)->stop(); |
84f26f52 |
s->client->disconnect();
if (s->decoder_component)
av_freep(&s->decoder_component);
av_freep(&s->dummy_buf);
av_freep(&s->end_frame);
// Reset the extradata back to the original mp4 format, so that
// the next invocation (both when decoding and when called from
// av_find_stream_info) get the original mp4 format extradata.
av_freep(&avctx->extradata);
avctx->extradata = s->orig_extradata;
avctx->extradata_size = s->orig_extradata_size;
delete s->in_queue;
delete s->out_queue; |
7b915a40 |
delete s->ts_map; |
84f26f52 |
delete s->client; |
e7f9aa6f |
delete s->decoder;
delete s->source; |
84f26f52 |
pthread_mutex_destroy(&s->in_mutex);
pthread_mutex_destroy(&s->out_mutex);
pthread_cond_destroy(&s->condition);
av_bitstream_filter_close(s->bsfc);
return 0;
}
AVCodec ff_libstagefright_h264_decoder = {
"libstagefright_h264", |
ea403716 |
NULL_IF_CONFIG_SMALL("libstagefright H.264"), |
84f26f52 |
AVMEDIA_TYPE_VIDEO, |
7a72695c |
AV_CODEC_ID_H264, |
ea403716 |
CODEC_CAP_DELAY,
NULL, //supported_framerates
NULL, //pix_fmts
NULL, //supported_samplerates
NULL, //sample_fmts
NULL, //channel_layouts
0, //max_lowres
NULL, //priv_class
NULL, //profiles |
84f26f52 |
sizeof(StagefrightContext), |
ea403716 |
NULL, //next
NULL, //init_thread_copy
NULL, //update_thread_context
NULL, //defaults
NULL, //init_static_data |
84f26f52 |
Stagefright_init,
NULL, //encode |
ea403716 |
NULL, //encode2 |
84f26f52 |
Stagefright_decode_frame, |
ea403716 |
Stagefright_close, |
84f26f52 |
}; |