mirror of
https://github.com/xenia-project/FFmpeg.git
synced 2024-11-24 20:19:55 +00:00
2ba8301769
Makes no difference for gcc but at least icc can put them in .rodata then. Originally committed as revision 19789 to svn://svn.ffmpeg.org/ffmpeg/trunk
510 lines
14 KiB
C
510 lines
14 KiB
C
/*
|
|
* PNM image format
|
|
* Copyright (c) 2002, 2003 Fabrice Bellard
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
#include "avcodec.h"
|
|
#include "bytestream.h"
|
|
#include "pnm.h"
|
|
|
|
|
|
static av_cold int common_init(AVCodecContext *avctx){
|
|
PNMContext *s = avctx->priv_data;
|
|
|
|
avcodec_get_frame_defaults((AVFrame*)&s->picture);
|
|
avctx->coded_frame= (AVFrame*)&s->picture;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pnm_decode_frame(AVCodecContext *avctx,
|
|
void *data, int *data_size,
|
|
AVPacket *avpkt)
|
|
{
|
|
const uint8_t *buf = avpkt->data;
|
|
int buf_size = avpkt->size;
|
|
PNMContext * const s = avctx->priv_data;
|
|
AVFrame *picture = data;
|
|
AVFrame * const p= (AVFrame*)&s->picture;
|
|
int i, n, linesize, h, upgrade = 0;
|
|
unsigned char *ptr;
|
|
|
|
s->bytestream_start=
|
|
s->bytestream= buf;
|
|
s->bytestream_end= buf + buf_size;
|
|
|
|
if(ff_pnm_decode_header(avctx, s) < 0)
|
|
return -1;
|
|
|
|
if(p->data[0])
|
|
avctx->release_buffer(avctx, p);
|
|
|
|
p->reference= 0;
|
|
if(avctx->get_buffer(avctx, p) < 0){
|
|
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
|
return -1;
|
|
}
|
|
p->pict_type= FF_I_TYPE;
|
|
p->key_frame= 1;
|
|
|
|
switch(avctx->pix_fmt) {
|
|
default:
|
|
return -1;
|
|
case PIX_FMT_RGB48BE:
|
|
n = avctx->width * 6;
|
|
goto do_read;
|
|
case PIX_FMT_RGB24:
|
|
n = avctx->width * 3;
|
|
goto do_read;
|
|
case PIX_FMT_GRAY8:
|
|
n = avctx->width;
|
|
if (s->maxval < 255)
|
|
upgrade = 1;
|
|
goto do_read;
|
|
case PIX_FMT_GRAY16BE:
|
|
case PIX_FMT_GRAY16LE:
|
|
n = avctx->width * 2;
|
|
if (s->maxval < 65535)
|
|
upgrade = 2;
|
|
goto do_read;
|
|
case PIX_FMT_MONOWHITE:
|
|
case PIX_FMT_MONOBLACK:
|
|
n = (avctx->width + 7) >> 3;
|
|
do_read:
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if(s->bytestream + n*avctx->height > s->bytestream_end)
|
|
return -1;
|
|
for(i = 0; i < avctx->height; i++) {
|
|
if (!upgrade)
|
|
memcpy(ptr, s->bytestream, n);
|
|
else if (upgrade == 1) {
|
|
unsigned int j, f = (255*128 + s->maxval/2) / s->maxval;
|
|
for (j=0; j<n; j++)
|
|
ptr[j] = (s->bytestream[j] * f + 64) >> 7;
|
|
} else if (upgrade == 2) {
|
|
unsigned int j, v, f = (65535*32768 + s->maxval/2) / s->maxval;
|
|
for (j=0; j<n/2; j++) {
|
|
v = be2me_16(((uint16_t *)s->bytestream)[j]);
|
|
((uint16_t *)ptr)[j] = (v * f + 16384) >> 15;
|
|
}
|
|
}
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
break;
|
|
case PIX_FMT_YUV420P:
|
|
{
|
|
unsigned char *ptr1, *ptr2;
|
|
|
|
n = avctx->width;
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if(s->bytestream + n*avctx->height*3/2 > s->bytestream_end)
|
|
return -1;
|
|
for(i = 0; i < avctx->height; i++) {
|
|
memcpy(ptr, s->bytestream, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
ptr1 = p->data[1];
|
|
ptr2 = p->data[2];
|
|
n >>= 1;
|
|
h = avctx->height >> 1;
|
|
for(i = 0; i < h; i++) {
|
|
memcpy(ptr1, s->bytestream, n);
|
|
s->bytestream += n;
|
|
memcpy(ptr2, s->bytestream, n);
|
|
s->bytestream += n;
|
|
ptr1 += p->linesize[1];
|
|
ptr2 += p->linesize[2];
|
|
}
|
|
}
|
|
break;
|
|
case PIX_FMT_RGB32:
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if(s->bytestream + avctx->width*avctx->height*4 > s->bytestream_end)
|
|
return -1;
|
|
for(i = 0; i < avctx->height; i++) {
|
|
int j, r, g, b, a;
|
|
|
|
for(j = 0;j < avctx->width; j++) {
|
|
r = *s->bytestream++;
|
|
g = *s->bytestream++;
|
|
b = *s->bytestream++;
|
|
a = *s->bytestream++;
|
|
((uint32_t *)ptr)[j] = (a << 24) | (r << 16) | (g << 8) | b;
|
|
}
|
|
ptr += linesize;
|
|
}
|
|
break;
|
|
}
|
|
*picture= *(AVFrame*)&s->picture;
|
|
*data_size = sizeof(AVPicture);
|
|
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
static int pnm_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){
|
|
PNMContext *s = avctx->priv_data;
|
|
AVFrame *pict = data;
|
|
AVFrame * const p= (AVFrame*)&s->picture;
|
|
int i, h, h1, c, n, linesize;
|
|
uint8_t *ptr, *ptr1, *ptr2;
|
|
|
|
if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){
|
|
av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
|
|
return -1;
|
|
}
|
|
|
|
*p = *pict;
|
|
p->pict_type= FF_I_TYPE;
|
|
p->key_frame= 1;
|
|
|
|
s->bytestream_start=
|
|
s->bytestream= outbuf;
|
|
s->bytestream_end= outbuf+buf_size;
|
|
|
|
h = avctx->height;
|
|
h1 = h;
|
|
switch(avctx->pix_fmt) {
|
|
case PIX_FMT_MONOWHITE:
|
|
c = '4';
|
|
n = (avctx->width + 7) >> 3;
|
|
break;
|
|
case PIX_FMT_GRAY8:
|
|
c = '5';
|
|
n = avctx->width;
|
|
break;
|
|
case PIX_FMT_GRAY16BE:
|
|
c = '5';
|
|
n = avctx->width * 2;
|
|
break;
|
|
case PIX_FMT_RGB24:
|
|
c = '6';
|
|
n = avctx->width * 3;
|
|
break;
|
|
case PIX_FMT_RGB48BE:
|
|
c = '6';
|
|
n = avctx->width * 6;
|
|
break;
|
|
case PIX_FMT_YUV420P:
|
|
c = '5';
|
|
n = avctx->width;
|
|
h1 = (h * 3) / 2;
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
snprintf(s->bytestream, s->bytestream_end - s->bytestream,
|
|
"P%c\n%d %d\n",
|
|
c, avctx->width, h1);
|
|
s->bytestream += strlen(s->bytestream);
|
|
if (avctx->pix_fmt != PIX_FMT_MONOWHITE) {
|
|
snprintf(s->bytestream, s->bytestream_end - s->bytestream,
|
|
"%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE && avctx->pix_fmt != PIX_FMT_RGB48BE) ? 255 : 65535);
|
|
s->bytestream += strlen(s->bytestream);
|
|
}
|
|
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
for(i=0;i<h;i++) {
|
|
memcpy(s->bytestream, ptr, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
|
|
if (avctx->pix_fmt == PIX_FMT_YUV420P) {
|
|
h >>= 1;
|
|
n >>= 1;
|
|
ptr1 = p->data[1];
|
|
ptr2 = p->data[2];
|
|
for(i=0;i<h;i++) {
|
|
memcpy(s->bytestream, ptr1, n);
|
|
s->bytestream += n;
|
|
memcpy(s->bytestream, ptr2, n);
|
|
s->bytestream += n;
|
|
ptr1 += p->linesize[1];
|
|
ptr2 += p->linesize[2];
|
|
}
|
|
}
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
static int pam_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){
|
|
PNMContext *s = avctx->priv_data;
|
|
AVFrame *pict = data;
|
|
AVFrame * const p= (AVFrame*)&s->picture;
|
|
int i, h, w, n, linesize, depth, maxval;
|
|
const char *tuple_type;
|
|
uint8_t *ptr;
|
|
|
|
if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){
|
|
av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
|
|
return -1;
|
|
}
|
|
|
|
*p = *pict;
|
|
p->pict_type= FF_I_TYPE;
|
|
p->key_frame= 1;
|
|
|
|
s->bytestream_start=
|
|
s->bytestream= outbuf;
|
|
s->bytestream_end= outbuf+buf_size;
|
|
|
|
h = avctx->height;
|
|
w = avctx->width;
|
|
switch(avctx->pix_fmt) {
|
|
case PIX_FMT_MONOWHITE:
|
|
n = (w + 7) >> 3;
|
|
depth = 1;
|
|
maxval = 1;
|
|
tuple_type = "BLACKANDWHITE";
|
|
break;
|
|
case PIX_FMT_GRAY8:
|
|
n = w;
|
|
depth = 1;
|
|
maxval = 255;
|
|
tuple_type = "GRAYSCALE";
|
|
break;
|
|
case PIX_FMT_RGB24:
|
|
n = w * 3;
|
|
depth = 3;
|
|
maxval = 255;
|
|
tuple_type = "RGB";
|
|
break;
|
|
case PIX_FMT_RGB32:
|
|
n = w * 4;
|
|
depth = 4;
|
|
maxval = 255;
|
|
tuple_type = "RGB_ALPHA";
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
snprintf(s->bytestream, s->bytestream_end - s->bytestream,
|
|
"P7\nWIDTH %d\nHEIGHT %d\nDEPTH %d\nMAXVAL %d\nTUPLETYPE %s\nENDHDR\n",
|
|
w, h, depth, maxval, tuple_type);
|
|
s->bytestream += strlen(s->bytestream);
|
|
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
|
|
if (avctx->pix_fmt == PIX_FMT_RGB32) {
|
|
int j;
|
|
unsigned int v;
|
|
|
|
for(i=0;i<h;i++) {
|
|
for(j=0;j<w;j++) {
|
|
v = ((uint32_t *)ptr)[j];
|
|
bytestream_put_be24(&s->bytestream, v);
|
|
*s->bytestream++ = v >> 24;
|
|
}
|
|
ptr += linesize;
|
|
}
|
|
} else {
|
|
for(i=0;i<h;i++) {
|
|
memcpy(s->bytestream, ptr, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
}
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
#if 0
|
|
static int pnm_probe(AVProbeData *pd)
|
|
{
|
|
const char *p = pd->buf;
|
|
if (pd->buf_size >= 8 &&
|
|
p[0] == 'P' &&
|
|
p[1] >= '4' && p[1] <= '6' &&
|
|
pnm_space(p[2]) )
|
|
return AVPROBE_SCORE_MAX - 1; /* to permit pgmyuv probe */
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
static int pgmyuv_probe(AVProbeData *pd)
|
|
{
|
|
if (match_ext(pd->filename, "pgmyuv"))
|
|
return AVPROBE_SCORE_MAX;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
static int pam_probe(AVProbeData *pd)
|
|
{
|
|
const char *p = pd->buf;
|
|
if (pd->buf_size >= 8 &&
|
|
p[0] == 'P' &&
|
|
p[1] == '7' &&
|
|
p[2] == '\n')
|
|
return AVPROBE_SCORE_MAX;
|
|
else
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
|
|
#if CONFIG_PGM_DECODER
|
|
AVCodec pgm_decoder = {
|
|
"pgm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PGM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
NULL,
|
|
NULL,
|
|
pnm_decode_frame,
|
|
CODEC_CAP_DR1,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"),
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PGM_ENCODER
|
|
AVCodec pgm_encoder = {
|
|
"pgm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PGM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"),
|
|
};
|
|
#endif // CONFIG_PGM_ENCODER
|
|
|
|
#if CONFIG_PGMYUV_DECODER
|
|
AVCodec pgmyuv_decoder = {
|
|
"pgmyuv",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PGMYUV,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
NULL,
|
|
NULL,
|
|
pnm_decode_frame,
|
|
CODEC_CAP_DR1,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"),
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PGMYUV_ENCODER
|
|
AVCodec pgmyuv_encoder = {
|
|
"pgmyuv",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PGMYUV,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"),
|
|
};
|
|
#endif // CONFIG_PGMYUV_ENCODER
|
|
|
|
#if CONFIG_PPM_DECODER
|
|
AVCodec ppm_decoder = {
|
|
"ppm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PPM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
NULL,
|
|
NULL,
|
|
pnm_decode_frame,
|
|
CODEC_CAP_DR1,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB48BE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"),
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PPM_ENCODER
|
|
AVCodec ppm_encoder = {
|
|
"ppm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PPM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB48BE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"),
|
|
};
|
|
#endif // CONFIG_PPM_ENCODER
|
|
|
|
#if CONFIG_PBM_DECODER
|
|
AVCodec pbm_decoder = {
|
|
"pbm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PBM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
NULL,
|
|
NULL,
|
|
pnm_decode_frame,
|
|
CODEC_CAP_DR1,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_MONOWHITE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"),
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PBM_ENCODER
|
|
AVCodec pbm_encoder = {
|
|
"pbm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PBM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_MONOWHITE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"),
|
|
};
|
|
#endif // CONFIG_PBM_ENCODER
|
|
|
|
#if CONFIG_PAM_DECODER
|
|
AVCodec pam_decoder = {
|
|
"pam",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PAM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
NULL,
|
|
NULL,
|
|
pnm_decode_frame,
|
|
CODEC_CAP_DR1,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"),
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PAM_ENCODER
|
|
AVCodec pam_encoder = {
|
|
"pam",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PAM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pam_encode_frame,
|
|
.pix_fmts= (const enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, PIX_FMT_NONE},
|
|
.long_name= NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"),
|
|
};
|
|
#endif // CONFIG_PAM_ENCODER
|