mirror of
https://github.com/jellyfin/jellyfin-ffmpeg.git
synced 2024-11-23 13:59:45 +00:00
a37e362e0f
Signed-off-by: nyanmisaka <nst799610810@gmail.com>
464 lines
17 KiB
C
464 lines
17 KiB
C
/*
|
|
* Copyright (c) 2012 Fredrik Mellbin
|
|
* Copyright (c) 2013 Clément Bœsch
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/pixdesc.h"
|
|
#include "libavutil/timestamp.h"
|
|
#include "avfilter.h"
|
|
#include "filters.h"
|
|
#include "internal.h"
|
|
|
|
#define INPUT_MAIN 0
|
|
#define INPUT_CLEANSRC 1
|
|
|
|
struct qitem {
|
|
AVFrame *frame;
|
|
int64_t maxbdiff;
|
|
int64_t totdiff;
|
|
};
|
|
|
|
typedef struct DecimateContext {
|
|
const AVClass *class;
|
|
struct qitem *queue; ///< window of cycle frames and the associated data diff
|
|
int fid; ///< current frame id in the queue
|
|
int filled; ///< 1 if the queue is filled, 0 otherwise
|
|
AVFrame *last; ///< last frame from the previous queue
|
|
AVFrame **clean_src; ///< frame queue for the clean source
|
|
int got_frame[2]; ///< frame request flag for each input stream
|
|
int64_t last_pts; ///< last output timestamp
|
|
int64_t last_duration; ///< last output duration
|
|
int64_t start_pts; ///< base for output timestamps
|
|
uint32_t eof; ///< bitmask for end of stream
|
|
int hsub, vsub; ///< chroma subsampling values
|
|
int depth;
|
|
int nxblocks, nyblocks;
|
|
int bdiffsize;
|
|
int64_t *bdiffs;
|
|
AVRational in_tb; // input time-base
|
|
AVRational nondec_tb; // non-decimated time-base
|
|
AVRational dec_tb; // decimated time-base
|
|
|
|
/* options */
|
|
int cycle;
|
|
double dupthresh_flt;
|
|
double scthresh_flt;
|
|
int64_t dupthresh;
|
|
int64_t scthresh;
|
|
int blockx, blocky;
|
|
int ppsrc;
|
|
int chroma;
|
|
int mixed;
|
|
} DecimateContext;
|
|
|
|
#define OFFSET(x) offsetof(DecimateContext, x)
|
|
#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
|
|
|
|
static const AVOption decimate_options[] = {
|
|
{ "cycle", "set the number of frame from which one will be dropped", OFFSET(cycle), AV_OPT_TYPE_INT, {.i64 = 5}, 2, 25, FLAGS },
|
|
{ "dupthresh", "set duplicate threshold", OFFSET(dupthresh_flt), AV_OPT_TYPE_DOUBLE, {.dbl = 1.1}, 0, 100, FLAGS },
|
|
{ "scthresh", "set scene change threshold", OFFSET(scthresh_flt), AV_OPT_TYPE_DOUBLE, {.dbl = 15.0}, 0, 100, FLAGS },
|
|
{ "blockx", "set the size of the x-axis blocks used during metric calculations", OFFSET(blockx), AV_OPT_TYPE_INT, {.i64 = 32}, 4, 1<<9, FLAGS },
|
|
{ "blocky", "set the size of the y-axis blocks used during metric calculations", OFFSET(blocky), AV_OPT_TYPE_INT, {.i64 = 32}, 4, 1<<9, FLAGS },
|
|
{ "ppsrc", "mark main input as a pre-processed input and activate clean source input stream", OFFSET(ppsrc), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS },
|
|
{ "chroma", "set whether or not chroma is considered in the metric calculations", OFFSET(chroma), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS },
|
|
{ "mixed", "set whether or not the input only partially contains content to be decimated", OFFSET(mixed), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS },
|
|
{ NULL }
|
|
};
|
|
|
|
AVFILTER_DEFINE_CLASS(decimate);
|
|
|
|
static void calc_diffs(const DecimateContext *dm, struct qitem *q,
|
|
const AVFrame *f1, const AVFrame *f2)
|
|
{
|
|
int64_t maxdiff = -1;
|
|
int64_t *bdiffs = dm->bdiffs;
|
|
int plane, i, j;
|
|
|
|
memset(bdiffs, 0, dm->bdiffsize * sizeof(*bdiffs));
|
|
|
|
for (plane = 0; plane < (dm->chroma && f1->data[2] ? 3 : 1); plane++) {
|
|
int x, y, xl;
|
|
const int linesize1 = f1->linesize[plane];
|
|
const int linesize2 = f2->linesize[plane];
|
|
const uint8_t *f1p = f1->data[plane];
|
|
const uint8_t *f2p = f2->data[plane];
|
|
int width = plane ? AV_CEIL_RSHIFT(f1->width, dm->hsub) : f1->width;
|
|
int height = plane ? AV_CEIL_RSHIFT(f1->height, dm->vsub) : f1->height;
|
|
int hblockx = dm->blockx / 2;
|
|
int hblocky = dm->blocky / 2;
|
|
|
|
if (plane) {
|
|
hblockx >>= dm->hsub;
|
|
hblocky >>= dm->vsub;
|
|
}
|
|
|
|
for (y = 0; y < height; y++) {
|
|
int ydest = y / hblocky;
|
|
int xdest = 0;
|
|
|
|
#define CALC_DIFF(nbits) do { \
|
|
for (x = 0; x < width; x += hblockx) { \
|
|
int64_t acc = 0; \
|
|
int m = FFMIN(width, x + hblockx); \
|
|
for (xl = x; xl < m; xl++) \
|
|
acc += abs(((const uint##nbits##_t *)f1p)[xl] - \
|
|
((const uint##nbits##_t *)f2p)[xl]); \
|
|
bdiffs[ydest * dm->nxblocks + xdest] += acc; \
|
|
xdest++; \
|
|
} \
|
|
} while (0)
|
|
if (dm->depth == 8) CALC_DIFF(8);
|
|
else CALC_DIFF(16);
|
|
|
|
f1p += linesize1;
|
|
f2p += linesize2;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < dm->nyblocks - 1; i++) {
|
|
for (j = 0; j < dm->nxblocks - 1; j++) {
|
|
int64_t tmp = bdiffs[ i * dm->nxblocks + j ]
|
|
+ bdiffs[ i * dm->nxblocks + j + 1]
|
|
+ bdiffs[(i + 1) * dm->nxblocks + j ]
|
|
+ bdiffs[(i + 1) * dm->nxblocks + j + 1];
|
|
if (tmp > maxdiff)
|
|
maxdiff = tmp;
|
|
}
|
|
}
|
|
|
|
q->totdiff = 0;
|
|
for (i = 0; i < dm->bdiffsize; i++)
|
|
q->totdiff += bdiffs[i];
|
|
q->maxbdiff = maxdiff;
|
|
}
|
|
|
|
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
|
|
{
|
|
int scpos = -1, duppos = -1;
|
|
int drop = INT_MIN, i, lowest = 0, ret;
|
|
AVFilterContext *ctx = inlink->dst;
|
|
AVFilterLink *outlink = ctx->outputs[0];
|
|
DecimateContext *dm = ctx->priv;
|
|
AVFrame *prv;
|
|
|
|
/* update frames queue(s) */
|
|
if (FF_INLINK_IDX(inlink) == INPUT_MAIN) {
|
|
dm->queue[dm->fid].frame = in;
|
|
dm->got_frame[INPUT_MAIN] = 1;
|
|
} else {
|
|
dm->clean_src[dm->fid] = in;
|
|
dm->got_frame[INPUT_CLEANSRC] = 1;
|
|
}
|
|
if (!dm->got_frame[INPUT_MAIN] || (dm->ppsrc && !dm->got_frame[INPUT_CLEANSRC]))
|
|
return 0;
|
|
dm->got_frame[INPUT_MAIN] = dm->got_frame[INPUT_CLEANSRC] = 0;
|
|
|
|
if (dm->ppsrc)
|
|
in = dm->queue[dm->fid].frame;
|
|
|
|
if (in) {
|
|
/* update frame metrics */
|
|
prv = dm->fid ? dm->queue[dm->fid - 1].frame : dm->last;
|
|
if (!prv) {
|
|
dm->queue[dm->fid].maxbdiff = INT64_MAX;
|
|
dm->queue[dm->fid].totdiff = INT64_MAX;
|
|
} else {
|
|
calc_diffs(dm, &dm->queue[dm->fid], prv, in);
|
|
}
|
|
if (++dm->fid != dm->cycle)
|
|
return 0;
|
|
av_frame_free(&dm->last);
|
|
dm->last = av_frame_clone(in);
|
|
dm->fid = 0;
|
|
|
|
/* we have a complete cycle, select the frame to drop */
|
|
lowest = 0;
|
|
for (i = 0; i < dm->cycle; i++) {
|
|
if (dm->queue[i].totdiff > dm->scthresh)
|
|
scpos = i;
|
|
if (dm->queue[i].maxbdiff < dm->queue[lowest].maxbdiff)
|
|
lowest = i;
|
|
}
|
|
if (dm->queue[lowest].maxbdiff < dm->dupthresh)
|
|
duppos = lowest;
|
|
|
|
if (dm->mixed && duppos < 0) {
|
|
drop = -1; // no drop if mixed content + no frame in cycle below threshold
|
|
} else {
|
|
drop = scpos >= 0 && duppos < 0 ? scpos : lowest;
|
|
}
|
|
}
|
|
|
|
/* metrics debug */
|
|
if (av_log_get_level() >= AV_LOG_DEBUG) {
|
|
av_log(ctx, AV_LOG_DEBUG, "1/%d frame drop:\n", dm->cycle);
|
|
for (i = 0; i < dm->cycle && dm->queue[i].frame; i++) {
|
|
av_log(ctx, AV_LOG_DEBUG," #%d: totdiff=%08"PRIx64" maxbdiff=%08"PRIx64"%s%s%s%s\n",
|
|
i + 1, dm->queue[i].totdiff, dm->queue[i].maxbdiff,
|
|
i == scpos ? " sc" : "",
|
|
i == duppos ? " dup" : "",
|
|
i == lowest ? " lowest" : "",
|
|
i == drop ? " [DROP]" : "");
|
|
}
|
|
}
|
|
|
|
/* push all frames except the drop */
|
|
ret = 0;
|
|
for (i = 0; i < dm->cycle && dm->queue[i].frame; i++) {
|
|
if (i == drop) {
|
|
if (dm->ppsrc)
|
|
av_frame_free(&dm->clean_src[i]);
|
|
av_frame_free(&dm->queue[i].frame);
|
|
} else {
|
|
AVFrame *frame = dm->queue[i].frame;
|
|
dm->queue[i].frame = NULL;
|
|
if (frame->pts != AV_NOPTS_VALUE && dm->start_pts == AV_NOPTS_VALUE)
|
|
dm->start_pts = av_rescale_q(frame->pts, dm->in_tb, outlink->time_base);
|
|
|
|
if (dm->ppsrc) {
|
|
av_frame_free(&frame);
|
|
frame = dm->clean_src[i];
|
|
if (!frame)
|
|
continue;
|
|
dm->clean_src[i] = NULL;
|
|
}
|
|
|
|
frame->pts = dm->last_duration ? dm->last_pts + dm->last_duration :
|
|
(dm->start_pts == AV_NOPTS_VALUE ? 0 : dm->start_pts);
|
|
frame->duration = dm->mixed ? av_div_q(drop < 0 ? dm->nondec_tb : dm->dec_tb, outlink->time_base).num : 1;
|
|
dm->last_duration = frame->duration;
|
|
dm->last_pts = frame->pts;
|
|
ret = ff_filter_frame(outlink, frame);
|
|
if (ret < 0)
|
|
break;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int activate(AVFilterContext *ctx)
|
|
{
|
|
DecimateContext *dm = ctx->priv;
|
|
AVFrame *frame = NULL;
|
|
int ret = 0, status;
|
|
int64_t pts;
|
|
|
|
FF_FILTER_FORWARD_STATUS_BACK_ALL(ctx->outputs[0], ctx);
|
|
|
|
if ((dm->got_frame[INPUT_MAIN] == 0) && !(dm->eof & (1 << INPUT_MAIN)) &&
|
|
(ret = ff_inlink_consume_frame(ctx->inputs[INPUT_MAIN], &frame)) > 0) {
|
|
ret = filter_frame(ctx->inputs[INPUT_MAIN], frame);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
if (ret < 0)
|
|
return ret;
|
|
if (dm->ppsrc &&
|
|
(dm->got_frame[INPUT_CLEANSRC] == 0) && !(dm->eof & (1 << INPUT_CLEANSRC)) &&
|
|
(ret = ff_inlink_consume_frame(ctx->inputs[INPUT_CLEANSRC], &frame)) > 0) {
|
|
ret = filter_frame(ctx->inputs[INPUT_CLEANSRC], frame);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
if (ret < 0) {
|
|
return ret;
|
|
} else if (dm->eof == ((1 << INPUT_MAIN) | (dm->ppsrc << INPUT_CLEANSRC))) {
|
|
ff_outlink_set_status(ctx->outputs[0], AVERROR_EOF, dm->last_pts);
|
|
return 0;
|
|
} else if (!(dm->eof & (1 << INPUT_MAIN)) && ff_inlink_acknowledge_status(ctx->inputs[INPUT_MAIN], &status, &pts)) {
|
|
if (status == AVERROR_EOF) { // flushing
|
|
dm->eof |= 1 << INPUT_MAIN;
|
|
if (dm->ppsrc)
|
|
filter_frame(ctx->inputs[INPUT_CLEANSRC], NULL);
|
|
filter_frame(ctx->inputs[INPUT_MAIN], NULL);
|
|
ff_outlink_set_status(ctx->outputs[0], AVERROR_EOF, dm->last_pts);
|
|
return 0;
|
|
}
|
|
} else if (dm->ppsrc && !(dm->eof & (1 << INPUT_CLEANSRC)) && ff_inlink_acknowledge_status(ctx->inputs[INPUT_CLEANSRC], &status, &pts)) {
|
|
if (status == AVERROR_EOF) { // flushing
|
|
dm->eof |= 1 << INPUT_CLEANSRC;
|
|
filter_frame(ctx->inputs[INPUT_MAIN], NULL);
|
|
filter_frame(ctx->inputs[INPUT_CLEANSRC], NULL);
|
|
ff_outlink_set_status(ctx->outputs[0], AVERROR_EOF, dm->last_pts);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (ff_inlink_queued_frames(ctx->inputs[INPUT_MAIN]) > 0 && (!dm->ppsrc ||
|
|
(dm->ppsrc && ff_inlink_queued_frames(ctx->inputs[INPUT_CLEANSRC]) > 0))) {
|
|
ff_filter_set_ready(ctx, 100);
|
|
} else if (ff_outlink_frame_wanted(ctx->outputs[0])) {
|
|
if (dm->got_frame[INPUT_MAIN] == 0)
|
|
ff_inlink_request_frame(ctx->inputs[INPUT_MAIN]);
|
|
if (dm->ppsrc && (dm->got_frame[INPUT_CLEANSRC] == 0))
|
|
ff_inlink_request_frame(ctx->inputs[INPUT_CLEANSRC]);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int decimate_init(AVFilterContext *ctx)
|
|
{
|
|
DecimateContext *dm = ctx->priv;
|
|
AVFilterPad pad = {
|
|
.name = "main",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
};
|
|
int ret;
|
|
|
|
if ((ret = ff_append_inpad(ctx, &pad)) < 0)
|
|
return ret;
|
|
|
|
if (dm->ppsrc) {
|
|
pad.name = "clean_src";
|
|
pad.config_props = NULL;
|
|
if ((ret = ff_append_inpad(ctx, &pad)) < 0)
|
|
return ret;
|
|
}
|
|
|
|
if ((dm->blockx & (dm->blockx - 1)) ||
|
|
(dm->blocky & (dm->blocky - 1))) {
|
|
av_log(ctx, AV_LOG_ERROR, "blockx and blocky settings must be power of two\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
dm->start_pts = AV_NOPTS_VALUE;
|
|
dm->last_duration = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold void decimate_uninit(AVFilterContext *ctx)
|
|
{
|
|
int i;
|
|
DecimateContext *dm = ctx->priv;
|
|
|
|
av_frame_free(&dm->last);
|
|
av_freep(&dm->bdiffs);
|
|
if (dm->queue) {
|
|
for (i = 0; i < dm->cycle; i++)
|
|
av_frame_free(&dm->queue[i].frame);
|
|
}
|
|
av_freep(&dm->queue);
|
|
if (dm->clean_src) {
|
|
for (i = 0; i < dm->cycle; i++)
|
|
av_frame_free(&dm->clean_src[i]);
|
|
}
|
|
av_freep(&dm->clean_src);
|
|
}
|
|
|
|
static const enum AVPixelFormat pix_fmts[] = {
|
|
#define PF_NOALPHA(suf) AV_PIX_FMT_YUV420##suf, AV_PIX_FMT_YUV422##suf, AV_PIX_FMT_YUV444##suf
|
|
#define PF_ALPHA(suf) AV_PIX_FMT_YUVA420##suf, AV_PIX_FMT_YUVA422##suf, AV_PIX_FMT_YUVA444##suf
|
|
#define PF(suf) PF_NOALPHA(suf), PF_ALPHA(suf)
|
|
PF(P), PF(P9), PF(P10), PF_NOALPHA(P12), PF_NOALPHA(P14), PF(P16),
|
|
AV_PIX_FMT_YUV440P10, AV_PIX_FMT_YUV440P12,
|
|
AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P,
|
|
AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P,
|
|
AV_PIX_FMT_YUVJ411P,
|
|
AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY9, AV_PIX_FMT_GRAY10, AV_PIX_FMT_GRAY12, AV_PIX_FMT_GRAY14,
|
|
AV_PIX_FMT_GRAY16,
|
|
AV_PIX_FMT_NONE
|
|
};
|
|
|
|
static int config_output(AVFilterLink *outlink)
|
|
{
|
|
AVFilterContext *ctx = outlink->src;
|
|
DecimateContext *dm = ctx->priv;
|
|
const AVFilterLink *inlink = ctx->inputs[INPUT_MAIN];
|
|
AVRational fps = inlink->frame_rate;
|
|
int max_value;
|
|
const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(inlink->format);
|
|
const int w = inlink->w;
|
|
const int h = inlink->h;
|
|
|
|
dm->hsub = pix_desc->log2_chroma_w;
|
|
dm->vsub = pix_desc->log2_chroma_h;
|
|
dm->depth = pix_desc->comp[0].depth;
|
|
max_value = (1 << dm->depth) - 1;
|
|
dm->scthresh = (int64_t)(((int64_t)max_value * w * h * dm->scthresh_flt) / 100);
|
|
dm->dupthresh = (int64_t)(((int64_t)max_value * dm->blockx * dm->blocky * dm->dupthresh_flt) / 100);
|
|
dm->nxblocks = (w + dm->blockx/2 - 1) / (dm->blockx/2);
|
|
dm->nyblocks = (h + dm->blocky/2 - 1) / (dm->blocky/2);
|
|
dm->bdiffsize = dm->nxblocks * dm->nyblocks;
|
|
dm->bdiffs = av_malloc_array(dm->bdiffsize, sizeof(*dm->bdiffs));
|
|
dm->queue = av_calloc(dm->cycle, sizeof(*dm->queue));
|
|
dm->in_tb = inlink->time_base;
|
|
dm->nondec_tb = av_inv_q(fps);
|
|
dm->dec_tb = av_mul_q(dm->nondec_tb, (AVRational){dm->cycle, dm->cycle - 1});
|
|
|
|
if (!dm->bdiffs || !dm->queue)
|
|
return AVERROR(ENOMEM);
|
|
|
|
if (dm->ppsrc) {
|
|
dm->clean_src = av_calloc(dm->cycle, sizeof(*dm->clean_src));
|
|
if (!dm->clean_src)
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
if (!fps.num || !fps.den) {
|
|
av_log(ctx, AV_LOG_ERROR, "The input needs a constant frame rate; "
|
|
"current rate of %d/%d is invalid\n", fps.num, fps.den);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (dm->mixed) {
|
|
outlink->time_base = av_gcd_q(dm->nondec_tb, dm->dec_tb, AV_TIME_BASE / 2, AV_TIME_BASE_Q);
|
|
av_log(ctx, AV_LOG_VERBOSE, "FPS: %d/%d -> VFR (use %d/%d if CFR required)\n",
|
|
fps.num, fps.den, outlink->time_base.den, outlink->time_base.num);
|
|
} else {
|
|
outlink->time_base = dm->dec_tb;
|
|
outlink->frame_rate = av_inv_q(outlink->time_base);
|
|
av_log(ctx, AV_LOG_VERBOSE, "FPS: %d/%d -> %d/%d\n",
|
|
fps.num, fps.den, outlink->frame_rate.num, outlink->frame_rate.den);
|
|
}
|
|
outlink->sample_aspect_ratio = inlink->sample_aspect_ratio;
|
|
if (dm->ppsrc) {
|
|
outlink->w = ctx->inputs[INPUT_CLEANSRC]->w;
|
|
outlink->h = ctx->inputs[INPUT_CLEANSRC]->h;
|
|
} else {
|
|
outlink->w = inlink->w;
|
|
outlink->h = inlink->h;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static const AVFilterPad decimate_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.config_props = config_output,
|
|
},
|
|
};
|
|
|
|
const AVFilter ff_vf_decimate = {
|
|
.name = "decimate",
|
|
.description = NULL_IF_CONFIG_SMALL("Decimate frames (post field matching filter)."),
|
|
.init = decimate_init,
|
|
.activate = activate,
|
|
.uninit = decimate_uninit,
|
|
.priv_size = sizeof(DecimateContext),
|
|
FILTER_OUTPUTS(decimate_outputs),
|
|
FILTER_PIXFMTS_ARRAY(pix_fmts),
|
|
.priv_class = &decimate_class,
|
|
.flags = AVFILTER_FLAG_DYNAMIC_INPUTS,
|
|
};
|