HEX
Server: LiteSpeed
System: Linux ip-172-31-76-142.ec2.internal 4.14.158-129.185.amzn2.x86_64 #1 SMP Tue Dec 24 03:15:32 UTC 2019 x86_64
User: 69b4844ae61d4e92bf26ad98af552775 (1065)
PHP: 7.2.27
Disabled: exec,passthru,shell_exec,system,eval
Upload Files
File: //lib/python2.7/site-packages/awscli/examples/rekognition/start-content-moderation.rst
**To start the recognition of unsafe content in a stored video**

The following ``start-content-moderation`` command starts a job to detect unsafe content in the specified video file stored in an Amazon S3 bucket. ::

    aws rekognition start-content-moderation \
        --video "S3Object={Bucket=MyVideoS3Bucket,Name=MyVideoFile.mpg}"

Output::

    {
        "JobId": "1234567890abcdef1234567890abcdef1234567890abcdef1234567890abcdef"
    }

For more information, see `Detecting Unsafe Stored Videos <https://docs.aws.amazon.com/rekognition/latest/dg/procedure-moderate-videos.html>`__ in the *Amazon Rekognition Developer Guide*.