My script does not longer work as desired. I guess it’s today simpler to update the command list manually. In addition, it was used mainly by FFCommand_Engine, which is no longer maintained.
A new section was added specifically to address how to deinterlace video
to frames, preserving the visual cadence of the source material.
An animated example of field separation was also provided for the
documentation.
For many years, hardware devices and online streaming platforms
have been capable of 50p or 60p output, so the older advice of
deinterlacing 25i/30i to 25p/30p is no longer necessary. For matching
the fidelity of interlaced sources as they were originally viewed on
period hardware, each field should be output to its own frame, producing
50p content from 25i sources, and 60p content from 30i sources. These
updates to ffmprovisr reflect that.
Additionally, the deinterlacing filter advice of yadif or w3fdif has
been changed to bwdif, which combines the best attributes of both older
filters.
Took out the audio example.
Took out mini example.
Took out hardcoded attributes.
Added IN_POINT, OUT_POINT, DURATION to the time in seconds (expressed as integers) per kfrn suggestion
Hopefully I got all the uses of "frames" out. (let me know if one snuck in)
And I used the example that was used in the stackoverflow forum.
This is for Issue: OG #386 / First attempt #422 / Now this attempt #424
Previously, it would expand the recipe if currently collapsed, and vice versa.
Now, if any recipes are currently expanded, it will collapse all; otherwise, all recipes will be expanded.
The 'expand all' functionality is necessary for full-text searching.
fix for ticket 399: meta tag validation error. w3 doesn't like chartset in the same tag as viewport so I split them out based on what was described in the ticket
2019-11-19 18:07:41 -05:00
19 changed files with 496 additions and 178 deletions
* [ ] I've referred to the [Guidelines for contributing](https://github.com/amiaopensource/ffmprovisr/blob/gh-pages/readme.md#guidelines-for-contributing)
<p>FFmpeg is a powerful tool for manipulating audiovisual files. Unfortunately, it also has a steep learning curve, especially for users unfamiliar with a command line interface. This app helps users through the command generation process so that more people can reap the benefits of FFmpeg.</p>
<p>Each button displays helpful information about how to perform a wide variety of tasks using FFmpeg. To use this site, click on the task you would like to perform. A new window will open up with a sample command and a description of how that command works. You can copy this command and understand how the command works with a breakdown of each of the flags.</p>
<p>Each button displays helpful information about how to perform a wide variety of tasks using FFmpeg. To use this site, click on the task you would like to perform. You will jump to a single command or a list of related commands. Click on a command description, and the site will display a sample command as well as an explanation of how that command works with a breakdown of each of its flags (or options).</p>
<p>This page does not have search functionality, but you can open all recipes (second option in the sidebar) and use your browser's search tool (often ctrl+f or cmd+f) to perform a keyword search through all recipes.</p>
<spanclass="intro-lead">Tutorials</span>
<h3>Tutorials</h3>
<p>For FFmpeg basics, check out the program’s <ahref="https://ffmpeg.org/"target="_blank">official website</a>.</p>
<p>For instructions on how to install FFmpeg on Mac, Linux, and Windows, refer to Reto Kromer’s <ahref="https://avpres.net/FFmpeg/#ch1"target="_blank">installation instructions</a>.</p>
<p>For instructions on how to install FFmpeg on macOS, Linux, and Windows, refer to Reto Kromer’s <ahref="https://avpres.net/FFmpeg/#ch1"target="_blank">installation instructions</a>.</p>
<p>For Bash and command line basics, try the <ahref="https://learnpythonthehardway.org/book/appendixa.html"target="_blank">Command Line Crash Course</a>. For a little more context presented in an ffmprovisr style, try <ahref="https://explainshell.com/"target="_blank">explainshell.com</a>!</p>
This work is licensed under a <ahref="https://creativecommons.org/licenses/by/4.0/"target="_blank">Creative Commons Attribution 4.0 International License</a>.
</p>
<spanclass="intro-lead">Sister projects</span>
<h3>Sibling projects</h3>
<p><ahref="https://dd388.github.io/crals/"target="_blank">Script Ahoy</a>: Community Resource for Archivists and Librarians Scripting</p>
<p><ahref="https://datapraxis.github.io/sourcecaster/"target="_blank">The Sourcecaster</a>: an app that helps you use the command line to work through common challenges that come up when working with digital primary sources.</p>
<p><ahref="https://pugetsoundandvision.github.io/micropops/"target="_blank">Micropops</a>: One liners and automation tools from Moving Image Preservation of Puget Sound</p>
<p><ahref="https://amiaopensource.github.io/cable-bible/"target="_blank">Cable Bible</a>: A Guide to Cables and Connectors Used for Audiovisual Tech</p>
<p><ahref="https://eaasi.gitlab.io/qemu-qed/"target="_blank">QEMU QED</a>: instructions for using QEMU (Quick EMUlator), a command line application for computer emulation and virtualization</p>
<p><ahref="https://eaasi.gitlab.io/program_docs/qemu-qed/"target="_blank">QEMU QED</a>: instructions for using QEMU (Quick EMUlator), a command line application for computer emulation and virtualization</p>
<p><ahref="https://amiaopensource.github.io/ffmpeg-artschool/"target="_blank">ffmpeg-artschool</a>: An AMIA workshop featuring scripts, exercises, and activities to make art using FFmpeg</p>
</div>
<divclass="well">
@@ -104,10 +105,61 @@
<pclass="link"></p>
</div>
<!-- End Streaming vs. Saving -->
<!-- Troubleshooting Common Issues -->
<labelclass="recipe"for="troubleshooting">Troubleshooting Common Issues</label>
<inputtype="checkbox"id="troubleshooting">
<divclass="hiding">
<h5>Troubleshooting Common Issues</h5>
<p>There are many common issues you may encounter when using FFmpeg. This section is here to provide some possible explanations and solutions. For more general solutions to issues regarding download, installation, and other possible questions or issues, please refer to the <ahref="https://ffmpeg.org/faq.html">FFmpeg FAQ</a>.
<p>Before delving into some specific error messages and what may cause them, there are a few general tips you can follow to help make sure your files are doing what you want them to.
First, always check that the output file actually opens, plays, looks, and/or sounds the way you intended it to. It is possible that FFmpeg will create an output file that, on the surface, looks correct, but upon further inspection, there may be inconsistencies with the codec, extension, or other file information, depending on what you intended to do with the file.
Second, always be certain that the file extension you are using for your output files is the intended extension. Within FFmpeg, certain extensions default to certain codecs, such as <code>.mp4</code> encoding to H.264. You always want to make sure that your extension coincides with your intended codec. A list of extensions and codec defaults can be found <ahref="https://amiaopensource.github.io/ffmprovisr/#codec-defaults">here</a>.</p>
<h5>General Error Messages</h5>
<h5>"Error: No such file or directory"</h5>
<p><imgsrc="img/directory_error_example.png"alt="no such file or directory error message"></p>
<p>There are a few reasons why this error may be appearing. First, check to make sure that you are in the correct directory. It can be easy to lose track of where your files are actually stored, especially on a personal computer. If you are working on a personal computer, the easiest way to do this is to check where your files are. When working with macOS, for example, files are often stored in the Downloads or Documents folders. To change your directory from home to this specific directory, use the command <code>cd</code> followed by your desired folder. For a folder named Downloads, it would look like this: <code>cd $HOME/Downloads</code></p>
<p>After you have ensured that you are in the correct directory, and the error is still occurring, check that your argument is formatted correctly. If you are copying and pasting from a separate document or ffmprovisr directly, there may be typos or missing spaces.
If the error message includes other parts of the command, such as a flag, that indicates where your typo or missing space may be, as in this example, where the <code>-c</code> flag has not been separated from the file name:</p>
<p><imgsrc="img/typo_space_error.png"alt="missing space typo error message"></p>
<p>Sometimes, when copying and pasting, even if the command is free of typos and correctly formatted, the error may occur. When this happens, it is best to attempt to rewrite the command from scratch in the command line.</p>
<p>This error message comes when attempting to rewrap a file that has audio or video codecs that are incompatible with the output files extension. This is discussed briefly at the bottom of the rewrap recipe, where it is suggested that you look through possible transcoding recipes. Being able to transcode the file into the correct codec for rewrapping before you run the rewrap command is ideal, but if you are not fully familiar with transcoding and are sure that the extension you are using for your output file will automatically encode the streams with the correct codecs, removing the <code>-c copy</code> flag will solve this error.</p>
<h5>"Killed"</h5>
<p>"Killed" does not appear as a normal error message, but will show up during the output of the command, and the point at which it appears can vary. Sometimes, after you run a command it will be "killed" immediately, only making it through one frame before terminating, as in this example:</p>
<p><imgsrc="img/first_frame_killed.png"alt="command killed at first frame example"></p>
<p>Killed may appear when the file you are working with exceeds the memory capacity of your server. If you are working on a remote server, this can be solved by increasing your memory capacity. If you are working on a local server that should have enough storage and it is still occurring, check that your version of FFmpeg is up to date, and try to avoid running commands with other programs that also require a large amount of memory space at the same time.</p>
<h5>Filters</h5>
<p>When using filters, there are a handful of errors that may pop up. While this section certainly does not cover them all, below you will find a couple of possible errors and what could be causing them.</p>
<h5>"Error splitting argument list: option not found"</h5>
<p>This error often comes from a mistake in the formatting of the command. If you are trying to use certain filters, such as <b>“hflip,vflip,”</b> you may need to remove the quotation marks, as they are not mandatory for certain filter arguments. This is not always the case, as quotation marks are used with filtergraphs to indicate the beginning and end of a filtergraph. Forgetting to remove the quotation marks on certain filters may prompt FFmpeg to run the command as a filtergraph, resulting in an error message like this one:</p>
<p>To avoid this error, check whether the filter you are using requires quotation marks or not. Similar to other errors, oftentimes writing out the entire command, rather than copying and pasting, is beneficial to check for these formatting errors.</p>
<p>Also between the two example errors above, there is one difference that can be helpful to pinpoint when trying to troubleshoot. For the first error, “option not found” refers to the use of quotation marks in a filter that did not require them, thus making the command an unidentifiable option to FFmpeg and making it impossible to differentiate between certain flags in the argument. For the second error, “filter not found” also refers to the use of quotation marks in a command that did not require them, but this time FFmpeg read the argument as a filtergraph due to the quotation marks. For more information regarding filtergraphs, check out the <ahref="https://amiaopensource.github.io/ffmprovisr/#filtergraphs">filtergraph section</a>.</p>
<h5>Streams</h5>
<p>In addition to the possible errors encountered due to typos and improper formatting, some error messages can occur due to inconsistencies with the file streams. Inside each file, there can be numerous streams for audio, video, subtitles, and other file information.</p>
<p>It is possible, when creating output files, to accidentally attempt to use a filter while also asking FFmpeg to copy the file streams directly. Be mindful of where you are trying to put a file’s streams, as filtering and streaming cannot be done together due to the re-encoding of streams that filters must do.</p>
<p><imgsrc="img/filter_and_stream_error.png"alt="filter and stream error message"></p>
<p>To help you further understand issues regarding streams, check out the <ahref="https://amiaopensource.github.io/ffmprovisr/#stream-mapping">stream mapping section.</a></p>
<!-- End Troubleshooting Common Issues -->
</div>
<divclass="well">
<h2id="concepts">Learn about more advanced FFmpeg concepts</h2>
<!-- Loop usage explanation -->
<labelclass="recipe"for="batch-loop">Batch and Loop script usage</label>
<p>One of the frequent uses of FFmpeg is to run batch commands within loops to, for example, generate access files for an entire collection at once.</p>
<p>When running an FFmpeg command within a loop it is often necessary to use the <code>-nostdin</code> flag prior to the input in order to ensure successful execution of the commands. This is needed to override FFmpeg's default behavior of enabling interaction on standard input which can result in errors as loop inputs are fed to the ongoing command.</p>
<p>When creating or transcoding files with FFmpeg, it is important to consider codec settings for both audio and video, as the default options may not be desirable in your particular context. The following is a brief list of codec defaults for some common file types:</p>
<ul>
<li><code>.avi</code>: Audio Codec: mp3, Video Codec: mpeg4</li>
<li><code>.mkv</code>: Audio Codec: ac3, Video Codec: h.264</li>
<li><code>.mov</code>: Audio Codec: aac, Video Codec: h.264</li>
<li><code>.mp4</code>: Audio Codec: aac, Video Codec: h.264</li>
<li><code>.mkv</code>: Audio Codec: ac3, Video Codec: H.264</li>
<li><code>.mov</code>: Audio Codec: AAC, Video Codec: H.264</li>
<li><code>.mp4</code>: Audio Codec: AAC, Video Codec: H.264</li>
<li><code>.mpg</code>: Audio Codec: mp2, Video Codec: mpeg1video</li>
<li><code>.mxf</code>: Audio Codec: pcm_s16le, Video Codec: mpeg2video</li>
<li><code>.wav</code>: Audio Codec: pcm_s16le (16 bit PCM)</li>
</ul>
<pclass="link"></p>
</div>
<!-- End Codec Defaults -->
@@ -183,7 +236,7 @@
<p>To map <em>all</em> streams in the input file to the output file, use <code>-map 0</code>. However, note that not all container formats can include all stream types: for example, .mp4 cannot contain timecode.</p>
<h4>Mapping with a failsafe</h4>
<p>To safely process files that may or may not contain given a type of stream, you can add a trailing <code>?</code> to your map commands: for example, <code>-map 0:a?</code> instead of <code>-map 0:a</code>.</p>
<p>This makes the map optional: audio streams will be mapped over if they are present in the file—but if the file contains no audio streams, the transcode will precede as usual, minus the audio stream mapping. Without adding the trailing <code>?</code>, FFmpeg will exit with an error on that file.</p>
<p>This makes the map optional: audio streams will be mapped over if they are present in the file—but if the file contains no audio streams, the transcode will proceed as usual, minus the audio stream mapping. Without adding the trailing <code>?</code>, FFmpeg will exit with an error on that file.</p>
<p>This is especially recommended when batch processing video files: it ensures that all files in your batch will be transcoded, whether or not they contain audio streams.</p>
<p>For more information, check out the FFmpeg wiki <ahref="https://trac.ffmpeg.org/wiki/Map"target="_blank">Map</a> page, and the official FFmpeg <ahref="https://ffmpeg.org/ffmpeg.html#Advanced-options"target="_blank">documentation on <code>-map</code></a>.</p>
<pclass="link"></p>
@@ -212,12 +265,42 @@
The new container you are rewrapping to is defined by the filename extension used here, e.g. .mkv, .mp4, .mov.</dd>
</dl>
<h4>Important caveat</h4>
<p>It may not be possible to rewrap a file's contents to a new container without re-encoding one or more of the streams within (that is, the video, audio, and subtitle tracks). Some containers can only contain streams of a certain encoding type: for example, the .mp4 container does not support uncompressed audio tracks. (In practice .mp4 goes hand-in-hand with a H.264-encoded video stream and an AAC-encoded video stream, although other types of video and audio streams are possible). Another example is that the Matroska container does not allow data tracks; see the <ahref="#mkv-to-mp4">MKV to MP4 recipe</a>.</p>
<p>It may not be possible to rewrap a file's contents to a new container without re-encoding one or more of the streams within (that is, the video, audio, and subtitle tracks). Some containers can only contain streams of a certain encoding type: for example, the .mp4 container does not support uncompressed audio tracks. (In practice .mp4 goes hand-in-hand with a H.264-encoded video stream and an AAC-encoded video stream, although other types of video and audio streams are possible). Another example is that the Matroska container does not allow data tracks.</p>
<p>In such cases, FFmpeg will throw an error. If you encounter errors of this kind, you may wish to consult the <ahref="#transcode">list of transcoding recipes</a>.</p>
<pclass="link"></p>
</div>
<!-- End Basic rewrap command -->
<!-- BWF -->
<labelclass="recipe"for="bwf">Convert to (or create) Broadcast WAV</label>
<p>This command will write a file in the Broadcast Wave Format (BWF) containing a BEXT chunk with related metadata.</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-i <em>input_file.wav</em></dt><dd>path and name of the input file</dd>
<dt>-c copy</dt><dd>this will copy the encoding/sample rate etc from the input. If not using a WAV as the input file you will have to specify codec settings in place of this.</dd>
<dt>-write_bext 1</dt><dd>tells FFmpeg to write a BEXT chunk, the part of the file where BWF metadata is stored.</dd>
<dt>-metadata field_name='Content'</dt><dd>This is where you can specify which BEXT fields to write, and what information to fill them with by replacing <code>field_name</code> and <code>'Content'</code> respectively. See below for additional details.</dd>
</dl>
<p>Notes: You can choose which fields to write by repeating <code>-metadata field_name='Content'</code> for each desired field. Flags for commonly used fields (such as those recommended by the <ahref="http://www.digitizationguidelines.gov/audio-visual/documents/Embed_Guideline_20120423.pdf">FADGI guidelines</a>) are as follows:</p>
<p>Additionally, users should be aware that BWF metadata fields are limited by characters, with some such as OriginatorReference maxing out at 32. Specific information can be found in the <ahref="https://tech.ebu.ch/docs/tech/tech3285.pdf">Broadcast Wave Format specification</a>. Additional examples of BWF metadata usage can be found in the <ahref="http://www.dlib.indiana.edu/projects/sounddirections/papersPresent/sd_bp_07.pdf">Sound Directions report</a> by Indiana University and Harvard.</p>
<pclass="link"></p>
</div>
<!-- ends BWF -->
<!-- Rewrap DV -->
<labelclass="recipe"for="rewrap-dv">Rewrap DV video to .dv file</label>
<inputtype="checkbox"id="rewrap-dv">
@@ -245,7 +328,7 @@
<inputtype="checkbox"id="to_prores">
<divclass="hiding">
<h5>Transcode into a deinterlaced Apple ProRes LT</h5>
<p>This command transcodes an input file into a deinterlaced Apple ProRes 422 LT file with 16-bit linear PCM encoded audio. The file is deinterlaced using the yadif filter (Yet Another De-Interlacing Filter).</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
@@ -261,7 +344,7 @@
<dt>-vf yadif</dt><dd>Runs a deinterlacing video filter (yet another deinterlacing filter) on the new file. <code>-vf</code> is an alias for <code>-filter:v</code>.</dd>
<dt>-c:a pcm_s16le</dt><dd>tells FFmpeg to encode the audio stream in 16-bit linear PCM</dd>
<dt><em>output_file</em></dt><dd>path, name and extension of the output file<br>
The extension for the QuickTime container is <code>.mov</code>.</dd>
There are currently three possible containers for ProRes 422 and 4444 which are all supported by FFmpeg: QuickTime (<code>.mov</code>), Matroska (<code>.mkv</code>) and Material eXchange Format (<code>.mxf</code>).</dd>
</dl>
<p>FFmpeg comes with more than one ProRes encoder:</p>
<ul>
@@ -301,11 +384,44 @@
<dt>-crf <em>18</em></dt><dd>Specifying a lower CRF will make a larger file with better visual quality. For H.264 files being encoded with a 4:2:0 chroma subsampling scheme (i.e., using <code>-pix_fmt yuv420p</code>), the scale ranges between 0-51 for 8-bit content, with 0 being lossless and 51 the worst possible quality.<br>
If no crf is specified, <code>libx264</code> will use a default value of 23. 18 is often considered a “visually lossless” compression.</dd>
</dl>
<p>By default, this recipe will include one track of each type (e.g. audio, video) in the output file. If you wish to include more tracks, consult the <ahref="#stream-mapping">entry on stream mapping</a>.</p>
<p>For more information, see the <ahref="https://trac.ffmpeg.org/wiki/Encode/H.264"target="_blank">FFmpeg and H.264 Encoding Guide</a> on the FFmpeg wiki.</p>
<pclass="link"></p>
</div>
<!-- ends Transcode to H.264 -->
<!-- Transcode to H.264 or H.265 using the GPU -->
<labelclass="recipe"for="transcode_gpu">Transcode to H.264/H.265 using the GPU</label>
<p>This command takes an input file and transcodes it to H.264 using the encoding functionality of an Nvidia GPU (without transcoding the audio). If you're using H.264 with AAC or AC3 audio, you can output to an .mp4 file; if you're using HEVC and/or more exotic audio, you should output to .mkv. While Nvidia's fixed-function hardware can be 10x as performant as encoding on the CPU, it requires a few more parameters in order to optimize quality at lower bitrates.</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-c:v <em>h264_nvenc</em></dt><dd>tells FFmpeg to encode the video stream as H.264 using Nvidia's encoder.</dd>
<dt>-preset <em>llhq</em></dt><dd>uses the "low latency, high quality" encoding preset, a good default when working with nvenc.</dd>
<dt>-rc:v <em>vbr_hq</em></dt><dd>means "variable bitrate, high quality," allowing you to set a minimum and maximum bitrate for the encode.</dd>
<dt>-cq:v <em>19</em></dt><dd>is the same as the CRF quality level specified using x264 or other CPU-based encoders, where 0 is lossless, 51 is the worst possible quality, and values from 18-23 are typical.</dd>
<dt>-b:v <em>8000k -maxrate:v 12000k</em></dt><dd>corresponds to a minimum bitrate of 8 megabits (8000k) per second, and a maximum of 12 megabits per second. nvenc is not as good at estimating bitrates as CPU-based encoders, and without this data, will occasionally choose a visibly lower bitrate. The 8-12 mbit range is generally a good one for high-quality 1080p h264.</dd>
<dt>-profile:v <em>high</em></dt><dd>uses the "high quality" profile of h264, something that's been baked in to the spec for a long time so that older players can declare compatibility; almost all h264 video now uses high.</dd>
<dt>-c:a <em>copy</em></dt><dd>will skip reencoding the audio stream, and copy the audio from the source file.</dd>
<dt><em>output_file</em></dt><dd>path, name and extension of the output file</dd>
</dl>
<p>In order to encode to HEVC instead, and optionally transcode the audio, you can try changing the command like this:</p>
<dt>-c:v <em>hevc_nvenc</em></dt><dd>encodes to HEVC (also called H.265), a more efficient codec supported on GPUs from approximately 2015 and newer.</dd>
<dt>-b:v <em>5000k -maxrate:v 8000k</em></dt><dd>specifies a slightly lower bitrate than when using h264, per HEVC's greater efficiency.</dd>
<dt>-profile:v <em>main10</em></dt><dd>declares the "main10" profile for working with HEVC; one of the primary advantages of this codec is better support for 10-bit video, enabling consumer HDR.</dd>
<dt>-c:a <em>aac</em></dt><dd>reencodes the audio to AAC with default parameters, a very common and widely supported format for access copies.</dd>
</dl>
<p>Much of the information in this entry was taken from <ahref="https://superuser.com/a/1236387"target="_blank">this superuser.com post</a> provided by an Nvidia developer, one of the best sources of information on the ffmpeg Nvidia encoders.</p>
<pclass="link"></p>
</div>
<!-- ends Transcode to H.264 or H.265 using the GPU -->
<!-- H.264 from DCP -->
<labelclass="recipe"for="dcp_to_h264">Transcode from DCP to an H.264 access file</label>
<p>This command takes an input file and transcodes it to H.265/HEVC in an .mp4 wrapper, keeping the audio codec the same as in the original file.</p>
<p><strong>Note:</strong> FFmpeg must be compiled with libx265, the library of the H.265 codec, for this script to work. (Add the flag <code>--with-x265</code> if using the <code>brew install ffmpeg</code> method).</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-i <em>input file</em></dt><dd>path, name and extension of the input file</dd>
<p>This command takes an input file and transcodes it to Ogg/Theora in an .ogv wrapper with 690k video bitrate.</p>
<p><strong>Note:</strong> FFmpeg must be installed with support for Ogg Theora. If you are using Homebrew, you can check with <code>brew info ffmpeg</code> and then update it with <code>brew upgrade ffmpeg --with-theora --with-libvorbis</code> if necessary.</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-i <em>input file</em></dt><dd>path, name and extension of the input file</dd>
@@ -599,26 +713,30 @@
For silent videos you can replace <code>-c:a copy</code> with <code>-an</code>.</dd>
<dt><em>output_file</em></dt><dd>path, name and extension of the output file</dd>
</dl>
<p>If your source is interlaced, you will want to deinterlace prior to scaling. In that case, your command would look like this:</p>
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-c:a copy</dt><dd>Copy all mapped audio streams.</dd>
<dt>-c:v copy</dt><dd>Copy all mapped video streams.</dd>
<dt>-aspect 4:3</dt><dd>Change Display Aspect Ratio to <code>4:3</code>. Experiment with other aspect ratios such as <code>16:9</code>. If used together with <code>-c:v copy</code>, it will affect the aspect ratio stored at container level, but not the aspect ratio stored in encoded frames, if it exists.</dd>
<dt><em>output_file</em></dt><dd>path, name and extension of the output file</dd>
</dl>
<pclass="link"></p>
</div>
<!-- ends Change display aspect ratio without re-encoding video -->
<!-- ends Change display aspect ratio without re-encoding -->
<!-- Convert colorspace -->
<labelclass="recipe"for="convert-colorspace">Convert colorspace of video</label>
@@ -690,7 +808,7 @@
<li>In the sound filter <code>atempo</code> the numerator <code>output_fps</code> sets the output speed and the denominator <code>input_fps</code> sets the input speed; both values are given in frames per second.</li>
</ul>
The different filters in a complex filter can be divided either by comma or semicolon. The quotation marks allow to insert a space between the filters for readability.</dd>
<dt>-map "[v]"</dt><dd>maps the video stream and:</dd>
<dt>-map "[v]"</dt><dd>maps the video stream and</dd>
<dt>-map "[a]"</dt><dd>maps the audio stream together into:</dd>
<dt><em>output_file</em></dt><dd>path, name and extension of the output file</dd>
</dl>
@@ -698,6 +816,26 @@
</div>
<!-- ends Modify speed -->
<!-- Fade both video and audio streams -->
<labelclass="recipe"for="fade_streams">Fade both video and audio streams</label>
<p>This command fades your video in and out. Change IN_POINT, OUT_POINT, and DURATION to the time in seconds (expressed as integers).</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-filter:v "fade=in:st=IN_POINT:d=DURATION, fade=out:st=OUT_POINT:d=DURATION"</dt><dd>applies a video filter that fades your video in and out. <code>st</code> sets the start and <code>d</code> sets the duration.</dd>
<dt>-filter:a "afade=in:st=IN_POINT:d=DURATION, afade=out:st=OUT_POINT:d=DURATION"</dt><dd>applies an audio filter that fades your video in and out. <code>st</code> sets the start and <code>d</code> sets the duration.</dd>
<dt>-c:v <em>video_codec</em></dt><dd>as a video filter is used, it is not possible to use <code>-c copy</code>. The video must be re-encoded with whatever video codec is chosen, e.g. <code>ffv1</code>, <code>v210</code> or <code>prores</code>.</dd>
<dt>-c:a <em>audio_codec</em></dt><dd>as an audio filter is used, it is not possible to use <code>-c copy</code>. The audio must be re-encoded with whatever audio codec is chosen, e.g. <code>aac</code>.</dd>
<dt><em>output_file</em></dt><dd>path, name and extension of the output_file</dd>
</dl>
<pclass="link"></p>
</div>
<!-- ends Fade both video and audio streams -->
<!-- Synchronize video and audio streams -->
<labelclass="recipe"for="sync_streams">Synchronize video and audio streams</label>
<inputtype="checkbox"id="sync_streams">
@@ -773,13 +911,13 @@
<p>It's also possible to specify the crop position by adding the x and y coordinates representing the top left of your cropped area to your crop filter, as such:</p>
<p>The original frame, a screenshot of the SMPTE colorbars:</p>
<p>The original frame, a screenshot of Maggie Cheung in the film <i>Hero</i>:</p>
<imgclass="sample-image"src="img/crop_example_orig.png"alt="VLC screenshot of Maggie Cheung">
<p>Result of the command <code>ffmpeg -i <em>smpte_colorsbars.mov</em> -vf "crop=500:500" <em>output_file</em></code>:</p>
<p>Result of the command <code>ffmpeg -i <em>maggie.mov</em> -vf "crop=500:500" <em>output_file</em></code>:</p>
<imgclass="sample-image-small"src="img/crop_example_aftercrop1.png"alt="VLC screenshot of Maggie Cheung, cropped from original">
<p>Result of the command <code>ffmpeg -i <em>smpte_colorsbars.mov</em> -vf "crop=500:500:0:0" <em>output_file</em></code>, appending <code>:0:0</code> to crop from the top left corner:</p>
<p>Result of the command <code>ffmpeg -i <em>maggie.mov</em> -vf "crop=500:500:0:0" <em>output_file</em></code>, appending <code>:0:0</code> to crop from the top left corner:</p>
<imgclass="sample-image-small"src="img/crop_example_aftercrop2.png"alt="VLC screenshot of Maggie Cheung, cropped from original">
<p>Result of the command <code>ffmpeg -i <em>smpte_colousbars.mov</em> -vf "crop=500:300:500:30" <em>output_file</em></code>:</p>
<p>Result of the command <code>ffmpeg -i <em>maggie.mov</em> -vf "crop=500:300:500:30" <em>output_file</em></code>:</p>
<imgclass="sample-image-small"src="img/crop_example_aftercrop3.png"alt="VLC screenshot of Maggie Cheung, cropped from original">
<p>This filter calculates and outputs loudness information in json about an input file (labeled input) as well as what the levels would be if loudnorm were applied in its one pass mode (labeled output). The values generated can be used as inputs for a 'second pass' of the loudnorm filter allowing more accurate loudness normalization than if it is used in a single pass.</p>
<p>These instructions use the loudnorm defaults, which align well with PBS recommendations for target loudness. More information can be found at the <ahref="https://ffmpeg.org/ffmpeg-filters.html#loudnorm"target="_blank">loudnorm documentation</a>.</p>
<p>Information about PBS loudness standards can be found in the <ahref="https://www-tc.pbs.org/capt/Producing/TOS-2012-Pt2-Distribution.pdf"target="_blank">PBS Technical Operating Specifications</a> document. Information about EBU loudness standards can be found in the <ahref="https://tech.ebu.ch/docs/r/r128-2014.pdf"target="_blank">EBU R 128</a> recommendation document.</p>
<p>Information about PBS loudness standards can be found in the <ahref="http://bento.cdn.pbs.org/hostedbento-prod/filer_public/PBS_About/Producing/Red%20Book/TOS%20Pt%201%20Submission%202016.pdf"target="_blank">PBS Technical Operating Specifications</a> document. Information about EBU loudness standards can be found in the <ahref="https://tech.ebu.ch/docs/r/r128-2014.pdf"target="_blank">EBU R 128</a> recommendation document.</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt><em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<p>This command takes two or more files of the same file type and joins them together to make a single file. All that the program needs is a text file with a list specifying the files that should be joined. However, it only works properly if the files to be combined have the exact same codec and technical specifications. Be careful, FFmpeg may appear to have successfully joined two video files with different codecs, but may only bring over the audio from the second file or have other weird behaviors. Don’t use this command for joining files with different codecs and technical specs and always preview your resulting video file!</p>
<p>This command takes two or more files of the same file type and joins them together to make a single file. All that the program needs is a text file with a list specifying the files that should be joined. If possible, run the command from the same directory where the files and the text file reside. Otherwise you'll have to use <code>-safe 0</code>, see below for more information. However, it only works properly if the files to be combined have the exact same codec and technical specifications. Be careful, FFmpeg may appear to have successfully joined two video files with different codecs, but may only bring over the audio from the second file or have other weird behaviors. Don’t use this command for joining files with different codecs and technical specs and always preview your resulting video file!</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-f concat</dt><dd>forces ffmpeg to concatenate the files and to keep the same file format</dd>
<dt>-i <em>mylist.txt</em></dt><dd>path, name and extension of the input file. Per the <ahref="https://ffmpeg.org/ffmpeg-formats.html#Options"target="_blank">FFmpeg documentation</a>, it is preferable to specify relative rather than absolute file paths, as allowing absolute file paths may pose a security risk.<br>
This text file contains the list of files to be concatenated and should be formatted as follows:
<pre>file '<em>./first_file.ext</em>'
file '<em>./second_file.ext</em>'
This text file contains the list of files (without their absolute path) to be concatenated and should be formatted as follows:
<pre>
file '<em>first_file.ext</em>'
file '<em>second_file.ext</em>'
. . .
file '<em>./last_file.ext</em>'</pre>
file '<em>last_file.ext</em>'
</pre>
In the above, <strong>file</strong> is simply the word "file". Straight apostrophes ('like this') rather than curved quotation marks (‘like this’) must be used to enclose the file paths.<br>
<strong>Note:</strong> If specifying absolute file paths in the .txt file, add <code>-safe 0</code> before the input file.<br>
<p>This command takes an interlaced input file and outputs a deinterlaced H.264 MP4, with each field separated into its own frame. This is preferred for interlaced video that contains a lot of motion, as the double-rate output preserves the visual cadence of the source material.</p>
<dl>
<dt>ffmpeg</dt><dd>starts the command</dd>
<dt>-i <em>input file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-c:v libx264</dt><dd>tells FFmpeg to encode the video stream as H.264</dd>
<dt>-vf</dt><dd>video filtering will be used (<code>-vf</code> is an alias of <code>-filter:v</code>)</dd>
<dt>"</dt><dd>start of filtergraph (see below)</dd>
<dt>idet</dt><dd>detect interlaced video field order<br>
<ahref="https://ffmpeg.org/ffmpeg-filters.html#idet"target="_blank">idet</a> will try to detect if the video is interlaced, and if so, what the order of the fields are (top-field-first, or bottom-field-first). This is done to ensure the output of the deinterlacing filter is correct.</dd>
By default, <ahref="https://ffmpeg.org/ffmpeg-filters.html#bwdif-1"target="_blank">bwdif</a> will output one frame for each field, matching the visual cadence of interlaced video. </dd>
<dt>,</dt><dd>separates filters</dd>
<dt>format=yuv420p</dt><dd>chroma subsampling set to 4:2:0<br>
By default, <code>libx264</code> will use a chroma subsampling scheme that is the closest match to that of the input. This can result in Y′C<sub>B</sub>C<sub>R</sub> 4:2:0, 4:2:2, or 4:4:4 chroma subsampling. QuickTime and most other non-FFmpeg based players can’t decode H.264 files that are not 4:2:0, therefore it’s advisable to specify 4:2:0 chroma subsampling.</dd>
<dt>"</dt><dd>end of filtergraph</dd>
<dt><em>output file</em></dt><dd>path, name and extension of the output file</dd>
</dl>
<p><code>"idet,bwdif,format=yuv420p"</code> is an FFmpeg <ahref="https://trac.ffmpeg.org/wiki/FilteringGuide#FiltergraphChainFilterrelationship"target="_blank">filtergraph</a>. Here the filtergraph is made up of one filter chain, which is itself made up of the three filters (separated by the comma).<br>
The enclosing quote marks are necessary when you use spaces within the filtergraph, e.g. <code>-vf "idet, bwdif, format=yuv420p"</code>, and are included above as an example of good practice.</p>
<p><strong>Note:</strong> bwdif also supports the older method of outputting one frame for each frame (thereby halving the number of output frames per second) with the syntax <code>bwdif=mode=send_frame</code>. This can be used when the presentation device is not capable of reproducing 50 (PAL) or 60 (NTSC) frames per second.</p>
<p>For more H.264 encoding options, see the latter section of the <ahref="#transcode_h264">encode H.264 command</a>.</p>
<divclass="sample-image">
<h2>Example</h2>
<p>Before and after deinterlacing with bwdif:</p>
<imgsrc="img/interlaced_video_fields.png"alt="VLC screenshot of original interlaced video">
<imgsrc="img/deinterlaced_video_frames.png"alt="VLC screenshot of deinterlaced video">
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-filter:v idet</dt><dd>This calls the <ahref="https://ffmpeg.org/ffmpeg-filters.html#idet"target="_blank">idet (detect video interlacing type) filter</a>.</dd>
<dt>-f null</dt><dd>Video is decoded with the <code>null</code> muxer. This allows video decoding without creating an output file.</dd>
<dt>-</dt><dd>FFmpeg syntax requires a specified output, and <code>-</code>is just a place holder. No file is actually created.</dd>
<dt>-</dt><dd>The FFmpeg syntax requires a specified output, and <code>-</code>prints the output to the screen (STDOUT), rather than creating a file.</dd>
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-f null</dt><dd>Video is decoded with the <code>null</code> muxer. This allows video decoding without creating an output file.</dd>
<dt>-</dt><dd>FFmpeg syntax requires a specified output, and <code>-</code>is just a place holder. No file is actually created.</dd>
<dt>-</dt><dd>The FFmpeg syntax requires a specified output, and <code>-</code>prints the output to the screen (STDOUT), rather than creating a file.</dd>
</dl>
<pclass="link"></p>
</div>
@@ -1904,7 +2081,7 @@
<dt>-report</dt><dd>Dump full command line and console output to a file named <em>ffmpeg-YYYYMMDD-HHMMSS.log</em> in the current directory. It also implies <code>-loglevel verbose</code>.</dd>
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-f null</dt><dd>Video is decoded with the <code>null</code> muxer. This allows video decoding without creating an output file.</dd>
<dt>-</dt><dd>FFmpeg syntax requires a specified output, and <code>-</code>is just a place holder. No file is actually created.</dd>
<dt>-</dt><dd>The FFmpeg syntax requires a specified output, and <code>-</code>prints the output to the screen (STDOUT), rather than creating a file.</dd>
</dl>
<pclass="link"></p>
</div>
@@ -1991,7 +2168,7 @@
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-map 0:v:0</dt><dd>designated the first video stream as the stream on which to perform this hash generation operation. <code>-map 0</code> can be used to run the operation on all streams.</dd>
<dt>-f hash -hash md5</dt><dd>produce a checksum hash, and set the hash algorithm to md5. See the official <ahref="https://ffmpeg.org/ffmpeg-formats.html#hash"target="_blank">documentation on hash</a> for other algorithms.</dd>
<dt>-</dt><dd>FFmpeg syntax requires a specified output, and <code>-</code>is just a place holder. No file is actually created.</dd>
<dt>-</dt><dd>The FFmpeg syntax requires a specified output, and <code>-</code>prints the output to the screen (STDOUT), rather than creating a file.</dd>
</dl>
<pclass="link"></p>
</div>
@@ -2011,7 +2188,7 @@
<dt>-i <em>input_file</em></dt><dd>path, name and extension of the input file</dd>
<dt>-map 0</dt><dd>map ALL streams from input file to output. If you omit this, ffmpeg chooses only the first "best" (*) stream: 1 for audio, 1 for video (not all streams).</dd>
<dt>-f streamhash -hash md5</dt><dd>produce a checksum hash per-stream, and set the hash algorithm to md5. See the official <ahref="https://www.ffmpeg.org/ffmpeg-formats.html#streamhash-1"target="_blank">documentation on streamhash</a> for other algorithms and more details.</dd>
<dt>-</dt><dd>FFmpeg syntax requires a specified output, and <code>-</code>is just a place holder. No file is actually created. Choose an output filename to write the hashcode lines into a textfile.</dd>
<dt>-</dt><dd>The FFmpeg syntax requires a specified output, and <code>-</code>prints the output to the screen (STDOUT), rather than creating a file. Choose an output filename to write the hashcode lines into a textfile.</dd>
<dt>-v quiet</dt><dd>(Optional) Disables FFmpeg's processing output. With this option it's easier to see the text output of the hashes.</dd>
</dl>
<p>The output looks like this, for example (1 video, 2 audio streams):
@@ -2282,7 +2459,6 @@
<inputtype="checkbox"id="ocr_on_top">
<divclass="hiding">
<h5>Plays video with OCR on top</h5>
<p>Note: ffmpeg must be compiled with the tesseract library for this script to work (<code>--with-tesseract</code> if using the <code>brew install ffmpeg</code> method).</p>
<p>Note: FFmpeg must be compiled with the tesseract library for this script to work (<code>--with-tesseract</code> if using the <code>brew install ffmpeg</code> method)</p>
<dt>-flags +global_header</dt><dd>Don't place extra data in every keyframe</dd>
<dt>-vf scale="1280:-1"</dt><dd>Scale to 1280 width, maintain aspect ratio.</dd>
<dt>-pix_fmt yuv420p</dt><dd>convert to 4:2:0 chroma subsampling scheme</dd>
<dt>-level 3.1</dt><dd>H264 Level (defines some thresholds for bitrate)</dd>
<dt>-level 3.1</dt><dd>H.264 Level (defines some thresholds for bitrate)</dd>
<dt>-vsync passthrough</dt><dd>Each frame is passed with its timestamp from the demuxer to the muxer.</dd>
<dt>-crf 26</dt><dd>Constant rate factor - basically the quality</dd>
<dt>-g 50</dt><dd>GOP size.</dd>
@@ -2812,5 +2987,6 @@ Change the above data-target field, the hover-over description, the button text,
<p>Made with ♥ at <ahref="https://wiki.curatecamp.org/index.php/Association_of_Moving_Image_Archivists_%26_Digital_Library_Federation_Hack_Day_2015"target="_blank">AMIA #AVhack15</a>! Contribute to the project via <ahref="https://github.com/amiaopensource/ffmprovisr"target="_blank">our GitHub page</a>!</p>
$(id).next('div').find('.link').append("<small>Link to this command: <a href='https://amiaopensource.github.io/ffmprovisr/index.html"+id+"'>https://amiaopensource.github.io/ffmprovisr/index.html"+id+"</a></small>");
link.innerHTML=("<small>Link to this command: <a href='https://amiaopensource.github.io/ffmprovisr/index.html#"+id+"'>https://amiaopensource.github.io/ffmprovisr/index.html#"+id+"</a></small>")
@@ -5,17 +5,18 @@ Repository of useful FFmpeg command lines for archivists!
* [What is this?](#what-is-this)
* [How do I see it?](#how-do-i-see-it)
* [How do I contribute?](#how-do-i-contribute)
* [Guidelines for contributing](#guidelines-for-contributing)
* [Code of conduct](#code-of-conduct)
* [Maintainers](#maintainers)
* [Contributors](#contributors)
* [AVHack Team](#avhack-team)
* [Sister projects](#sister-projects)
* [Awards and mentions](#articles-and-mentions)
* [Articles and mentions](#articles-and-mentions)
* [License](#license)
## What is this?
#### Project Objective
### Project Objective
To facilitate better understanding of FFmpeg through collaborative sharing of useful scripts and detailed flag-level description of how each script works, so archivists can copy-paste and produce their own scripts, but also understand how and why they work.
@@ -34,7 +35,7 @@ ffmprovisr
```
This works currently under macOS, Linux and the Linux apps on Windows (Ubuntu and Debian tested). On classic Windows you can install the last [release](https://github.com/amiaopensource/ffmprovisr/releases) manually and the open `index.html` in a browser.
#### Parseable list of the commands
### Parseable list of the commands
A list of all recipes in an easily parseable [ASCII text](recipes.txt) format is provided as well. It contains for each recipe its title and command in the following format:
@@ -52,21 +53,34 @@ ffmpeg command n-1
ffmpeg command n
```
The used [one-liner](scripts/get_recipe_list) is in the `scripts` folder.
## How do I contribute?
You are welcome to edit the codebase yourself, or just supply the information and ask it to be added to the site.
#### Edit codebase
### Edit codebase
To contribute to this project directly (and more quickly), clone this repository and create a new branch (`git checkout -b your-branch-name`) and add or modify a new block in `index.html`. Then [submit a pull request](https://github.com/amiaopensource/ffmprovisr/pulls) and the maintainers will review and integrate your code. There is a commented-out sample block available at the bottom of `index.html` that can be as a guideline for your command.
To contribute to this project directly (and more quickly), clone this repository and create a new branch (`git checkout -b your-branch-name`) and add or modify a new block in `index.html`. Then [submit a pull request](https://github.com/amiaopensource/ffmprovisr/pulls) and the maintainers will review and integrate your code. There is a commented-out sample block available at the bottom of `index.html` that can be a guideline for your command.
#### Make a request
#### Guidelines for contributing
* Recipes should ideally do just one thing or solve one problem, to keep things as user-friendly as possible and to avoid unintended side-effects
* Explanations and examples for recipes should be as generic as possible, to allow users to alter the command for their own use-case.
* This may involve the use of `VARIABLE_NAMES`
* Example: [Fade video and audio streams](https://amiaopensource.github.io/ffmprovisr/#fade_streams)
* Some recipes may benefit from including a GIF that shows the output
* Example: [Plays a graphical output showing decibel levels of an input file](https://amiaopensource.github.io/ffmprovisr/#astats)
* Some recipes require an explanatory section which may not fit gracefully into the recipe itself
* Example: [Rewrap a file](https://amiaopensource.github.io/ffmprovisr/#basic-rewrap)
* Recipes involving `-filter_complex` can be some of the most verbose and difficult to understand, so breaking these down as much as possible into their relevant sections is ideal
* Example: [Generate two access MP3s from input](https://amiaopensource.github.io/ffmprovisr/#append_mp3)
* Some recipes may refer to a specific standard or vocabulary, and it's useful to link to these so that the user can scale the recipe to their use case
If you are having trouble with coding it yourself or with GitHub, feel free to [submit an issue](https://github.com/amiaopensource/ffmprovisr/issues) with the kind of command you would like to see added to the site.
#### General help
### General help
If you want to help but don't have a new script to add, you can help us by testing out the scripts available, by refining or clarifying the documentation, or [creating an issue](https://github.com/amiaopensource/ffmprovisr/issues) for anything that sounds confusing and requires clarification.
@@ -83,31 +97,36 @@ You can read our contributor code of conduct [here](https://github.com/amiaopens
*Code Contributors*:
ablwr (Ashley)
alavigne12 (A. Lavigne)
Anushka-codergirl (Anushka Raj)
bastibeckr (Basti Becker)
b00giehead (Joanna White)
bturkus
dericed (Dave Rice)
digitensions (Joanna White)
edsu (Ed Summers)
jamessam (Jim Sam)
jfarbowitz (Jonathan Farbowitz)
kfrn (Katherine Frances Nagels)
kgrons (Kathryn Gronsbell)
kieranjol (Kieran O'Leary)
llogan (Lou)
llogan (Lou Logan)
macasaurusrex (Maura)
mgiraldo (Mauricio Giraldo)
pjotrek-b (Peter B.)
privatezero (Andrew Weaver)
retokromer (Reto Kromer)
rfraimow
*All Contributors*:
All Contributors:
ablwr (Ashley)
alavigne12 (A. Lavigne)
Anushka-codergirl (Anushka Raj)
audiovisualopen
bastibeckr (Basti Becker)
b00giehead (Joanna White)
brainwane (Sumana Harihareswara)
bturkus
dericed (Dave Rice)
digitensions (Joanna White)
drodz11 (Dave Rodriguez)
edsu (Ed Summers)
EG-tech (Ethan Gates)
@@ -122,7 +141,9 @@ kellyhaydon (metacynic)
kfrn (Katherine Frances Nagels)
kgrons (Kathryn Gronsbell)
kieranjol (Kieran O'Leary)
llogan (Lou)
llogan (Lou Logan)
macasaurusrex (Maura)
mercuryswitch
mgiraldo (Mauricio Giraldo)
mulvya
nkrabben (Nick Krabbenhoeft)
@@ -132,12 +153,13 @@ retokromer (Reto Kromer)
rfraimow
richardpl (Paul B Mahol)
ross-spencer (Ross Spencer)
taschenbach (Tommy Aschenbach)
todrobbins (Tod Robbins)
Repo: amiaopensource/ffmprovisr
Code Contributors: 17
All Contributors: 32
Last updated: 2019-02-11
GitHub Contributors: 20
All Contributors: 37
Last updated: 2019-12-11
## AVHack Team
@@ -148,7 +170,8 @@ Last updated: 2019-02-11
## Sister projects
[The Cable Bible](https://amiaopensource.github.io/cable-bible/): A Guide to Cables and Connectors Used for Audiovisual Tech
[QEMU QED](https://eaasi.gitlab.io/qemu-qed): instructions for using QEMU (Quick EMUlator), a command line application for computer emulation and virtualization
[FFCommand_Engine](https://github.com/ColorlabMD/FFCommand_Engine): a tool for easier use of FFmpeg binaries
[QEMU QED](https://eaasi.gitlab.io/program_docs/qemu-qed/): instructions for using QEMU (Quick EMUlator), a command line application for computer emulation and virtualization
[Script Ahoy](http://dd388.github.io/crals/): Community Resource for Archivists and Librarians Scripting
[sourcecaster](https://datapraxis.github.io/sourcecaster/): helps you use the command line to work through common challenges that come up when working with digital primary sources.
curl https://amiaopensource.github.io/ffmprovisr/ -s | grep -E '<h5>.*</h5>|<p><code>.*</code></p>' | sed 's/.*<code>\(.*\)<\/code>/\1/' | sed 's/.*<h5>\(.*\)<\/h5>/# \1/' | grep -v '\*\*\*' | sed -e 's/<[^>]*>//g'
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.