[FFmpeg-cvslog] doc: update documentation to use avconv

Luca Barbato git at videolan.org
Sat Dec 10 02:08:51 CET 2011


ffmpeg | branch: master | Luca Barbato <lu_zero at gentoo.org> | Thu Dec  8 17:39:28 2011 +0100| [d5837d7fe922883d4b7609ec33802d44f9af83b5] | committer: Luca Barbato

doc: update documentation to use avconv

> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=d5837d7fe922883d4b7609ec33802d44f9af83b5
---

 doc/avserver.texi            |   24 +++++++++++-----------
 doc/avtools-common-opts.texi |   10 ++++----
 doc/bitstream_filters.texi   |    6 ++--
 doc/indevs.texi              |   46 +++++++++++++++++++++---------------------
 doc/muxers.texi              |   38 +++++++++++++++++-----------------
 doc/protocols.texi           |   44 ++++++++++++++++++++--------------------
 6 files changed, 84 insertions(+), 84 deletions(-)

diff --git a/doc/avserver.texi b/doc/avserver.texi
index 7ac4a13..9838028 100644
--- a/doc/avserver.texi
+++ b/doc/avserver.texi
@@ -33,19 +33,19 @@ debug mode or a NoDaemon option is specified in the configuration
 file.
 
 This documentation covers only the streaming aspects of avserver /
-ffmpeg. All questions about parameters for ffmpeg, codec questions,
-etc. are not covered here. Read @file{ffmpeg-doc.html} for more
+avconv. All questions about parameters for avconv, codec questions,
+etc. are not covered here. Read @file{avconv.html} for more
 information.
 
 @section How does it work?
 
-avserver receives prerecorded files or FFM streams from some ffmpeg
+avserver receives prerecorded files or FFM streams from some avconv
 instance as input, then streams them over RTP/RTSP/HTTP.
 
 An avserver instance will listen on some port as specified in the
-configuration file. You can launch one or more instances of ffmpeg and
+configuration file. You can launch one or more instances of avconv and
 send one or more FFM streams to the port where avserver is expecting
-to receive them. Alternately, you can make avserver launch such ffmpeg
+to receive them. Alternately, you can make avserver launch such avconv
 instances at startup.
 
 Input streams are called feeds, and each one is specified by a <Feed>
@@ -107,11 +107,11 @@ LAME is important as it allows for streaming audio to Windows Media Player.
 Don't ask why the other audio types do not work.
 
 As a simple test, just run the following two command lines where INPUTFILE
-is some file which you can decode with ffmpeg:
+is some file which you can decode with avconv:
 
 @example
 ./avserver -f doc/avserver.conf &
-./ffmpeg -i INPUTFILE http://localhost:8090/feed1.ffm
+./avconv -i INPUTFILE http://localhost:8090/feed1.ffm
 @end example
 
 At this point you should be able to go to your Windows machine and fire up
@@ -130,7 +130,7 @@ The same is true of AVI files.
 @section What happens next?
 
 You should edit the avserver.conf file to suit your needs (in terms of
-frame rates etc). Then install avserver and ffmpeg, write a script to start
+frame rates etc). Then install avserver and avconv, write a script to start
 them up, and off you go.
 
 @section Troubleshooting
@@ -138,13 +138,13 @@ them up, and off you go.
 @subsection I don't hear any audio, but video is fine.
 
 Maybe you didn't install LAME, or got your ./configure statement wrong. Check
-the ffmpeg output to see if a line referring to MP3 is present. If not, then
+the avconv output to see if a line referring to MP3 is present. If not, then
 your configuration was incorrect. If it is, then maybe your wiring is not
 set up correctly. Maybe the sound card is not getting data from the right
 input source. Maybe you have a really awful audio interface (like I do)
 that only captures in stereo and also requires that one channel be flipped.
 If you are one of these people, then export 'AUDIO_FLIP_LEFT=1' before
-starting ffmpeg.
+starting avconv.
 
 @subsection The audio and video loose sync after a while.
 
@@ -250,7 +250,7 @@ Use @file{configfile} instead of @file{/etc/avserver.conf}.
 @item -n
 Enable no-launch mode. This option disables all the Launch directives
 within the various <Stream> sections. Since avserver will not launch
-any ffmpeg instances, you will have to launch them manually.
+any avconv instances, you will have to launch them manually.
 @item -d
 Enable debug mode. This option increases log verbosity, directs log
 messages to stdout and causes avserver to run in the foreground
@@ -265,7 +265,7 @@ rather than as a daemon.
 
 @c man begin SEEALSO
 
-avconv(1), avplay(1), avprobe(1), the @file{ffmpeg/doc/avserver.conf}
+avconv(1), avplay(1), avprobe(1), the @file{avserver.conf}
 example and the Libav HTML documentation
 @c man end
 
diff --git a/doc/avtools-common-opts.texi b/doc/avtools-common-opts.texi
index ba552fc..8892e9e 100644
--- a/doc/avtools-common-opts.texi
+++ b/doc/avtools-common-opts.texi
@@ -132,8 +132,8 @@ libavcodec libraries. To see the list of available AVOptions, use the
 @option{-help} option. They are separated into two categories:
 @table @option
 @item generic
-These options can be set for any container, codec or device. Generic options are
-listed under AVFormatContext options for containers/devices and under
+These options can be set for any container, codec or device. Generic options
+are listed under AVFormatContext options for containers/devices and under
 AVCodecContext options for codecs.
 @item private
 These options are specific to the given container, device or codec. Private
@@ -144,14 +144,14 @@ For example to write an ID3v2.3 header instead of a default ID3v2.4 to
 an MP3 file, use the @option{id3v2_version} private option of the MP3
 muxer:
 @example
-ffmpeg -i input.flac -id3v2_version 3 out.mp3
+avconv -i input.flac -id3v2_version 3 out.mp3
 @end example
 
 All codec AVOptions are obviously per-stream, so the chapter on stream
 specifiers applies to them
 
-Note -nooption syntax cannot be used for boolean AVOptions, use -option
-0/-option 1.
+Note @option{-nooption} syntax cannot be used for boolean AVOptions,
+use @option{-option 0}/@option{-option 1}.
 
 Note2 old undocumented way of specifying per-stream AVOptions by prepending
 v/a/s to the options name is now obsolete and will be removed soon.
diff --git a/doc/bitstream_filters.texi b/doc/bitstream_filters.texi
index c33fca6..1fbd40a 100644
--- a/doc/bitstream_filters.texi
+++ b/doc/bitstream_filters.texi
@@ -34,7 +34,7 @@ JPEG image. The individual frames can be extracted without loss,
 e.g. by
 
 @example
-ffmpeg -i ../some_mjpeg.avi -vcodec copy frames_%d.jpg
+avconv -i ../some_mjpeg.avi -c:v copy frames_%d.jpg
 @end example
 
 Unfortunately, these chunks are incomplete JPEG images, because
@@ -57,9 +57,9 @@ stream (carrying the AVI1 header ID and lacking a DHT segment) to
 produce fully qualified JPEG images.
 
 @example
-ffmpeg -i mjpeg-movie.avi -vcodec copy -vbsf mjpeg2jpeg frame_%d.jpg
+avconv -i mjpeg-movie.avi -c:v copy -vbsf mjpeg2jpeg frame_%d.jpg
 exiftran -i -9 frame*.jpg
-ffmpeg -i frame_%d.jpg -vcodec copy rotated.avi
+avconv -i frame_%d.jpg -c:v copy rotated.avi
 @end example
 
 @section mjpega_dump_header
diff --git a/doc/indevs.texi b/doc/indevs.texi
index 87dee70..4405a5b 100644
--- a/doc/indevs.texi
+++ b/doc/indevs.texi
@@ -42,10 +42,10 @@ specify card number or identifier, device number and subdevice number
 To see the list of cards currently recognized by your system check the
 files @file{/proc/asound/cards} and @file{/proc/asound/devices}.
 
-For example to capture with @file{ffmpeg} from an ALSA device with
+For example to capture with @command{avconv} from an ALSA device with
 card id 0, you may run the command:
 @example
-ffmpeg -f alsa -i hw:0 alsaout.wav
+avconv -f alsa -i hw:0 alsaout.wav
 @end example
 
 For more information see:
@@ -72,14 +72,14 @@ For more detailed information read the file
 Documentation/fb/framebuffer.txt included in the Linux source tree.
 
 To record from the framebuffer device @file{/dev/fb0} with
- at file{ffmpeg}:
+ at command{avconv}:
 @example
-ffmpeg -f fbdev -r 10 -i /dev/fb0 out.avi
+avconv -f fbdev -r 10 -i /dev/fb0 out.avi
 @end example
 
 You can take a single screenshot image with the command:
 @example
-ffmpeg -f fbdev -vframes 1 -r 1 -i /dev/fb0 screenshot.jpeg
+avconv -f fbdev -frames:v 1 -r 1 -i /dev/fb0 screenshot.jpeg
 @end example
 
 See also @url{http://linux-fbdev.sourceforge.net/}, and fbset(1).
@@ -109,10 +109,10 @@ To list the JACK clients and their properties you can invoke the command
 @file{jack_lsp}.
 
 Follows an example which shows how to capture a JACK readable client
-with @file{ffmpeg}.
+with @command{avconv}.
 @example
-# Create a JACK writable client with name "ffmpeg".
-$ ffmpeg -f jack -i ffmpeg -y out.wav
+# Create a JACK writable client with name "libav".
+$ avconv -f jack -i libav -y out.wav
 
 # Start the sample jack_metro readable client.
 $ jack_metro -b 120 -d 0.2 -f 4000
@@ -123,11 +123,11 @@ system:capture_1
 system:capture_2
 system:playback_1
 system:playback_2
-ffmpeg:input_1
+libav:input_1
 metro:120_bpm
 
-# Connect metro to the ffmpeg writable client.
-$ jack_connect metro:120_bpm ffmpeg:input_1
+# Connect metro to the avconv writable client.
+$ jack_connect metro:120_bpm libav:input_1
 @end example
 
 For more information read:
@@ -145,10 +145,10 @@ The filename to provide to the input device is the device node
 representing the OSS input device, and is usually set to
 @file{/dev/dsp}.
 
-For example to grab from @file{/dev/dsp} using @file{ffmpeg} use the
+For example to grab from @file{/dev/dsp} using @command{avconv} use the
 command:
 @example
-ffmpeg -f oss -i /dev/dsp /tmp/oss.wav
+avconv -f oss -i /dev/dsp /tmp/oss.wav
 @end example
 
 For more information about OSS see:
@@ -248,10 +248,10 @@ The filename to provide to the input device is the device node
 representing the sndio input device, and is usually set to
 @file{/dev/audio0}.
 
-For example to grab from @file{/dev/audio0} using @file{ffmpeg} use the
+For example to grab from @file{/dev/audio0} using @command{avconv} use the
 command:
 @example
-ffmpeg -f sndio -i /dev/audio0 /tmp/oss.wav
+avconv -f sndio -i /dev/audio0 /tmp/oss.wav
 @end example
 
 @section video4linux and video4linux2
@@ -290,7 +290,7 @@ avplay -f video4linux2 /dev/video0
 # Grab and record the input of a video4linux2 device, autoadjust size,
 # frame rate value defaults to 0/0 so it is read from the video4linux2
 # driver.
-ffmpeg -f video4linux2 -i /dev/video0 out.mpeg
+avconv -f video4linux2 -i /dev/video0 out.mpeg
 @end example
 
 @section vfwcap
@@ -326,12 +326,12 @@ Check the X11 documentation (e.g. man X) for more detailed information.
 Use the @file{dpyinfo} program for getting basic information about the
 properties of your X11 display (e.g. grep for "name" or "dimensions").
 
-For example to grab from @file{:0.0} using @file{ffmpeg}:
+For example to grab from @file{:0.0} using @command{avconv}:
 @example
-ffmpeg -f x11grab -r 25 -s cif -i :0.0 out.mpg
+avconv -f x11grab -r 25 -s cif -i :0.0 out.mpg
 
 # Grab at position 10,20.
-ffmpeg -f x11grab -r 25 -s cif -i :0.0+10,20 out.mpg
+avconv -f x11grab -r 25 -s cif -i :0.0+10,20 out.mpg
 @end example
 
 @subsection @var{follow_mouse} AVOption
@@ -348,10 +348,10 @@ zero) to the edge of region.
 
 For example:
 @example
-ffmpeg -f x11grab -follow_mouse centered -r 25 -s cif -i :0.0 out.mpg
+avconv -f x11grab -follow_mouse centered -r 25 -s cif -i :0.0 out.mpg
 
 # Follows only when the mouse pointer reaches within 100 pixels to edge
-ffmpeg -f x11grab -follow_mouse 100 -r 25 -s cif -i :0.0 out.mpg
+avconv -f x11grab -follow_mouse 100 -r 25 -s cif -i :0.0 out.mpg
 @end example
 
 @subsection @var{show_region} AVOption
@@ -367,10 +367,10 @@ being grabbed if only a portion of the screen is grabbed.
 
 For example:
 @example
-ffmpeg -f x11grab -show_region 1 -r 25 -s cif -i :0.0+10,20 out.mpg
+avconv -f x11grab -show_region 1 -r 25 -s cif -i :0.0+10,20 out.mpg
 
 # With follow_mouse
-ffmpeg -f x11grab -follow_mouse centered -show_region 1  -r 25 -s cif -i :0.0 out.mpg
+avconv -f x11grab -follow_mouse centered -show_region 1  -r 25 -s cif -i :0.0 out.mpg
 @end example
 
 @c man end INPUT DEVICES
diff --git a/doc/muxers.texi b/doc/muxers.texi
index bcba77c..17b1de3 100644
--- a/doc/muxers.texi
+++ b/doc/muxers.texi
@@ -35,20 +35,20 @@ CRC=0x at var{CRC}, where @var{CRC} is a hexadecimal number 0-padded to
 For example to compute the CRC of the input, and store it in the file
 @file{out.crc}:
 @example
-ffmpeg -i INPUT -f crc out.crc
+avconv -i INPUT -f crc out.crc
 @end example
 
 You can print the CRC to stdout with the command:
 @example
-ffmpeg -i INPUT -f crc -
+avconv -i INPUT -f crc -
 @end example
 
-You can select the output format of each frame with @file{ffmpeg} by
+You can select the output format of each frame with @command{avconv} by
 specifying the audio and video codec and format. For example to
 compute the CRC of the input audio converted to PCM unsigned 8-bit
 and the input video converted to MPEG-2 video, use the command:
 @example
-ffmpeg -i INPUT -acodec pcm_u8 -vcodec mpeg2video -f crc -
+avconv -i INPUT -c:a pcm_u8 -c:v mpeg2video -f crc -
 @end example
 
 See also the @ref{framecrc} muxer.
@@ -71,21 +71,21 @@ number 0-padded to 8 digits containing the CRC of the decoded frame.
 For example to compute the CRC of each decoded frame in the input, and
 store it in the file @file{out.crc}:
 @example
-ffmpeg -i INPUT -f framecrc out.crc
+avconv -i INPUT -f framecrc out.crc
 @end example
 
 You can print the CRC of each decoded frame to stdout with the command:
 @example
-ffmpeg -i INPUT -f framecrc -
+avconv -i INPUT -f framecrc -
 @end example
 
-You can select the output format of each frame with @file{ffmpeg} by
+You can select the output format of each frame with @command{avconv} by
 specifying the audio and video codec and format. For example, to
 compute the CRC of each decoded input audio frame converted to PCM
 unsigned 8-bit and of each decoded input video frame converted to
 MPEG-2 video, use the command:
 @example
-ffmpeg -i INPUT -acodec pcm_u8 -vcodec mpeg2video -f framecrc -
+avconv -i INPUT -c:a pcm_u8 -c:v mpeg2video -f framecrc -
 @end example
 
 See also the @ref{crc} muxer.
@@ -119,26 +119,26 @@ The pattern "img%%-%d.jpg" will specify a sequence of filenames of the
 form @file{img%-1.jpg}, @file{img%-2.jpg}, ..., @file{img%-10.jpg},
 etc.
 
-The following example shows how to use @file{ffmpeg} for creating a
+The following example shows how to use @command{avconv} for creating a
 sequence of files @file{img-001.jpeg}, @file{img-002.jpeg}, ...,
 taking one image every second from the input video:
 @example
-ffmpeg -i in.avi -r 1 -f image2 'img-%03d.jpeg'
+avconv -i in.avi -vsync 1 -r 1 -f image2 'img-%03d.jpeg'
 @end example
 
-Note that with @file{ffmpeg}, if the format is not specified with the
+Note that with @command{avconv}, if the format is not specified with the
 @code{-f} option and the output filename specifies an image file
 format, the image2 muxer is automatically selected, so the previous
 command can be written as:
 @example
-ffmpeg -i in.avi -r 1 'img-%03d.jpeg'
+avconv -i in.avi -vsync 1 -r 1 'img-%03d.jpeg'
 @end example
 
 Note also that the pattern must not necessarily contain "%d" or
 "%0 at var{N}d", for example to create a single image file
 @file{img.jpeg} from the input video you can employ the command:
 @example
-ffmpeg -i in.avi -f image2 -vframes 1 img.jpeg
+avconv -i in.avi -f image2 -frames:v 1 img.jpeg
 @end example
 
 @section mpegts
@@ -171,7 +171,7 @@ and @code{service_name}. If they are not set the default for
 @code{service_name} is "Service01".
 
 @example
-ffmpeg -i file.mpg -acodec copy -vcodec copy \
+avconv -i file.mpg -c copy \
      -mpegts_original_network_id 0x1122 \
      -mpegts_transport_stream_id 0x3344 \
      -mpegts_service_id 0x5566 \
@@ -189,19 +189,19 @@ Null muxer.
 This muxer does not generate any output file, it is mainly useful for
 testing or benchmarking purposes.
 
-For example to benchmark decoding with @file{ffmpeg} you can use the
+For example to benchmark decoding with @command{avconv} you can use the
 command:
 @example
-ffmpeg -benchmark -i INPUT -f null out.null
+avconv -benchmark -i INPUT -f null out.null
 @end example
 
 Note that the above command does not read or write the @file{out.null}
-file, but specifying the output file is required by the @file{ffmpeg}
+file, but specifying the output file is required by the @command{avconv}
 syntax.
 
 Alternatively you can write the command as:
 @example
-ffmpeg -benchmark -i INPUT -f null -
+avconv -benchmark -i INPUT -f null -
 @end example
 
 @section matroska
@@ -264,7 +264,7 @@ Both eyes laced in one Block, Right-eye view is first
 
 For example a 3D WebM clip can be created using the following command line:
 @example
-ffmpeg -i sample_left_right_clip.mpg -an -vcodec libvpx -metadata STEREO_MODE=left_right -y stereo_clip.webm
+avconv -i sample_left_right_clip.mpg -an -c:v libvpx -metadata STEREO_MODE=left_right -y stereo_clip.webm
 @end example
 
 @c man end MUXERS
diff --git a/doc/protocols.texi b/doc/protocols.texi
index f06546c..225fda8 100644
--- a/doc/protocols.texi
+++ b/doc/protocols.texi
@@ -67,10 +67,10 @@ File access protocol.
 
 Allow to read from or read to a file.
 
-For example to read from a file @file{input.mpeg} with @file{ffmpeg}
+For example to read from a file @file{input.mpeg} with @command{avconv}
 use the command:
 @example
-ffmpeg -i file:input.mpeg output.mpeg
+avconv -i file:input.mpeg output.mpeg
 @end example
 
 The ff* tools default to the file protocol, that is a resource
@@ -109,10 +109,10 @@ be used to test muxers without writing an actual file.
 Some examples follow.
 @example
 # Write the MD5 hash of the encoded AVI file to the file output.avi.md5.
-ffmpeg -i input.flv -f avi -y md5:output.avi.md5
+avconv -i input.flv -f avi -y md5:output.avi.md5
 
 # Write the MD5 hash of the encoded AVI file to stdout.
-ffmpeg -i input.flv -f avi -y md5:
+avconv -i input.flv -f avi -y md5:
 @end example
 
 Note that some formats (typically MOV) require the output protocol to
@@ -134,18 +134,18 @@ pipe (e.g. 0 for stdin, 1 for stdout, 2 for stderr).  If @var{number}
 is not specified, by default the stdout file descriptor will be used
 for writing, stdin for reading.
 
-For example to read from stdin with @file{ffmpeg}:
+For example to read from stdin with @command{avconv}:
 @example
-cat test.wav | ffmpeg -i pipe:0
+cat test.wav | avconv -i pipe:0
 # ...this is the same as...
-cat test.wav | ffmpeg -i pipe:
+cat test.wav | avconv -i pipe:
 @end example
 
-For writing to stdout with @file{ffmpeg}:
+For writing to stdout with @command{avconv}:
 @example
-ffmpeg -i test.wav -f avi pipe:1 | cat > test.avi
+avconv -i test.wav -f avi pipe:1 | cat > test.avi
 # ...this is the same as...
-ffmpeg -i test.wav -f avi pipe: | cat > test.avi
+avconv -i test.wav -f avi pipe: | cat > test.avi
 @end example
 
 Note that some formats (typically MOV), require the output protocol to
@@ -219,9 +219,9 @@ meaning as specified for the RTMP native protocol.
 See the librtmp manual page (man 3 librtmp) for more information.
 
 For example, to stream a file in real-time to an RTMP server using
- at file{ffmpeg}:
+ at command{avconv}:
 @example
-ffmpeg -re -i myfile -f flv rtmp://myserver/live/mystream
+avconv -re -i myfile -f flv rtmp://myserver/live/mystream
 @end example
 
 To play the same stream using @file{avplay}:
@@ -249,7 +249,7 @@ The required syntax for a RTSP url is:
 rtsp://@var{hostname}[:@var{port}]/@var{path}
 @end example
 
-The following options (set on the @file{avconv}/@file{avplay} command
+The following options (set on the @command{avconv}/@file{avplay} command
 line, or set in code via @code{AVOption}s or in @code{avformat_open_input}),
 are supported:
 
@@ -310,7 +310,7 @@ avplay -rtsp_transport http rtsp://server/video.mp4
 To send a stream in realtime to a RTSP server, for others to watch:
 
 @example
-ffmpeg -re -i @var{input} -f rtsp -muxdelay 0.1 rtsp://server/live.sdp
+avconv -re -i @var{input} -f rtsp -muxdelay 0.1 rtsp://server/live.sdp
 @end example
 
 @section sap
@@ -362,19 +362,19 @@ Example command lines follow.
 To broadcast a stream on the local subnet, for watching in VLC:
 
 @example
-ffmpeg -re -i @var{input} -f sap sap://224.0.0.255?same_port=1
+avconv -re -i @var{input} -f sap sap://224.0.0.255?same_port=1
 @end example
 
 Similarly, for watching in avplay:
 
 @example
-ffmpeg -re -i @var{input} -f sap sap://224.0.0.255
+avconv -re -i @var{input} -f sap sap://224.0.0.255
 @end example
 
 And for watching in avplay, over IPv6:
 
 @example
-ffmpeg -re -i @var{input} -f sap sap://[ff0e::1:2:3:4]
+avconv -re -i @var{input} -f sap sap://[ff0e::1:2:3:4]
 @end example
 
 @subsection Demuxer
@@ -420,7 +420,7 @@ tcp://@var{hostname}:@var{port}[?@var{options}]
 Listen for an incoming connection
 
 @example
-ffmpeg -i @var{input} -f @var{format} tcp://@var{hostname}:@var{port}?listen
+avconv -i @var{input} -f @var{format} tcp://@var{hostname}:@var{port}?listen
 avplay tcp://@var{hostname}:@var{port}
 @end example
 
@@ -472,21 +472,21 @@ For receiving, this gives the benefit of only receiving packets from
 the specified peer address/port.
 @end table
 
-Some usage examples of the udp protocol with @file{ffmpeg} follow.
+Some usage examples of the udp protocol with @command{avconv} follow.
 
 To stream over UDP to a remote endpoint:
 @example
-ffmpeg -i @var{input} -f @var{format} udp://@var{hostname}:@var{port}
+avconv -i @var{input} -f @var{format} udp://@var{hostname}:@var{port}
 @end example
 
 To stream in mpegts format over UDP using 188 sized UDP packets, using a large input buffer:
 @example
-ffmpeg -i @var{input} -f mpegts udp://@var{hostname}:@var{port}?pkt_size=188&buffer_size=65535
+avconv -i @var{input} -f mpegts udp://@var{hostname}:@var{port}?pkt_size=188&buffer_size=65535
 @end example
 
 To receive over UDP from a remote endpoint:
 @example
-ffmpeg -i udp://[@var{multicast-address}]:@var{port}
+avconv -i udp://[@var{multicast-address}]:@var{port}
 @end example
 
 @c man end PROTOCOLS



More information about the ffmpeg-cvslog mailing list