+

WO2006009275A1 - Procede et systeme d'edition de fichier audiovisuels - Google Patents

Procede et systeme d'edition de fichier audiovisuels Download PDF

Info

Publication number
WO2006009275A1
WO2006009275A1 PCT/JP2005/013540 JP2005013540W WO2006009275A1 WO 2006009275 A1 WO2006009275 A1 WO 2006009275A1 JP 2005013540 W JP2005013540 W JP 2005013540W WO 2006009275 A1 WO2006009275 A1 WO 2006009275A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
audiovisual
frames
tab
video
Prior art date
Application number
PCT/JP2005/013540
Other languages
English (en)
Inventor
Pi-Chung Hsu
Hon-Wen Pon
Original Assignee
Matsushita Electric Industrial Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Matsushita Electric Industrial Co., Ltd. filed Critical Matsushita Electric Industrial Co., Ltd.
Publication of WO2006009275A1 publication Critical patent/WO2006009275A1/fr

Links

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/32Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23424Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/23614Multiplexing of additional data and video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/2368Multiplexing of audio and video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/242Synchronization processes, e.g. processing of PCR [Program Clock References]

Definitions

  • the invention relates to a method and system for editing audiovisual files, more particularly to a method and system for editing audiovisual files which can effectively prevent accumulation and propagation of stream errors during repeated editing.
  • Motion Pictures Experts Group is a system set up by the International Standards Organization (ISO) specifically for digital video and audio compression.
  • ISO International Standards Organization
  • bit streams When bit streams are encoded, the system requires a minimum number of rules to be followed during encoding so that the receiver can unambiguously decode the encoded bit streams received thereby.
  • a bit stream generally includes a video component, an audio component, and a system component.
  • the system component defines information of how the video and audio components in a single bit stream can be combined and synchronized. Therefore, the MPEG standard defines a system for compressing encoded video and audio bit streams.
  • the audio components can hardly be synchronized.
  • the audio frames may possibly change in size. Therefore, the synchronization problem is partly due to the fact that there is rarely a one-to-one correlation between the audio frames and the video frames.
  • the number of audio frames corresponding to the identified video frames will not be a predetermined value. Accordingly, when a video segment is copied from a file and is joined to another copied segment, the audio components in the copied segment might no longer be synchronized with the corresponding video frames.
  • an error representing a synchronization discrepancy between the video and audio frames will be introduced into the resultant bit stream.
  • This error can be expressed in terms of the number or percentage of the audio frames.
  • a synchronization error that is introduced by the joining of two bit stream segments may be as small as a fraction of an audio frame or as large as several audio frames.
  • the error associated with the joining of only two bit stream segments may be merely a few audio frames under certain conditions, when a number of bit stream segments are joined in a more complicated editing task, the errors of the connected segments are summed. Therefore, the errors thus generated may be very large, and the resultant audio frames may be seriously un-synchronized that they cannot be played back.
  • U.S. Patent No. 6,262,777 discloses a conventional method for synchronizing edited audiovisual files. This method can prevent bit stream errors from exceeding half an audio frame regardless of the number of joined segments after a continuous copying operation so that the video frames are substantially synchronized with the audio frames.
  • tabs the tab-in audio frames and the tab-out audio frames are referred to as tabs.
  • the tab processing flow in Figure 1 will be illustrated by way of an example with reference to Figures 2 and 3 (i.e., Figures 14 and 15 of the aforesaid U.S. Patent No. 6,262,777).
  • Figures 2 and 3 i.e., Figures 14 and 15 of the aforesaid U.S. Patent No. 6,262,777.
  • four segments SEGMENT A, SEGMENT B, SEGMENT C, and SEGMENT D are to be connected together.
  • step 602 the existing stream errors present prior to processing of the tab 706 are determined.
  • SEGMENT A is the first segment, and there is no previous tab that introduces an existing stream error. Therefore, the existing stream error is 0.
  • step 604 is executed to determine the tab error of the tab 706. As shown in Figure 3, the tab error of the tab 706 is assumed to be 0.2 (i.e., 20% of an audio frame).
  • step 606 it is determined whether summation (herein referred to as cumulative error) of the existing stream error and the tab error is greater than half a frame (i.e., 0.5). If it is determined in step 606 that the summation is greater than half a frame, step 610 is executed to drop the tab processed in the aforesaid step 604. Conversely, if it is determined in step 606 that the summation is not greater than half a frame, step 608 is executed to retain the tab processed in the aforesaid step 604. Step 612 is executed after steps 608 and 610 are ended. In step 606, since the cumulative error of the tab 706 is 0.2, which is smaller than 0.5, step 608 is executed to retain the tab 706.
  • step 612 it is determined whether there are other tabs. If it is determined in step 612 that there are other tabs, the flow skips back to step 602. Otherwise, the tab processing flow is ended. In this example, since the aforementioned steps are concerned with the processing of the tab 706 only, it is determined in step 612 that there are other tabs, and the flow skips back to step 602.
  • step 602 It is determined in step 602 that the existing stream error of the tab 708 is 0.2, whereas it is determined in step 604 that the tab error of the tab 708 is 0.5.
  • step 606 the flow will skip to step 610 to drop the tab 708 because the cumulative error of the tab 708 is 0.7.
  • steps 602-612 are repeated to continue processing of the tabs 710, 712, 714, 716, 718, and 720 until all of the tabs have been processed. After tab processing, all the segments SEGMENT A, SEGMENT B, SEGMENT C, and SEGMENT D can be connected together.
  • the conventional method utilizes the tab processing operation as described above to ensure that the stream error does not exceed 0.5 audio frame so that the video frames are substantially synchronized with the audio frames regardless of the number of segments that are joined during a copying operation.
  • the conventional method can effectively prevent accumulation of errors during a single editing task of the audiovisual segments, growth and propagation of cumulative stream errors are hardly avoided when these edited audiovisual segments are subjected to repeated editing tasks.
  • the original audiovisual segment (s) shown in Figure 4 is connected to another audiovisual segment (X).
  • the audio frame (a) will be dropped, and it is assumed that the current existing stream error is -0.5 audio frame. Therefore, in Figure 5, the tab- processed audiovisual segment (s) is joined to the audiovisual segment (X), and all the audio frames will be shifted to the left by 0.5 audio frame beginning from (b) so as to form a new audiovisual segment (s 1 ).
  • the second editing is commenced to cut an audiovisual segment beginning from a video frame 2 from the audiovisual segment (s') for joining to another audiovisual segment (Y).
  • the audio frame (d) will be dropped, and the current existing stream error is set to -0.5 audio frame.
  • the cut audiovisual segment is connected to the audiovisual segment (Y)
  • all the audio frames will be shifted to the left by 0.5 audio frame beginning from the audio frame (e) so as to form a new audiovisual segment (s"). Therefore, the error of the audiovisual segment that has been edited for a second time becomes one audio frame, i.e.,
  • 1.
  • an audiovisual segment beginning from the video frame 4 is cut from the audiovisual segment (s"'), and the audio frame (j) is dropped and the existing stream error is set to -0.5 during tab processing.
  • the processed audiovisual segment is connected to a further audiovisual segment (V)
  • the audio frames as a whole will be shifted to the left by 0.5 audio frame beginning from the audio frame 5 so as to form a new audiovisual segment (s" ⁇ ).
  • the stream error thus accumulated after four editing operations is increased to 2, i.e.,
  • 2.
  • the audio frames which lie below the video frame 5 in the unedited audiovisual segment (s) are three audio frames 5.
  • two of the three audio frames 5 of the audiovisual segment (s"" have shifted to the left to below the video frame 4 after the fourth editing operation.
  • the conventional method is still unable to inhibit the growth and propagation of stream errors during repeated editing, so that the audio components in the bit stream that was edited repeatedly might no longer be synchronized with the proper video components.
  • the inventors of this application contemplated the logging or recording of the amount of shifting of audio frames during joining of audiovisual segments so that, during repeated editing, the audio frames are first restored to their positions prior to the joining of the audiovisual segments based on the recorded amount of the shifting of the audio frames before further tab processing and segment joining are performed, thereby avoiding accumulation of stream errors during repeated editing.
  • the object of the present invention is to provide a method and system for editing audiovisual files, which can effectively avoid accumulation and propagation of stream errors during repeated editing.
  • a method for editing audiovisual files of this invention is adapted to process an input bit stream according to a copy task.
  • the input bit stream has a plurality of video frames and a plurality of audio frames.
  • the method comprises the following steps:
  • This invention utilizes the synchronization error table to correct the positional relationship between the video and audio frames before performing tab processing and segment joining so as to achieve the effects of effectively avoiding accumulation and propagation of stream errors during repeated editing.
  • FIG. 1 is a flowchart of tab processing according to U.S. Patent No. 6,262,777;
  • Figure 2 is a schematic diagram of an example of a plurality of audiovisual segments to be joined together according to the method of U.S. Patent No. 6,262,777;
  • Figure 3 is a table showing calculation results during processing of a plurality of the tabs shown in Figure 2;
  • Figures 4 to 8 are schematic diagrams of the audiovisual segments when the audiovisual segments are subjected to repeated editing according to the method of U.S. Patent No. 6,262,777;
  • Figure 9 is a schematic diagram of the preferred embodiment of a system for editing audiovisual files according to this invention.
  • Figure 10 is a schematic diagram of a bit stream in this invention
  • Figure 11 is a flowchart of the preferred embodiment of a method for editing audiovisual files according to this invention
  • Figure 12 is a flowchart of a copying operation in Figure 1 1 ;
  • Figure 13 is a schematic diagram showing a bit stream cut from that of Figure 10 according to the flowchart of Figure 12;
  • Figure 14 is a schematic diagram illustrating copying of an audiovisual segment according to the flowchart of Figure 12;
  • Figure 15 is a flowchart of tab processing in the flowchart of Figure 11 ;
  • Figure 16 is a schematic view showing how a plurality of audiovisual segments are joined together during tab processing in this preferred embodiment.
  • Figures 17 to 22 are schematic diagrams of audiovisual segments that are subjected to repeated editing according to this invention.
  • an editing engine 11 for executing audiovisual file editing, and a storage medium 12 such as a hard disk.
  • the editing engine 11 is used to edit one or more input bit streams.
  • the editing engine 11 is exemplified herein as one that is used for editing two input bit streams, i.e., A.MPEG and B. MPEG.
  • the editing engine 11 can also be used to edit one or more input bit streams, and is not limited to that which is disclosed in this embodiment.
  • an audiovisual bit stream 13 generally has a video bit stream 131 and an audio bit stream 132.
  • the video bit stream 131 has a plurality of video frames, each of which has an exclusive serial number.
  • the audio bit stream 132 has a plurality of audio frames.
  • the storage medium 12 stores an edit list 121.
  • the edit list 121 records a plurality of copy tasks inputted by an operator. Each copy task contains a specific input bit stream, a start serial number, and an end serial number, e.g., COPY 10...25 A.MPEG.
  • the storage medium 12 is also provided for access of data by the editing engine 11.
  • the system for editing audiovisual files in this embodiment is an open system, which receives the aforesaid input bit streams from the outside and which outputs edited output bit streams to the outside. Therefore, in step 21 , the editing engine 11 receives the edit list 121 inputted by the operator, and stores the edit list 121 temporarily in the storage medium 12. Next, in step 22, the editing engine 11 generates a brand new synchronization error table 122 for use in subsequent processing.
  • step 23 the editing engine 11 executes each of the copy tasks in the edit list 121 in sequence so as to generate a corresponding copying operation 111 , such as an object, to decide the number of audio frames to be copied, i.e., the number of audio frames having a tab-in and a tab-out, so as to generate an audiovisual segment, thereby forming an output bit stream.
  • the first audio frame i.e., the tab-in audio frame
  • the first video frame i.e., a mark-in video frame
  • step 24 is performed to insert the synchronization error table 122 into the output bit stream.
  • the synchronization error table 122 is used to record stream errors of the audio frames that have undergone tab processing.
  • the job executed by the copying operation 111 as generated by the editing engine 11 generally includes copying and tab processing.
  • the task flowchart of the copying operation in step 23 will be first described in detail with reference to Figure 12.
  • the copying operation 111 selects a mark-in position and a mark-out position from the video bit stream 131 based on the start serial number and end serial number specified in the copy task so as to identify a mark-in video frame and a mark-out video frame to thereby decide the number of video frames to be copied (i.e., the video frames from the mark-in video position to the mark-out video position).
  • the start serial number and the end serial number are specified as 10 and 25, respectively
  • the mark-in position is video frame 10
  • the mark-out position is video frame 25.
  • the video frame 10 is the mark-in video frame
  • the video frame 25 is the mark-out video frame.
  • the segment to be copied includes the frames from the video frame 10 to the video frame 25.
  • the audiovisual segment to be copied will also include the audio frames associated with the above-identified video frames. These audio frames lie below their associated video frames.
  • the synchronization error table 122 will be used to record the stream errors, and the synchronization error table 122 will be inserted into the bit stream at a suitable position (as in step 24), e.g., the position of a system component in the bit stream or a header or the like in an audiovisual segment, such as PES_private_date in PES_header.
  • step 32 it is first determined whether a synchronization error table is present in the input bit stream. If it is determined in step 32 that there is a synchronization error table, this means that the input bit stream has been previously edited, such as an input bit stream A. MPEG, and there are stream errors between the audio bit stream 132 and the video bit stream 131. Therefore, step 33 is first executed. If it is determined in step 32 that there is not a synchronization error table, this means that the input bit stream has yet to be edited, such as an input bit stream B. MPEG, and step 36 is executed right away.
  • step 33 the copying operation 111 first loads the synchronization error table present in the input bit stream.
  • This synchronization error table is different from the synchronization error table 122 generated in step 22.
  • the synchronization error table in this example records the numbers of the audio frames that were shifted in the previous editing, and the stream errors after the shifting thereof.
  • only those audio frames that are tab audio frames are recorded therein. This is because, for an audio segment that is selected for editing, shifting of the entire audio segment starts from a tab audio frame thereof. Since the amount of shifting of each of the audio frames in the entire audio segment will be consistent, it will be sufficient to record only the tab audio frames.
  • the audiovisual bit stream 13 has been edited, and may be composed of a plurality of interconnected audiovisual segments such that the synchronization error table records several entries of audio frames and the corresponding stream errors thereof.
  • the video bit stream 131 is divided into a plurality of sub-audiovisual segments based on the serial numbers of the audio frames recorded in the synchronization error table. This means that the audiovisual bit stream 13 is split at the positions of the audio frames recorded in the synchronization error table.
  • the video frame located above the recorded audio frame is separated from an immediately preceding video frame so as to form a sub-audiovisual segment. If there are two video frames above the audio frame, these two video frames are separated.
  • the copying operation 111 selects an associated correction tab audio frame for and based on each of the two adjacent video frames that are to be split.
  • the split two adjacent video frames refer to the video frame that is located above the audio frame recorded in the synchronization error table and its immediately preceding video frame, or the two video frames that are located above the recorded audio frame.
  • the length of the audio segment preferably exceeds or is equal to the length of the video segment, if the video frame is located at an earlier time, the end time of the selected audio frame is required to be the same as or later than the end time of the video frame that is at the earlier time. If the video frame is located at a later time, the start time of the selected audio frame must be the same as or earlier than the start time of the video frame that is at the later time.
  • the audio frame located below the splitting position may be duplicated so that the two video frames are associated with the aforesaid audio frame.
  • step 35 the positional relationship between the video and audio frames in each sub-audiovisual segment is corrected based on the corresponding stream error of each recorded audio frame.
  • Step 36 is executed after step 35 is ended.
  • step 34 as shown in Figure 13, since video frame 9 is above the audio frame 134 and since video frames 19, 20 are above audio frame 135, the copying operation 111 will execute splitting of the audiovisual bit stream 13 between video frames 8, 9 and between video frames 19, 20 to thereby form three sub-audiovisual segments 136, 137, 138.
  • the audio frame at the start position of the sub-audiovisual segment 136 which serves as a correction tab audio frame, is the audio frame 139 that precedes the audio frame 134; the audio frame at the end position of the sub- audiovisual segment 136 is a duplicated audio frame 135; and the audio frame at the start position of the sub-audiovisual segment 138 is also the audio frame 135.
  • the audio frame 135 is duplicated.
  • step 35 the positional relationship between the video and audio frames in each sub-audiovisual segment 137, 138 is corrected according to the corresponding stream error of each audio frame 134, 135. Specifically, since the first sub-audiovisual segment 136 does not have any corresponding error information, the relationship between the video and audio frames therein is maintained. Since the stream error of the first audio frame 134 in the audio bit stream in the second sub- audiovisual segment 137 is 0.3 audio frame, the entire audio bit stream in the sub-audiovisual segment 137 is shifted to the left by 0.3 audio frame to eliminate the stream error caused to the audiovisual segment 137 during the previous editing.
  • the audio frame 135 in the third sub-audiovisual segment 138 since the stream error is -0.2 audio frame, the entire audio bit stream in the sub-audiovisual segment 138 is shifted to the right by 0.2 audio frame to correct the positional relationship between the video and audio frames.
  • the correction of the positional relationship between the video frames and the audio frames using the stream error information in the synchronization error table in the input bit stream in steps 33-35 is to restore the audio frames to their positions prior to their being edited.
  • a tab-in audio frame and a tab-out audio frame are selected based on the mark-in video frame and the mark-out video frame so as to determine the size of the audio segment to be copied.
  • the start time of the tab-in audio frame which serves as the initial audio frame 143, is preferably equal to or earlier than the start time 141 of the mark-in video frame 10
  • the start time of the tab-out audio frame which serves as the end audio frame 144, is preferably equal to or earlier than the end time 142 of the mark-out video frame 25.
  • the tab-in audio frame 143 has an earlier start time compared to the start time 141 of the mark-in video frame 10
  • the tab-out audio frame 144 has an earlier start time compared to the end time 142 of the mark-out video frame 25.
  • step 37 the selected audiovisual segment (i.e., the segment containing the selected video frames 10-25 and audio frames 143-144) is outputted after tab processing (to be described in detail hereinafter).
  • the copying operation 111 After the copying operation 111 has executed step 37, the copying operation 111 will be terminated.
  • the stream error generated in the previous editing operation can be eliminated beforehand so as to achieve the effect of effectively avoiding accumulation of stream errors due to repeated editing.
  • the determination of the presence of the synchronization error table and the corresponding position correcting process (as in steps 32-35) in this embodiment are performed after selecting the video frame segment to be copied (as in step 31 ), it should be apparent to those skilled in the art that the determination of the presence of the synchronization error table and the corresponding position correcting process can be executed before selecting the video frame segment to be copied.
  • the operation of selecting the correction tab audio frame in step 34 may be executed together with the selection of audio frames in step 36, and should not be limited to that which is disclosed in this embodiment.
  • the editing engine 11 may generate a new copying operation 1 11 according to a next copy task in the edit list 121 so as to execute the next copy task.
  • the copying operation 111 executes the copy task, it will be terminated after outputting the copied audiovisual segment so that the editing engine 11 continues to generate copying operations 111 to execute the next copy task. If the audiovisual segment in the subsequent copy task needs to be joined to the audiovisual segment of the previous copy task, the stream error of the previous audiovisual segment will be considered during tab processing of the audiovisual segment generated by the subsequent copying operation. Therefore, in the following description, the tab processing mentioned in step 37 of the aforesaid copying operation will be described by way of an example with reference to Figures 15 and 16.
  • each of the sub-audiovisual segments 15, 17 has a tab-in audio frame (hereinafter referred to as tab) 151 , 171 , and a correction tab audio frame (hereinafter referred to as tab) 152, 172 at the start and the end of the audio segment, respectively.
  • the audio segment in each sub-audiovisual segment 16, 18 also has a correction tab audio frame 161 , 181 and a tab-out audio frame 162, 182 (hereinafter referred to as tabs) at the start and the end thereof, respectively.
  • step 41 the existing stream error of the current tab 151 is determined. As shown in Figure 15, the current tab 151 belongs to the first segment 15, and there is no previous tab to introduce an existing stream error. Therefore, the existing stream error is 0.
  • step 42 the tab error of the tab 151 is determined.
  • the tab error of the tab 151 is assumed to be 0.3.
  • step 43 it is determined whether sum of the existing stream error and the tab error (herein referred to as cumulative error) is greater than half a frame (i.e., 0.5). If so determined in step 43, step 45 is executed to drop the tab 151 processed in the previous step 42. Otherwise, step 44 is executed to retain the tab 151 processed in the previous step 42. According to this example, in step 43, since the cumulative error of the tab 151 is 0.3, which is smaller than 0.5, step 44 is executed to retain the tab 151. In addition, the cumulative error of the tab 151 is also referred to as a new stream error. After executing step 44 or 45, step 46 is executed.
  • step 46 the new stream error of the tab 151 is recorded in the synchronization error table 122 generated in the aforesaid step 22 so as to update the synchronization error table 122.
  • the new stream error of the tab 151 is recorded in the synchronization error table 122.
  • the shifting (i.e., stream error) of the two tabs which are respectively at the start and end positions of the same audiovisual segment will be the same during joining, it is only necessary to record at least one of the two tabs in order to permit restoration of the positional relationship between the audio segment and the video segment.
  • the stream error of the tab 151 at the start position is selected for recording in the synchronization error table 122.
  • Each entry of the recorded stream error is about 4-6 bytes.
  • step 47 it is determined in step 47 whether there are other tabs. If it is determined that there are other tabs in step 47, the flow returns to step 41. Otherwise, the tab processing flow is ended. In this example, since only the tab 151 was processed using the aforesaid steps, it will be determined in step 47 that there are other tabs, and the flow will return to step 41.
  • step 41 It is determined in step 41 that the existing stream error of the tab
  • step 47 since the tab 152 is one at the end position of the segment, the synchronization error table 122 is not updated, and execution of step 47 is continued to determine the next tab 161. Thereafter, steps 41 -47 are repeated until all the tabs have been processed.
  • the existing stream error of the tab 161 in step 41 is - 0.2
  • the tab error of the tab 161 is determined to be 0.3 in step 42. Since the cumulative error of the tab 161 is 0.1 , the tab 161 will be retained, and the cumulative error 0.1 (i.e., the new stream error) of the tab 161 will be recorded in the synchronization error table 122 (as shown in Table 2) in step 46.
  • the tabs 171 , 181 which are at the start positions of the third and fourth segments 17, 18, respectively, will be recorded in the synchronization error table 122.
  • the synchronization error table 122 can be concurrently updated to record the amount of shifting (i.e., stream error) of the audio frames in each of the sub-audiovisual segments 15, 16, 17, and 18 in the synchronization error table 122.
  • the editing engine 11 When the editing engine 11 subsequently outputs the processed bit stream (e.g., the joined sub- audiovisual segment 15, 16, 17, and 18), the editing engine 11 will insert the synchronization error table 122 into the output bit stream (as in step 24 of Figure 11 ) so that when the bit stream is re-edited, the stream error information in the synchronization error table 122 can be used to first correct the relationship between the audio and video frames before proceeding with the editing job so as to effectively avoid accumulation and propagation of stream errors.
  • the synchronization error table 122 into the output bit stream (as in step 24 of Figure 11 ) so that when the bit stream is re-edited, the stream error information in the synchronization error table 122 can be used to first correct the relationship between the audio and video frames before proceeding with the editing job so as to effectively avoid accumulation and propagation of stream errors.
  • the copying operation 111 will output the output bit stream after accomplishing processing of the audiovisual segment which was generated in response to the copy task.
  • the sub-audiovisual segments 15, 16 will be outputted to the output bit stream after tab processing, and the sub-audiovisual segments 17, 18 will be outputted to the output bit stream after tab processing in the next copying operation. How this embodiment can avoid stream error accumulation will be described by way of an example hereinbelow.
  • the original audiovisual segment 19 of Figure 17 is joined to another audiovisual segment (X).
  • the audio frame (a) is dropped, and the currently existing stream error is assumed to be -0.5 audio frame.
  • the audiovisual segment 19, which has been tab processed is joined to the audiovisual segment (X), and all the audio frames are shifted to the left by 0.5 audio frame beginning from the audio frame (b) so as to form a new audiovisual segment 19'.
  • the amount of shifting (i.e., stream error) of the audio frame (b), i.e., -0.5 will be recorded in the synchronization error table 122, such as (b), i.e., -0.5.
  • the editing engine 11 will insert the synchronization error table 122 into the bit stream.
  • a second editing operation is commenced to cut an audiovisual segment beginning from the video frame 2 from the audiovisual segment (s 1 ) for joining to another audiovisual segment (Y).
  • the synchronization error table present in the bit stream will be loaded before the selection of the audiovisual segment to be cut so that the stream error information can be used to correct the positional relationship between the audio and video frames.
  • the audio frame (c) will be dropped, and the currently existing stream error is set to 0.4 audio frame.
  • an audiovisual segment beginning from the video frame 3 is cut from the audiovisual segment 19" for joining to another audiovisual segment (Z).
  • an audiovisual segment is cut from the audiovisual segment 19"
  • the positional relationship between the audio and video frames will be first corrected using the synchronization error table. Then, in the subsequent tab processing operation, the audio frame (f) will be retained, and the stream error will be assumed to be 0.4.
  • the system for editing audiovisual files in the aforesaid embodiment is an open system so that the synchronization error table is located in the bit stream to facilitate communication with the outside
  • the synchronization error table 122 can be stored in the storage medium 12 if the system for editing audiovisual files is a closed system, and that it is only necessary to load the synchronization error table 122 into the storage medium 12 in order to correct the positional relationship between the video and audio frames.
  • the synchronization error table 122 in the storage medium 12 can be updated directly during tab processing, which differs from the aforesaid embodiment. In a closed system for editing audiovisual files, it is not necessary to insert the synchronization error table 122 into the bit stream.
  • the present invention utilizes a synchronization error table 122 to record stream error information so that, during repeated editing, the stream error information can be used to correct the positional relationship between the audio and video frames before proceeding with editing. Moreover, the new stream error information will be recorded in the newly generated synchronization error table after editing, and the synchronization error table recording the new stream error information is inserted into the edited output bit stream or stored in the storage medium 12, thereby achieving the effect of effectively avoiding accumulation and propagation of stream errors due to repeated editing.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Television Signal Processing For Recording (AREA)
  • Management Or Editing Of Information On Record Carriers (AREA)

Abstract

Un procédé d'édition de fichiers audiovisuels est utilisé pour éditer un flux binaire d'entrée en fonction d'au moins une tâche de copie. Le flux binaire d'entrée comprend une pluralité de trames vidéo et une pluralité de trames audio. Le procédé se caractérise en ce que, pendant l'exécution de la tâche de copie, une table d'erreurs de synchronisation est utilisée pour corriger tout d'abord la relation positionnelle entre les trames vidéo et les trames audio avant de poursuivre avec la sélection de la copie de manière à ce qu'on évite effectivement l'accumulation et la propagation des erreurs de flux au cours d'une édition répétée.
PCT/JP2005/013540 2004-07-19 2005-07-19 Procede et systeme d'edition de fichier audiovisuels WO2006009275A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN200410069912.0 2004-07-19
CN 200410069912 CN1725360A (zh) 2004-07-19 2004-07-19 编辑影音档案的方法及系统

Publications (1)

Publication Number Publication Date
WO2006009275A1 true WO2006009275A1 (fr) 2006-01-26

Family

ID=35106675

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2005/013540 WO2006009275A1 (fr) 2004-07-19 2005-07-19 Procede et systeme d'edition de fichier audiovisuels

Country Status (2)

Country Link
CN (1) CN1725360A (fr)
WO (1) WO2006009275A1 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1933323A1 (fr) 2006-12-15 2008-06-18 Fujitsu Limited Procédé et dispositif d'édition de fichier de contenu composite et appareil de reproduction
EP2104105A1 (fr) * 2008-03-20 2009-09-23 British Telecommunications Public Limited Company Codage numérique de clip audio et vidéo
EP2104103A1 (fr) * 2008-03-20 2009-09-23 British Telecommunications Public Limited Company Assemblage numérique des clip audio et vidéo

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101101753B (zh) * 2006-07-07 2011-04-20 乐金电子(昆山)电脑有限公司 音频帧识别方法

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6201924B1 (en) * 1990-09-28 2001-03-13 Adobe Systems Incorporated Disk-assisted editing for recorded video material
US6262777B1 (en) * 1996-11-15 2001-07-17 Futuretel, Inc. Method and apparatus for synchronizing edited audiovisual files
US20020047937A1 (en) * 1997-05-30 2002-04-25 Nicholas Dominic Wells Video and audio signal processing

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6201924B1 (en) * 1990-09-28 2001-03-13 Adobe Systems Incorporated Disk-assisted editing for recorded video material
US6262777B1 (en) * 1996-11-15 2001-07-17 Futuretel, Inc. Method and apparatus for synchronizing edited audiovisual files
US20020047937A1 (en) * 1997-05-30 2002-04-25 Nicholas Dominic Wells Video and audio signal processing

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1933323A1 (fr) 2006-12-15 2008-06-18 Fujitsu Limited Procédé et dispositif d'édition de fichier de contenu composite et appareil de reproduction
US8090682B2 (en) 2006-12-15 2012-01-03 Fujitsu Limited Method and device for editing composite content file and reproduction apparatus
US8433678B2 (en) 2006-12-15 2013-04-30 Fujitsu Limited Method and device for editing composite content file and reproduction apparatus
EP2104105A1 (fr) * 2008-03-20 2009-09-23 British Telecommunications Public Limited Company Codage numérique de clip audio et vidéo
EP2104103A1 (fr) * 2008-03-20 2009-09-23 British Telecommunications Public Limited Company Assemblage numérique des clip audio et vidéo
WO2009115801A1 (fr) * 2008-03-20 2009-09-24 British Telecommunications Public Limited Company Codage de clip audio et vidéo numérique
US20110007797A1 (en) * 2008-03-20 2011-01-13 Randall-Reilly Publishing Company, Llc Digital Audio and Video Clip Encoding

Also Published As

Publication number Publication date
CN1725360A (zh) 2006-01-25

Similar Documents

Publication Publication Date Title
EP1329108B1 (fr) Systeme et procede de traitement de flux mpeg pour l'insertion d'informations relatives a l'index des dossiers
US6262777B1 (en) Method and apparatus for synchronizing edited audiovisual files
US8285113B2 (en) Method for generating additional information for guaranteeing seamless playback between data streams, recording medium storing the information, and recording, editing and/or playback apparatus using the same
JP4285512B2 (ja) 記録装置、記録方法、再生装置、再生方法、記録再生装置、記録再生方法、撮像記録装置及び撮像記録方法
US20020034255A1 (en) System and method of processing MPEG streams for timecode packet insertion
US20020018645A1 (en) Information processing apparatus and method, and recording medium
JP3768662B2 (ja) オーディオビジュアル・セグメントを貼り合わせる方法および装置、オーディオビジュアル・セグメントを接合する方法、ならびにコンピュータが読み取り可能な媒体
AU7143391A (en) Media pipeline system
US7460766B2 (en) Recording medium having recorded thereon a data structure capable of linking and reproducing audio data and video data, and method and apparatus of recording/reproducing using the data structure
JP3665456B2 (ja) 映像情報の記録再生システム及び同システムに適用する映像編集方法
US20040141722A1 (en) Apparatus and method for video edition
US7215627B2 (en) Method and apparatus of processing audio information having system header
WO2006009275A1 (fr) Procede et systeme d'edition de fichier audiovisuels
EP1883072A2 (fr) Appareil pour l'édition d'un flux de données
CN1476719A (zh) 游艺时间表改编器
US20070263982A1 (en) Resume Reproduction System
US7526724B2 (en) Data processing apparatus and data processing method
CA2451997C (fr) Modification de signaux video par conversion en images non intra
US7020382B2 (en) Modifying video by inserting shadow intra pictures
JP2006324848A (ja) 情報処理装置及び情報処理方法
US8630533B2 (en) Editing apparatus, editing method and program
US20110008021A1 (en) Method for rapid archiving and restoring of a video store
US20090110364A1 (en) Reproduction apparatus and reproduction method
US20090214176A1 (en) Information processing apparatus, information processing method, and program
JP2001285800A (ja) データ信号記録再生装置およびデータ信号記録再生方法

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

NENP Non-entry into the national phase

Ref country code: DE

121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase
点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载