US20130124202A1 - Method and apparatus for processing scripts and related data - Google Patents
Method and apparatus for processing scripts and related data Download PDFInfo
- Publication number
- US20130124202A1 US20130124202A1 US12/789,708 US78970810A US2013124202A1 US 20130124202 A1 US20130124202 A1 US 20130124202A1 US 78970810 A US78970810 A US 78970810A US 2013124202 A1 US2013124202 A1 US 2013124202A1
- Authority
- US
- United States
- Prior art keywords
- script
- words
- sub
- dialogue
- ordered
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/44—Colour synchronisation
- H04N9/475—Colour synchronisation for mutually locking different synchronisation sources
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/57—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for processing of video signals
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/414—Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
- H04N21/4143—Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance embedded in a Personal Computer [PC]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
- H04N21/43072—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/435—Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
- H04N21/4394—Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/440236—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by media transcoding, e.g. video is transformed into a slideshow of still pictures, audio is converted into text
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47205—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8547—Content authoring involving timestamps for synchronizing content
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
Definitions
- a script serves as a roadmap to when and how elements of a movie/video will be produced.
- scripts are a rich source of additional metadata and include numerous references to characters, people, places, and things.
- directors, editors, sound engineers, set designers, marketing, advertisers, and other production personnel are interested in knowing which people, places, and things occurred or will occur in certain scenes.
- This information is often present in the script but is not typically directly correlated to the corresponding video content (e.g., video and audio) because timing information is missing from the script. That is, elements of the script are not correlated with a time in which they appear in the corresponding video content.
- script elements e.g., spoken dialogue
- production personnel may know that a character speaks a certain line of dialogue in a scene based on the script, the production personnel may not be able to readily determine the precise time in the working or final video when the particular line was spoken.
- a full script can include several thousand script elements or entities. If one were to try to find the actual point in time when a particular event (e.g., when a line was spoken) in a corresponding movie/video, the video content may have to be manually searched by a viewer to locate the event such that the corresponding timecode can be manually recorded. Thus, production personnel may not be able to easily to search or index their scripts and video content.
- the script text is said to be “aligned” with the recorded dialogue, and the resulting script may be referred to as an “aligned script.”
- Aligned scripts may be useful as production personnel often desire to search or index video/audio content based on the text provided in the script.
- production personnel may desire to generate closed caption text that is synchronized to actual spoken dialogue in video content.
- time aligning is a difficult task to automate.
- time-aligning textual scripts and metadata to actual video content is a tedious task that is accomplished by a manual process that can be expensive and time-consuming.
- a person may have to view and listen to video content and manually transcribe the corresponding audio to generate an index of what took place and when, or to generate closed captioning text that is synchronized to the video.
- To manually locate and record a timecode for even a small fraction of the dialogue words and script elements within a full-length movie often requires several hours of manual work, and doing this for the entire script might require several days or more. Similar difficulties may be encountered while creating video descriptions for the hearing impaired.
- a movie may be manually searched to identify gaps in dialogue for the insertion of video description narrations that describe visual elements (e.g., actions, settings) and a more complete description of what is taking place on screen.
- a method that includes providing script data that includes ordered script words indicative of dialogue and providing audio data corresponding to at least a portion of the dialogue.
- the audio data includes timecodes associated with dialogue.
- the method includes correlating the script data with the audio data, and generating time-aligned script data that includes time-aligned words indicative of dialogue spoken in the audio data and corresponding timecodes for time-aligned words.
- a computer implemented method that includes providing video content data corresponding to the script data including ordered script words indicative of dialogue.
- the video content data includes audio data includes a transcript including transcript words corresponding to at least a portion of the dialogue and timecodes associated with the transcript words.
- the method also includes correlating the script data with the video content data, and generating time-aligned script data that includes time-aligned words indicative of words spoken in the video content and corresponding timecodes for time-aligned words.
- a computer implemented method that includes receiving script data including ordered script words of a script, wherein the ordered script words are indicative of dialogue words to be spoken, receiving audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a matrix of the ordered script words versus the dialogue words, performing an alignment of the matrix to determine hard alignment points, including matching consecutive sequences of ordered script words with corresponding sequences of dialogue words.
- the method also includes partitioning the matrix of ordered script words into sub-matrices, wherein the bounds of each of the sub-matrices are defined by adjacent hard-alignment points, and wherein the sub-matrices include a sub-set of the ordered script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, performing an alignment of each of the sub-matrices.
- the alignment of the sub-matrices including: matching ordered script words of the sub-subset of ordered script words of the respective sub-matrix with dialogue words of the sub-subset of dialogue words of the respective sub-matrix, assigning, to the matched ordered script words, timecodes associated with corresponding matching dialogue words, and determining timecodes for the unmatched ordered script words of sub-set of ordered script words of the sub-matrix using interpolation based on the timecodes associated with the matching ordered script words.
- the method also includes generating time-aligned script data including the ordered script words of the script and their corresponding timecodes.
- a non-transitory computer readable storage medium having program instructions stored thereon, wherein the program instructions are executable to cause a computer system to perform a method that includes receiving script data including ordered script words of a script, wherein the ordered script words are indicative of dialogue words to be spoken, receiving audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a matrix of the ordered script words versus the dialogue words, performing an alignment of the matrix to determine hard alignment points, including matching consecutive sequences of ordered script words with corresponding sequences of dialogue words.
- the method also includes partitioning the matrix of ordered script words into sub-matrices, wherein the bounds of each of the sub-matrices are defined by adjacent hard-alignment points, and wherein the sub-matrices include a sub-set of the ordered script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, performing an alignment of each of the sub-matrices.
- the alignment of the sub-matrices including: matching ordered script words of the sub-subset of ordered script words of the respective sub-matrix with dialogue words of the sub-subset of dialogue words of the respective sub-matrix, assigning, to the matched ordered script words, timecodes associated with corresponding matching dialogue words, and determining timecodes for the unmatched ordered script words of sub-set of ordered script words of the sub-matrix using interpolation based on the timecodes associated with the matching ordered script words.
- the method also includes generating time-aligned script data including the ordered script words of the script and their corresponding timecodes.
- a computer system for receiving script data including ordered script words of a script, wherein the ordered script words are indicative of dialogue words to be spoken, receiving audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a matrix of the ordered script words versus the dialogue words, performing an alignment of the matrix to determine hard alignment points, including matching consecutive sequences of ordered script words with corresponding sequences of dialogue words.
- the method also includes partitioning the matrix of ordered script words into sub-matrices, wherein the bounds of each of the sub-matrices are defined by adjacent hard-alignment points, and wherein the sub-matrices include a sub-set of the ordered script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, performing an alignment of each of the sub-matrices.
- the alignment of the sub-matrices including: matching ordered script words of the sub-subset of ordered script words of the respective sub-matrix with dialogue words of the sub-subset of dialogue words of the respective sub-matrix, assigning, to the matched ordered script words, timecodes associated with corresponding matching dialogue words, and determining timecodes for the unmatched ordered script words of sub-set of ordered script words of the sub-matrix using interpolation based on the timecodes associated with the matching ordered script words.
- the method also includes generating time-aligned script data including the ordered script words of the script and their corresponding timecodes.
- FIG. 1A is a block diagram that illustrates components and dataflow for document time-alignment in accordance with one or more embodiments of the present technique.
- FIG. 1B is text that illustrates exemplary script data in accordance with one or more embodiments of the present technique.
- FIG. 1C is text that illustrates exemplary transcript data in accordance with one or more embodiments of the present technique.
- FIG. 1D is text that illustrates exemplary time-aligned script data in accordance with one or more embodiments of the present technique.
- FIG. 2 is a block diagram that illustrates components and dataflow for script time-alignment in accordance with one or more embodiments of the present technique.
- FIG. 3 is a flowchart that illustrates a script time-alignment method in accordance with one or more embodiments of the present technique.
- FIG. 4 is a flowchart that illustrates a script synchronization method in accordance with one or more embodiments of the present technique.
- FIG. 5A is a depiction of an exemplary alignment matrix in accordance with one or more embodiments of the present technique.
- FIG. 5B is a depiction of an exemplary alignment sub-matrix in accordance with one or more embodiments of the present technique.
- FIG. 6 is a depiction of an exemplary graphical user interface sequence in accordance with one or more embodiments of the present technique.
- FIG. 7A is a depiction of multiple lines of text that include a script phrase, a transcript phrase and a corresponding representation of alignment in accordance with one or more embodiments of the present technique.
- FIG. 7B is a depiction of multiple lines of text that include a script phrase, a transcript phrase and a corresponding representation of alignment in accordance with one or more embodiments of the present technique.
- FIG. 7C is a depiction of a line of text and corresponding in/out ranges in accordance with one or more embodiments of the present technique.
- FIGS. 8A and 8B are block diagrams that illustrate components and dataflow of a script time-alignment technique in accordance with one or more embodiments of the present technique.
- FIG. 9A is a depiction of an exemplary script document in accordance with one or more embodiments of the present technique.
- FIG. 9B is a depiction of a portion of an exemplary video description script in accordance with one or more embodiments of the present technique.
- FIG. 9C is a flowchart that illustrates a method of generating a video description in accordance with one or more embodiments of the present technique.
- FIG. 10 is a block diagram that illustrates an example computer system in accordance with one or more embodiments of the present technique.
- Speech-To-Text a process by which source audio containing dialogue or narrative is automatically transcribed to a textual representation of the dialogue or narrative.
- the source audio may also contain music, noise, and/or sound effects that generally contribute to lower transcription accuracy.
- STT transcript a document generated by a STT transcription engine containing the transcription of the dialogue or narrative of the audio source.
- Each word in the transcript may include an associated timecode which indicates precisely when the audio content associated with each word of the dialogue or narrative occurred. Timecodes are typically provided in hours, minutes, seconds and frames.
- Script a document that outlines all of the visual, audio, behavioral, and spoken elements required to tell the story in a corresponding video or movie. Dramatic scripts are often referred to as a “screenplay”. Scripts may not include timecode data, such that they may not provide information about when an element of the script actually occurs within corresponding video content (e.g., a script may not provide a relative time within the video content that indicates precisely when the audio content associated with each word of the dialogue or narrative occurred).
- Script dialogue/narrative the script lines to be spoken in a corresponding video or movie. Each script line may include text that includes one or more words.
- Script alignment a process by which a set of words of a dialogue or narrative in a script are matched to corresponding transcribed words of video content.
- Script alignment may include providing an output that is indicative of a relative time within the video content that words of dialogue or narrative contained in the script are spoken.
- Aligned Script a script that outlines all of the visual, audio, behavioral, and spoken elements required to tell the story in a corresponding video or movie and includes timecode data indicative of when elements of the script actually occur within corresponding video content (e.g., a time aligned script may include a relative time within the video content that indicates precisely when the audio content associated with each word of the dialogue or narrative occurred).
- Timecodes are typically provided in hours, minutes, seconds and frames.
- Feature films are typically shot at 24 frames per second, thus twelve frames is about 1 ⁇ 2 second in duration.
- Word n-gram a consecutive subsequence of N words from a given sequence. For example, (The, rain, in), (rain, in, Spain) and (in, Spain, falls) are valid 3-grams from the sentence, “The rain in Spain falls mainly on the plain.”
- Alignment matrix a mathematical structure used to represent how the words from a script source will align with the transcribed words of a transcript (e.g., an STT transcript generated via a speech-to-text (STT) process).
- a vertical axis of the matrix may be formed of words in a script in the sequence/order in which they occur (e.g., ordered script words)
- a horizontal axis of the matrix may be formed of words in the transcript in the sequence/order in which they occur (e.g., ordered transcript words).
- Each matrix cell at the intersection of a corresponding row/column may indicate the accumulated number of word insert, update or delete operations needed to match the sequence of ordered script words to the sequence of ordered transcript words to the (row, col) entry.
- a path with the lowest score through the matrix is indicative of the best word alignment.
- NLP Natural Language Processing
- Program a visual and audio production that is recorded and played back to an audience, such as a movie, television show, documentary, etc.
- Dialogue the words spoken by actors or other on-screen talent during a program.
- Video Description an audio track in a program containing descriptions of the setting and action.
- the video description may be inserted into the natural pauses in dialogue or between critical sound elements.
- a video description often includes narration to fill in the story gaps for the blind or visually impaired by helping to describe visual elements and provide a more complete description of what's happening (e.g., visually) in the program.
- Describer a person who develops the description to be recorded by the voicer. In some cases, the describer is also the voicer.
- SAP Secondary Audio Program
- Digital Television broadcasting (DTV)—Analog broadcasting ceased in the U.S. in 2009 and was replaced by DTV.
- a document includes at least a portion of a script document, such as a movie or speculative script (e.g., dramatic screenplay), that outlines visual, audio, behavioral, and spoken elements required to tell a story.
- video content includes video and/or audio data that corresponds to at least a portion of the script document.
- the audio data of the video content is transcribed into a textual format (e.g., spoken dialogue/narration is translated into words).
- the transcription is provided via a speech-to-text (STT) engine that automatically generates a transcript of words that correspond to the audio data of the video content.
- the transcript includes timing information that is indicative of a point in time within the video content that one or more words were actually spoken.
- the words of the transcript (“transcript words”) are aligned with corresponding words of the script (“script words”).
- aligning the transcript words with corresponding script words includes implementation of various processing techniques, such as matching sequences of words, assessing confidence/probabilities that the words identified are in fact correct, and substitution/replacement of script/transcript word with transcript/script words.
- the resulting output includes time-aligned script data.
- the script data includes a time-aligned script document including accurate representation of each of the words actually spoken in the video content, and timing information that is indicative of when the word of the script were actually spoken within the video content (e.g., a timecode associated with each word of dialogue/narration).
- time-aligned data may include timecodes for other elements of the script, such as scene headings, action elements, character names, parentheticals, transitions, shot elements, and the like.
- two source inputs are provided: (1) a script (e.g., plain dialogue text or a Hollywood Spec. Script/Dramatic screenplay) and (2) an audio track dialogue (e.g., an audio track dialogue from video content corresponding to the script).
- a coarse-grain alignment of blocks of text is performed by first matching identical or near identical N-gram sequences of words to generate corresponding “hard alignment points”.
- the hard-alignment points may include matches between portions of the script and transcript (e.g., N-gram matches of a sequence of script words with a sequence of transcript words) which are used to partition an initial single alignment matrix (e.g., providing a correspondence of all ordered script words vs.
- the soft alignment points may define multiple non-overlapping interpolation intervals.
- unmatched words may be located between the matched words (e.g., between the hard alignment points and/or the soft alignment points).
- an interpolation e.g., linear or non-linear interpolation
- an interpolation may be performed to determine timecodes for each of the non-matched words (e.g., words that have not been assigned timecode information) occurring between the matched points.
- all words e.g., matched and unmatched
- the timecode information may be merged with the words of the script and/or transcript documents to generate a time-aligned script document that includes all of the words spoken and their corresponding timecode information to indicate when each of the words was actually spoken within the video content.
- Such a technique may benefit from combining the accuracy of the script words and the timecodes of the transcript words.
- the techniques described herein may provide techniques by which all textual elements (e.g., dialogue/narration) of a script (e.g., a Hollywood movie script or dramatic screenplay script) can be automatically time-aligned to the specific points in time within corresponding video content, to identify when specific dialogue, text, or actions within the script actually occur within the video content. This enables identifying and locating when dialogue and important semantic metadata provided in a script actually occurs within corresponding production video content.
- time alignment may be applied to all elements of the script (e.g., scene headings, action elements, etc.) to enable a user to readily identify where various elements, not just dialogue words, occur within the script.
- the timecode information may also be used to identify gaps in dialogue for the insertion of video description content that includes narrations to fill in the story gaps for the blind or visually impaired, thereby helping to describe visual elements and provide a more complete description of what's happening (e.g., visually) in the program
- the techniques described herein may be employed to automatically and accurately synchronize the written movie script (e.g., which may contain accurate text, but no time information) to a corresponding audio transcript (e.g., which contains accurate time information but may include very noisy or erroneous text).
- techniques may employ the transcript to identify actual words/phrases spoken that vary from the text of the script.
- the accuracy of the words in the script or transcript may, thus, be combined with accurate timing information in the transcript to provide an accurate time aligned script.
- the techniques described herein may demonstrate good tolerance to noisy transcripts or transcripts that have a large number of errors. By partitioning the alignment matrix into many smaller sub-matrices, the techniques described herein may also provide improved performance including increased processing speeds while maintaining significantly higher overall accuracy.
- FIG. 1 is a block diagram that illustrates system components and dataflow of a system for implementing time-alignment (system) 100 in accordance with one or more embodiments of the present technique.
- system 100 implements a synchronization module 102 to analyze a document 104 and corresponding video content 106 . Based on the analysis, system 100 generates time-aligned data (e.g., time aligned script document) 116 that associates various portions of document 104 with corresponding portions of video content 106 .
- Time aligned data 116 may provide the specific points in time within video content 106 that elements (e.g., specific dialogue, text, or actions) defined in document 104 actually occur.
- document 104 (e.g., a script) is provided to a document extractor 108 .
- Document extractor 108 may generate a corresponding document data 110 , such as a structured/tagged document.
- a structured/tagged document may include embedded script data that is provided to synchronization module 102 for processing.
- document 104 may include a script document, such as a movie script (e.g., a Hollywood script), a speculative script, a shooting script (e.g., a Hollywood shooting script), a closed caption (SRT) video transcript or the like.
- a script document such as a movie script (e.g., a Hollywood script), a speculative script, a shooting script (e.g., a Hollywood shooting script), a closed caption (SRT) video transcript or the like.
- a script document such as a movie script (e.g., a Hollywood script), a speculative script, a shooting script (e.g., a Hollywood shooting script), a closed caption (SRT) video transcript or the like.
- a movie script e.g., a Hollywood script
- a shooting script e.g., a Hollywood shooting script
- SRT closed caption
- a movie script may include a document that outlines all of the visual, audio, behavioral, and spoken elements required to tell a story.
- a speculative (“spec”) script or screenplay may include a preliminary script used in both film and television industries.
- a spec script for film generally includes an original screenplay and may be a unique plot idea, an adaptation of a book, or a sequel to an existing movie.
- a “television” spec script is typically written for an existing show using characters and storylines that have already been established.
- a “pilot” spec script typically includes an original idea for a new show.
- a television spec script is typically 20-30 pages for a half hour of programming, 40-60 pages for a full hour of programming, or 80-120 pages for two hours of programming.
- Script 104 may include a full script including several thousand script elements or entities, for instance, or a partial script including only a portion of the full script, such as a few lines, a full scene, or several scenes.
- script 104 may include a portion of a script that corresponds to a clip provided as video content 106 . Since film production is a highly collaborative process, the director, cast, editors, and production crew may use various forms of the script to interpret the underlying story during the production process. Further, since numerous individuals are involved in the making of a film, it is generally desirable that a script conform to specific standards and conventions that all involved parties understand (e.g., it will use a specific format w.r.t. layout, margins, notation, and other production conventions).
- a script document is intended to structure all of the script elements used in a screenplay into a consistent layout.
- Scripts generally include script elements embedded in the script document. Script elements often include a title, author name(s), scene headings, action elements, character names, parentheticals, transitions, shot elements, dialogue/narrations, and the like.
- An exemplary portion of a script segment 130 is depicted in FIG. 1B .
- Script segment 130 includes a scene heading 130 a , action elements 130 b , character names 130 c , dialogues 130 d , and parentheticals 130 e.
- Document (script) extractor 108 may process script 104 to provide document (script) data 110 , such as a structured/tagged script document. Words contained in the document (script) data may be referred to as script words.
- a structured/tagged (script) document may include a sequential listing of the lines of the document in accordance with their order in script 104 , along with a corresponding tag (e.g., tags—“TRAN”, “SCEN”, “ACTN”, “CHAR”, “DIAG”, “PARN” or the like) identifying a determined element type associated with some, substantially all, or all of each of the lines or groupings of the lines.
- a structured/tagged document may include an Extensible Markup Language (XML) format, such as *.ASTX format used by certain products, such as those produced by Adobe Systems, Inc., having headquarters in San Jose, Calif. (hereinafter “Adobe”).
- XML Extensible Markup Language
- Adobe an Extensible Markup Language
- document extractor 108 may obtain script 104 (e.g., a layout preserved version of the document), perform a statistical analysis and/or feature matching of features contained within the document, identify document elements based on the statistical analysis and/or the feature matching, pass the identified document elements through a finite state machine to assess/determine/verify the identified document elements, assess whether or not document elements are incorrectly identified, and, if it is determined that there are incorrectly identified document elements, re-performing at least a portion of the identification steps, or, if it is determined that there are no (or sufficiently few) incorrectly identified document elements, and generate/store/output a structured/tagged (script) document or other forms of document (script) data 110 that is provided to synchronization module 102 .
- script 104 e.g., a layout preserved version of the document
- identify document elements based on the statistical analysis and/or the feature matching
- pass the identified document elements through a finite state machine to assess/determine/verify the identified document elements, assess whether or not document elements are
- document extractor 108 may employ various techniques for extracting and transcribing audio data, such as those described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein.
- video content 106 is provided to an audio extractor 112 .
- Audio extractor 112 may generate a corresponding transcript 114 .
- Video content 106 may include video image data and corresponding audio soundtracks that include dialogue (e.g., character's spoken words or narrations), sound effects, music, and the like.
- Video content 106 for a movie may be produced in segments (e.g., clips) and then assembled together to form the final movie or video product during the editing process.
- a movie may include several scenes, and each scene may include a sequence of several different shots that typically specify a location and a sequence of actions and dialogue for the characters of the scene.
- the sequence of shots may include several video clips that are assembled into a scene, and multiple scenes may be combined to form the final movie product.
- a clip, including video content 106 may be recorded for each shot of a scene, resulting in a large number of clips for the movie.
- Tools such as Adobe Premiere Pro by Adobe Systems, Inc., may be used for editing and assembling clips from a collection of shots or video segments.
- audio content e.g., without corresponding video content may be provided).
- audio content such as that of a radio show
- Audio extractor 112 may process video content 106 to generate a corresponding transcript that includes an interpretation of words (e.g., dialogue or narration) spoken in video content 106 .
- Transcript 114 may be provided as a transcribed document or transcribed data that is capable of being provided to other portions of system 100 for subsequent processing.
- audio extractor 112 includes a speech-to-text engine that takes an audio segment from video content 106 containing spoken dialogue, and uses speech-to-text (STT) technology to generate a time-code transcript of the dialogue.
- transcript 114 may indicate the timecode and duration for each spoken word that is identified by the audio extractor. Words of transcript 114 may be referred to as transcript words.
- speech-to-text (STT) technology may implement a custom language model such as that described herein.
- speech-to-text (STT) technology may implement a custom language model and/or an enhanced multicore STT transcription engine such as those described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009 and/or U.S. patent application Ser. No. 12/332,309 entitled “MULTI-CORE PROCESSING FOR PARALLEL SPEECH-TO-TEXT PROCESSING”, filed Dec. 10, 2008, which are hereby incorporated by reference as though fully set forth herein.
- a transcript 114 generated by audio extractor 112 may include a raw transcript.
- An exemplary raw transcript (e.g., STT transcript) 132 is depicted in FIG. 1C .
- Raw transcript 132 includes a sequential listing of identified transcript words having associated time code, duration, STT word estimate and additional comments regarding the transcription.
- the timecode may indicate at what point in time within the video content the word was spoken (e.g., transcript word “dad” was spoken 7165.21 seconds from the beginning of the associated video content), the duration may indicate the amount of time the word was spoken from start to finish (e.g., it took about 0.27 sec to say the word “dad”), and comments may indicate potential problems (e.g., that noise in the audio data may have generated an error).
- the raw transcript information may also include a confidence value that indicates the probability that the interpreted/indicated word is accurate.
- the raw transcript information may not include additional text features, such as punctuation, capitalization, and the like.
- document extraction and audio extraction may occur in parallel.
- document extractor 108 receives script 104 and generates script data 110 independent of audio extractor 112 receiving video content 106 and generating transcript 114 . Accordingly, these two processes may be performed in parallel with one another.
- document extraction and audio extraction may occur in series. For example, document extractor 108 may receive document 104 and generate document data 110 prior to audio extractor 112 receiving video content 106 and generating transcript 114 , or vice versa.
- Synchronization module 102 may generate time-aligned data 116 .
- Time-aligned data 116 may be provided as a document or raw data that is capable of being provided to other portions of system 100 for subsequent processing.
- Time-aligned data 116 may be based on script information (e.g., document data 110 ) and video content information (e.g., transcript 114 ).
- script information e.g., document data 110
- transcript 114 video content information
- synchronization module 102 may compare transcript words in transcript 114 to script words in the document (script) data 110 to determine whether or not the transcribed words are accurate. The comparison may use various indicators to assess the accuracy. For example, a plurality of words and phrases with exact matches between transcript 114 and document data 110 may have high probabilities of being correct, and may be referred to as “hard reference points”.
- Words and phrases with partial matches may have a lower probability of being correct, and may be referred to as “soft reference points”.
- Words and phrases that do not appear to have matches may have a low probability of being correct.
- Words and phrases with a low probability of being correct may be subject to additional amounts of processing. For example, low probability matches may be subject to interpolation based on the hard and soft reference points. Words that are part of hard or soft reference pints may be referred to as words having a match, whereas words that are not part of a hard or soft reference point may be referred to as unmatched words or words not having a match.
- the hard-alignment points may be used to partition the document data and the transcript into smaller segments that correspond to one another, and additional processing may be performed on the smaller segments in substantial isolation.
- the timecodes and other information associated with matched words may be used to derive (e.g., interpolate) timecode and other information about the unmatched words.
- Time aligned data 116 may include words (e.g., from the script words or transcript words) having a specific timecode associated therewith.
- time aligned data 116 may include words from both document data 110 and transcript data 114 used to generate a single script that accurately identifies words actually spoken in video content 106 along with corresponding timecode information for each spoken word of dialogue or other elements.
- the timecode for each word may be obtained directly from matching words of the transcript, or may generated (e.g., via interpolation).
- Time aligned data 116 may be stored at a storage medium 118 (e.g., a database), displayed at a display device 120 (e.g., a graphical display viewable by a user), or provided to other modules 122 for processing.
- An exemplary time-aligned script data/document 134 is depicted in FIG. 1D .
- time-aligned data/document 134 includes spoken words 136 grouped with other spoken words of their respective script elements 137 , and provided along with their associated timecodes 138 .
- a start time 140 for each element grouping of lines is also provided.
- each of the script elements (and text of the script elements) is also assigned a corresponding time code.
- FIG. 2 is a block diagram that illustrates components and dataflow of system 100 in accordance with one or more embodiments of the present technique.
- synchronization module 102 includes a script reader 200 , a script analyzer 202 , a Speech-to-Text (STT) reader 204 , an STT analyzer 206 , a matrix aligner 208 , an interval generator/interpolator 210 , and a time-coded script generator 212 .
- STT Speech-to-Text
- FIG. 3 is a flowchart that illustrates a script time-alignment method 300 according to one or more embodiments of the present technique.
- Method 300 may provide alignment techniques using components and dataflow implemented at system 100 .
- method 300 includes providing script content, as depicted at block 302 , providing audio content, as depicted at block 304 , aligning the script content and audio content, as depicted at block 306 , and providing time-coded script data, as depicted at block 308 .
- providing script content includes inputting or otherwise providing a script 104 , such as a Hollywood Spec. Movie Script or dramatic screenplay script, to system 100 .
- a plain text document such as a raw script document
- script extractor 108 which processes script 104 (e.g., to identify, structure, and extract the text of script 104 ) to generate script data 110 , such as a structured/tagged script document.
- Script extractor 108 may employ techniques for converting documents, such as those described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, U.S. patent application Ser.
- Document data 110 may be provided to synchronization module 102 for subsequent processing, as described in more detail below.
- providing audio content includes inputting or otherwise providing video content 106 , such as a clip/shot of a Hollywood movie, having associated audio content that corresponds to a script 104 , to system 100 .
- Audio data may be extracted from video content 106 using various techniques. For example, an audio data track may be extracted from video content 106 using a Speech-to-Text (STT) engine and/or a custom language model.
- STT Speech-to-Text
- audio extractor 112 may employ an STT engine and/or custom language model to generate transcript 114 that includes a transcription of spoken words (e.g., audio dialogue or narration) of the Hollywood movie or other audio data.
- Audio extractor 112 may employ various techniques for extracting and transcribing audio data, such as those described below and/or those techniques described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, and/or U.S. patent application Ser. No. 12/332,309 entitled “MULTI-CORE PROCESSING FOR PARALLEL SPEECH-TO-TEXT PROCESSING”, filed Dec. 10, 2008, which are both hereby incorporated by reference as though fully set forth herein.
- a resulting transcript 114 may be provided to synchronization module 102 for subsequent processing, as described in more detail below.
- aligning the script and audio content includes employing a matching technique to align the script words (e.g., dialogue or narrations) of script 104 to elements of the video content 106 . This may include aligning script words to corresponding transcript words.
- alignment includes synchronization module 102 implementing a two-level word matching system to align script words of script 110 to corresponding transcript words of transcript 114 .
- a first matching routine is executed to partition a matrix of script words vs. transcript words into a sub-matrix. For example, an N-gram matching scheme may be used to identify high probability matches of a sequence of multiple words.
- N-gram matching may include attempting to exactly (or at least partially) match phrases of multiple transcript words with script words.
- the matched sequence of words may be referred to as hard-alignment points.
- the hard alignment points may include several matched words, and may be used to define boundaries of each sub-matrix. Thus, the hard-alignment points may define smaller matrices of script words vs. transcript words.
- Each of the smaller sub-matrices may, then, be processed (e.g., in series or parallel) using additional matching techniques to identify word matches within each of the sub-matrices.
- processing may be provided via multiple processors. For example, processing in series or parallel may be performed using multiple processors of one or more hosted services or cloud computing environments.
- each of the sub-matrix is processed independent of (e.g., in substantial isolation from) processing of the other sub-matrices.
- These resulting additional word matches may be referred to as soft alignment points.
- the timecode information associated with the words of the hard and soft alignment points may be used to assess timecode information for the unmatched words (e.g., via interpolation). For example, timecodes associated with the words that make up the matched points at the end and beginning of an interval of time may be used as references to interpolate time values for unmatched words that fall within the interval between the matched words. Alignment techniques that may be implemented by synchronization module 102 are discussed in more detail below. Further, techniques for matching are discussed in more detail below with respect to FIGS. 8A and 8B .
- providing time-coded script data includes providing timecodes assigned to all dialogue and other script element types.
- synchronization module 102 after synchronization module 102 aligns word N-grams from script 110 with corresponding word N-grams of transcript 114 , it may output (e.g., to a client application) time information in the form of time-coded script data (e.g., time-aligned script data 116 ) that contains timecodes assigned to some or all dialogue and to some and/or all other script element types associated with script 104 .
- the data may be stored, displayed/presented or processed.
- a script e.g., a Hollywood Spec.
- Time-aligned script data 116 may be processed and used by other applications, such as the Script Align feature of Adobe Premiere Pro.
- processing may be implemented to time-align script elements other than audio (e.g., scene headings, action description words, etc.) directly to the video scene or full video content.
- timecodes of the script words may be used to determine a timecode of the script element.
- each of the script elements may be provided in the time-aligned script data in association with a timecode, as discussed above with regard to FIG. 1D .
- Providing time-coded script data may include providing the resulting time-aligned data 116 to a storage medium, display device, or other modules for processing, as described above with regard to FIG. 1A .
- FIG. 4 is a flowchart that illustrates a time-alignment method 400 according to one or more embodiments of the present technique.
- Method 400 may provide alignment techniques using components and dataflow implemented at synchronization module 102 .
- method 400 generally includes reading a script (SCR) file and a speech-to-text (STT) file, and processing the SCR and STT files using various techniques to generate an output that includes time-aligned script data.
- SCR script
- STT speech-to-text
- method 400 includes reading an SCR file, as depicted at block 402 .
- This may include reading script data, such as script data 110 , described above with respect to block 302 .
- reading an SCR file may include script reader 200 reading a generated SCR file (e.g., document data 110 ).
- the SCR file may include a record-format representation of a source Hollywood spec. script of dramatic screenplay script. Records contained in the SCR file may each include one complete script element.
- Script reader 200 may extract script element type and data values from each record and place these into an internal representation (e.g., a structured/tagged script document).
- method 400 includes reading an STT file, as depicted at block 404 .
- This may include reading STT data, such as transcript 114 , as described above with respect to block 304 .
- Transcript 114 may include an STT file having transcribed data, such as that of the STT word transcript data 132 depicted in FIG. 1C .
- the STT data may provide a timecode for each spoken word in the audio sound track which corresponds in time to video content 106 .
- method 400 includes building a SCR N-gram dictionary, as depicted at block 406 .
- building an SCR N-gram dictionary includes identifying all possible sequences of a given number of consecutive words.
- the number of words in the sequence may be represented by a number “N”.
- N is set to a value of 3: (The, rain, in), (rain, in, Spain), (in, Spain, falls), (Spain, falls, mainly), (falls, mainly, on), (mainly, on, the), and (on, the, plain).
- additional N-gram word sequences may be generated based on words that precede or follow a phrase. For example, where the first word of a following sentence is “Why”, an additional 3-gram may include (the, plain, why).
- the value of N may be set by a user. In some embodiments, the value of N is set to a predetermined value, such as four. For example, N may be automatically set to a default value of four, and the user may have the option to change the value of N to something other than four (e.g., one, two, three, five, etc.).
- script analyzer 202 may build a word N-gram “dictionary” of all words from script 110 and may record their relative positions within script 110 and/or STT analyzer 206 may build a word N-gram “dictionary” of all words from transcript 114 and may record their relative positions within transcript 114 .
- the resulting N-gram dictionaries may include an ordered table of 1-gram, 2-gram, 3-gram, or N-gram word sequences.
- method 400 includes matching N-grams, as depicted at block 408 .
- matching N-grams may include attempting to match N-grams of the script 110 to corresponding N-grams of transcript 114 .
- SCR analyzer 202 and/or STT analyzer 206 may attempt to match all word N-grams of the N-gram dictionaries and may store the matches (e.g., in an internal table) in association with corresponding timecode information associated with the respective transcript word(s).
- the stored matching N-grams may indicate the potential for a matched sequence of words, and may be referred to as “candidate” N-grams for merging.
- a phrase from the script N-gram dictionary may be matched with a corresponding phrase the transcript N-gram dictionary, however, due to the phrase being repeated several time within the script/video content, the match may not be accepted until the relative positions can be verified.
- method 400 includes merging N-grams, as depicted at block 410 .
- merging of N-grams may be provided by SCR analyzer 202 and/or STT analyzer 206 .
- merging N-grams includes merge some or all sequential N-gram matches into longer matched N-grams. For example, where two consecutive matching N-grams are identified, such as two consecutive 3-grams of (The, rain, in) and (rain in Spain), they may be merged together to form a single N-gram, referred to as a single 4-gram of (The, rain, in, Spain). Such a technique may result in merged N-grams of length N+1 after each iteration.
- the technique may be repeated (e.g., iteratively) to merge all consecutive N-grams to provide N-grams having higher values of N.
- N-grams with higher values of N may have higher probabilities of being an accurate match.
- the iterative process may continue until no additional N-gram matches are identified. For example, where there are at most ten consecutive words identified as matching, increasing to an 11-gram length may yield no matching results, thereby terminating the merging process. Further, techniques for N-gram matching are discussed in more detail below with respect to FIGS. 8A and 8B .
- the resulting set of merged N-grams may provide a set of “hard alignment points”.
- each separate N-gram may indicate with relatively high certainty that a sequence of words in script 110 precisely matches a sequence of words in transcript 114 .
- the sequence of words may identify a hard-alignment point.
- a hard alignment point may include a series of matched words.
- the hard alignment points may include a series of words that each soft-align.
- timing data for each of the words of the matching N-grams may be correlated with the corresponding script words.
- timing data for other words e.g., unmatched words or words having low probabilities of accurate matches
- may be assessed and determined based on the timecode data of words associated with matched words e.g., words that make up one or at least a portion of one or more alignment points. For example, interpolation may be used to assess and determine the position of a script word that occurs between matched script words (e.g., script words associated with alignment points).
- method 400 includes generating a sub-matrix, as depicted at block 412 .
- each hard alignment point may define a block of script text (e.g. a sequence of words in script 110 ) and a timecode indicative of where the hard alignment point occurs in the video.
- script and transcript words associated with hard alignment points may be associated with timecode data, other script words (e.g., unmatched words between each hard alignment point) may still need to be aligned to corresponding transcript words to assess and determine their respective timecode.
- each successive pair of hard/soft alignment points is used to create an alignment sub-matrix.
- the alignment sub-matrix may include script words (e.g., sub-set of script words) that occur between matched script words (e.g., script words associated with hard alignment points) and intermediate transcript words (e.g., a sub-set of transcript words) that occur between matched transcript words (e.g., transcript words associated with hard alignment points).
- the script words may be provided along one axis (e.g., the y or x-axis) of the sub-matrix, and the intermediate transcript words may be provided along the other axis (e.g., the x or y-axis) of the sub-matrix.
- FIG. 5A depicts an exemplary (full) alignment matrix 500 in accordance with one or more embodiments of the present technique.
- Alignment matrix 500 may include some or all of the script words aligned in sequence along the y-axis and all of some of the transcript words aligned in sequence along the x-axis, or vice versa.
- script words and transcript words would match exactly, resulting in a substantially straight line having a slope of about one or negative one.
- hard alignment points 502 are identified. Between each of the hard-alignment points 502 are a number of soft alignment points 504 (denoted by squares) and/or interpolated alignment points 506 (denoted by X's).
- Hard alignment points 502 may be determined as a result of matching/merging N-gram sequences as discussed above with respect to blocks 408 and 410 .
- Soft alignment points 504 may be determined as a result of additional processing, such as use of a standard/optimized Levenshtein algorithm, discussed in more detail below.
- Interpolated alignment points 506 may be determined as a result of additional processing, such as linear or non-linear interpolation between hard and/or soft alignment points, discussed in more detail below.
- Interpolation intervals 507 extend between adjacent soft alignments points 504 .
- alignment matrix 500 may include one or more alignment sub-matrices 508 a - 508 g (referred to collectively as sub-matrices 508 ).
- Sub-matrices 508 a - 508 g may be defined by the set of points (e.g., script words and transcript words) that are located between adjacent, respective, hard alignment points 502 .
- matrix 500 includes seven sub-matrices 508 a - 508 g .
- An exemplary sub-matrix 508 e is also depicted in detail in FIG. 5B .
- method 400 includes pre-processing a sub-matrix, as depicted at block 414 .
- Pre-processing of the sub-matrix may be provided at matrix aligner 208 .
- pre-processing the sub-matrix may include identifying the range of a particular sub-matrix (e.g., the range/sequence of associated script words and transcript words associated with the axis of the particular sub-matrix).
- script and transcript words that fall between two words contained in adjacent hard alignment points 502 may be identified as a matrix sub-set of script words (SCR word sub-set) 510 (represented by outlined triangles) and a corresponding matrix sub-set of transcript words (STT word sub-set) 512 (represented by solid triangles), as depicted in FIG. 5B with respect to sub-matrix 508 e .
- SCR word sub-set represented by outlined triangles
- STT word sub-set represented by solid triangles
- a timecode and position offset data structure used for booking is initialized prior to words of SCR word sub-set 510 being aligned to words of STT word sub-set 512 of sub-matrix 508 e .
- all special symbols and punctuation are removed from SCR word sub-set 510 . This may provide for a more accurate alignment as both symbols and punctuations are typically not present in a transcript 114 , and, are, thus, not present in STT word sub-set 512 .
- sub-matrices 508 of the initial alignment matrix 500 are sequentially processed (e.g., in order of their location along the diagonal of the alignment matrix 500 ) to find the best time alignment for words between each pair of hard reference points 502 that define each respective sub-matrix 508 a - 508 g .
- system 100 includes a single core system used to process the sub-matrices
- alignment of the sub-matrices 508 may be processed sequentially (e.g., in series—one after the other).
- system 100 includes a multi-core system used to process sub-matrices
- alignment of some or all of sub-matrices 508 may be processed in parallel (e.g., simultaneously). Such parallel processing may be possible as the processing of each sub-matrix is independent of all of the other sub-matrices due to the bounding of the matrices with hard alignment points that are assumed to be accurate and that include known timecode information.
- method 400 includes aligning the sub-matrix, as depicted at block 416 .
- Aligning the sub-matrix may be provided at matrix aligner 208 .
- a sub-matrix may be aligned using an algorithm.
- An algorithm may employ a dynamic programming technique to assess multiple potential alignments for a sub-matrix, to determine the best fit alignment of the potential alignments, and employ the best fit alignment for the given sub-matrix.
- an algorithm may identify several possible solutions within the sub-matrix, and may select the solution having the lowest indication of possible error.
- the algorithm may include a Levenshtein Word Edit Distance algorithm.
- a dynamic programming algorithm for computing the Levenshtein distance may require the use of an (n+1) ⁇ (m+1) matrix, where n and m are the lengths of the two respective word sets (e.g., the SCR word set and the STT word set).
- the algorithm may be based on the Wagner-Fischer algorithm for edit distance.
- an alignment path defines a potential sequence of words that may be used between hard alignment points.
- aligning the sub-matrix may include breaking alignment paths within each sub-matrix into discrete sections during processing to more accurately assess individual portions of the alignment path. Based on match probabilities/strengths of various portions of the alignment path, a single alignment path may be broken into separate discrete intervals that are assessed individually. For example, where an alignment path within a sub-matrix includes a first portion having a relatively high match probability and an adjacent second portion having a relatively low match probability, the first and second portions can be separated.
- the first portion may be identified as a sequence of words having a high probability of a match
- the second portion may be identified as a sequence of words having a low probability of a match. Accordingly, the first portion may be identified as an accurate match that can be relied on in subsequent processing and the second portion may be identified as an inaccurate match that should not be relied on in subsequent processing.
- Such a technique may be used in place of merely identifying a mediocre match of the entire alignment path that may or may not be reliable for use in subsequent processing.
- aligning the sub-matrix may include weighting various processing operations to reflect operations that may be indicative of inaccuracies.
- aligning the sub-matrix may include assessing weighting penalties for matched words that are subject to an insert, delete, or substitute operation. Such a technique may help to adapt to false-positive word identifications produced by an STT engine.
- the algorithm may be modified in an attempt to improve alignment. For example, in some embodiments, timecode information recorded with each word of an STT word set is correlated with a matching word of a corresponding SCR word set.
- the matching word may include a single word or a continuous sequence of words, wherein the sequence of words includes less than the number (“N”) of words required by the selected N-gram.
- N the number of words required by the selected N-gram.
- an algorithm such as a Levenshtein Word Edit Distance algorithm, may be used to identify soft-alignment points.
- the soft designation is used to indicate that because of noise, error artifacts, and the like in STT transcript 114 , these alignments may have a lower probability of being accurate than the multi-word, hard-alignment points that define the range/partition of the respective sub-matrix.
- soft-alignment points may be determined using heuristic and/or phonetic matching.
- aligning the sub-matrix may include heuristic filtering.
- Heuristic filtering of noise may include filtering (e.g., ignoring or removing) “stop words” (e.g., short articles such as “a”, “the”, etc.) that are typically inserted into an STT transcript when the STT engine is confused or otherwise unable to decipher the audio track.
- stop words e.g., short articles such as “a”, “the”, etc.
- STT engines often insert articles such as “a”, “the”, etc. while various events other than dialogue occur, such as the presence of noise, music or sound effects.
- Such articles may also be inserted when dialogue is present but cannot be deciphered by the STT engine, such as when noise, music or sound effects drown out dialogue or narration.
- the STT transcript may include a sequence of “the the the the . . . ” indicative of a duration when music or other such events occur in the audio content.
- heuristics may be used to identity portion transcript words that should be ignored. For example, transcript words that should not be considered in the alignment process, and/or should not be included in the resulting time-aligned script data.
- heuristics may be used to identify repetitive sequences of words, and to determine which of the repeated sequence of words, if any need to be included or ignored in the resulting script document. For example, where a clip includes repetitive dialogue, such as where an actor repeats their lines several times in an attempt to get the line correct, transcript 114 may include several repetitions (e.g., “i'll be back i'll be back i'll be back). A corresponding portion of script 110 may include a single recitation of the line (e.g., “I'll be back.”).
- heuristics may be implemented to identify the repeated phrases, to identify one of the phrases of the transcript for use in aligning with script words, and to align the corresponding script words to the selected phrase of transcript 114 .
- the timecodes for words of one of the three phrases in transcript 114 may be associated with the corresponding script words of the phrase “I'll be back”.
- the other repeated phrases are ignored/deleted.
- ignored/deleted transcript words may not be considered in the alignment process, and/or may not be included in the resulting time-aligned script data. Ignoring/deleting the phrases may help to ensure that they do not create errors in aligning other portions of script 110 .
- alignment may attempt to match the other two repeated phrases (e.g., those not selected) with phrases preceding or following the corresponding phrase of script 110 .
- they can also be aligned as “alternate takes”. For example, it may not know which take will eventually be used in a finished edit, so regardless of which take is used, the correct script text and timing information may flow through to that portion of the recorded clip in use.
- a single portion script text may be aligned to each of the repeated portions of the transcript text.
- aligning the sub-matrix may include matching based at least partially on phonetic characteristics of words. For example, a word/phrase of the SCR word set may be considered a match to a word/phrase of the STT word set when the two words/phrases sound similar.
- a special phonetic word comparator may be used to assess word/phrase matches.
- a phonetic comparator may include “fuzzy” encodings that provide for matching script words/phrases that may sound similar to a word identified in the STT transcript. Thus, a word/phrase may be considered a match if they fall within a specific phonetic match threshold.
- a script word may be considered a match to a transcript word if the transcript word is a word identified as being an phonetic equivalent to the word in script 110 , or vice versa.
- the terms “their” and “there” may be identified as phonetic matches although the terms do not exactly match one another.
- Such a technique may account for variations in spoken language (e.g., dialects) that may not be readily identified by an STT engine.
- Use of phonetic matching may be used in place of or in combination with an exact word/phrase match for each word/phrase.
- method 400 includes generating and/or interpolating intervals, as depicted at block 418 .
- Generating and/or interpolating intervals may be provided at interval generator/interpolator 210 .
- generating and/or interpolating intervals may include identifying intervals between identified matched words (e.g., words of hard and/or soft reference points), interpolating the relative position of un-matched words between the matched words.
- An interpolated timecode for the un-matched words may be based on their interpolated position between the matched words and the known timecodes of the matched words.
- the sub-matrices are combined to form a list including script words and corresponding transcript words for each word associated with a hard or soft alignment point.
- all possible word alignment correspondences have been identified, leaving only unmatched script dialogue words (e.g., words that are not associated with hard nor soft reference points), and non-dialogue words within the script such as scene action descriptions and other information.
- unmatched script dialogue words e.g., words that are not associated with hard nor soft reference points
- non-dialogue words within the script such as scene action descriptions and other information.
- the timecode information for the unmatched script words is provided via linear timecode interpolation.
- Linear time code interpolation may include defining an interval that extends between two adjacent reference points, and spacing each of the unmatched words that occur between the two reference points across equal time spacing (e.g., sub-interpolation intervals) within the interval.
- a sub-interpolation interval may be defined as:
- sub_interpolaton ⁇ _interval t ⁇ ⁇ 1 - t ⁇ ⁇ 2 n + 1 ( 1 )
- t 1 is a timecode of a first reference point defining a first end of an interpolation interval
- t 2 is a timecode of second reference point defining a second end of the interpolation interval
- n is the number of unmatched words.
- a first of the unmatched words may be determined to occur at 1.25 seconds
- a second of the unmatched words may be determined to occur at 1.50 seconds
- a third of the unmatched words may be determined to occur at 1.75 seconds.
- the sub-interpolation interval is equal to (2 sec-1 sec)/(3+1), or 0.25 sec.
- FIG. 5B illustrates interpolated points 506 for unmatched script words that are evenly spaced between soft alignment points in accordance with the above described linear interpolation technique. A similar technique may be repeated for each respective interpolation interval between hard/soft alignment points.
- method 400 includes assigning timecodes, as depicted at block 420 .
- Assigning timecodes may be provided at time-coded script generator 212 .
- assigning time codes includes assigning times for each of the script words based on the reference points and interpolated points. For example, in some embodiments, the entire list of soft alignment points is scanned and each successive pair of soft alignment points defines an interpolation interval. Upon defining each interpolation interval, sub-interpolation intervals are determined, and timecode data aligning with the sub-interpolation intervals is assigned to all of the script words of the respective script word set. For example, the unmatched words of the above described interpolation interval may be assigned timecodes of 1.25 seconds, 1.50 seconds, and 1.75 seconds, respectively. Further, techniques for interpolating are discussed in more detail below with respect to FIGS. 8A and 8B .
- a non-linear interpolation technique may be employed to assess and determine timecode information associated with words/phrases within a script document.
- non-linear interpolation or similar matching techniques may be used in place of or in combination with linear interpolation techniques employed to determine timecodes for script words.
- Non-linear interpolation may be useful to account for words that were not spoken at even rate between alignment points. For example, where two alignment points define an interval having matched words on either end and several unmatched words between them, linear interpolation may assign timecode information to the unmatched words assuming an even spacing across the interval as discussed above. The resulting timecodes may be reflective of someone speaking at a constant cadence across the interval. Unfortunately, the resulting timecode information may be inaccurate due to different rates of speech across the interval, pauses within the interval, or the like.
- non-linear interpolation of timecode information may include assessing an expected rate (or cadence) for spoken words and applying that expected rate to assess and determine timecode information for the unmatched words.
- non-linear interpolation may include, for a given script word, determining a rate of speaking for matched script words proximate the script word, and applying the rate of speaking to determine a timecode for the script word.
- FIG. 7A illustrates alignment of a script phrase 700 (e.g., a portion of script data 110 ) with a spoken phrase 701 (e.g., a portion of transcript 114 ) that may be accomplished using non-linear interpolation in accordance with one or more embodiments of the present technique.
- script phrase 700 is illustrated in association with an alignment 702 .
- Phrase 700 includes, “What is your answer to my question? I need to know your answer now!”
- Alignment 702 includes a series of word-match indicators (e.g., word associated with a hard alignment point (H) and words associated with a soft alignment point (S)) and words that are unmatched (U).
- the dots/points between the unmatched representations of “question” and “I” may indicate a pause between speaking of the words (e.g., a pause that would be indicated by timecode information differential between transcript words “position” and “eye” of spoken phrase 701 ).
- the sequence of four words “What is your answer to” and “know your answer now” include matches, and the words, “my”, “question”, “I”, “need” and “to” are unmatched.
- rates of speaking matched words proximate/adjacent (e.g., before or after) unmatched words may be used to assess and determine timecode information for the unmatched words.
- the rate of speaking “What is your answer to” may be used to assess and determine timecode information for the words “my” and “question.” That is, if it is determined that “What is your answer to” is spoken at a rate of one word every 0.1 seconds (e.g., via timecode information provided in the transcript and/or prior alignment/matching), the following words “my question” may be assigned timecode information in accordance with the rate of 0.1 words per second.
- timecodes associated with twenty-one minutes and one-tenth of a second (21:00.1) and twenty-one minutes and two-tenths of a second (21:00.2) may be assigned to the words “my” and “question”, respectively, in aligned script data 116 , for example.
- punctuation within the script may also be used to assess and determine timecode information.
- punctuation indicative of the end of a phrase may be used to determine the presence of a pause between words or phrases.
- the presence of the question mark in phrase 700 may indicate that the phrases “What is your answer to my question?” and “I need to know your answer now!” may be separated by a pause and, thus may each be spoken at different rates.
- Such a technique may be employed to assure that non-linear interpolation is applied to the individual phrases within a sub-matrix to account for an expected pause.
- the rate of speaking “know your answer now” may be used to assess and determine timecode information for the words “I”, “need” and “to”. That is, if it is determined that “know your answer now” was spoken at a rate of one word every 0.2 seconds (e.g., via timecode information provided in transcript 114 ), the preceding words “I need to” may be assigned timecode information in accordance with the rate of 0.2 words per second.
- the word “know” is determined to have been spoken at exactly twenty-one minutes and ten seconds (21:10.00) within a movie
- the word “need” was spoken at twenty-one minutes nine and six-tenths of a second (21:09.6)
- the word “to” was spoken twenty-one minutes nine and eight-tenths of a second (21:09.8).
- Timecodes associated with twenty-one minutes nine and four-tenths of a second (21:09.4), twenty-one minutes nine and six-tenths of a second (21:09.6), and twenty-one minutes nine and eight-tenths of a second (21:09.8) may be assigned to the words “I”, “need”, and “know”, respectively, in aligned script data 116 , for example. Accordingly, punctuation may be used to identify pauses or similar breakpoints that can be used to break words or phrases into discrete intervals such that respective rates of speaking (e.g., cadence) can be appropriately applied to each of the discrete intervals. Other indicators may be used to indicate characteristics of the spoken words. For example, “stopwords” present in the transcript may be indicative of a pause or break in speaking and may be interpreted as a pause and implemented as discussed above.
- the unmatched words may be assigned timecode information based on even spacing between the matched words, and thus, may not account for the pause or similar variations.
- the first of the words “to” is determined to have been spoken at exactly twenty-one minutes (21:00.0) and the word “know” is determined to have been spoken at exactly twenty-one minutes and ten seconds (21:10.0)
- the five unmatched words “my”, “question”, “I”, “need” and “to” would be evenly spaced across the ten second interval at 1.67 second intervals, not accounting for the pause. Although minor in these small increments, this could lead to increased alignment errors where a pause in dialogue occurs for several minutes, for example.
- a rate of speech may be based on machine learning. For example, a rate of speech may be based on other words spoken proximate to the words in question. In some embodiments, a rate of speech may be determined based on elements of the script. For example, a long description of an action item may be indicative of a long pause in the actual dialogue spoken.
- words of the script that occur proximate/between reference points may be aligned with unmatched words of the transcript that also occur proximate/between the same reference points.
- the four unmatched words “my”, “question”, “I” and “need” of script phrase 700 fall within in the interval between matched words “to” and “know”.
- the timecodes associated with the unmatched words of transcript phrase 701 may be assigned to the four unmatched words “my”, “question”, “I” and “need” of script phrase 700 , respectively. That is the timecode of the first unmatched transcript word in the interval may be assigned to the first unmatched script word in the interval, the timecode of the second unmatched transcript word in the interval may be assigned to the second unmatched script word in the interval, and so forth.
- punctuation and/or capitalization from script text may be used to improve alignment. For example, if the first alignment point (hard or soft) occurs in the middle of the first sentence of the clip, it may be determined that the script words and transcript words preceding the alignment point in the script text and the corresponding transcript text should align with one another.
- the timecodes for the script words may be interpolated (e.g., linearly or non-linearly) across the time interval that extends from the beginning of speaking of the corresponding transcript words in the scene to the corresponding alignment point.
- the corresponding script words and transcript words may have a one-to one correspondence, and, thus, timecode information may be directly correlated.
- the first script word of the sentence may be associated with the timecode information of the first transcript word of the clip
- the second script word of the sentence may be associated with the timecode information of the second transcript word of the clip
- the beginning of a sentence may be identified by a capitalized word and the end of a sentence may be identified by a period, exclamation point, question mark, or the like.
- FIG. 7B is a depiction of multiple lines of text that include a script phrase, a transcript phrase and a corresponding representation of alignment in accordance with one or more embodiments of the present technique. More specifically, FIG. 7B illustrates alignment of a script text 703 (e.g., a portion of script 110 ) with a spoken dialog 704 (e.g., a portion of transcript 114 ) that may be accomplished with the aid of capitalization and punctuation in accordance with one or more embodiments of the present technique.
- Script text 703 includes a portion of a script that is spoken throughout a clip/scene.
- script text 703 includes the first sentence of the clip/scene (e.g., “It is good to see you again”) and the last sentence of the clip/scene (e.g., “I will talk to you later tonight”).
- Spoken dialog 704 may include transcript text of a corresponding clip (e.g., “get it could to see you again” and “i will talk with you house get gator flight”).
- script text 703 and transcript text 704 is illustrated in association with an alignment 705 .
- Alignment 705 includes a series of word-match indicators (e.g., word associated with a hard alignment point (H) and words associated with a soft alignment point (S)) and words that are unmatched (U).
- timecode for the script words at the beginning of the scene/clip that precede the first alignment point may be interpolated across the time interval that extends from the beginning of speaking of the corresponding transcript words in the scene/clip to the corresponding alignment point (e.g., interpolated between the timecode of the transcript words “get” and “to” in the transcript phrase 704 ).
- the number of corresponding unmatched script words and transcript words has a one-to-one correspondence, and, thus, timecode information may be directly correlated.
- the first three script words (“It”, “is” and “good”) may each be assigned timecodes of the first three transcript words (“get”, “it” and “could”), respectively.
- the location of the alignment points in the middle of the last sentence may enable the unmatched words “about”, “it”, “later”, and “tonight” that are located between the last alignment point of the scene/clip and the period indicative of the end of the scene/clip, to be interpolated across the interval between the transcript words “you” and “flight” and/or to each be assigned timecode information corresponding to transcript words “house”, “get”, “gator”, and “flight”, respectively.
- script elements may be used to identify the beginning or end of a sentence. For example, if between two lines of dialog, there is a parenthetical script element that corresponds to a sound effect, such as a car crash, the presence of the sound effect, indicated by a pause or stop words, may be used to identify the beginning or end of adjacent lines of dialog.
- the techniques described with regard to alignment points in the middle of a sentence at the beginning or end of a scene/clip may be employed.
- the timecodes for the unmatched words that occur between the alignment point and the identifiable script element may be interpolated across the corresponding interval or otherwise be determined. That is, the intermediate script element may be used in the same manner as capitalization and/or punctuation is used as described above.
- the density of the words in the transcript may be used to assess and determine timecode information associated with the words in the script. For example, in the illustrated embodiment of FIG. 7 , there are four unmatched transcript words in the interval of phrase 701 between matched words (e.g., “two” and “know”) and there are five unmatched words (e.g., “my”, “question”, “I”, “need” and “to”) in the corresponding interval of phrase 700 between matched words (e.g., “to” and “know”). Based on the timecode information for the transcript words in the interval, it may be determined that two of the four unmatched transcript words are spoken at the beginning of the interval and that two of the four unmatched transcript words are spoken at the end of the interval.
- timecode information for the transcript words in the interval it may be determined that two of the four unmatched transcript words are spoken at the beginning of the interval and that two of the four unmatched transcript words are spoken at the end of the interval.
- a corresponding percentage of the script words (e.g., approximately equal to the percentage of transcript words) will be provided over the respective portions of the interval. For example, in the embodiment of FIG.
- the word “to” (in the first portion of the phrase 700 ) that defines a start of the interval is determined to have been spoken at exactly twenty-one minutes (21:00.0)
- the word “know” defining an end of the interval is determined to have been spoken at exactly twenty-one minutes and ten seconds (21:10.0)
- the word “position” is determined to have been spoken at exactly twenty-one minutes and ten and two-tenths seconds (21:00.2)
- the word “eye” is determined to have been spoken at exactly twenty-one minutes and nine and four-tenths seconds (21:09.4)
- the two unmatched script words “my” and “question” may be evenly spaced over the first portion of the interval from twenty-one minutes (21:00.0) to twenty-one minutes and ten and two-tenths seconds (21:00.2)
- the three unmatched words “I”, “need” and “to” may be evenly spaced across the third portion of the interval from twenty-one minutes and nine and four-tenths seconds (21:09.4) to twenty-one minutes
- the distribution of script words within the interval is approximately equivalent to the distribution of transcript words in the corresponding interval. That is, about fifty percent of the script words in the interval are time aligned across the first portion of the interval before the pause and about fifty percent of the script words in the interval are time aligned across the third portion of the interval after the pause.
- a plurality of script words may be accepted for use in the time-aligned script data based on a confidence (e.g., high probability/density of word matches that were previously determined).
- a confidence e.g., high probability/density of word matches that were previously determined.
- Such a technique may enable blocks of text to be verified/imported from the script data to the time-aligned script data when matches within the blocks are indicative of a high probability that the corresponding script words are accurate. That is, the script data will be the text used in the time-aligned script data for those respective words of the script/dialogue.
- a block of script words may be imported when word matches (e.g., hard alignment points and/or soft alignment points) meet a threshold level.
- verifying/importing blocks of text may include using some individual script words having a match (e.g., associated with hard and/or soft alignment points) with words of the script, while importing/using unmatched transcript words (e.g., that are not associated with a soft and/or hard alignment points).
- verifying/importing script words may include importing text characteristics, such as capitalization, punctuation, and the like. In the embodiment of FIG.
- the script text may be used for the entire block of text in the aligned script document, including matched and unmatched words for use in the script-aligned data. For example, the block of corresponding script text “What is your answer to my question? I need to know your answer now!” may be used in the aligned script although all of the words do not have a match.
- the imported script words have incorporated the capitalization and punctuation of the corresponding text of the script document.
- Timecode information may be associated with each of the script and transcript words using any of the techniques described herein to properly time align the unmatched words of the phrase (e.g., to provide timecodes for the words “my question? I need to”).
- the transcript words including those not matched
- the transcript words may be used in the resulting time-aligned script. Accordingly, if the transcripts words of the phrase “What is your answer to by position eye do know your answer now!” have a high confidence leave but are not all matched, the phrase may be used in the resulting text of the time-aligned script data. Note that both, the matched and unmatched words of the raw STT have been imported.
- Such a technique may facilitate use of transcript words in place of script words where the actor ad-libs or otherwise does not recite the exacting wording of the script.
- a user could choose for themselves whether to use the Script word(s) or SST transcript word(s), based on an indication, such as confidence level. For example, even if the confidence level assumes one is more accurate than the other, it may not be so, and the user may be provided an opportunity to correct this by switching use of one or the other in the script data. Also, the user can manually edit in a correction, and this correction could be automatically stamped with a 100% confidence label. In some embodiments, the automated changes/imports may be marked such that a user can readily identify them, and modify them as needed.
- confidence/probability information provided during STT operations may be employed to assess whether or not a word or block of words in a transcript meets threshold criteria, such that the transcript words may be used in the time-aligned script data in place of the corresponding script words.
- Such an embodiment may resolve discrepancies by using the transcript word in the aligned script data 116 where there is a high confidence that the transcript word is accurate and the corresponding script word is not (e.g., where an actor ad-libs a line such that the actual words spoken are different from the words in the script).
- an STT engine may provide a high confidence level (e.g., above 90%) for a given transcript word, and, thus, the transcript word is considered to meet the threshold criteria (e.g., 85% or above). That is, the word in the transcript may be more accurate than corresponding script words. As a result, the transcript word is provided in the aligned script data, in place of a corresponding script word.
- a confidence/probability provided by an STT operation may be used in combination with matching criteria.
- the transcript word may be provided in the aligned script data, in place of a corresponding script word.
- a high confidence level e.g., above 90%
- the STT engine provides a low confidence level (e.g., below 50%) for a corresponding transcript word
- the script word may be provided in the aligned script data, in place of a corresponding transcript word.
- a portion of the script may be longer than a corresponding clip.
- the portion of the script that is actually spoken may be time aligned appropriately, and the unspoken portions of the script may be bunched together between aligned points.
- the bunching of words may result in timecode information being associated with the bunched words that indicates them being spoken at an extremely high rate, when in fact they may not have been spoken at all.
- a threshold is applied to ignore or delete words that appear to have been spoken too quickly such that bunched words may be ignored or deleted.
- a threshold word rate may be set to a value that is indicative of the fastest reasonable rate for a person to speak (e.g., about six words per second).
- the threshold word rate may be set to a default value, may be determined automatically, or may user selected.
- a speaking rate may be customized based on the character speaking the dialogue. For example, one actor may speak slowly whereas another actor may speak much faster, and thus the slower speaking character's dialogue may be associated with a lower threshold rate, where as the faster speaking character's dialogue may be associated with a higher threshold rate.
- Automatically determining a threshold word rate may include sampling other spoken portions of a script (e.g., other lines delivered by the same character) to determine a reasonable rate for words that are actually spoken, and the threshold rate may be set at that value or based off of that value.
- a maximum word rate threshold may be set to approximately twenty percent greater than that value (e.g., about six words per second). Such a cushion may account for natural variations in speaking rate that may occur while still identifying unlikely variations in speaking rate. In some embodiments, words having spacing that do not fall within the maximum word rate threshold are ignored or deleted, such that they are not aligned. For example, a script may read:
- words that were bunched at the beginning or end of dialogue may be identified and removed.
- the following lines at the beginning of the dialogue were linearly interpolated:
- the bunched words are deleted/ignored such that they are not included or indicated as being aligned in the resulting aligned script data.
- interpolated alignment of text that is located at the beginning or end of dialogue and that is bunched into a short duration may be deleted/ignored.
- ignoring/deleting words that appear to exceed a maximum threshold rate may also help to eliminate “stopwords” generated by an STT engine from being considered for alignment. For example, where an STT engine inserts a plurality of “the, the, the, . . . ” in place of music or sound effects, the high frequency of the words “the” may be identified and they may be ignored/deleted such that they are not aligned to words in the script.
- the stopwords may be flagged (e.g., not recognized) so that a user can take further action if desired.
- a clip may include audio content having extraneous spoken words that are not intended to be aligned with corresponding script words.
- extraneous words and phrases may include an operator calling out “Speed!” shortly before starting the camera rolling while audio is already being recorded, the director calling out “Action!” shortly before the characters beginning to speak lines of dialogue, the director calling out “Cut!” at the end of a take, or conversations inadvertently recorded shortly before, after, or even in the middle of a take. These cues typically occur at the beginning and end of shots, and, thus, processing may be able to recognize these words based on their location and/or their audio-waveforms that are recognized and provided in a corresponding STT transcript.
- synchronization module 102 may align the extraneous words of the transcript to script words, resulting in numerous errors.
- User defined words such as “Speed”, “Action” and “Cut” may be defined and can be recognized by their audio waveforms and provided in a corresponding STT transcript. The user defined words may be automatically flagged for the user or deleted.
- only a defined range of recorded dialogue is aligned to script text. Such a technique may be useful to ignore or eliminate extraneous recorded audio from the alignment analysis. For example, defining a range of recorded dialog may enable the analysis to ignore extraneous conversations or spoken words that are incidentally recorded just before or after a take for a given scene.
- an in/out range defines the portion of the audio that is aligned to a corresponding portion of the script.
- Defining an in/out range may define discrete portions of the script (e.g., script word) and/or audio content (e.g., transcript words) to analyze while also defining discrete portions of the audio content data to ignore during the alignment of transcript words with corresponding script words, thereby preventing extraneous words (e.g., transcript words) from inadvertently being aligned with script words.
- FIG. 7C is a depiction of a line of text and corresponding in/out ranges in accordance with one or more embodiments of the present technique. More specifically, FIG. 7C illustrates an exemplary in-range 710 and out-ranges 711 .
- the in-range 710 and out-ranges 711 limits analysis to only audio content of in-range 710 , referred to herein as audio content of interest 712 , and excludes audio content not located within in-range 710 (e.g., content located in out-ranges 711 ).
- Audio content of interest 712 may include the dialogue or narration spoken during the respective clip that falls within one or more specified in/out-ranges.
- Extraneous audio content 714 may include words captured on the audio that are not intended to be aligned with a corresponding portion of script document, and, thus, fall outside of the one or more specified in/out-ranges.
- audio content of interest 712 includes the transcribed phrase “hello mike . . .
- extraneous audio content 714 includes the phrases/words “are we ready speed action” spoken at the head of the clip, just before audio content of interest 712 and “cut how did that look” spoken at the tail of the clip, just after audio content of interest 712 .
- in range 710 is defined by an in-marker 710 a and an out-marker 710 b .
- In-marker 710 a defines a beginning of audio content of interest 712
- out-marker 710 b defines an end of audio content of interest 712 .
- extraneous content 714 at the head and tail of the clip is ignored during analysis, as indicated by the grayed out bar in FIG. 7C .
- embodiments may include multiple discrete ranges defined within a single clip.
- two additional in/out markers may be added within in-range 710 , thereby dividing it into two discrete in-ranges and providing an additional out-range embedded therein.
- the use of in/out-ranges may be employed to resolve issues normally associated with multiple takes of a given scene or clip.
- an out-range may be located at any portion of the clip.
- the in/out-ranges may be swapped, thereby ignoring extraneous audio data in the middle of the clip, while analyzing audio content of interest at the head and tail of the clip.
- markers 710 a and 710 b may be user defined. For example, a user may be presented with a display similar to that of FIG. 7C and may use a slider-type control to move markers 710 a and 710 b , thereby windowing in/out-ranges 710 and 711 . Thus, a user may view some or all of the text and may cut-out the extraneous audio content 714 using in/out-ranges.
- markers 710 a and 710 b may be defined as an offset of a given duration of time or number of words. For example, an offset of ten-seconds may exclude ten seconds of audio data at the head or tail of the clip.
- Such a technique may be of particular use where there is a consistent delay at the beginning or end of filming a clip.
- An offset of five words may exclude the first and/or last five words of spoken dialog at the head or tail of the clip.
- Such a technique may be of particular use where there is a consistent phrase or series of words spoken at the beginning or end of filming a clip.
- the offsets may be predetermined and/or user selectable. For example, a default offset value may be employed, but may be editable by a user (e.g. via a sliding window as described above).
- portions of the audio content may include extraneous audio other than spoken words, such as music or sound effects. If analyzed, the extraneous audio may create an additional processing burden on the system. For example, synchronization module 102 may dedicate processing in an attempt to match/align extraneous transcript words (e.g., stop words) to script words. In some embodiments, the extraneous audio content may be identified and ignored during alignment. Such a technique may enable processing to focus on dialogue portions of audio content, while skipping over segments of extraneous audio. In some embodiments, the audio content may be processed to classify segments of the audio content into one of a plurality of discrete audio content types.
- segments of the audio content identified as including dialogue may be classified as dialogue type audio
- segments of the audio content identified as including music may be classified as music type audio
- segments of the audio content identified as including sound effects may be classified as sound effect type audio.
- segments of transcript words that include a series of different words occurring one after another (e.g., how are you doing) and/or that are not indicative of stop words may be classified as a dialogue type audio
- segments of transcript words that include a series of stop words of a long duration e.g., the the the the . . .
- segments of transcript words that include a series of stop words of a short duration e.g., the the the the
- segments of the audio content that cannot be identified as one of dialogue, music or sound effect type audio may be categorized as unclassified type audio.
- each of the segments may or may not be subject to alignment or related processing based on their classification. For example, during alignment of transcript words to script words, the segments associated with dialogue type audio may be processed, whereas the segments associated with music and sound effect type audio may be ignored. By ignoring music and sound effect type segments, processing resources may be focused on the dialogue segments, and, thus, are not wasted attempting to align the transcript words associated with the music and sound effect to script words.
- unclassified type audio may be considered for alignment or may be ignored.
- what classifications are processed and what classifications are ignored may include a default setting and/or may be user selectable.
- a weighting value is assigned to each word based on the alignment type (e.g., interpolation, hard alignment, or soft alignment). Stronger alignments (e.g., hard and soft alignments) may have higher weighting than weaker alignments (e.g., interpolation).
- a total weighting is assessed for a window/interval that includes several consecutive words. The interval of several words is a sliding window that is moved to assess adjacent intervals/windows of words.
- the total weighting e.g., sum of weightings
- timecodes may be assigned to one or more of the words, thereby, not ignoring/deleting the words in the window.
- Such a technique may be provided at the beginning and end of a set of dialogue to assess and determine the start and stop of the actual spoken dialogue and to ignore/delete the script dialogue that preceded/followed the spoken dialogue in the script, but was not actually spoken (e.g., the script text that was linearly interpolated as was bunched before or after the dialogue actually spoken).
- processing may be implemented to time-align script elements other than dialogue (e.g., scene headings, action description words, etc.) directly to the video scene or full video content.
- script elements other than dialogue (e.g., scene headings, action description words, etc.) directly to the video scene or full video content.
- a script element, other than dialogue e.g., a scene heading
- timecodes of the words may be used to determine a timecode of the intervening script element.
- a script element occurring in the script between the two words may be assigned a timecode between 21:00.00 and 21:10.00, such as 21:05.00.
- one or more script elements may have their timecodes determined via linear and/or non-linear interpolation, similar to that described above.
- the amount of content e.g., the number of lines or number of words
- script elements may be used to assess a timecode for a given script element or plurality of script elements.
- first script element between two words having timecodes includes half the amount of content of a second script element also located between the two words
- the first script element may be assigned a timecode of 21:03.00 and the second script element may be assigned a time code of 21:05.00, thereby reflecting the smaller content and potentially shorter duration of the first element relative to the second element.
- some or all of the script elements may be provided in the time-aligned script data in association with a timecode.
- timecodes are first assigned to the dialogue words during initial alignment, and timecodes are assigned to the other script elements in a subsequent alignment process based on the timecodes of the dialogue determined in the initial alignment (e.g., via interpolation).
- the resulting time aligned data 116 may include timecodes for some or all of the script elements of script 104 .
- method 400 includes generating a time-aligned script output, as depicted at block 422 , as discussed above.
- Generating time-aligned script output may be provided via time-coded script generator 212 .
- each word or element of the script and/or transcript may be associated with a corresponding timecode.
- the complete list of script word and/or transcript words that are associated with hard, soft and interpolated timecodes may be used to generate time-aligned data 116 , including a final TimeCodedScript (TCS) data file which contains some or all of the script elements with assigned time codes.
- TCS TimeCodedScript
- the TCS data file may be provided to another application, such as the Adobe Script Align and Replace feature of Adobe Premiere Pro, for additional processing.
- time-aligned data 116 may be stored in a database for use by other applications, such as the Script Align feature of Abode Premiere Pro.
- a graphical user interface may provide a graphical display that indicates where matches (e.g., hard and/or soft alignment points) or non-matches occur within a user interface.
- the user interface may include symbols or color coding to enable a user to readily identify various characteristics of the alignment. For example, hard alignments may be provided in red (or green) to indicate a good/high confidence, soft alignments in blue (or yellow) to indicate a lower confidence, and interpolated points in yellow (or red) to indicate an even lower confidence level.
- the user interface may enable a user to quickly scan the results to assess and determine where inaccuracies are most likely to have occurred.
- a user may commit resources for review and proofing efforts on portions of a time-aligned script that may be susceptible to errors (e.g., where no or few matches occur) and may not commit resources for review and proofing efforts on portions of a time-aligned script that may not be susceptible to errors (e.g., where a large number of matches occur).
- a user may be presented with a chart, such as that illustrated in FIG. 5A .
- the chart may enable a user to readily identify portions of the script that do not include a high percentage of matches (e.g., the sub-matrix 508 located at the uppermost left portion of the chart).
- high confidence areas may include a similar visual indicator (e.g., grayed out) and portions that may require attention may have appropriate visual indicators (e.g., bright colors—not grayed out).
- a user may be provided the option to select whether or not to use the text from the raw STT analysis or the text from the written script. For example, a user may be provided a selection in association with the sub-matrix 508 located at the uppermost left portion of the chart that enables all, some, or individual words contained in the sub-matrix to use the text from the raw STT analysis or the text from the written script.
- the information may be returned to synchronization module 102 and processed in accordance with the user input. For example, where a user opts to use STT text in place of script text, synchronization module 102 may conduct additional processing to provide the corresponding time-aligned script data. In some embodiments, the user may be prompted for input while synchronization module 102 is performing the time alignment. For example, as the synchronization module 102 encounters a decision point, it may prompt the user for input.
- speech-to-text analysis may provide the option of creating a custom dictionary (e.g., custom language model).
- a custom dictionary may be generated for a given clip based on one or more reference scripts that have content that is the same or similar to the given script, or based on a single reference script that at least partially corresponds to the video content or exactly matches the audio portions of the video content.
- some or all words of the reference script may be used to define a custom dictionary
- a raw speech analysis may be performed to generate a transcript using words of the custom dictionary to transcribe words of the audio content, transcript words may then be matched against the script words of the reference script to find alignment points, and the words of the reference script text may be paired with the corresponding timecodes, thereby providing a time-aligned/coded version of the reference script.
- a custom language model is generated for one or more portions of video content. For example, where a movie or scene includes a plurality of clips, a custom language module may be provided for each clip to improve speech recognition accuracy.
- a custom language model is provided to a STT engine such that the STT engine may be provided with terms that are likely to be used in the clip that is being analyzed by the STT engine. For example, during STT transcription, the STT engine may at least partially rely on terms or speech patterns defined in the custom language model.
- a custom language model may be directed toward a certain sub-set of language. For example, the custom language model may specify a language (e.g., English, German, Spanish, French, etc.).
- the custom language model may specify a certain language segment.
- the custom language module may be directed to a certain profession or industry (e.g., a custom language module including common medical terms and phrases may be used for clips from a medical television series).
- the STT engine may weight words/phrases found in the associated custom language module over the standard language model. For example, if the STT engine associates a word with a word that is present in the associated custom language model and a word that is present in a standard/default language model, the STT engine may select the word associated custom language model as opposed to the word present in the standard/default language model.
- a word identified in a transcript that is found in the selected custom language model may be assigned a higher confidence level than a similar word that is only found in the standard/default language model.
- a custom language model is generated from script text.
- script data 110 may include embedded script text (e.g., words and phrases) that can be extracted and used to define a custom language model.
- embedded metadata may be provided using various techniques, such as those described in described in U.S. patent application Ser. No. 12/168,522 entitled “SYSTEMS AND METHODS FOR ASSOCIATING METADATA WITH MEDIA USING METADATA PLACEHOLDERS”, filed Jul. 7, 2008, which is hereby incorporated by reference as though fully set forth herein.
- a custom language model may include a word frequency table (e.g., how often each of the words in the custom language model is used within a given portion of the script) and a word tri-graph (e.g., indicative of other words that precede and followed a given word in a given portion of the script).
- all or some of the text identified in the script may be used to populate the custom language model.
- Such a technique may be particularly accurate because the script and resulting language model should include all or at least a majority of the words that are expected to be spoken in the clip.
- speech-to-text (STT) technology may implement a custom language model as described in U.S.
- metadata included in the script may be used to further improve accuracy of the STT analysis.
- the script includes a clip identifier, such as a scene number
- the scene number may be associated with the clip such that a particular custom language model is used for STT analysis of video content that corresponds to the associated portion of the script.
- a first portion of the script is associated with scene one and a second portion of the script is associated with scene two
- a first custom language model may be extracted from the first portion of the script
- a second custom language model may be extract from the second portions of the script.
- the STT engine may automatically use the first custom language model
- the STT engine may automatically use the second custom language model.
- a clip when a clip contains only a few lines of dialogue in a short scene out of a very long script, knowing that the clip contains a specific scene number (e.g., harvested from the script metadata) allows focusing on the text in the script for that scene, and not having to assess the entire script.
- a specific scene number e.g., harvested from the script metadata
- FIG. 6 depicts a sequence of dialogs 600 in accordance with one or more embodiments of the present technique.
- a user may select a clip or group of clips, then chooses “Analyze Content” from a Clip menu, initiating the sequence of dialogs 600 .
- the Analyze Content dialog may allow a user to use embedded Adobe Story Script text if present for the speech analysis, or to add a reference script which will be used to improve speech analysis accuracy.
- the sequence of dialogs 600 includes content analysis dialogs that allow users to import a reference script to create a custom dictionary/language model for speech analysis.
- a reference script may include a text document containing dialogue text similar to the recorded content in the project (e.g., a series of nature documentary scripts, or a collection of scripts from a client's previous training videos).
- a user may choose Add from the Reference Script menu.
- the File Open dialog 604 a user may navigate to the reference script text file, select it and click OK.
- the Add Reference Script dialog 606 may open, where a user can name the reference script, choose a language, and view the text of the file below in a scrolling window.
- the “Script Text Matches Recorded Dialogue” option may be selected if the imported script exactly matches the recorded dialogue in the clips (e.g., a script the actors read their lines from).
- the analysis engine automatically sets the weighting of the reference script vs. the base language model based on length, frequency of key words, etc.
- a user may click the OK button, the Import Script dialog closes, and the analysis of the reference script may begin.
- the reference script is selected in the Analyze Content's Reference Script menu.
- the OK button the selected clip's speech content is analyzed.
- the reference script matches the recorded dialogue exactly (e.g., the script that was written for the project or transcriptions of interview sound bites).
- a user may select the “Script Text Matches Recorded Dialogue” option in the Add Reference Script dialog 606 , as discussed above. This may override the automatic weighting against the base language model and give the selected reference script a much higher weighting.
- Significantly higher accuracy can be achieved using matching reference scripts, although accuracy may be primarily dependent on the clarity of the spoken words and the quality of the recorded dialogue.
- an Adobe Story to Adobe OnLocation workflow may be used to embed the dialogue from each scene into a clip's metadata.
- a script written in Adobe Story may be imported into OnLocation, which may produce a list of shot placeholders for each scene.
- These placeholders may be recorded direct to disk using OnLocation during production or merged with clips that are imported into OnLocation after they were recorded on another device.
- the text for each scene from the original script may be embedded in the metadata of all the clips that were shot for that scene.
- Embedded metadata may be provided using various techniques, such as those described in described in U.S. patent application Ser. No. 12/168,522 entitled “SYSTEMS AND METHODS FOR ASSOCIATING METADATA WITH MEDIA USING METADATA PLACEHOLDERS”, filed Jul. 7, 2008, which is hereby incorporated by reference as though fully set forth herein.
- the script text embedded in each of the clips may be automatically used as a reference script and, then, aligned with the recorded speech during the analysis.
- the analyzed speech text is replaced with the script text embedded in the source clip's extensible metadata platform (XMP) metadata.
- XMP extensible metadata platform
- the “Use Embedded Adobe Story Script Option” of Analyze Content dialog 602 when the “Use Embedded Adobe Story Script Option” of Analyze Content dialog 602 is selected, Adobe Story script text embedded in an XMP will be used for analysis, and the Reference Script popup menu may be disabled. If the selected clip contains Adobe Story script embedded text, the “Use Embedded Adobe Story Script Option” may be checked by default. For mixed states in the selection (e.g., where at least one clip has Adobe Story script text embedded, and at least one clip does not), the dialog will open with the “Use Embedded Adobe Story Script Option” checkbox indicating a mixed state and the Reference Script popup menu may be enabled.
- the clip with the Adobe Story script embedded will be analyzed using the Adobe Story script and the clip without the Adobe Story script embedded will be analyzed using the reference script. Selecting the mixed state may generate a check in the “Use Embedded Adobe Story Script Option” checkbox and disable the “Reference Script” menu. If the analysis is run in this state, the result may be the same as above. Selecting the checkbox again may remove the check mark at the “Use Embedded Adobe Story Script Option” checkbox and may re-enable the “Reference Script” menu. If the analysis is run in this state, all clips may use the assigned reference script, and ignore any embedded Story Script text that may be in one or more of the selected clips.
- an STT engine may require that a custom language model include a minimum number of words (e.g., a minimum word count). That is, an STT engine may return an error and/or ignore a custom language model if the model does not include a minimum number of words. For example, if a portion of a script includes only ten words, a corresponding custom language model may include only the ten words. If the STT engine required a minimum of twenty-five words, the STT may not be able to use the custom language model having only ten words. In some embodiments, the words in the custom language model may be duplicated to meet the minimum word count.
- a minimum number of words e.g., a minimum word count
- the ten words may be repeated two additional times in an associated document or file that defines the custom language model to generate a total of thirty words, thereby enabling the resulting custom language model to meet the minimum word requirement of twenty-five words. It is noted that if all of the words are replicated the same number of times, the word frequency table (e.g., how often each of the words in the custom language model is used), and the word tri-graph (e.g., indicative of other words that precede and followed a given word) of the custom language model should remain accurate. That is the frequencies and words that precede or follow a given word remain the same.
- entities e.g., dialogue and events
- users e.g., marketing personnel, advertisers, and legal personnel
- users may be interested in identifying and locating when specific people, places, or things occur in the final production video or film to enable, for example, identifying prominent entities that occur in a scene in order to perform contextual advertising (e.g., an advertisement showing a certain type of car ad if the car appears in a crucial segment.)
- contextual advertising e.g., an advertisement showing a certain type of car ad if the car appears in a crucial segment.
- the processed script, extracted entities, and time-aligned dialogue/entity metadata may enable third-parties applications (e.g., contextual advertisers) to perform high relevancy ad placement.
- a method for identifying and aligning some or all entities within a script includes receiving script data, processing the script data, receiving video content data (e.g., video and audio data), processing the video content data, and synchronizing the script data with the video content data to generate time-aligned script data, and categorizing each regular or proper noun entity within the time-aligned script data.
- receiving and processing script data and receiving and processing video content data are performed in series or parallel prior to performing synchronizing the script data with the video content data which is flowed by categorizing each regular or proper noun entity within the time-aligned script data.
- Receiving script data may include processes similar to those above described with respect to document extractor 108 .
- receiving script data may include accepting a Hollywood “Spec.” Movie Script or dramatic screenplay script document (e.g., document 104 ), converting this script into specific structured and tagged representation (e.g., document data 110 ) via systematically extracting and tagging all key script elements (e.g., Scene Headings, Action Descriptions, Dialogue Lines), and then storing these elements as objects in a specialized document object model (DOM) (e.g., a structured/tagged document) for subsequent processing.
- DOM specialized document object model
- Processing the script data may include extracting specific portions of the script. Extracted portions may include noun items.
- processing script data may include processing the objects (e.g., entire sentences tagged by script section) within the script DOM using an NLP engine that identifies, extracts, and tags the noun items identified by the system for each sentence. The extracted and tagged noun elements are then recorded into a specialized metadata database.
- Receiving video content data may include processes similar to those described above with respect to audio extractor 112 .
- receiving video content data may include receiving a video or audio file (e.g., video content 112 ) that contains spoken dialogue that closely but not necessarily exactly corresponds to the dialogue sections of the input script (e.g., document 104 ).
- the audio track in the provided video or audio file is then processed using a Speech-to-Text engine (e.g., audio extractor 112 ) to generate a transcription of the spoken dialogue (e.g., transcript 114 ).
- the transcription may include extremely accurate timecode information but potentially higher error rates due to noise and language model artifacts. All spoken words and timecode information of the transcript that indicates at exactly what point in time in the video or audio the words were spoken, is stored.
- Synchronizing the script data with the video content data to generate time-aligned script data may include processes similar to those described above with respect to synchronization module 102 .
- synchronizing the script data with the video content data to generate time-aligned script data may include analyzing and synchronizing the structured (but untimed) information in a tagged script document (e.g., document data 110 ) and the text resulting from the STT transcription stored in metadata repository (e.g., transcript 114 ) to generate a time-aligned script data (e.g., time aligned script data 116 ).
- the time-aligned script data is provided to a named Entity Recognition system to categorize each regular or proper noun entity contained within the time-aligned script data.
- FIGS. 8A and 8B are block diagrams that illustrates components of and dataflow in a document time-alignment technique in accordance with one or more embodiments of the present technique. Note, the dashed lines indicate potential communication paths between various portions of the two block diagrams.
- System 800 may include features similar to that of previously described system 100 .
- script data is provided to system 800 .
- Script document/data 802 may be similar to document 104 .
- movie script documents, closed caption data, and source transcripts are presented as inputs to the system 100 .
- Movie scripts may be represented using a semi-structured Hollywood “Spec.” or dramatic screenplay format which provides descriptions of all scene, action, and dialogue events within a movie.
- script data 802 may be provided to a script converter 804 .
- Script converter 804 may be similar to document extractor 108 .
- script elements may be systematically extracted and imported into a standard structured (e.g., XML, ASTX, etc.).
- Script converter 804 may enable all script elements (e.g., Scenes, Shots, Action, Characters, Dialogue, Parentheticals, and Camera transitions) to be accessible as metadata to applications (e.g., Adobe Story, Adobe OnLocation, and Adobe Premiere Pro) enabling indexing, searching, and organization of video by textual content.
- applications e.g., Adobe Story, Adobe OnLocation, and Adobe Premiere Pro
- Script converter 804 may enable scripts to be captured from a wide variety of sources including: professional screenwriters using word processing or script writing tools, from fan-transcribed scripts of film and television content, and from legacy script archives captured by OCR.
- Script converter 804 may employ various techniques for extracting and transcribing audio data, such as those described in described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein.
- converted script data 805 (e.g., an ASTX format movie script) from script converter 804 may be provided to a script parser 806 .
- parser may be implemented as a portion of document extractor 108 .
- Spec. scripts captured and converted into a standard (e.g., Adobe) script format may be parsed by script parser 806 to identify and tag specific script elements such as scenes, actions, camera transitions, dialogue, and parenthetical.
- the ability to capture, analyze, and generate structured movie scripts may be used in certain time-alignment workflows (e.g., Adobe Pro “Script Align” feature where dialogue text within a movie script is automatically synchronized to the audio dialogue portion of video content).
- parsed script data is processed by a natural language (processing) engine (NLP) 808 .
- NLP natural language
- a filter 808 a analyzes dialogue and action text from the parsed script data. For example, the input text is normalized and then broken into individual sentences for further processing. Each sentence may form a basic information unit for lines of the script, such as lines of dialogue in the script, or descriptive sentences that describe the setting of a scene or the action within a scene.
- grammatical units of each sentence are tagged at a part-of-speech (POS) tagger 808 b .
- POS tagger 808 b may use a transformational grammar rules technique to first induce and learn a set of lexical and contextual grammar rules from an annotated and tagged reference corpus, and then apply the learned runs for performing the POS tagging step of submitted script sentences.
- tagged verb and noun phrases are submitted to a Named Entity Recognition (NER) system 808 c .
- NER system 808 c may then identify and classify entities and actions within each verb or noun phrase.
- NER 808 c may employ one or more external world-knowledge ontologies (API's) to perform the final entity tagging and classification.
- API's world-knowledge ontologies
- some or all extracted entities from NER system 808 c are then represented using a script Entity-Relationship (E-R) data model 810 that includes Scripts, Movie Sets, Scenes, Actions, Transitions, Characters, Parentheticals, Dialogue, and/or Entities.
- E-R script Entity-Relationship
- the instantiated model 810 may be physically stored into a relational database 812 .
- the instantiated model 810 may be mapped into an RDF-Triplestore 814 (see FIG. 8B ).
- a specialized relational database schema may be provided for certain application (e.g., for Adobe Story).
- script metadata may be used to record all script metadata and entities and the interrelationships between all entities.
- a relational database to RDF mapping processor 816 may then used automatically processes the relational database schema representation of the E-R model 810 to transfer all script entities in relational database table rows into the RDF-Triplestore 814 .
- Mapping may include RDF mapping system and process techniques, such as those described in described in U.S. patent application Ser. No. 12/507,746 entitled “CONVERSION OF RELATIONAL DATABASES INTO TRIPLESTORES”, filed Jul. 22, 2009, which is hereby incorporated by reference as though fully set forth herein.
- E-R model 810 may be saved to relational database 812 .
- Relational database 812 may implement E-R model 810 though a set of specially defined tables and primary key/foreign key referential integrity constraints between tables.
- an RDF-Triplestore 820 may be used to store to the mapped relational database 812 using output of relational database to RDF mapping processor 816 .
- RDF-Triplestore 820 may represent the relational information as a directed acyclic graph and may enable both sub-graph and inference chain queries needed by movie or script query applications that retrieve script metadata.
- Use of RDF-Triplestore 820 may allow video scene entities to be queried using an RDF query language such as SPARQL or a logic programming language, like Prolog.
- RDF-Triplestore enables certain kinds of limited machine reasoning and inferences on the script entities (e.g., finding prop objects common to specific movie sets, classifying a scene entity using its IS_A generalization chain for a particular prop, or determining the usage and ownership rights to specific cartoon characters within a movie, for example.
- Script dialogue data may be stored within RDF-Triplestore 820 .
- an application server 822 may be used to process incoming job requests and then communicate RDF-Triplestore data back to one or more client applications 824 , such as Adobe Story.
- Application server 822 may contain a workflow engine along with one or more optional web-servers. Script analysis requests or queries for video and script metadata may be processed by server 822 , and then dispatched to a workflow engine which invokes either the NLP analysis engine 808 or a multimodal video query engine 826 .
- Application server 822 may include a Triad/Metasky web server.
- client application 824 may be used to implement further processing.
- Adobe Story is a product that a client may use to leverage outputs of the workflows described herein to allow script writers to edit and collaborate on movie scripts, to extract, index, and to tag script entities such as people, places, and objects mentioned in the dialogue and action sections of a script.
- Adobe story may include a script editing service.
- the above described steps may describe certain aspects of text processing.
- the following described steps may describe certain aspects of video and audio processing.
- video/audio content 830 is input and accepted by the workflow system 800 .
- Video/audio content 830 may be similar to that of video content 106 .
- Video/audio content 830 may provide video footage and corresponding dialogue sound tracks.
- the audio data may be analyzed and transcribed into text using an STT engine, such as those described herein.
- a resulting generated STT transcript (e.g., similar to transcript 114 ) may be aligned with converted textual movie scripts 805 .
- the STT transcript may be processed by the natural language analysis and entity extraction components for keyword searching of the video. Natural language analysis and entity extraction components for keyword searching of the video may use multimodal video search techniques, such as those described in U.S.
- audio content is provided.
- input audio dialogue tracks may be directly provided by television or movie studios, or extracted from the provided video files using standard known extraction methods.
- the extracted audio may be converted to a mono channel format that uses 16-bit samples with a 16 kHz frequency response.
- an STT engine 832 is modified by use of a custom language model (CLM).
- STT engine 832 may employ transcription based at least partially or completely on a provided CLM.
- the CLM may be provided/built using certain methods, such as those described herein.
- STT engine 832 includes a multicore STT engine.
- the multicore STT engine may segment the source audio data, may provide STT transcriptions using parallel processing.
- speech-to-text (STT) technology may implement a custom language model and/or an enhanced multicore STT transcription engine such as those described in U.S.
- a metadata time synchronization service 834 aligns elements of transcript 832 with corresponding portions of script data 802 to generate time-aligned script data.
- Metadata time synchronization service 834 may be similar to synchronization module 102 .
- metadata time synchronization service 834 implements a specialized STT/Script alignment component to provide time alignment of non-timecoded words in the script with timecoded words in the STT transcript using a hybrid two-level alignment process, such as that described herein with regard to synchronization module 102 . For example, in level one processing, smaller regions or partitions of text and STT transcription keywords are accurately identified and prepared for detailed alignment.
- each script word may be assigned an accurate video timecode. This facilitates keyword search and time-indexing of the video by client applications such as the multimodal video search engine 826 , or other applications.
- a modified Viterbi and/or phonetic/text comparator is implemented by metadata time synchronization service 834 .
- the alignment process may also implement special override rules to resolve alignment option ties. As described herein, a decision as to whether or not an alignment is made may not rely only on precise text matches between the transcribed STT word and the script word, but rather, may rely on how closely words sound to each other; this may be provided for using a specialize phonetic encoding of the STT words and script words. Such a technique may be applicable to supplement a wide variety of STT alignment applications.
- data relating to the user is provided a graphical display that presents source script dialogue, the resulting time aligned words, and/or video content in association with one another.
- a GUI/visualization element of an application e.g., CS5 Premiere Pro Script Align feature
- a user may search a video based on the corresponding words in the time-aligned script data.
- a multimodal video search engine may allow a user to search for specific segments of video based on provided query keywords.
- the search feature may implement various techniques, such as those described in U.S. patent application Ser. No. 12/618,353 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, which is hereby incorporated by reference as though fully set forth herein.
- locations for the insertion of video descriptions can be located, video description content can be extracted from the script and automatically inserted into a time aligned script and/or audio track using time aligned script data (e.g., time aligned script data 116 as described with respect to FIGS. 1 and 2 ) provided by system 100 .
- Video descriptions may include an audio track in a movie or television program containing descriptions of the setting and action. Video description narrations fill in the story gaps by describing visual elements and provide a more complete description of what's happening in the program. This may be of particular value to the blind or visually impaired by helping to describe visual elements that they cannot view.
- the video description may be inserted into the natural pauses in dialogue or between critical sound elements, or the video and audio may be modified to enable insertion of video descriptions that may other wise be too long for the natural pauses.
- Video description content may be generated by extracting descriptive information and narrative content from a script written for the project, syncing and editing it to the video program for playback.
- Video description content may be extracted directly from descriptive text embedded in the script. For example, location settings, actor movements, non-verbal events, etc. that may be provided in script elements (e.g., title, author name(s), scene headings, action elements, character names, parentheticals, transitions, shot elements, dialogue/narrations, and the like) may be extracted as the video description content, aligned to the correct portion of scenes (e.g., to pauses in dialogue) using time alignment data, and the video description content may be manually or automatically edited (if needed) to fit into the spaces available between dialogue segments.
- script elements e.g., title, author name(s), scene headings, action elements, character names, parentheticals, transitions, shot elements, dialogue/narrations, and the like
- the video description content may be manually or automatically edited (if needed) to fit into
- the time aligned data acquired using system 100 may be used to identify the location of pauses within the audio content for embedding narrative content (e.g., action elements).
- the locations of the pauses in the audio content may be provided to a user as locations for inserting video description content.
- narrative content e.g., action element descriptions embedded in the script
- narrative content may be automatically inserted into corresponding pauses within the dialogue of the audio track to provide the corresponding video description content.
- the resulting video description content may be reviewable and editable by a user.
- a text version of the video description content can be used as a blueprint for recording by a human voiceover talent.
- the video description track can be created automatically using synthesized speech to read the video description content (e.g., without necessarily requiring any or at least a significant amount of human labor).
- a script may include a variety of script elements such as a scene heading, action, character, parenthetical, dialogue, transition, or other text that cannot be classified. Any or all of these and other script elements can be used to generate useful information for a video description track.
- a scene heading (also referred to as a “slugline”) includes a description of where the scene physically occurs.
- a scene heading may indicate that the scene takes place indoors (e.g., INT.) or outdoors (e.g., EXT.), or possibly both indoors and outdoors (e.g., INT./EXT.)
- a location name follows the description of where the scene physically occurs.
- “INT./EXT.” may be immediately followed by a more detailed description of where the scene occurs. (e.g., INT. KITCHEN, INT. LIVING ROOM, EXT. BASEBALL STADIUM, INT. AIRPLANE, etc.).
- the scene heading may also include the time of day (e.g., NIGHT, DAY, DAWN, EVENING, etc.). This information embedded in the script helps to “set the scene.”
- the scene type is typically designated as internal (INT.) or external (EXT.), and includes a period following the INT or EXT designation.
- a hyphen is typically used between other elements of the scene heading. For example, a complete scene heading may read, “INT. FERRY TERMINAL BAR—DAY” or “EXT. MAROON MOVIE STUDIO—DAY”.
- An action element typically describes the setting of the scene and introduces the characters in a scene. Action elements may also describe what will actually happen during the scene.
- a character name element may include an actual name (e.g., MS. SUTTER), description (e.g., BIG MAN) or occupation (e.g., BARTENDER) of a character. Sequence numbers are typically used to differentiate similar characters (e.g., COP #1 and COP #2). A character name is almost always inserted prior to a character speaking (e.g., just before dialog element), to indicate that the character's dialogue follows.
- a dialog element indicates what a character says when anyone on screen or off screen speaks. This may include conversation between characters, when a character speaks out loud to themselves, or when a character is off-screen and only their voice is heard (e.g., in a narration). Dialog elements may also include voice-overs or narration when the speaker is on screen but is not actively speaking on screen.
- a parenthetical typically includes a remark that indicates an attitude in dialog delivery, and/or specifies a verbal direction or action direction for the actor who is speaking the part of a character.
- Parentheticals are typically short, concise and descriptive statements located under the characters name.
- a transition typically includes a notation indicating an editing transition within the telling of a story.
- “DISSOLVE TO:” means the action seems to blur and refocus into another scene, as generally used to denote a passage of time. Transitions almost always follow an action element and precede a scene heading. Common transitions include: “DISSOLVE TO:”, “CUT TO:”, “SMASH CUT:”, “QUICK CUT:”, “FADE IN:”, “FADE OUT:”, and “FADE TO:”.
- a shot element typically indicates what the camera sees.
- a shot element that recites “TRACKING SHOT” generally indicates the camera should follow a character as he walks in a scene.
- “WIDE SHOT” generally indicates that every character appears in the scene.
- a SHOT tells the reader the focal point within a scene has changed.
- Example of shot elements include: “ANGLE ON . . . ”, “PAN TO . . . ”, “EXTREME CLOSE UP . . . ”, “FRANKIE′S POV . . . ”, and “REVERSE ANGLE . . . ”.
- script elements may be identified and extracted as described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein.
- the script elements may be time aligned to provide time-aligned data 116 as described herein.
- the time aligned data may include dialogue as well as other script elements having corresponding timecodes that identify when each of the respective words/elements occur within the video/audio corresponding to the script.
- FIG. 9A illustrates an exemplary script document 900 in accordance with one or more embodiments of the present technique.
- Script document 900 depicts an exemplary layout of the above described script elements.
- script document 900 includes a transition element 902 , a scene heading element 904 , action elements 906 a , 906 b and 906 c , character name elements 908 , dialog elements 910 , parenthetical elements 912 , and shot element 914 .
- Script writers and describers often have closely aligned goals to describe onscreen actions succinctly, vividly and imaginatively.
- the action element text may be the most useful for creating video description content, as action elements typically provide the descriptions that clearly describe what has happened, is happening, or about to happen in a scene.
- long text passages in a script describing major changes in the setting or complex action sequences translate to longer spaces between dialogue in the recorded program (often filled with music and sound effects) and provide opportunities for including longer segments of video description content.
- the action described under the scene heading 904 and action element 906 a is a wide establishing shot that follows the character out onto a busy studio lot.
- a user may have control over which script elements to use in creating a video description. For example, a user may select to use only action elements and shot elements and to ignore other elements of the script. In some embodiments, the selection may be done before or after the video description is generated. For example, a user may allow the system to generate a video description using all or some of the script elements, and may subsequently pick-and-choose which elements to keep after the initial video description is generated.
- FIG. 9B illustrates an exemplary portion of a video description script 920 that corresponds to the portion of script 900 of FIG. 9A .
- Video description script 920 includes a video description track 922 broken into discrete segments (1-9) provided relative to gaps and dialogue of an audio track (e.g., main audio program recorded dialogue) 924 that corresponds to spoken words of dialogue content of script 920 .
- the content of video description track 922 corresponds to action element text of action elements 906 a , 906 b and 906 c of script 900 of FIG. 9A .
- Each corresponding pause/gap in dialogue of audio track 922 is identified with a time of duration (e.g., “00:00:28:00 Gap” indicating a gap of twenty-eight seconds prior to the beginning of the script dialogue of segment 2 ).
- the corresponding content of video description 922 is provided adjacent the gap/pause, and is identified with a time of duration for the video description content (e.g., “00:00:27:00” indicating twenty-seven seconds for the video description content to be spoken) where applicable.
- the content of video description 922 may be modified to fit within the corresponding gap. For example, in the illustrated embodiment, a portion of the first segment of video description content is removed to enable the resulting video description content to fit within the duration of the gap when spoken.
- the entire video description content may be deleted or ignored where there is not a gap of sufficient length for the video description content.
- the video description content of segment 3 was deleted/ignored as the corresponding pause in dialogue was only about twelve frames (or 1 ⁇ 2 a second) in duration—too short for the insertion of the corresponding video description content.
- Video description script 920 and video description content 922 can be used as a blueprint for recording by a human voiceover talent. Thus, a voicer may simply have to read the corresponding narration content as opposed to having to manually search through a program, manually identify breaks in the dialog, and derive/record narrations to describe the video.
- the video description track can be created automatically using synthesized speech to read the video description content 922 (e.g., without necessarily requiring any or at least a significant amount of human labor).
- FIG. 9C is a flowchart that illustrates a method 950 of generating a video description in accordance with one or more embodiments of the present technique.
- Method 950 may provide video description techniques using components and dataflow implemented at system 100 .
- Method 950 generally includes identifying script elements, time aligning the script, identifying gaps/pauses in dialogue, aligning video description content to the gaps/pauses, generating a script with video description content, and generating a video description.
- Method 950 may include identifying script elements, as depicted at block 952 .
- Identifying script elements may include identifying some or all of the script elements contained within a script from which a video description is to be generated. For example, a script may be analyzed to provide script metadata that identifies a variety of script elements, such as scene headings, actions, characters, parentheticals, dialogue, transitions, or other text that cannot be classified.
- script elements may be identified and extracted as described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein.
- the identification of the elements may not actually be performed but may simply be provided or retrieved for analysis.
- Method 950 may also include time aligning the script, as depicted at block 954 .
- Time aligning the script may include using techniques, such as those described herein with regard to system 100 , to provide a timecode for some or all elements of the corresponding script.
- a script may be processed to provide a timecode for some or all of the words within the script, including dialogue or other script elements.
- the timecode information may provide stop and start time for various elements, including dialogue, which enables the identification of pauses between spoken words of dialogue.
- the time alignment may not actually be performed but may simply be provided. For example, a system generating a video description may be provided or retrieve time aligned script data 116 .
- Method 950 may also include identifying gaps/pauses in dialogue, as depicted at block 956 .
- identifying gaps/pauses in dialogue may include assessing timecode information for each word of spoken dialogue to identify the beginning and end of spoken lines of dialogue, as well as any pauses in the spoken lines of dialogue that may provide gaps for the insertion of video description content. For example, in video description script 920 of FIG. 9B , a pause of twenty-eight seconds was identified at segment 1 , prior to the start of recorded dialogue of segment 2 , a pause of 0.12 seconds was identified at segment 3 , and a pause of 4.06 seconds was identified at segment 7 .
- a gap threshold may be used to identify what pauses are of sufficient length to constitute a gap that may be of sufficient length to be used for inserting video description content. For example, a gap threshold of three seconds may be set, thereby ignoring all pauses of less than three seconds and identifying only pauses equal to or greater than three-seconds as gaps of sufficient length to be used for inserting video description content. Such a technique may be useful to ignore normal pauses in speech (e.g., between spoken words) or short breaks between characters lines of dialogue that may be so short that it would be difficult to provide any substantive video description within the pause.
- the gap threshold value may be user selectable. As depicted in FIG.
- segment 3 of recorded dialogue 924 includes an inserted statement of “No gap available”, and the corresponding action text was deleted/ignored (as indicated by the strikethrough).
- the gap may be detected, but may be ignored.
- the user may be alerted to the gap, thereby enabling them to readily identify gaps that could be used for the insertion of additional video description content.
- video descriptions may be inserted into any available gaps, even out of sequence with their corresponding location in the script, according to rules or preferences provided by the user.
- segment 3 there was no available gap for the video description that would normally be inserted at that point according to the script. However, if there were another available gap within a prescribed number of seconds before or after that segment (e.g., segment 3 ), the video description could be inserted at that other location nearby within the prescribed number of seconds before or after that segment (e.g., segment 3 ).
- Method 950 may also include aligning video description content to gaps/pauses, as depicted at block 958 .
- Aligning the video description content may include aligning the script elements with dialogue relative to where they occur within the script.
- each of the action elements 906 a , 906 b and 906 c are aligned relative to dialogue that occurs before or after the respective action elements.
- aligning video description content includes modifying the video description content and/or the recorded dialogue for merging of the video description content with the recorded dialogue where possible. For example, as depicted in FIG. 9B the script action elements have been aligned to the recorded dialog and the action element text from the script has been aligned with the available gaps when possible.
- Two gaps were identified at segments 1 and 7 for the insertion of corresponding video description content and one action element text segment was deleted because a gap/pause of sufficient length was not available between the lines of dialogue where it was located in the script.
- the user may be provided the opportunity to edit, rewrite, move, or delete the video description content, or the video description content may be automatically modified to fit within the provided gap or deleted.
- a user may have control over the resulting video description. For example, a user may modify a video description at their choosing, or may be provided an opportunity to select how to truncate a video description that does not fit within a gap. For example, in the illustrated embodiment of FIG. 9B , a user may select to remove the text of segment 1 (as indicated by the strikethrough) in an effort to make the video description fit within the corresponding gap.
- video description content may be automatically modified to fit within a given gap. If a gap is too short to fit the corresponding video description content, the video description content may be automatically truncated using rules of grammar.
- the last word(s) or entire last sentence(s) may be incrementally truncated/removed until the remaining video content description is short enough to fit within the gap.
- the last sentence “Maroon is leading an entourage of ASSISTANTS trying to keep up” may have been automatically removed, relieving the user of the need to manually modify the content.
- the user may have the opportunity to approve or modify the changes.
- the duration may be updated dynamically to indicate to the user whether the revised description will fit within an available gap.
- a gap in the recorded program may be created or the duration of a gap may be modified to provide for the insertion of video description content.
- the gap in the recorded audio may be increased (e.g., by inserting an additional amount of pause in the audio track between the end of segment 2 and the beginning of segment 4 ) to five seconds to enable the action element text to be fit within the resulting gap.
- Such a technique may be automatically applied at some or all instances where a gap is too short in duration to fit the corresponding video description content.
- modifications of the dialogue may introduce delays or pauses within the corresponding video and, thus, may modify the video and dialogue of a traditional program, it may be particularly helpful in the context of audio-only programs. For example, for books-on-tape or similar audio tracks produced for the blind or visually impaired.
- video description content may be allowed to overlap certain portions of the audio track.
- a user may have the option of modifying the video description content to overlap seemingly less important portions of the dialogue, music, sound effects, or the like.
- the main audio recorded dialogue, music, sound effects, or the like may be dipped (e.g., reduced) in volume so that the video description may be heard more clearly. For example, the volume of music may be lowered while the video description content is being recited.
- Method 950 may also include generating a script with video description content, as depicted at block 960 .
- Generating a script with video content may include generating a script document that includes video description content; script/recorded dialogue, and/or other script elements aligned with respect to one another.
- FIG. 9B illustrates an exemplary video description script 920 that includes video description content 922 and recorded dialogue 924 .
- the modifications to the video description content are displayed.
- a “clean” version of the video description script may be provided.
- clean video description script may incorporate some or all of the modifications that are not visible.
- a text version of the video description content can be used as a blueprint for recording by a human voiceover talent.
- a voicer may simply have to read the corresponding narration content as opposed to having to manually search through a program, manually identify breaks in the dialog, compose appropriate video descriptions of correct lengths, and/or derive/record narrations to describe the program.
- Method 950 may also include generating a video description, as depicted at block 962 .
- Generating the video description may include recording a reading of the video description content. For example, a reading by a voicer and/or a synthesized reading of the video description content may be recorded to generate a video description track.
- the video description track may be merged with the original audio of the program to generate a program containing both the original audio and the video description audio.
- FIG. 10 Various components of embodiments of a document time-alignment technique as described herein may be executed on one or more computer systems, which may interact with various other devices.
- computer system 1000 includes one or more processors 1010 coupled to a system memory 1020 via an input/output (I/O) interface 1030 .
- Computer system 1000 further includes a network interface 1040 coupled to I/O interface 1030 , and one or more input/output devices 1050 , such as cursor control device 1060 , keyboard 1070 , audio device 1090 , and display(s) 1080 .
- embodiments may be implemented using a single instance of computer system 1000 , while in other embodiments multiple such systems, or multiple nodes making up computer system 1000 , may be configured to host different portions or instances of embodiments.
- some elements may be implemented via one or more nodes of computer system 1000 that are distinct from those nodes implementing other elements.
- computer system 1000 may be a uniprocessor system including one processor 1010 , or a multiprocessor system including several processors 1010 (e.g., two, four, eight, or another suitable number).
- processors 1010 may be any suitable processor capable of executing instructions.
- processors 1010 may be general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, PowerPC, SPARC, or MIPS ISAs, or any other suitable ISA.
- ISAs instruction set architectures
- each of processors 1010 may commonly, but not necessarily, implement the same ISA.
- At least one processor 1010 may be a graphics processing unit.
- a graphics processing unit or GPU may be considered a dedicated graphics-rendering device for a personal computer, workstation, game console or other computer system.
- Modern GPUs may be very efficient at manipulating and displaying computer graphics and their highly parallel structure may make them more effective than typical CPUs for a range of complex graphical algorithms.
- a graphics processor may implement a number of graphics primitive operations in a way that makes executing them much faster than drawing directly to the screen with a host central processing unit (CPU).
- the methods disclosed herein for layout-preserved text generation may be implemented by program instructions configured for execution on one of, or parallel execution on two or more of, such GPUs.
- the GPU(s) may implement one or more application programmer interfaces (APIs) that permit programmers to invoke the functionality of the GPU(s).
- APIs application programmer interfaces
- Suitable GPUs may be commercially available from vendors such as NVIDIA Corporation having headquarters in Santa Clara, Calif., ATI Technologies of AMD having headquarters in Sunnyvale, Calif., and others.
- System memory 1020 may be configured to store program instructions and/or data accessible by processor 1010 .
- System memory 1020 may include tangible a non-transitory storage medium for storing program instructions and other data thereon.
- system memory 1020 may be implemented using any suitable memory technology, such as static random access memory (SRAM), synchronous dynamic RAM (SDRAM), nonvolatile/Flash-type memory, or any other type of memory.
- SRAM static random access memory
- SDRAM synchronous dynamic RAM
- Flash-type memory any other type of memory.
- program instructions and data implementing desired functions are shown stored within system memory 1020 as program instructions 1025 and data storage 1035 , respectively.
- program instructions and/or data may be received, sent or stored upon different types of computer-accessible media or on similar media separate from system memory 1020 or computer system 1000 .
- a computer-accessible medium may include storage media or memory media such as magnetic or optical media, e.g., disk or CD/DVD-ROM coupled to computer system 1000 via I/O interface 1030 .
- Program instructions and data stored via a computer-accessible medium may be transmitted by transmission media or signals such as electrical, electromagnetic, or digital signals, which may be conveyed via a communication medium such as a network and/or a wireless link, such as may be implemented via network interface 1040 .
- I/O interface 1030 may be configured to coordinate I/O traffic between processor 1010 , system memory 1020 , and any peripheral devices in the device, including network interface 1040 or other peripheral interfaces, such as input/output devices 1050 .
- I/O interface 1030 may perform any necessary protocol, timing or other data transformations to convert data signals from one component (e.g., system memory 1020 ) into a format suitable for use by another component (e.g., processor 1010 ).
- I/O interface 1030 may include support for devices attached through various types of peripheral buses, such as a variant of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard, for example.
- PCI Peripheral Component Interconnect
- USB Universal Serial Bus
- I/O interface 1030 may be split into two or more separate components. In addition, in some embodiments some or all of the functionality of I/O interface 1030 , such as an interface to system memory 1020 , may be incorporated directly into processor 1010 .
- Network interface 1040 may be configured to allow data to be exchanged between computer system 1000 and other devices attached to a network, such as other computer systems, or between nodes of computer system 1000 .
- network interface 1040 may support communication via wired or wireless general data networks, such as any suitable type of Ethernet network, for example; via telecommunications/telephony networks such as analog voice networks or digital fiber communications networks; via storage area networks such as Fibre Channel SANs, or via any other suitable type of network and/or protocol.
- Input/output devices 1050 may, in some embodiments, include one or more display terminals, keyboards, keypads, touchpads, scanning devices, voice or optical recognition devices, or any other devices suitable for entering or retrieving data by one or more computer system 1000 .
- Multiple input/output devices 1050 may be present in computer system 1000 or may be distributed on various nodes of computer system 1000 .
- similar input/output devices may be separate from computer system 1000 and may interact with one or more nodes of computer system 1000 through a wired or wireless connection, such as over network interface 1040 .
- memory 1020 may include program instructions 1025 , configured to implement embodiments of a layout-preserved text generation method as described herein, and data storage 1035 , comprising various data accessible by program instructions 1025 .
- program instructions 1025 may include software elements of a layout-preserved text generation method illustrated in the above Figures.
- Data storage 1035 may include data that may be used in embodiments, for example input PDF documents or output layout-preserved text documents. In other embodiments, other or different software elements and/or data may be included.
- computer system 1000 is merely illustrative and is not intended to limit the scope of a layout-preserved text generation method as described herein.
- the computer system and devices may include any combination of hardware or software that can perform the indicated functions, including computers, network devices, internet appliances, PDAs, wireless phones, pagers, etc.
- Computer system 1000 may also be connected to other devices that are not illustrated, or instead may operate as a stand-alone system.
- the functionality provided by the illustrated components may in some embodiments be combined in fewer components or distributed in additional components. Similarly, in some embodiments, the functionality of some of the illustrated components may not be provided and/or other additional functionality may be available.
- instructions stored on a computer-accessible medium separate from computer system 1000 may be transmitted to computer system 1000 via transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as a network and/or a wireless link.
- Various embodiments may further include receiving, sending or storing instructions and/or data implemented in accordance with the foregoing description upon a computer-accessible medium. Accordingly, the present invention may be practiced with other computer system configurations.
- portions of the techniques described herein e.g., preprocessing of script and metadata may be hosted in a cloud computing infrastructure.
- a computer-accessible storage medium may include a non-transitory storage media or memory media such as magnetic or optical media, e.g., disk or DVD/CD-ROM, volatile or non-volatile media such as RAM (e.g. SDRAM, DDR, RDRAM, SRAM, etc.), ROM, etc., as well as transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as network and/or a wireless link.
- a non-transitory storage media or memory media such as magnetic or optical media, e.g., disk or DVD/CD-ROM, volatile or non-volatile media such as RAM (e.g. SDRAM, DDR, RDRAM, SRAM, etc.), ROM, etc., as well as transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as network and/or a wireless link.
- RAM e.g. SDRAM, DDR, RDRAM, SRAM, etc.
- ROM etc
- such quantities may take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared or otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to such signals as bits, data, values, elements, symbols, characters, terms, numbers, numerals or the like. It should be understood, however, that all of these or similar terms are to be associated with appropriate physical quantities and are merely convenient labels. Unless specifically stated otherwise, as apparent from the discussion, it is appreciated that throughout this specification discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining” or the like refer to actions or processes of a specific apparatus, such as a special purpose computer or a similar special purpose electronic computing device.
- a special purpose computer or a similar special purpose electronic computing device is capable of manipulating or transforming signals, typically represented as physical electronic or magnetic quantities within memories, registers, or other information storage devices, transmission devices, or display devices of the special purpose computer or similar special purpose electronic computing device.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Machine Translation (AREA)
Abstract
Description
- This patent application claims priority to U.S. Provisional Patent Application No. 61/323,121 entitled “Method and Apparatus for Time Synchronized Script Metadata” by Jerry R. Scoggins II, et. al, filed Apr. 12, 2010, which is hereby incorporated by reference as though fully set forth herein.
- In a video production environment, a script serves as a roadmap to when and how elements of a movie/video will be produced. In addition to specifying dialogue to be recorded, scripts are a rich source of additional metadata and include numerous references to characters, people, places, and things. During the production process, directors, editors, sound engineers, set designers, marketing, advertisers, and other production personnel are interested in knowing which people, places, and things occurred or will occur in certain scenes. This information is often present in the script but is not typically directly correlated to the corresponding video content (e.g., video and audio) because timing information is missing from the script. That is, elements of the script are not correlated with a time in which they appear in the corresponding video content. Thus, it may be difficult to link script elements (e.g., spoken dialogue) with the time when they actually occur within the corresponding video. For example, although production personnel may know that a character speaks a certain line of dialogue in a scene based on the script, the production personnel may not be able to readily determine the precise time in the working or final video when the particular line was spoken. A full script can include several thousand script elements or entities. If one were to try to find the actual point in time when a particular event (e.g., when a line was spoken) in a corresponding movie/video, the video content may have to be manually searched by a viewer to locate the event such that the corresponding timecode can be manually recorded. Thus, production personnel may not be able to easily to search or index their scripts and video content.
- When a known, written script text is time-matched to raw speech transcript produced from an analysis of recorded dialogue, the script text is said to be “aligned” with the recorded dialogue, and the resulting script may be referred to as an “aligned script.” Aligned scripts may be useful as production personnel often desire to search or index video/audio content based on the text provided in the script. Moreover, production personnel may desire to generate closed caption text that is synchronized to actual spoken dialogue in video content. However, due to variations in spoken dialogue versus the corresponding written text, as well as gaps, pauses, sound effects, music, etc. in the recorded dialogue, time aligning is a difficult task to automate. Typically, the task of time-aligning textual scripts and metadata to actual video content is a tedious task that is accomplished by a manual process that can be expensive and time-consuming. For example, a person may have to view and listen to video content and manually transcribe the corresponding audio to generate an index of what took place and when, or to generate closed captioning text that is synchronized to the video. To manually locate and record a timecode for even a small fraction of the dialogue words and script elements within a full-length movie often requires several hours of manual work, and doing this for the entire script might require several days or more. Similar difficulties may be encountered while creating video descriptions for the hearing impaired. For example, a movie may be manually searched to identify gaps in dialogue for the insertion of video description narrations that describe visual elements (e.g., actions, settings) and a more complete description of what is taking place on screen.
- Although some automated techniques for time-synchronizing scripts and corresponding video have been implemented, such as using a word alignment matrix (e.g., script words vs. transcript words), they are traditionally slow and error-prone. These techniques often require a great deal of processing and may contain a large number of errors, rendering the output inaccurate. For example, due to noise or other non-dialogue artifacts, in speech-to-text transcripts the wrong time values, off by several minutes or more, are often assigned to script text. As a result, the output may not be reliable, thereby requiring additional time to identify and correct the errors, or causing users to shy away from its use altogether.
- Accordingly, it is desirable to provide a technique for providing efficient and accurate time-alignment of a script document and corresponding video content.
- Various embodiments of methods and apparatus for time aligning documents (e.g., scripts) to associated video/audio content (e.g., movies) are described. In some embodiments, provided is a method that includes providing script data that includes ordered script words indicative of dialogue and providing audio data corresponding to at least a portion of the dialogue. The audio data includes timecodes associated with dialogue. The method includes correlating the script data with the audio data, and generating time-aligned script data that includes time-aligned words indicative of dialogue spoken in the audio data and corresponding timecodes for time-aligned words.
- In some embodiments, provided is a computer implemented method that includes providing video content data corresponding to the script data including ordered script words indicative of dialogue. The video content data includes audio data includes a transcript including transcript words corresponding to at least a portion of the dialogue and timecodes associated with the transcript words. The method also includes correlating the script data with the video content data, and generating time-aligned script data that includes time-aligned words indicative of words spoken in the video content and corresponding timecodes for time-aligned words.
- In some embodiments, provided is a computer implemented method that includes receiving script data including ordered script words of a script, wherein the ordered script words are indicative of dialogue words to be spoken, receiving audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a matrix of the ordered script words versus the dialogue words, performing an alignment of the matrix to determine hard alignment points, including matching consecutive sequences of ordered script words with corresponding sequences of dialogue words. The method also includes partitioning the matrix of ordered script words into sub-matrices, wherein the bounds of each of the sub-matrices are defined by adjacent hard-alignment points, and wherein the sub-matrices include a sub-set of the ordered script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, performing an alignment of each of the sub-matrices. The alignment of the sub-matrices including: matching ordered script words of the sub-subset of ordered script words of the respective sub-matrix with dialogue words of the sub-subset of dialogue words of the respective sub-matrix, assigning, to the matched ordered script words, timecodes associated with corresponding matching dialogue words, and determining timecodes for the unmatched ordered script words of sub-set of ordered script words of the sub-matrix using interpolation based on the timecodes associated with the matching ordered script words. The method also includes generating time-aligned script data including the ordered script words of the script and their corresponding timecodes.
- Provided in some embodiments is a non-transitory computer readable storage medium having program instructions stored thereon, wherein the program instructions are executable to cause a computer system to perform a method that includes receiving script data including ordered script words of a script, wherein the ordered script words are indicative of dialogue words to be spoken, receiving audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a matrix of the ordered script words versus the dialogue words, performing an alignment of the matrix to determine hard alignment points, including matching consecutive sequences of ordered script words with corresponding sequences of dialogue words. The method also includes partitioning the matrix of ordered script words into sub-matrices, wherein the bounds of each of the sub-matrices are defined by adjacent hard-alignment points, and wherein the sub-matrices include a sub-set of the ordered script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, performing an alignment of each of the sub-matrices. The alignment of the sub-matrices including: matching ordered script words of the sub-subset of ordered script words of the respective sub-matrix with dialogue words of the sub-subset of dialogue words of the respective sub-matrix, assigning, to the matched ordered script words, timecodes associated with corresponding matching dialogue words, and determining timecodes for the unmatched ordered script words of sub-set of ordered script words of the sub-matrix using interpolation based on the timecodes associated with the matching ordered script words. The method also includes generating time-aligned script data including the ordered script words of the script and their corresponding timecodes.
- Provided in some embodiments is a computer system for receiving script data including ordered script words of a script, wherein the ordered script words are indicative of dialogue words to be spoken, receiving audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a matrix of the ordered script words versus the dialogue words, performing an alignment of the matrix to determine hard alignment points, including matching consecutive sequences of ordered script words with corresponding sequences of dialogue words. The method also includes partitioning the matrix of ordered script words into sub-matrices, wherein the bounds of each of the sub-matrices are defined by adjacent hard-alignment points, and wherein the sub-matrices include a sub-set of the ordered script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, performing an alignment of each of the sub-matrices. The alignment of the sub-matrices including: matching ordered script words of the sub-subset of ordered script words of the respective sub-matrix with dialogue words of the sub-subset of dialogue words of the respective sub-matrix, assigning, to the matched ordered script words, timecodes associated with corresponding matching dialogue words, and determining timecodes for the unmatched ordered script words of sub-set of ordered script words of the sub-matrix using interpolation based on the timecodes associated with the matching ordered script words. The method also includes generating time-aligned script data including the ordered script words of the script and their corresponding timecodes.
-
FIG. 1A is a block diagram that illustrates components and dataflow for document time-alignment in accordance with one or more embodiments of the present technique. -
FIG. 1B is text that illustrates exemplary script data in accordance with one or more embodiments of the present technique. -
FIG. 1C is text that illustrates exemplary transcript data in accordance with one or more embodiments of the present technique. -
FIG. 1D is text that illustrates exemplary time-aligned script data in accordance with one or more embodiments of the present technique. -
FIG. 2 is a block diagram that illustrates components and dataflow for script time-alignment in accordance with one or more embodiments of the present technique. -
FIG. 3 is a flowchart that illustrates a script time-alignment method in accordance with one or more embodiments of the present technique. -
FIG. 4 is a flowchart that illustrates a script synchronization method in accordance with one or more embodiments of the present technique. -
FIG. 5A is a depiction of an exemplary alignment matrix in accordance with one or more embodiments of the present technique. -
FIG. 5B is a depiction of an exemplary alignment sub-matrix in accordance with one or more embodiments of the present technique. -
FIG. 6 is a depiction of an exemplary graphical user interface sequence in accordance with one or more embodiments of the present technique. -
FIG. 7A is a depiction of multiple lines of text that include a script phrase, a transcript phrase and a corresponding representation of alignment in accordance with one or more embodiments of the present technique. -
FIG. 7B is a depiction of multiple lines of text that include a script phrase, a transcript phrase and a corresponding representation of alignment in accordance with one or more embodiments of the present technique. -
FIG. 7C is a depiction of a line of text and corresponding in/out ranges in accordance with one or more embodiments of the present technique. -
FIGS. 8A and 8B are block diagrams that illustrate components and dataflow of a script time-alignment technique in accordance with one or more embodiments of the present technique. -
FIG. 9A is a depiction of an exemplary script document in accordance with one or more embodiments of the present technique. -
FIG. 9B is a depiction of a portion of an exemplary video description script in accordance with one or more embodiments of the present technique. -
FIG. 9C is a flowchart that illustrates a method of generating a video description in accordance with one or more embodiments of the present technique. -
FIG. 10 is a block diagram that illustrates an example computer system in accordance with one or more embodiments of the present technique. - While the invention is described herein by way of example for several embodiments and illustrative drawings, those skilled in the art will recognize that the invention is not limited to the embodiments or drawings described. It should be understood, that the drawings and detailed description thereto are not intended to limit the invention to the particular form disclosed, but on the contrary, the intention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention. The headings used herein are for organizational purposes only and are not meant to be used to limit the scope of the description. As used throughout this application, the word “may” is used in a permissive sense (i.e., meaning having the potential to), rather than the mandatory sense (i.e., meaning must). Similarly, the words “include”, “including”, and “includes” mean including, but not limited to. As used throughout this application, the singular forms “a”, “an” and “the” include plural referents unless the content clearly indicates otherwise. Thus, for example, reference to “an element” includes a combination of two or more elements.
- In the following detailed description, numerous specific details are set forth to provide a thorough understanding of claimed subject matter. However, it will be understood by those skilled in the art that claimed subject matter may be practiced without these specific details. In other instances, methods, apparatuses or systems that would be known by one of ordinary skill have not been described in detail so as not to obscure claimed subject matter.
- Speech-To-Text (STT)—a process by which source audio containing dialogue or narrative is automatically transcribed to a textual representation of the dialogue or narrative. The source audio may also contain music, noise, and/or sound effects that generally contribute to lower transcription accuracy.
- STT transcript—a document generated by a STT transcription engine containing the transcription of the dialogue or narrative of the audio source. Each word in the transcript may include an associated timecode which indicates precisely when the audio content associated with each word of the dialogue or narrative occurred. Timecodes are typically provided in hours, minutes, seconds and frames.
- Script—a document that outlines all of the visual, audio, behavioral, and spoken elements required to tell the story in a corresponding video or movie. Dramatic scripts are often referred to as a “screenplay”. Scripts may not include timecode data, such that they may not provide information about when an element of the script actually occurs within corresponding video content (e.g., a script may not provide a relative time within the video content that indicates precisely when the audio content associated with each word of the dialogue or narrative occurred).
- Script dialogue/narrative—the script lines to be spoken in a corresponding video or movie. Each script line may include text that includes one or more words.
- Script alignment—a process by which a set of words of a dialogue or narrative in a script are matched to corresponding transcribed words of video content. Script alignment may include providing an output that is indicative of a relative time within the video content that words of dialogue or narrative contained in the script are spoken.
- Aligned Script—a script that outlines all of the visual, audio, behavioral, and spoken elements required to tell the story in a corresponding video or movie and includes timecode data indicative of when elements of the script actually occur within corresponding video content (e.g., a time aligned script may include a relative time within the video content that indicates precisely when the audio content associated with each word of the dialogue or narrative occurred). Timecodes are typically provided in hours, minutes, seconds and frames. Feature films are typically shot at 24 frames per second, thus twelve frames is about ½ second in duration.
- Word n-gram—a consecutive subsequence of N words from a given sequence. For example, (The, rain, in), (rain, in, Spain) and (in, Spain, falls) are valid 3-grams from the sentence, “The rain in Spain falls mainly on the plain.”
- Alignment matrix—a mathematical structure used to represent how the words from a script source will align with the transcribed words of a transcript (e.g., an STT transcript generated via a speech-to-text (STT) process). For example, a vertical axis of the matrix may be formed of words in a script in the sequence/order in which they occur (e.g., ordered script words), and a horizontal axis of the matrix may be formed of words in the transcript in the sequence/order in which they occur (e.g., ordered transcript words). Each matrix cell at the intersection of a corresponding row/column may indicate the accumulated number of word insert, update or delete operations needed to match the sequence of ordered script words to the sequence of ordered transcript words to the (row, col) entry. A path with the lowest score through the matrix is indicative of the best word alignment.
- Natural Language Processing (NLP)—a technique in which natural language text is input and then sentences, part-of-speech, noun and verb phrases, and other semantics are automatically extracted. NLP may be provided as a component in processing unstructured or semi-structured text where a large quantity of rich metadata can be found, (e.g., in spec. movie scripts and dramatic screenplays).
- Program—a visual and audio production that is recorded and played back to an audience, such as a movie, television show, documentary, etc.
- Dialogue—the words spoken by actors or other on-screen talent during a program.
- Video Description (or Audio Description)—an audio track in a program containing descriptions of the setting and action. The video description may be inserted into the natural pauses in dialogue or between critical sound elements. A video description often includes narration to fill in the story gaps for the blind or visually impaired by helping to describe visual elements and provide a more complete description of what's happening (e.g., visually) in the program.
- Describer—a person who develops the description to be recorded by the voicer. In some cases, the describer is also the voicer.
- Voicer (or Voice Talent)—a person who voices the Video Description.
- Secondary Audio Program (SAP)—an auxiliary audio channel for analog television that is broadcast or transmitted both over the air and by cable TV. It is often used for an alternate language or Descriptive Video Service.
- Digital Television broadcasting (DTV)—Analog broadcasting ceased in the U.S. in 2009 and was replaced by DTV.
- Various embodiments of methods and apparatus for aligning features of a script document with features of corresponding video content are provided. Embodiments described herein facilitate aligning script data to the video content data, and to use the script data to improve the accuracy of corresponding speech transcript (e.g., using the script data in place of the potentially inaccurate SST audio transcript from the video content data). In some embodiments, a document includes at least a portion of a script document, such as a movie or speculative script (e.g., dramatic screenplay), that outlines visual, audio, behavioral, and spoken elements required to tell a story. In certain embodiments, video content includes video and/or audio data that corresponds to at least a portion of the script document. In some embodiments, the audio data of the video content is transcribed into a textual format (e.g., spoken dialogue/narration is translated into words). In certain embodiments, the transcription is provided via a speech-to-text (STT) engine that automatically generates a transcript of words that correspond to the audio data of the video content. In some embodiments, the transcript includes timing information that is indicative of a point in time within the video content that one or more words were actually spoken. In certain embodiments, the words of the transcript (“transcript words”) are aligned with corresponding words of the script (“script words”). In some embodiments, aligning the transcript words with corresponding script words includes implementation of various processing techniques, such as matching sequences of words, assessing confidence/probabilities that the words identified are in fact correct, and substitution/replacement of script/transcript word with transcript/script words. In some embodiments, the resulting output includes time-aligned script data. In certain embodiments, the script data includes a time-aligned script document including accurate representation of each of the words actually spoken in the video content, and timing information that is indicative of when the word of the script were actually spoken within the video content (e.g., a timecode associated with each word of dialogue/narration). In some embodiments, time-aligned data may include timecodes for other elements of the script, such as scene headings, action elements, character names, parentheticals, transitions, shot elements, and the like.
- In some embodiments, two source inputs are provided: (1) a script (e.g., plain dialogue text or a Hollywood Spec. Script/Dramatic screenplay) and (2) an audio track dialogue (e.g., an audio track dialogue from video content corresponding to the script). In certain embodiments, a coarse-grain alignment of blocks of text is performed by first matching identical or near identical N-gram sequences of words to generate corresponding “hard alignment points”. The hard-alignment points may include matches between portions of the script and transcript (e.g., N-gram matches of a sequence of script words with a sequence of transcript words) which are used to partition an initial single alignment matrix (e.g., providing a correspondence of all ordered script words vs. all ordered transcript words) into a number of smaller sub-matrices (e.g., providing a correspondence of script words that occur between the hard alignment points vs. transcript words that occur at or between the hard alignment points). Using an algorithm, such as a standard or optimized Levenshtein word edit distance algorithm, additional words matches—between the words of the script and the transcrip—may be indentified as “soft alignment points” within each sub-matrix block of text. The soft alignment points may define multiple non-overlapping interpolation intervals. In some instances, unmatched words may be located between the matched words (e.g., between the hard alignment points and/or the soft alignment points). Knowing the time data (e.g., timecode) information for the matched words, an interpolation (e.g., linear or non-linear interpolation) may be performed to determine timecodes for each of the non-matched words (e.g., words that have not been assigned timecode information) occurring between the matched points. As a result, all words (e.g., matched and unmatched) are provide with corresponding timecode information, and the timecode information may be merged with the words of the script and/or transcript documents to generate a time-aligned script document that includes all of the words spoken and their corresponding timecode information to indicate when each of the words was actually spoken within the video content. Such a technique may benefit from combining the accuracy of the script words and the timecodes of the transcript words.
- As described in more detail below, the techniques described herein may provide techniques by which all textual elements (e.g., dialogue/narration) of a script (e.g., a Hollywood movie script or dramatic screenplay script) can be automatically time-aligned to the specific points in time within corresponding video content, to identify when specific dialogue, text, or actions within the script actually occur within the video content. This enables identifying and locating when dialogue and important semantic metadata provided in a script actually occurs within corresponding production video content. In some embodiments, time alignment may be applied to all elements of the script (e.g., scene headings, action elements, etc.) to enable a user to readily identify where various elements, not just dialogue words, occur within the script. In certain embodiments, the timecode information may also be used to identify gaps in dialogue for the insertion of video description content that includes narrations to fill in the story gaps for the blind or visually impaired, thereby helping to describe visual elements and provide a more complete description of what's happening (e.g., visually) in the program
- The techniques described herein may be employed to automatically and accurately synchronize the written movie script (e.g., which may contain accurate text, but no time information) to a corresponding audio transcript (e.g., which contains accurate time information but may include very noisy or erroneous text). In certain instances, techniques may employ the transcript to identify actual words/phrases spoken that vary from the text of the script. The accuracy of the words in the script or transcript may, thus, be combined with accurate timing information in the transcript to provide an accurate time aligned script. The techniques described herein may demonstrate good tolerance to noisy transcripts or transcripts that have a large number of errors. By partitioning the alignment matrix into many smaller sub-matrices, the techniques described herein may also provide improved performance including increased processing speeds while maintaining significantly higher overall accuracy.
-
FIG. 1 is a block diagram that illustrates system components and dataflow of a system for implementing time-alignment (system) 100 in accordance with one or more embodiments of the present technique. In some embodiments,system 100 implements asynchronization module 102 to analyze adocument 104 andcorresponding video content 106. Based on the analysis,system 100 generates time-aligned data (e.g., time aligned script document) 116 that associates various portions ofdocument 104 with corresponding portions ofvideo content 106. Time aligneddata 116 may provide the specific points in time withinvideo content 106 that elements (e.g., specific dialogue, text, or actions) defined indocument 104 actually occur. - In the illustrated embodiment, document 104 (e.g., a script) is provided to a
document extractor 108.Document extractor 108 may generate acorresponding document data 110, such as a structured/tagged document. A structured/tagged document may include embedded script data that is provided tosynchronization module 102 for processing. - In some embodiments,
document 104 may include a script document, such as a movie script (e.g., a Hollywood script), a speculative script, a shooting script (e.g., a Hollywood shooting script), a closed caption (SRT) video transcript or the like. For simplicity,document 104 may be referred to as a “script” although it will be appreciated thatdocument 104 may include other forms of documents including dialogue text, as described herein. - A movie script may include a document that outlines all of the visual, audio, behavioral, and spoken elements required to tell a story. A speculative (“spec”) script or screenplay may include a preliminary script used in both film and television industries. A spec script for film generally includes an original screenplay and may be a unique plot idea, an adaptation of a book, or a sequel to an existing movie. A “television” spec script is typically written for an existing show using characters and storylines that have already been established. A “pilot” spec script typically includes an original idea for a new show. A television spec script is typically 20-30 pages for a half hour of programming, 40-60 pages for a full hour of programming, or 80-120 pages for two hours of programming. It will be appreciated that once a spec script is purchased, it may undergo a series of complete rewrites or edits before it is put into production. Once in “production”, the script may evolve into a “Shooting Script” or “Production Script” having a more complex format. Numerous scripts exist and new scripts are continually created and sold.
-
Script 104 may include a full script including several thousand script elements or entities, for instance, or a partial script including only a portion of the full script, such as a few lines, a full scene, or several scenes. For example,script 104 may include a portion of a script that corresponds to a clip provided asvideo content 106. Since film production is a highly collaborative process, the director, cast, editors, and production crew may use various forms of the script to interpret the underlying story during the production process. Further, since numerous individuals are involved in the making of a film, it is generally desirable that a script conform to specific standards and conventions that all involved parties understand (e.g., it will use a specific format w.r.t. layout, margins, notation, and other production conventions). Thus, a script document is intended to structure all of the script elements used in a screenplay into a consistent layout. Scripts generally include script elements embedded in the script document. Script elements often include a title, author name(s), scene headings, action elements, character names, parentheticals, transitions, shot elements, dialogue/narrations, and the like. An exemplary portion of ascript segment 130 is depicted inFIG. 1B .Script segment 130 includes a scene heading 130 a,action elements 130 b,character names 130 c,dialogues 130 d, and parentheticals 130 e. - Document (script)
extractor 108 may process script 104 to provide document (script)data 110, such as a structured/tagged script document. Words contained in the document (script) data may be referred to as script words. A structured/tagged (script) document may include a sequential listing of the lines of the document in accordance with their order inscript 104, along with a corresponding tag (e.g., tags—“TRAN”, “SCEN”, “ACTN”, “CHAR”, “DIAG”, “PARN” or the like) identifying a determined element type associated with some, substantially all, or all of each of the lines or groupings of the lines. In some embodiments, a structured/tagged document may include an Extensible Markup Language (XML) format, such as *.ASTX format used by certain products, such as those produced by Adobe Systems, Inc., having headquarters in San Jose, Calif. (hereinafter “Adobe”). In some embodiments,document extractor 108 may obtain script 104 (e.g., a layout preserved version of the document), perform a statistical analysis and/or feature matching of features contained within the document, identify document elements based on the statistical analysis and/or the feature matching, pass the identified document elements through a finite state machine to assess/determine/verify the identified document elements, assess whether or not document elements are incorrectly identified, and, if it is determined that there are incorrectly identified document elements, re-performing at least a portion of the identification steps, or, if it is determined that there are no (or sufficiently few) incorrectly identified document elements, and generate/store/output a structured/tagged (script) document or other forms of document (script)data 110 that is provided tosynchronization module 102. In some embodiments,document extractor 108 may employ various techniques for extracting and transcribing audio data, such as those described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein. - In the illustrated embodiment,
video content 106 is provided to anaudio extractor 112.Audio extractor 112 may generate acorresponding transcript 114.Video content 106 may include video image data and corresponding audio soundtracks that include dialogue (e.g., character's spoken words or narrations), sound effects, music, and the like.Video content 106 for a movie may be produced in segments (e.g., clips) and then assembled together to form the final movie or video product during the editing process. For example, a movie may include several scenes, and each scene may include a sequence of several different shots that typically specify a location and a sequence of actions and dialogue for the characters of the scene. The sequence of shots may include several video clips that are assembled into a scene, and multiple scenes may be combined to form the final movie product. A clip, includingvideo content 106, may be recorded for each shot of a scene, resulting in a large number of clips for the movie. Tools, such as Adobe Premiere Pro by Adobe Systems, Inc., may be used for editing and assembling clips from a collection of shots or video segments. In some embodiments, audio content (e.g., without corresponding video content may be provided). For example, audio content, such as that of a radio show) may be provided toaudio extractor 112 in place of or along with content that includes video. Although a number of embodiments described here refer tovideo content 106 as including both video data and audio data, the techniques described herein may be applied to audio content in a similar manner. -
Audio extractor 112 may processvideo content 106 to generate a corresponding transcript that includes an interpretation of words (e.g., dialogue or narration) spoken invideo content 106.Transcript 114 may be provided as a transcribed document or transcribed data that is capable of being provided to other portions ofsystem 100 for subsequent processing. In some embodiments,audio extractor 112 includes a speech-to-text engine that takes an audio segment fromvideo content 106 containing spoken dialogue, and uses speech-to-text (STT) technology to generate a time-code transcript of the dialogue. Thus,transcript 114 may indicate the timecode and duration for each spoken word that is identified by the audio extractor. Words oftranscript 114 may be referred to as transcript words. - In some embodiments, speech-to-text (STT) technology may implement a custom language model such as that described herein. In some embodiments, speech-to-text (STT) technology may implement a custom language model and/or an enhanced multicore STT transcription engine such as those described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009 and/or U.S. patent application Ser. No. 12/332,309 entitled “MULTI-CORE PROCESSING FOR PARALLEL SPEECH-TO-TEXT PROCESSING”, filed Dec. 10, 2008, which are hereby incorporated by reference as though fully set forth herein. A
transcript 114 generated byaudio extractor 112 may include a raw transcript. An exemplary raw transcript (e.g., STT transcript) 132 is depicted inFIG. 1C .Raw transcript 132 includes a sequential listing of identified transcript words having associated time code, duration, STT word estimate and additional comments regarding the transcription. The timecode may indicate at what point in time within the video content the word was spoken (e.g., transcript word “dad” was spoken 7165.21 seconds from the beginning of the associated video content), the duration may indicate the amount of time the word was spoken from start to finish (e.g., it took about 0.27 sec to say the word “dad”), and comments may indicate potential problems (e.g., that noise in the audio data may have generated an error). In some embodiments, the raw transcript information may also include a confidence value that indicates the probability that the interpreted/indicated word is accurate. The raw transcript information may not include additional text features, such as punctuation, capitalization, and the like. - In some embodiments, document extraction and audio extraction may occur in parallel. For example, in the illustrated embodiment,
document extractor 108 receivesscript 104 and generatesscript data 110 independent ofaudio extractor 112 receivingvideo content 106 and generatingtranscript 114. Accordingly, these two processes may be performed in parallel with one another. In some embodiments, document extraction and audio extraction may occur in series. For example,document extractor 108 may receivedocument 104 and generatedocument data 110 prior toaudio extractor 112 receivingvideo content 106 and generatingtranscript 114, or vice versa. -
Synchronization module 102 may generate time-aligneddata 116. Time-aligneddata 116 may be provided as a document or raw data that is capable of being provided to other portions ofsystem 100 for subsequent processing. Time-aligneddata 116 may be based on script information (e.g., document data 110) and video content information (e.g., transcript 114). For example,synchronization module 102 may compare transcript words intranscript 114 to script words in the document (script)data 110 to determine whether or not the transcribed words are accurate. The comparison may use various indicators to assess the accuracy. For example, a plurality of words and phrases with exact matches betweentranscript 114 anddocument data 110 may have high probabilities of being correct, and may be referred to as “hard reference points”. Words and phrases with partial matches (e.g., single words or only a few matched words) may have a lower probability of being correct, and may be referred to as “soft reference points”. Words and phrases that do not appear to have matches may have a low probability of being correct. Words and phrases with a low probability of being correct may be subject to additional amounts of processing. For example, low probability matches may be subject to interpolation based on the hard and soft reference points. Words that are part of hard or soft reference pints may be referred to as words having a match, whereas words that are not part of a hard or soft reference point may be referred to as unmatched words or words not having a match. As described in more detail below, the hard-alignment points may be used to partition the document data and the transcript into smaller segments that correspond to one another, and additional processing may be performed on the smaller segments in substantial isolation. Further, as described in more detail below, the timecodes and other information associated with matched words may be used to derive (e.g., interpolate) timecode and other information about the unmatched words. - The results of the comparison may be used to generate time aligned
data 116. Time aligneddata 116 may include words (e.g., from the script words or transcript words) having a specific timecode associated therewith. In some embodiments, time aligneddata 116 may include words from bothdocument data 110 andtranscript data 114 used to generate a single script that accurately identifies words actually spoken invideo content 106 along with corresponding timecode information for each spoken word of dialogue or other elements. The timecode for each word may be obtained directly from matching words of the transcript, or may generated (e.g., via interpolation). Time aligneddata 116 may be stored at a storage medium 118 (e.g., a database), displayed at a display device 120 (e.g., a graphical display viewable by a user), or provided toother modules 122 for processing. An exemplary time-aligned script data/document 134 is depicted inFIG. 1D . As depicted, time-aligned data/document 134 includes spokenwords 136 grouped with other spoken words of theirrespective script elements 137, and provided along with their associatedtimecodes 138. Astart time 140 for each element grouping of lines is also provided. In the depicted time-aligned data/document, each of the script elements (and text of the script elements) is also assigned a corresponding time code. -
FIG. 2 is a block diagram that illustrates components and dataflow ofsystem 100 in accordance with one or more embodiments of the present technique. In the illustrated embodiment,synchronization module 102 includes ascript reader 200, ascript analyzer 202, a Speech-to-Text (STT)reader 204, anSTT analyzer 206, amatrix aligner 208, an interval generator/interpolator 210, and a time-codedscript generator 212. -
FIG. 3 is a flowchart that illustrates a script time-alignment method 300 according to one or more embodiments of the present technique.Method 300 may provide alignment techniques using components and dataflow implemented atsystem 100. In the illustrated embodiment,method 300 includes providing script content, as depicted atblock 302, providing audio content, as depicted atblock 304, aligning the script content and audio content, as depicted atblock 306, and providing time-coded script data, as depicted atblock 308. - In some embodiments, providing script content (block 302) includes inputting or otherwise providing a
script 104, such as a Hollywood Spec. Movie Script or dramatic screenplay script, tosystem 100. For example, a plain text document, such as a raw script document, may be provided in an electronic format to scriptextractor 108 which processes script 104 (e.g., to identify, structure, and extract the text of script 104) to generatescript data 110, such as a structured/tagged script document.Script extractor 108 may employ techniques for converting documents, such as those described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, U.S. patent application Ser. No. 12/332,309 entitled “MULTI-CORE PROCESSING FOR PARALLEL SPEECH-TO-TEXT PROCESSING”, filed Dec. 10, 2008, and/or U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, are all hereby incorporated by reference as though fully set forth herein.Document data 110 may be provided tosynchronization module 102 for subsequent processing, as described in more detail below. - In some embodiments, providing audio content (block 304) includes inputting or otherwise providing
video content 106, such as a clip/shot of a Hollywood movie, having associated audio content that corresponds to ascript 104, tosystem 100. Audio data may be extracted fromvideo content 106 using various techniques. For example, an audio data track may be extracted fromvideo content 106 using a Speech-to-Text (STT) engine and/or a custom language model. In some embodiments,audio extractor 112 may employ an STT engine and/or custom language model to generatetranscript 114 that includes a transcription of spoken words (e.g., audio dialogue or narration) of the Hollywood movie or other audio data.Audio extractor 112 may employ various techniques for extracting and transcribing audio data, such as those described below and/or those techniques described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, and/or U.S. patent application Ser. No. 12/332,309 entitled “MULTI-CORE PROCESSING FOR PARALLEL SPEECH-TO-TEXT PROCESSING”, filed Dec. 10, 2008, which are both hereby incorporated by reference as though fully set forth herein. A resultingtranscript 114 may be provided tosynchronization module 102 for subsequent processing, as described in more detail below. - In some embodiments, aligning the script and audio content (block 306) includes employing a matching technique to align the script words (e.g., dialogue or narrations) of
script 104 to elements of thevideo content 106. This may include aligning script words to corresponding transcript words. In some embodiments, alignment includessynchronization module 102 implementing a two-level word matching system to align script words ofscript 110 to corresponding transcript words oftranscript 114. In some embodiments, a first matching routine is executed to partition a matrix of script words vs. transcript words into a sub-matrix. For example, an N-gram matching scheme may be used to identify high probability matches of a sequence of multiple words. N-gram matching may include attempting to exactly (or at least partially) match phrases of multiple transcript words with script words. The matched sequence of words may be referred to as hard-alignment points. The hard alignment points may include several matched words, and may be used to define boundaries of each sub-matrix. Thus, the hard-alignment points may define smaller matrices of script words vs. transcript words. Each of the smaller sub-matrices may, then, be processed (e.g., in series or parallel) using additional matching techniques to identify word matches within each of the sub-matrices. In some embodiments, processing may be provided via multiple processors. For example, processing in series or parallel may be performed using multiple processors of one or more hosted services or cloud computing environments. In some embodiments, each of the sub-matrix is processed independent of (e.g., in substantial isolation from) processing of the other sub-matrices. These resulting additional word matches may be referred to as soft alignment points. Where unmatched words remain between the hard and/or soft alignment points, the timecode information associated with the words of the hard and soft alignment points may be used to assess timecode information for the unmatched words (e.g., via interpolation). For example, timecodes associated with the words that make up the matched points at the end and beginning of an interval of time may be used as references to interpolate time values for unmatched words that fall within the interval between the matched words. Alignment techniques that may be implemented bysynchronization module 102 are discussed in more detail below. Further, techniques for matching are discussed in more detail below with respect toFIGS. 8A and 8B . - In some embodiments, providing time-coded script data includes providing timecodes assigned to all dialogue and other script element types. For example, in some embodiments, after
synchronization module 102 aligns word N-grams fromscript 110 with corresponding word N-grams oftranscript 114, it may output (e.g., to a client application) time information in the form of time-coded script data (e.g., time-aligned script data 116) that contains timecodes assigned to some or all dialogue and to some and/or all other script element types associated withscript 104. As described above, the data may be stored, displayed/presented or processed. In some embodiments, using the alignment processes described herein, a script (e.g., a Hollywood Spec. script or dramatic screenplay script) and a corresponding STT audio transcript are merged together by aligning script words with transcript words to provide resulting time-alignedscript data 116. Time-alignedscript data 116 may be processed and used by other applications, such as the Script Align feature of Adobe Premiere Pro. In some embodiments, processing may be implemented to time-align script elements other than audio (e.g., scene headings, action description words, etc.) directly to the video scene or full video content. For example, where a script element, other than dialogue (e.g., a scene heading) occurs between two script words, the timecodes of the script words may be used to determine a timecode of the script element. In some embodiments, each of the script elements may be provided in the time-aligned script data in association with a timecode, as discussed above with regard toFIG. 1D . Providing time-coded script data (block (308) may include providing the resulting time-aligneddata 116 to a storage medium, display device, or other modules for processing, as described above with regard toFIG. 1A . -
FIG. 4 is a flowchart that illustrates a time-alignment method 400 according to one or more embodiments of the present technique.Method 400 may provide alignment techniques using components and dataflow implemented atsynchronization module 102. In the illustrated embodiment,method 400 generally includes reading a script (SCR) file and a speech-to-text (STT) file, and processing the SCR and STT files using various techniques to generate an output that includes time-aligned script data. - In the illustrated embodiment,
method 400 includes reading an SCR file, as depicted atblock 402. This may include reading script data, such asscript data 110, described above with respect to block 302. For example, reading an SCR file may includescript reader 200 reading a generated SCR file (e.g., document data 110). The SCR file may include a record-format representation of a source Hollywood spec. script of dramatic screenplay script. Records contained in the SCR file may each include one complete script element.Script reader 200 may extract script element type and data values from each record and place these into an internal representation (e.g., a structured/tagged script document). - In the illustrated embodiment,
method 400 includes reading an STT file, as depicted atblock 404. This may include reading STT data, such astranscript 114, as described above with respect to block 304.Transcript 114 may include an STT file having transcribed data, such as that of the STTword transcript data 132 depicted inFIG. 1C . The STT data may provide a timecode for each spoken word in the audio sound track which corresponds in time tovideo content 106. - In the illustrated embodiment,
method 400 includes building a SCR N-gram dictionary, as depicted atblock 406. In some embodiments, building an SCR N-gram dictionary includes identifying all possible sequences of a given number of consecutive words. The number of words in the sequence may be represented by a number “N”. For example, the sentence, “The rain in Spain falls mainly on the plain” may be used to generate the following N-gram word sequences, where N is set to a value of 3: (The, rain, in), (rain, in, Spain), (in, Spain, falls), (Spain, falls, mainly), (falls, mainly, on), (mainly, on, the), and (on, the, plain). Note that additional N-gram word sequences may be generated based on words that precede or follow a phrase. For example, where the first word of a following sentence is “Why”, an additional 3-gram may include (the, plain, why). In some embodiments, the value of N may be set by a user. In some embodiments, the value of N is set to a predetermined value, such as four. For example, N may be automatically set to a default value of four, and the user may have the option to change the value of N to something other than four (e.g., one, two, three, five, etc.). - In some embodiments, some or all of the possible sequences of N number of consecutive words are identified for the script and/or the transcript, and the respective sequences are stored for use in processing. For example,
script analyzer 202 may build a word N-gram “dictionary” of all words fromscript 110 and may record their relative positions withinscript 110 and/orSTT analyzer 206 may build a word N-gram “dictionary” of all words fromtranscript 114 and may record their relative positions withintranscript 114. The resulting N-gram dictionaries may include an ordered table of 1-gram, 2-gram, 3-gram, or N-gram word sequences. - In the illustrated embodiment,
method 400 includes matching N-grams, as depicted atblock 408. In some embodiments, matching N-grams may include attempting to match N-grams of thescript 110 to corresponding N-grams oftranscript 114. For example,SCR analyzer 202 and/orSTT analyzer 206 may attempt to match all word N-grams of the N-gram dictionaries and may store the matches (e.g., in an internal table) in association with corresponding timecode information associated with the respective transcript word(s). The stored matching N-grams may indicate the potential for a matched sequence of words, and may be referred to as “candidate” N-grams for merging. For example, a phrase from the script N-gram dictionary may be matched with a corresponding phrase the transcript N-gram dictionary, however, due to the phrase being repeated several time within the script/video content, the match may not be accepted until the relative positions can be verified. - In the illustrated embodiment,
method 400 includes merging N-grams, as depicted atblock 410. In some embodiments, merging of N-grams may be provided bySCR analyzer 202 and/orSTT analyzer 206. In some embodiments, merging N-grams includes merge some or all sequential N-gram matches into longer matched N-grams. For example, where two consecutive matching N-grams are identified, such as two consecutive 3-grams of (The, rain, in) and (rain in Spain), they may be merged together to form a single N-gram, referred to as a single 4-gram of (The, rain, in, Spain). Such a technique may result in merged N-grams of length N+1 after each iteration. The technique may be repeated (e.g., iteratively) to merge all consecutive N-grams to provide N-grams having higher values of N. N-grams with higher values of N may have higher probabilities of being an accurate match. The iterative process may continue until no additional N-gram matches are identified. For example, where there are at most ten consecutive words identified as matching, increasing to an 11-gram length may yield no matching results, thereby terminating the merging process. Further, techniques for N-gram matching are discussed in more detail below with respect toFIGS. 8A and 8B . - With merging complete, the resulting set of merged N-grams may provide a set of “hard alignment points”. For example, each separate N-gram may indicate with relatively high certainty that a sequence of words in
script 110 precisely matches a sequence of words intranscript 114. The sequence of words may identify a hard-alignment point. Thus, a hard alignment point may include a series of matched words. In some case, the hard alignment points may include a series of words that each soft-align. - Due to the high probability of hard alignment points including accurate matches of words within
script 110 and words withintranscript 114, the timing data for each of the words of the matching N-grams (e.g., the corresponding timecode for transcript words) may be correlated with the corresponding script words. As discussed in more detail below, timing data for other words (e.g., unmatched words or words having low probabilities of accurate matches) may be assessed and determined based on the timecode data of words associated with matched words (e.g., words that make up one or at least a portion of one or more alignment points). For example, interpolation may be used to assess and determine the position of a script word that occurs between matched script words (e.g., script words associated with alignment points). - Hard alignment points may be found every 30-60 seconds within video content. In some embodiments, if hard alignment points are not found with N=4 (e.g., there are no matches of four consecutive words between the script and the transcript), N is decremented and the process repeated (e.g., returning to block 408). When N=1, words are matched one-to-one. In some embodiments, a default value of N=4 may be used, although the value of N may be modified.
- In the illustrated embodiment,
method 400 includes generating a sub-matrix, as depicted atblock 412. As noted above each hard alignment point may define a block of script text (e.g. a sequence of words in script 110) and a timecode indicative of where the hard alignment point occurs in the video. Although script and transcript words associated with hard alignment points may be associated with timecode data, other script words (e.g., unmatched words between each hard alignment point) may still need to be aligned to corresponding transcript words to assess and determine their respective timecode. In some embodiments, each successive pair of hard/soft alignment points is used to create an alignment sub-matrix. The alignment sub-matrix may include script words (e.g., sub-set of script words) that occur between matched script words (e.g., script words associated with hard alignment points) and intermediate transcript words (e.g., a sub-set of transcript words) that occur between matched transcript words (e.g., transcript words associated with hard alignment points). The script words may be provided along one axis (e.g., the y or x-axis) of the sub-matrix, and the intermediate transcript words may be provided along the other axis (e.g., the x or y-axis) of the sub-matrix. -
FIG. 5A depicts an exemplary (full)alignment matrix 500 in accordance with one or more embodiments of the present technique.Alignment matrix 500 may include some or all of the script words aligned in sequence along the y-axis and all of some of the transcript words aligned in sequence along the x-axis, or vice versa. In an ideal alignment match (which may rarely be the case) script words and transcript words would match exactly, resulting in a substantially straight line having a slope of about one or negative one. - As depicted in the illustrated embodiment, several (e.g., eight) hard alignment points 502 (denoted by circles) are identified. Between each of the hard-
alignment points 502 are a number of soft alignment points 504 (denoted by squares) and/or interpolated alignment points 506 (denoted by X's). Hard alignment points 502 may be determined as a result of matching/merging N-gram sequences as discussed above with respect toblocks Interpolation intervals 507 extend between adjacent soft alignments points 504. - As depicted,
alignment matrix 500 may include one or more alignment sub-matrices 508 a-508 g (referred to collectively as sub-matrices 508). Sub-matrices 508 a-508 g may be defined by the set of points (e.g., script words and transcript words) that are located between adjacent, respective, hard alignment points 502. For example, in the illustrated embodiment,matrix 500 includes seven sub-matrices 508 a-508 g. An exemplary sub-matrix 508 e is also depicted in detail inFIG. 5B . - In some embodiments,
method 400 includes pre-processing a sub-matrix, as depicted atblock 414. Pre-processing of the sub-matrix may be provided atmatrix aligner 208. In some embodiments, pre-processing the sub-matrix may include identifying the range of a particular sub-matrix (e.g., the range/sequence of associated script words and transcript words associated with the axis of the particular sub-matrix). For example, script and transcript words that fall between two words contained in adjacent hard alignment points 502 may be identified as a matrix sub-set of script words (SCR word sub-set) 510 (represented by outlined triangles) and a corresponding matrix sub-set of transcript words (STT word sub-set) 512 (represented by solid triangles), as depicted inFIG. 5B with respect to sub-matrix 508 e. It will be appreciated that the triangles ofFIGS. 5A and 5B represent only sub-sets of the script and transcript words, as each axis may represent all of the words for a particular portion of a clip, scene or entire movie being aligned. - In some embodiments, prior to words of
SCR word sub-set 510 being aligned to words ofSTT word sub-set 512 of sub-matrix 508 e, a timecode and position offset data structure used for booking is initialized. In some embodiments, all special symbols and punctuation are removed fromSCR word sub-set 510. This may provide for a more accurate alignment as both symbols and punctuations are typically not present in atranscript 114, and, are, thus, not present inSTT word sub-set 512. - In some embodiments, sub-matrices 508 of the
initial alignment matrix 500 are sequentially processed (e.g., in order of their location along the diagonal of the alignment matrix 500) to find the best time alignment for words between each pair ofhard reference points 502 that define each respective sub-matrix 508 a-508 g. Wheresystem 100 includes a single core system used to process the sub-matrices, alignment of the sub-matrices 508 may be processed sequentially (e.g., in series—one after the other). Wheresystem 100 includes a multi-core system used to process sub-matrices, alignment of some or all of sub-matrices 508 may be processed in parallel (e.g., simultaneously). Such parallel processing may be possible as the processing of each sub-matrix is independent of all of the other sub-matrices due to the bounding of the matrices with hard alignment points that are assumed to be accurate and that include known timecode information. - In the illustrated embodiment,
method 400 includes aligning the sub-matrix, as depicted atblock 416. Aligning the sub-matrix may be provided atmatrix aligner 208. In some embodiments, a sub-matrix may be aligned using an algorithm. An algorithm may employ a dynamic programming technique to assess multiple potential alignments for a sub-matrix, to determine the best fit alignment of the potential alignments, and employ the best fit alignment for the given sub-matrix. For example, an algorithm may identify several possible solutions within the sub-matrix, and may select the solution having the lowest indication of possible error. In some embodiments the algorithm may include a Levenshtein Word Edit Distance algorithm. Where a traditional Levenshtein algorithm is employed, a dynamic programming algorithm for computing the Levenshtein distance may require the use of an (n+1)×(m+1) matrix, where n and m are the lengths of the two respective word sets (e.g., the SCR word set and the STT word set). The algorithm may be based on the Wagner-Fischer algorithm for edit distance. - In some embodiments, an alignment path defines a potential sequence of words that may be used between hard alignment points. In some embodiments, aligning the sub-matrix may include breaking alignment paths within each sub-matrix into discrete sections during processing to more accurately assess individual portions of the alignment path. Based on match probabilities/strengths of various portions of the alignment path, a single alignment path may be broken into separate discrete intervals that are assessed individually. For example, where an alignment path within a sub-matrix includes a first portion having a relatively high match probability and an adjacent second portion having a relatively low match probability, the first and second portions can be separated. That is, the first portion may be identified as a sequence of words having a high probability of a match, and the second portion may be identified as a sequence of words having a low probability of a match. Accordingly, the first portion may be identified as an accurate match that can be relied on in subsequent processing and the second portion may be identified as an inaccurate match that should not be relied on in subsequent processing. Such a technique may be used in place of merely identifying a mediocre match of the entire alignment path that may or may not be reliable for use in subsequent processing.
- In some embodiments, aligning the sub-matrix may include weighting various processing operations to reflect operations that may be indicative of inaccuracies. For example, in some embodiments, aligning the sub-matrix may include assessing weighting penalties for matched words that are subject to an insert, delete, or substitute operation. Such a technique may help to adapt to false-positive word identifications produced by an STT engine.
- In some embodiments, the algorithm may be modified in an attempt to improve alignment. For example, in some embodiments, timecode information recorded with each word of an STT word set is correlated with a matching word of a corresponding SCR word set. The matching word may include a single word or a continuous sequence of words, wherein the sequence of words includes less than the number (“N”) of words required by the selected N-gram. The resulting alignments from this process are referred to as “soft alignment points.” In some embodiments, an algorithm, such as a Levenshtein Word Edit Distance algorithm, may be used to identify soft-alignment points. The soft designation is used to indicate that because of noise, error artifacts, and the like in
STT transcript 114, these alignments may have a lower probability of being accurate than the multi-word, hard-alignment points that define the range/partition of the respective sub-matrix. In some embodiments, soft-alignment points may be determined using heuristic and/or phonetic matching. - In some embodiments, aligning the sub-matrix may include heuristic filtering. Heuristic filtering of noise may include filtering (e.g., ignoring or removing) “stop words” (e.g., short articles such as “a”, “the”, etc.) that are typically inserted into an STT transcript when the STT engine is confused or otherwise unable to decipher the audio track. For example, STT engines often insert articles such as “a”, “the”, etc. while various events other than dialogue occur, such as the presence of noise, music or sound effects. Such articles may also be inserted when dialogue is present but cannot be deciphered by the STT engine, such as when noise, music or sound effects drown out dialogue or narration. As a result, the STT transcript may include a sequence of “the the the the . . . ” indicative of a duration when music or other such events occur in the audio content. Thus, heuristics may be used to identity portion transcript words that should be ignored. For example, transcript words that should not be considered in the alignment process, and/or should not be included in the resulting time-aligned script data.
- In some embodiments, heuristics may be used to identify repetitive sequences of words, and to determine which of the repeated sequence of words, if any need to be included or ignored in the resulting script document. For example, where a clip includes repetitive dialogue, such as where an actor repeats their lines several times in an attempt to get the line correct,
transcript 114 may include several repetitions (e.g., “i'll be back i'll be back i'll be back). A corresponding portion ofscript 110 may include a single recitation of the line (e.g., “I'll be back.”). In one embodiment, heuristics may be implemented to identify the repeated phrases, to identify one of the phrases of the transcript for use in aligning with script words, and to align the corresponding script words to the selected phrase oftranscript 114. For example, only the timecodes for words of one of the three phrases intranscript 114 may be associated with the corresponding script words of the phrase “I'll be back”. In some embodiments, the other repeated phrases are ignored/deleted. For example, ignored/deleted transcript words may not be considered in the alignment process, and/or may not be included in the resulting time-aligned script data. Ignoring/deleting the phrases may help to ensure that they do not create errors in aligning other portions ofscript 110. For example, if the additional phrases were not ignored/deleted, alignment may attempt to match the other two repeated phrases (e.g., those not selected) with phrases preceding or following the corresponding phrase ofscript 110. In some embodiments, instead of just throwing out (ignoring/deleting) the other repeated takes, they can also be aligned as “alternate takes”. For example, it may not know which take will eventually be used in a finished edit, so regardless of which take is used, the correct script text and timing information may flow through to that portion of the recorded clip in use. In some embodiments, a single portion script text may be aligned to each of the repeated portions of the transcript text. - In some embodiments, aligning the sub-matrix may include matching based at least partially on phonetic characteristics of words. For example, a word/phrase of the SCR word set may be considered a match to a word/phrase of the STT word set when the two words/phrases sound similar. In some embodiments, a special phonetic word comparator may be used to assess word/phrase matches. A phonetic comparator may include “fuzzy” encodings that provide for matching script words/phrases that may sound similar to a word identified in the STT transcript. Thus, a word/phrase may be considered a match if they fall within a specific phonetic match threshold. For example, a script word may be considered a match to a transcript word if the transcript word is a word identified as being an phonetic equivalent to the word in
script 110, or vice versa. For example, the terms “their” and “there” may be identified as phonetic matches although the terms do not exactly match one another. Such a technique may account for variations in spoken language (e.g., dialects) that may not be readily identified by an STT engine. Use of phonetic matching may be used in place of or in combination with an exact word/phrase match for each word/phrase. - In the illustrated embodiment,
method 400 includes generating and/or interpolating intervals, as depicted atblock 418. Generating and/or interpolating intervals may be provided at interval generator/interpolator 210. In some embodiments, generating and/or interpolating intervals may include identifying intervals between identified matched words (e.g., words of hard and/or soft reference points), interpolating the relative position of un-matched words between the matched words. An interpolated timecode for the un-matched words may be based on their interpolated position between the matched words and the known timecodes of the matched words. For example, after some or all of the sub-matrices are aligned, the sub-matrices are combined to form a list including script words and corresponding transcript words for each word associated with a hard or soft alignment point. At this stage of processing, all possible word alignment correspondences have been identified, leaving only unmatched script dialogue words (e.g., words that are not associated with hard nor soft reference points), and non-dialogue words within the script such as scene action descriptions and other information. These unmatched dialogue words still need to be assigned accurate timecodes to complete the script time-synchronization process. - In some embodiments, the timecode information for the unmatched script words is provided via linear timecode interpolation. Linear time code interpolation may include defining an interval that extends between two adjacent reference points, and spacing each of the unmatched words that occur between the two reference points across equal time spacing (e.g., sub-interpolation intervals) within the interval. A sub-interpolation interval may be defined as:
-
- Where t1 is a timecode of a first reference point defining a first end of an interpolation interval, t2 is a timecode of second reference point defining a second end of the interpolation interval, and n is the number of unmatched words.
- Where three unmatched words are identified in the script as being located between two matched words having timecodes of one second and two seconds, a first of the unmatched words may be determined to occur at 1.25 seconds, a second of the unmatched words may be determined to occur at 1.50 seconds, and a third of the unmatched words may be determined to occur at 1.75 seconds. In the above described embodiment, the sub-interpolation interval is equal to (2 sec-1 sec)/(3+1), or 0.25 sec.
FIG. 5B illustrates interpolatedpoints 506 for unmatched script words that are evenly spaced between soft alignment points in accordance with the above described linear interpolation technique. A similar technique may be repeated for each respective interpolation interval between hard/soft alignment points. - In the illustrated embodiment of
FIG. 4 ,method 400 includes assigning timecodes, as depicted atblock 420. Assigning timecodes may be provided at time-codedscript generator 212. In some embodiments, assigning time codes includes assigning times for each of the script words based on the reference points and interpolated points. For example, in some embodiments, the entire list of soft alignment points is scanned and each successive pair of soft alignment points defines an interpolation interval. Upon defining each interpolation interval, sub-interpolation intervals are determined, and timecode data aligning with the sub-interpolation intervals is assigned to all of the script words of the respective script word set. For example, the unmatched words of the above described interpolation interval may be assigned timecodes of 1.25 seconds, 1.50 seconds, and 1.75 seconds, respectively. Further, techniques for interpolating are discussed in more detail below with respect toFIGS. 8A and 8B . - In some embodiments, a non-linear interpolation technique may be employed to assess and determine timecode information associated with words/phrases within a script document. For example, non-linear interpolation or similar matching techniques may be used in place of or in combination with linear interpolation techniques employed to determine timecodes for script words. Non-linear interpolation may be useful to account for words that were not spoken at even rate between alignment points. For example, where two alignment points define an interval having matched words on either end and several unmatched words between them, linear interpolation may assign timecode information to the unmatched words assuming an even spacing across the interval as discussed above. The resulting timecodes may be reflective of someone speaking at a constant cadence across the interval. Unfortunately, the resulting timecode information may be inaccurate due to different rates of speech across the interval, pauses within the interval, or the like.
- In some embodiments, non-linear interpolation of timecode information may include assessing an expected rate (or cadence) for spoken words and applying that expected rate to assess and determine timecode information for the unmatched words. For example, non-linear interpolation may include, for a given script word, determining a rate of speaking for matched script words proximate the script word, and applying the rate of speaking to determine a timecode for the script word.
FIG. 7A illustrates alignment of a script phrase 700 (e.g., a portion of script data 110) with a spoken phrase 701 (e.g., a portion of transcript 114) that may be accomplished using non-linear interpolation in accordance with one or more embodiments of the present technique. In the illustrated embodiment, script phrase 700 is illustrated in association with analignment 702. Phrase 700 includes, “What is your answer to my question? I need to know your answer now!”Alignment 702 includes a series of word-match indicators (e.g., word associated with a hard alignment point (H) and words associated with a soft alignment point (S)) and words that are unmatched (U). The dots/points between the unmatched representations of “question” and “I” may indicate a pause between speaking of the words (e.g., a pause that would be indicated by timecode information differential between transcript words “position” and “eye” of spoken phrase 701). The sequence of four words “What is your answer to” and “know your answer now” include matches, and the words, “my”, “question”, “I”, “need” and “to” are unmatched. - In some embodiments, rates of speaking matched words proximate/adjacent (e.g., before or after) unmatched words may be used to assess and determine timecode information for the unmatched words. For example, in the illustrated embodiment, the rate of speaking “What is your answer to” may be used to assess and determine timecode information for the words “my” and “question.” That is, if it is determined that “What is your answer to” is spoken at a rate of one word every 0.1 seconds (e.g., via timecode information provided in the transcript and/or prior alignment/matching), the following words “my question” may be assigned timecode information in accordance with the rate of 0.1 words per second. For example, where the word “to” is determined to have been spoken at exactly twenty-one minutes (21:00.0) within a movie, it may be determined that the word “my” was spoken at twenty-one minutes and one-tenth of a second (21:00.1) and that the word “question” was spoken at twenty-one minutes and two-tenths of a second (21:00.2). Thus, timecodes associated with twenty-one minutes and one-tenth of a second (21:00.1) and twenty-one minutes and two-tenths of a second (21:00.2) may be assigned to the words “my” and “question”, respectively, in aligned
script data 116, for example. - In some embodiments, punctuation within the script may also be used to assess and determine timecode information. In one embodiment, for instance, punctuation indicative of the end of a phrase may be used to determine the presence of a pause between words or phrases. For example, the presence of the question mark in phrase 700 may indicate that the phrases “What is your answer to my question?” and “I need to know your answer now!” may be separated by a pause and, thus may each be spoken at different rates. Such a technique may be employed to assure that non-linear interpolation is applied to the individual phrases within a sub-matrix to account for an expected pause. For example, in the illustrated embodiment, the rate of speaking “know your answer now” may be used to assess and determine timecode information for the words “I”, “need” and “to”. That is, if it is determined that “know your answer now” was spoken at a rate of one word every 0.2 seconds (e.g., via timecode information provided in transcript 114), the preceding words “I need to” may be assigned timecode information in accordance with the rate of 0.2 words per second. For example, where the word “know” is determined to have been spoken at exactly twenty-one minutes and ten seconds (21:10.00) within a movie, it may be determined that the word “I” was spoken at twenty-one minutes nine and four-tenths of a second (21:09.4), that the word “need” was spoken at twenty-one minutes nine and six-tenths of a second (21:09.6), and the word “to” was spoken twenty-one minutes nine and eight-tenths of a second (21:09.8). Timecodes associated with twenty-one minutes nine and four-tenths of a second (21:09.4), twenty-one minutes nine and six-tenths of a second (21:09.6), and twenty-one minutes nine and eight-tenths of a second (21:09.8) may be assigned to the words “I”, “need”, and “know”, respectively, in aligned
script data 116, for example. Accordingly, punctuation may be used to identify pauses or similar breakpoints that can be used to break words or phrases into discrete intervals such that respective rates of speaking (e.g., cadence) can be appropriately applied to each of the discrete intervals. Other indicators may be used to indicate characteristics of the spoken words. For example, “stopwords” present in the transcript may be indicative of a pause or break in speaking and may be interpreted as a pause and implemented as discussed above. - It is noted that with some linear interpolation techniques, the unmatched words may be assigned timecode information based on even spacing between the matched words, and thus, may not account for the pause or similar variations. For example, in the embodiment of
FIG. 7A , where the first of the words “to” is determined to have been spoken at exactly twenty-one minutes (21:00.0) and the word “know” is determined to have been spoken at exactly twenty-one minutes and ten seconds (21:10.0), the five unmatched words “my”, “question”, “I”, “need” and “to” would be evenly spaced across the ten second interval at 1.67 second intervals, not accounting for the pause. Although minor in these small increments, this could lead to increased alignment errors where a pause in dialogue occurs for several minutes, for example. - In some embodiments, a rate of speech may be based on machine learning. For example, a rate of speech may be based on other words spoken proximate to the words in question. In some embodiments, a rate of speech may be determined based on elements of the script. For example, a long description of an action item may be indicative of a long pause in the actual dialogue spoken.
- In some embodiments, words of the script that occur proximate/between reference points may be aligned with unmatched words of the transcript that also occur proximate/between the same reference points. For example, in the illustrated embodiment of
FIG. 7A , the four unmatched words “my”, “question”, “I” and “need” of script phrase 700 fall within in the interval between matched words “to” and “know”. Where four unmatched words oftranscript phrase 701 also fall within the same interval, the timecodes associated with the unmatched words oftranscript phrase 701 may be assigned to the four unmatched words “my”, “question”, “I” and “need” of script phrase 700, respectively. That is the timecode of the first unmatched transcript word in the interval may be assigned to the first unmatched script word in the interval, the timecode of the second unmatched transcript word in the interval may be assigned to the second unmatched script word in the interval, and so forth. - In some embodiments, punctuation and/or capitalization from script text may be used to improve alignment. For example, if the first alignment point (hard or soft) occurs in the middle of the first sentence of the clip, it may be determined that the script words and transcript words preceding the alignment point in the script text and the corresponding transcript text should align with one another. In some embodiments, the timecodes for the script words may be interpolated (e.g., linearly or non-linearly) across the time interval that extends from the beginning of speaking of the corresponding transcript words in the scene to the corresponding alignment point. In some embodiments, the corresponding script words and transcript words may have a one-to one correspondence, and, thus, timecode information may be directly correlated. For example, the first script word of the sentence may be associated with the timecode information of the first transcript word of the clip, the second script word of the sentence may be associated with the timecode information of the second transcript word of the clip, and so forth. The beginning of a sentence may be identified by a capitalized word and the end of a sentence may be identified by a period, exclamation point, question mark, or the like.
-
FIG. 7B is a depiction of multiple lines of text that include a script phrase, a transcript phrase and a corresponding representation of alignment in accordance with one or more embodiments of the present technique. More specifically,FIG. 7B illustrates alignment of a script text 703 (e.g., a portion of script 110) with a spoken dialog 704 (e.g., a portion of transcript 114) that may be accomplished with the aid of capitalization and punctuation in accordance with one or more embodiments of the present technique.Script text 703 includes a portion of a script that is spoken throughout a clip/scene. More specifically, in the illustrated embodiment,script text 703 includes the first sentence of the clip/scene (e.g., “It is good to see you again”) and the last sentence of the clip/scene (e.g., “I will talk to you later tonight”).Spoken dialog 704 may include transcript text of a corresponding clip (e.g., “get it could to see you again” and “i will talk with you house get gator flight”). In the illustrated embodiment,script text 703 andtranscript text 704 is illustrated in association with analignment 705.Alignment 705 includes a series of word-match indicators (e.g., word associated with a hard alignment point (H) and words associated with a soft alignment point (S)) and words that are unmatched (U). As depicted, the first alignment point occurs midway though the first sentence of the scene/clip, and the first four words of the scene/clip are unmatched. In some embodiments, timecode for the script words at the beginning of the scene/clip that precede the first alignment point (e.g., “It is good”) may be interpolated across the time interval that extends from the beginning of speaking of the corresponding transcript words in the scene/clip to the corresponding alignment point (e.g., interpolated between the timecode of the transcript words “get” and “to” in the transcript phrase 704). In the illustrated embodiment, the number of corresponding unmatched script words and transcript words has a one-to-one correspondence, and, thus, timecode information may be directly correlated. For example, there are three words in each ofscript phrase 703 andtranscript phrase 704 that precede the first alignment point, and, thus, the first three script words (“It”, “is” and “good”) may each be assigned timecodes of the first three transcript words (“get”, “it” and “could”), respectively. Similarly, the location of the alignment points in the middle of the last sentence may enable the unmatched words “about”, “it”, “later”, and “tonight” that are located between the last alignment point of the scene/clip and the period indicative of the end of the scene/clip, to be interpolated across the interval between the transcript words “you” and “flight” and/or to each be assigned timecode information corresponding to transcript words “house”, “get”, “gator”, and “flight”, respectively. - In some embodiments, script elements may be used to identify the beginning or end of a sentence. For example, if between two lines of dialog, there is a parenthetical script element that corresponds to a sound effect, such as a car crash, the presence of the sound effect, indicated by a pause or stop words, may be used to identify the beginning or end of adjacent lines of dialog. In some embodiments, the techniques described with regard to alignment points in the middle of a sentence at the beginning or end of a scene/clip may be employed. For example, where the an alignment point within the dialog is preceded by or flowed by unmatched points and an identifiable script element (such as a sound effect), the timecodes for the unmatched words that occur between the alignment point and the identifiable script element may be interpolated across the corresponding interval or otherwise be determined. That is, the intermediate script element may be used in the same manner as capitalization and/or punctuation is used as described above.
- In some embodiments, the density of the words in the transcript may be used to assess and determine timecode information associated with the words in the script. For example, in the illustrated embodiment of
FIG. 7 , there are four unmatched transcript words in the interval ofphrase 701 between matched words (e.g., “two” and “know”) and there are five unmatched words (e.g., “my”, “question”, “I”, “need” and “to”) in the corresponding interval of phrase 700 between matched words (e.g., “to” and “know”). Based on the timecode information for the transcript words in the interval, it may be determined that two of the four unmatched transcript words are spoken at the beginning of the interval and that two of the four unmatched transcript words are spoken at the end of the interval. That is, about fifty percent of the spoken words were delivered in a first portion of the interval, no words were spoken in a second portion of the interval (e.g., during the pause) and about fifty percent of the words were spoken in a third portion of the interval. In one embodiment, a corresponding percentage of the script words (e.g., approximately equal to the percentage of transcript words) will be provided over the respective portions of the interval. For example, in the embodiment ofFIG. 7A , where the word “to” (in the first portion of the phrase 700) that defines a start of the interval is determined to have been spoken at exactly twenty-one minutes (21:00.0), the word “know” defining an end of the interval is determined to have been spoken at exactly twenty-one minutes and ten seconds (21:10.0), the word “position” is determined to have been spoken at exactly twenty-one minutes and ten and two-tenths seconds (21:00.2), and the word “eye” is determined to have been spoken at exactly twenty-one minutes and nine and four-tenths seconds (21:09.4), the two unmatched script words “my” and “question” may be evenly spaced over the first portion of the interval from twenty-one minutes (21:00.0) to twenty-one minutes and ten and two-tenths seconds (21:00.2), and the three unmatched words “I”, “need” and “to” may be evenly spaced across the third portion of the interval from twenty-one minutes and nine and four-tenths seconds (21:09.4) to twenty-one minutes and ten seconds (21:10.0). Thus, the distribution of script words within the interval is approximately equivalent to the distribution of transcript words in the corresponding interval. That is, about fifty percent of the script words in the interval are time aligned across the first portion of the interval before the pause and about fifty percent of the script words in the interval are time aligned across the third portion of the interval after the pause. - In some embodiments, a plurality of script words may be accepted for use in the time-aligned script data based on a confidence (e.g., high probability/density of word matches that were previously determined). Such a technique may enable blocks of text to be verified/imported from the script data to the time-aligned script data when matches within the blocks are indicative of a high probability that the corresponding script words are accurate. That is, the script data will be the text used in the time-aligned script data for those respective words of the script/dialogue. In some embodiments, a block of script words may be imported when word matches (e.g., hard alignment points and/or soft alignment points) meet a threshold level. For example, at least a portion of a block of words may be verified/imported for use in the aligned script when at least fifty percent of the words in the block are associated with a match (e.g., associated with hard and/or soft alignment points). In some embodiments, verifying/importing blocks of text may include using some individual script words having a match (e.g., associated with hard and/or soft alignment points) with words of the script, while importing/using unmatched transcript words (e.g., that are not associated with a soft and/or hard alignment points). In some embodiments, verifying/importing script words may include importing text characteristics, such as capitalization, punctuation, and the like. In the embodiment of
FIG. 7A , more than fifty-percent of the words of script phrase 700 are identified as having a hard and/or soft match. In some embodiments, upon determining that the script text and transcript text have a high enough percentage of matches (e.g., exceeding a block match threshold), the script text may be used for the entire block of text in the aligned script document, including matched and unmatched words for use in the script-aligned data. For example, the block of corresponding script text “What is your answer to my question? I need to know your answer now!” may be used in the aligned script although all of the words do not have a match. The imported script words have incorporated the capitalization and punctuation of the corresponding text of the script document. Timecode information may be associated with each of the script and transcript words using any of the techniques described herein to properly time align the unmatched words of the phrase (e.g., to provide timecodes for the words “my question? I need to”). As discussed in more detail below, where a high confidence for a block of transcript words is provided, the transcript words (including those not matched) may be used in the resulting time-aligned script. Accordingly, if the transcripts words of the phrase “What is your answer to by position eye do know your answer now!” have a high confidence leave but are not all matched, the phrase may be used in the resulting text of the time-aligned script data. Note that both, the matched and unmatched words of the raw STT have been imported. Such a technique may facilitate use of transcript words in place of script words where the actor ad-libs or otherwise does not recite the exacting wording of the script. - In some embodiments, a user could choose for themselves whether to use the Script word(s) or SST transcript word(s), based on an indication, such as confidence level. For example, even if the confidence level assumes one is more accurate than the other, it may not be so, and the user may be provided an opportunity to correct this by switching use of one or the other in the script data. Also, the user can manually edit in a correction, and this correction could be automatically stamped with a 100% confidence label. In some embodiments, the automated changes/imports may be marked such that a user can readily identify them, and modify them as needed.
- In some embodiments, confidence/probability information provided during STT operations may be employed to assess whether or not a word or block of words in a transcript meets threshold criteria, such that the transcript words may be used in the time-aligned script data in place of the corresponding script words. Such an embodiment may resolve discrepancies by using the transcript word in the aligned
script data 116 where there is a high confidence that the transcript word is accurate and the corresponding script word is not (e.g., where an actor ad-libs a line such that the actual words spoken are different from the words in the script). In one embodiment, an STT engine may provide a high confidence level (e.g., above 90%) for a given transcript word, and, thus, the transcript word is considered to meet the threshold criteria (e.g., 85% or above). That is, the word in the transcript may be more accurate than corresponding script words. As a result, the transcript word is provided in the aligned script data, in place of a corresponding script word. In some embodiments, a confidence/probability provided by an STT operation may be used in combination with matching criteria. For example, where a low confidence level (e.g., below 50%) is provided for a script word as a result of matching/merging, and the STT engine provides a high confidence level (e.g., above 90%) for a corresponding transcript words, the transcript word may be provided in the aligned script data, in place of a corresponding script word. Conversely, where a high confidence level (e.g., above 90%) is provided for a script word as a result of matching/merging, and the STT engine provides a low confidence level (e.g., below 50%) for a corresponding transcript word, the script word may be provided in the aligned script data, in place of a corresponding transcript word. - In some embodiments, a portion of the script may be longer than a corresponding clip. As a result, the portion of the script that is actually spoken may be time aligned appropriately, and the unspoken portions of the script may be bunched together between aligned points. The bunching of words may result in timecode information being associated with the bunched words that indicates them being spoken at an extremely high rate, when in fact they may not have been spoken at all. In some embodiments, a threshold is applied to ignore or delete words that appear to have been spoken too quickly such that bunched words may be ignored or deleted. For example, a threshold word rate may be set to a value that is indicative of the fastest reasonable rate for a person to speak (e.g., about six words per second). In some embodiments, the threshold word rate may be set to a default value, may be determined automatically, or may user selected. A speaking rate may be customized based on the character speaking the dialogue. For example, one actor may speak slowly whereas another actor may speak much faster, and thus the slower speaking character's dialogue may be associated with a lower threshold rate, where as the faster speaking character's dialogue may be associated with a higher threshold rate. Automatically determining a threshold word rate may include sampling other spoken portions of a script (e.g., other lines delivered by the same character) to determine a reasonable rate for words that are actually spoken, and the threshold rate may be set at that value or based off of that value. For example, where one portion of a script includes an average word rate of five words per second, a maximum word rate threshold may be set to approximately twenty percent greater than that value (e.g., about six words per second). Such a cushion may account for natural variations in speaking rate that may occur while still identifying unlikely variations in speaking rate. In some embodiments, words having spacing that do not fall within the maximum word rate threshold are ignored or deleted, such that they are not aligned. For example, a script may read:
-
- That's his name. Henry Jones, Junior.
- INDY
- I like Indiana more than the name Henry Jones, Junior.
- HENRY
- We named the dog Indiana.
The corresponding video content (e.g., clip) however, may only include an actor reciting Henry's lines, one after the other. Thus, the lines delivered for Henry may be provided accurate timecode information associated with the time periods in which the two lines are spoken, however, the line associated with Indy, that is not spoken, may be bunched into the pause between delivery of Henry's first and second lines. For example, if Henry's lines were delivered one-after the other, with a half-second pause in-between, the phrase “I like Indiana more than the name Henry Jones, Junior” may not be matched (because it was not actually spoken) and, thus, may be interpolated (e.g., linearly) over the half-second time frame between the lines in the script. Corresponding timecode information may indicate that “I like Indiana more than the name Henry Jones, Junior” was spoken at a rate of one word about every five one-hundredths of a second, or about twenty words per second. Where the maximum word threshold is set to about six words per second, the determined rate of about twenty words per second would exceed the maximum word threshold. Thus the phrase “I like Indiana more than the name Henry Jones, Junior” may be ignored/deleted, such that alignment may be provided for only the lines actually spoken (e.g., Henry's lines). The phrase “I like Indiana more than the name Henry Jones, Junior” may not be provided in the time-aligneddata 116.
- That's his name. Henry Jones, Junior.
- In some embodiments, words that were bunched at the beginning or end of dialogue (e.g., the script text that was linearly interpolated and bunched before or after the dialogue was actually spoken) may be identified and removed. For example, the following lines at the beginning of the dialogue were linearly interpolated:
-
01:58:00:02 1:5938 {circumflex over ( )}EXT./01:58:00:02 {circumflex over ( )}ENTRANCE/01:58:00:02 1:5939 Scene {circumflex over ( )}TO/01:58:00:02 {circumflex over ( )}MOUNTAIN/01:58:00:02 {circumflex over ( )}TEMPLE/01:58:00:02 {circumflex over ( )}-/01:58:00:02 {circumflex over ( )}Scene {circumflex over ( )}AFTERNOON/01:58:00:02
Bunching of the words is indicated by them each having been assigned the same timecode, which may be a result of linearly interpolating over a very short period of time (e.g., prior to the start of actual dialogue of “Indy” following the above lines at time 01:58:00:04). In some embodiments, the bunched words are deleted/ignored such that they are not included or indicated as being aligned in the resulting aligned script data. Thus, interpolated alignment of text that is located at the beginning or end of dialogue and that is bunched into a short duration may be deleted/ignored. - In some embodiments, ignoring/deleting words that appear to exceed a maximum threshold rate may also help to eliminate “stopwords” generated by an STT engine from being considered for alignment. For example, where an STT engine inserts a plurality of “the, the, the, . . . ” in place of music or sound effects, the high frequency of the words “the” may be identified and they may be ignored/deleted such that they are not aligned to words in the script. In some embodiments, the stopwords may be flagged (e.g., not recognized) so that a user can take further action if desired.
- In some instances, a clip may include audio content having extraneous spoken words that are not intended to be aligned with corresponding script words. For example, extraneous words and phrases may include an operator calling out “Speed!” shortly before starting the camera rolling while audio is already being recorded, the director calling out “Action!” shortly before the characters beginning to speak lines of dialogue, the director calling out “Cut!” at the end of a take, or conversations inadvertently recorded shortly before, after, or even in the middle of a take. These cues typically occur at the beginning and end of shots, and, thus, processing may be able to recognize these words based on their location and/or their audio-waveforms that are recognized and provided in a corresponding STT transcript. If the entire recorded audio from the clip were to be analyzed, the extraneous/incidental words may provide significant challenges during alignment. For example,
synchronization module 102 may align the extraneous words of the transcript to script words, resulting in numerous errors. User defined words, such as “Speed”, “Action” and “Cut” may be defined and can be recognized by their audio waveforms and provided in a corresponding STT transcript. The user defined words may be automatically flagged for the user or deleted. - In some embodiments, only a defined range of recorded dialogue is aligned to script text. Such a technique may be useful to ignore or eliminate extraneous recorded audio from the alignment analysis. For example, defining a range of recorded dialog may enable the analysis to ignore extraneous conversations or spoken words that are incidentally recorded just before or after a take for a given scene. In some embodiments, an in/out range defines the portion of the audio that is aligned to a corresponding portion of the script. Defining an in/out range may define discrete portions of the script (e.g., script word) and/or audio content (e.g., transcript words) to analyze while also defining discrete portions of the audio content data to ignore during the alignment of transcript words with corresponding script words, thereby preventing extraneous words (e.g., transcript words) from inadvertently being aligned with script words.
FIG. 7C is a depiction of a line of text and corresponding in/out ranges in accordance with one or more embodiments of the present technique. More specifically,FIG. 7C illustrates an exemplary in-range 710 and out-ranges 711. The in-range 710 and out-ranges 711 limits analysis to only audio content of in-range 710, referred to herein as audio content ofinterest 712, and excludes audio content not located within in-range 710 (e.g., content located in out-ranges 711). Audio content ofinterest 712 may include the dialogue or narration spoken during the respective clip that falls within one or more specified in/out-ranges. Extraneousaudio content 714 may include words captured on the audio that are not intended to be aligned with a corresponding portion of script document, and, thus, fall outside of the one or more specified in/out-ranges. In the illustrated embodiment, audio content ofinterest 712 includes the transcribed phrase “hello mike . . . I am doing well also” andextraneous audio content 714 includes the phrases/words “are we ready speed action” spoken at the head of the clip, just before audio content ofinterest 712 and “cut how did that look” spoken at the tail of the clip, just after audio content ofinterest 712. As depicted, inrange 710 is defined by an in-marker 710 a and an out-marker 710 b. In-marker 710 a defines a beginning of audio content ofinterest 712, and out-marker 710 b defines an end of audio content ofinterest 712. By specifying an in/out range, other portions of the dialog may be excluded from the analysis. For example, in the illustrated embodiment,extraneous content 714 at the head and tail of the clip is ignored during analysis, as indicated by the grayed out bar inFIG. 7C . In the illustrated embodiment, only a single in-range 710 is depicted, however, embodiments may include multiple discrete ranges defined within a single clip. For example, two additional in/out markers may be added within in-range 710, thereby dividing it into two discrete in-ranges and providing an additional out-range embedded therein. In some embodiments, the use of in/out-ranges may be employed to resolve issues normally associated with multiple takes of a given scene or clip. For example, a user could select the desired portion of the take by selecting an in-range that includes the desired take and/or selecting an out-range that excludes the undesired takes. In some embodiments, an out-range may be located at any portion of the clip. For example, in a case opposite from that depicted, the in/out-ranges may be swapped, thereby ignoring extraneous audio data in the middle of the clip, while analyzing audio content of interest at the head and tail of the clip. - In some embodiments,
markers FIG. 7C and may use a slider-type control to movemarkers ranges extraneous audio content 714 using in/out-ranges. In some embodiments,markers - In some embodiments, portions of the audio content may include extraneous audio other than spoken words, such as music or sound effects. If analyzed, the extraneous audio may create an additional processing burden on the system. For example,
synchronization module 102 may dedicate processing in an attempt to match/align extraneous transcript words (e.g., stop words) to script words. In some embodiments, the extraneous audio content may be identified and ignored during alignment. Such a technique may enable processing to focus on dialogue portions of audio content, while skipping over segments of extraneous audio. In some embodiments, the audio content may be processed to classify segments of the audio content into one of a plurality of discrete audio content types. For example, segments of the audio content identified as including dialogue may be classified as dialogue type audio, segments of the audio content identified as including music may be classified as music type audio, and segments of the audio content identified as including sound effects may be classified as sound effect type audio. For example, segments of transcript words that include a series of different words occurring one after another (e.g., how are you doing) and/or that are not indicative of stop words may be classified as a dialogue type audio, segments of transcript words that include a series of stop words of a long duration (e.g., the the the the . . . ) may be classified as a music type audio, and segments of transcript words that include a series of stop words of a short duration (e.g., the the the) may be classified as a sound effect type audio. In some embodiments, segments of the audio content that cannot be identified as one of dialogue, music or sound effect type audio may be categorized as unclassified type audio. During subsequent processing, each of the segments may or may not be subject to alignment or related processing based on their classification. For example, during alignment of transcript words to script words, the segments associated with dialogue type audio may be processed, whereas the segments associated with music and sound effect type audio may be ignored. By ignoring music and sound effect type segments, processing resources may be focused on the dialogue segments, and, thus, are not wasted attempting to align the transcript words associated with the music and sound effect to script words. In some embodiments, unclassified type audio may be considered for alignment or may be ignored. In some embodiments, what classifications are processed and what classifications are ignored may include a default setting and/or may be user selectable. - In some embodiments, a weighting value is assigned to each word based on the alignment type (e.g., interpolation, hard alignment, or soft alignment). Stronger alignments (e.g., hard and soft alignments) may have higher weighting than weaker alignments (e.g., interpolation). In some embodiments, a total weighting is assessed for a window/interval that includes several consecutive words. The interval of several words is a sliding window that is moved to assess adjacent intervals/windows of words. When the total weighting (e.g., sum of weightings) of the words in a given interval/window meets a threshold value, it may be determined that the words are not merely bunched words, and timecodes may be assigned to one or more of the words, thereby, not ignoring/deleting the words in the window. Such a technique may be provided at the beginning and end of a set of dialogue to assess and determine the start and stop of the actual spoken dialogue and to ignore/delete the script dialogue that preceded/followed the spoken dialogue in the script, but was not actually spoken (e.g., the script text that was linearly interpolated as was bunched before or after the dialogue actually spoken).
- In some embodiments, processing may be implemented to time-align script elements other than dialogue (e.g., scene headings, action description words, etc.) directly to the video scene or full video content. For example, where a script element, other than dialogue (e.g., a scene heading) occurs between two words having timecodes associated therewith (e.g., dialogue words in the time-aligned script data) the timecodes of the words may be used to determine a timecode of the intervening script element. For example, where a last word of a scene includes a timecode of 21:00.00 and the first word of the next scene includes a timecode of 21:10.00, a script element occurring in the script between the two words may be assigned a timecode between 21:00.00 and 21:10.00, such as 21:05.00. In some embodiments, one or more script elements may have their timecodes determined via linear and/or non-linear interpolation, similar to that described above. For example, the amount of content (e.g., the number of lines or number of words) within script elements may be used to assess a timecode for a given script element or plurality of script elements. Where a first script element between two words having timecodes includes half the amount of content of a second script element also located between the two words, the first script element may be assigned a timecode of 21:03.00 and the second script element may be assigned a time code of 21:05.00, thereby reflecting the smaller content and potentially shorter duration of the first element relative to the second element. In some embodiments, some or all of the script elements may be provided in the time-aligned script data in association with a timecode. In some embodiments, timecodes are first assigned to the dialogue words during initial alignment, and timecodes are assigned to the other script elements in a subsequent alignment process based on the timecodes of the dialogue determined in the initial alignment (e.g., via interpolation). The resulting time aligned
data 116 may include timecodes for some or all of the script elements ofscript 104. - In the illustrated embodiment,
method 400 includes generating a time-aligned script output, as depicted atblock 422, as discussed above. Generating time-aligned script output may be provided via time-codedscript generator 212. In some embodiments, each word or element of the script and/or transcript may be associated with a corresponding timecode. For example, the complete list of script word and/or transcript words that are associated with hard, soft and interpolated timecodes may be used to generate time-aligneddata 116, including a final TimeCodedScript (TCS) data file which contains some or all of the script elements with assigned time codes. In some embodiments, the TCS data file may be provided to another application, such as the Adobe Script Align and Replace feature of Adobe Premiere Pro, for additional processing. In some embodiments, time-aligneddata 116 may be stored in a database for use by other applications, such as the Script Align feature of Abode Premiere Pro. - In some embodiments, a graphical user interface may provide a graphical display that indicates where matches (e.g., hard and/or soft alignment points) or non-matches occur within a user interface. The user interface may include symbols or color coding to enable a user to readily identify various characteristics of the alignment. For example, hard alignments may be provided in red (or green) to indicate a good/high confidence, soft alignments in blue (or yellow) to indicate a lower confidence, and interpolated points in yellow (or red) to indicate an even lower confidence level. The user interface may enable a user to quickly scan the results to assess and determine where inaccuracies are most likely to have occurred. Thus, a user may commit resources for review and proofing efforts on portions of a time-aligned script that may be susceptible to errors (e.g., where no or few matches occur) and may not commit resources for review and proofing efforts on portions of a time-aligned script that may not be susceptible to errors (e.g., where a large number of matches occur). For example, a user may be presented with a chart, such as that illustrated in
FIG. 5A . The chart may enable a user to readily identify portions of the script that do not include a high percentage of matches (e.g., the sub-matrix 508 located at the uppermost left portion of the chart). In some embodiments, high confidence areas may include a similar visual indicator (e.g., grayed out) and portions that may require attention may have appropriate visual indicators (e.g., bright colors—not grayed out). - In some embodiments, a user may be provided the option to select whether or not to use the text from the raw STT analysis or the text from the written script. For example, a user may be provided a selection in association with the sub-matrix 508 located at the uppermost left portion of the chart that enables all, some, or individual words contained in the sub-matrix to use the text from the raw STT analysis or the text from the written script.
- In some embodiments, upon receiving a user input, the information may be returned to
synchronization module 102 and processed in accordance with the user input. For example, where a user opts to use STT text in place of script text,synchronization module 102 may conduct additional processing to provide the corresponding time-aligned script data. In some embodiments, the user may be prompted for input whilesynchronization module 102 is performing the time alignment. For example, as thesynchronization module 102 encounters a decision point, it may prompt the user for input. - Some embodiments may include additional features that help to improve the performance of
system 100. For example, in some embodiments, speech-to-text analysis (e.g.,audio extractor 112 and/or the method of block 304) may provide the option of creating a custom dictionary (e.g., custom language model). In some embodiments, a custom dictionary may be generated for a given clip based on one or more reference scripts that have content that is the same or similar to the given script, or based on a single reference script that at least partially corresponds to the video content or exactly matches the audio portions of the video content. In some embodiments, such as where the reference script exactly matches the audio content, some or all words of the reference script may be used to define a custom dictionary, a raw speech analysis may be performed to generate a transcript using words of the custom dictionary to transcribe words of the audio content, transcript words may then be matched against the script words of the reference script to find alignment points, and the words of the reference script text may be paired with the corresponding timecodes, thereby providing a time-aligned/coded version of the reference script. - In some embodiments, a custom language model is generated for one or more portions of video content. For example, where a movie or scene includes a plurality of clips, a custom language module may be provided for each clip to improve speech recognition accuracy. In some embodiments, a custom language model is provided to a STT engine such that the STT engine may be provided with terms that are likely to be used in the clip that is being analyzed by the STT engine. For example, during STT transcription, the STT engine may at least partially rely on terms or speech patterns defined in the custom language model. In some embodiments, a custom language model may be directed toward a certain sub-set of language. For example, the custom language model may specify a language (e.g., English, German, Spanish, French, etc.). In some embodiments, the custom language model may specify a certain language segment. For example, the custom language module may be directed to a certain profession or industry (e.g., a custom language module including common medical terms and phrases may be used for clips from a medical television series). In some embodiments, the STT engine may weight words/phrases found in the associated custom language module over the standard language model. For example, if the STT engine associates a word with a word that is present in the associated custom language model and a word that is present in a standard/default language model, the STT engine may select the word associated custom language model as opposed to the word present in the standard/default language model. In some embodiments, a word identified in a transcript that is found in the selected custom language model may be assigned a higher confidence level than a similar word that is only found in the standard/default language model.
- In some embodiments, a custom language model is generated from script text. For example,
script data 110 may include embedded script text (e.g., words and phrases) that can be extracted and used to define a custom language model. Embedded metadata may be provided using various techniques, such as those described in described in U.S. patent application Ser. No. 12/168,522 entitled “SYSTEMS AND METHODS FOR ASSOCIATING METADATA WITH MEDIA USING METADATA PLACEHOLDERS”, filed Jul. 7, 2008, which is hereby incorporated by reference as though fully set forth herein. A custom language model may include a word frequency table (e.g., how often each of the words in the custom language model is used within a given portion of the script) and a word tri-graph (e.g., indicative of other words that precede and followed a given word in a given portion of the script). In some embodiments, all or some of the text identified in the script may be used to populate the custom language model. Such a technique may be particularly accurate because the script and resulting language model should include all or at least a majority of the words that are expected to be spoken in the clip. In some embodiments, speech-to-text (STT) technology may implement a custom language model as described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, which is hereby incorporated by reference as though fully set forth herein - In some embodiments, metadata included in the script may be used to further improve accuracy of the STT analysis. For example, where the script includes a clip identifier, such as a scene number, the scene number may be associated with the clip such that a particular custom language model is used for STT analysis of video content that corresponds to the associated portion of the script. For example, where a first portion of the script is associated with scene one and a second portion of the script is associated with scene two, a first custom language model may be extracted from the first portion of the script, and a second custom language model may be extract from the second portions of the script. Then, during STT analysis of the first scene, the STT engine may automatically use the first custom language model, and during STT analysis of the second scene, the STT engine may automatically use the second custom language model.
- In some embodiments, when a clip contains only a few lines of dialogue in a short scene out of a very long script, knowing that the clip contains a specific scene number (e.g., harvested from the script metadata) allows focusing on the text in the script for that scene, and not having to assess the entire script.
-
FIG. 6 depicts a sequence ofdialogs 600 in accordance with one or more embodiments of the present technique. In some embodiments, a user may select a clip or group of clips, then chooses “Analyze Content” from a Clip menu, initiating the sequence ofdialogs 600. The Analyze Content dialog may allow a user to use embedded Adobe Story Script text if present for the speech analysis, or to add a reference script which will be used to improve speech analysis accuracy. The sequence ofdialogs 600 includes content analysis dialogs that allow users to import a reference script to create a custom dictionary/language model for speech analysis. A reference script may include a text document containing dialogue text similar to the recorded content in the project (e.g., a series of nature documentary scripts, or a collection of scripts from a client's previous training videos). In the Analyze Content dialog 602, a user may choose Add from the Reference Script menu. In theFile Open dialog 604, a user may navigate to the reference script text file, select it and click OK. The AddReference Script dialog 606 may open, where a user can name the reference script, choose a language, and view the text of the file below in a scrolling window. The “Script Text Matches Recorded Dialogue” option may be selected if the imported script exactly matches the recorded dialogue in the clips (e.g., a script the actors read their lines from). When a reference script is used that doesn't exactly match the recorded dialogue in the clips, the analysis engine automatically sets the weighting of the reference script vs. the base language model based on length, frequency of key words, etc. A user may click the OK button, the Import Script dialog closes, and the analysis of the reference script may begin. When analysis is complete, the reference script is selected in the Analyze Content's Reference Script menu. When a user clicks the OK button, the selected clip's speech content is analyzed. - Higher accuracy may be possible when the reference script matches the recorded dialogue exactly (e.g., the script that was written for the project or transcriptions of interview sound bites). In this scenario, a user may select the “Script Text Matches Recorded Dialogue” option in the Add
Reference Script dialog 606, as discussed above. This may override the automatic weighting against the base language model and give the selected reference script a much higher weighting. Significantly higher accuracy can be achieved using matching reference scripts, although accuracy may be primarily dependent on the clarity of the spoken words and the quality of the recorded dialogue. - High accuracy (e.g., up to 100%) may be achievable when additional associated software packages in the production workflow are used in conjunction with one another. For example, an Adobe Story to Adobe OnLocation workflow may be used to embed the dialogue from each scene into a clip's metadata. In such a workflow, a script written in Adobe Story may be imported into OnLocation, which may produce a list of shot placeholders for each scene. These placeholders may be recorded direct to disk using OnLocation during production or merged with clips that are imported into OnLocation after they were recorded on another device. In both cases, the text for each scene from the original script may be embedded in the metadata of all the clips that were shot for that scene. Embedded metadata may be provided using various techniques, such as those described in described in U.S. patent application Ser. No. 12/168,522 entitled “SYSTEMS AND METHODS FOR ASSOCIATING METADATA WITH MEDIA USING METADATA PLACEHOLDERS”, filed Jul. 7, 2008, which is hereby incorporated by reference as though fully set forth herein. When the clips are imported into Adobe Premiere Pro, the script text embedded in each of the clips may be automatically used as a reference script and, then, aligned with the recorded speech during the analysis. When enough hard alignment points reach a minimum accuracy threshold, the analyzed speech text is replaced with the script text embedded in the source clip's extensible metadata platform (XMP) metadata. This may result in speech analysis text that is at or near 100% accurate relative to the original script. Correct spelling, proper names and punctuation may also be carried over from the script. Accuracy in this workflow may be dictated by the closeness of the match between the reference script text and the recorded dialogue.
- With regard to
FIG. 6 , in some embodiments, when the “Use Embedded Adobe Story Script Option” of Analyze Content dialog 602 is selected, Adobe Story script text embedded in an XMP will be used for analysis, and the Reference Script popup menu may be disabled. If the selected clip contains Adobe Story script embedded text, the “Use Embedded Adobe Story Script Option” may be checked by default. For mixed states in the selection (e.g., where at least one clip has Adobe Story script text embedded, and at least one clip does not), the dialog will open with the “Use Embedded Adobe Story Script Option” checkbox indicating a mixed state and the Reference Script popup menu may be enabled. If the analysis is run in this mixed state, the clip with the Adobe Story script embedded will be analyzed using the Adobe Story script and the clip without the Adobe Story script embedded will be analyzed using the reference script. Selecting the mixed state may generate a check in the “Use Embedded Adobe Story Script Option” checkbox and disable the “Reference Script” menu. If the analysis is run in this state, the result may be the same as above. Selecting the checkbox again may remove the check mark at the “Use Embedded Adobe Story Script Option” checkbox and may re-enable the “Reference Script” menu. If the analysis is run in this state, all clips may use the assigned reference script, and ignore any embedded Story Script text that may be in one or more of the selected clips. - In some embodiments, an STT engine may require that a custom language model include a minimum number of words (e.g., a minimum word count). That is, an STT engine may return an error and/or ignore a custom language model if the model does not include a minimum number of words. For example, if a portion of a script includes only ten words, a corresponding custom language model may include only the ten words. If the STT engine required a minimum of twenty-five words, the STT may not be able to use the custom language model having only ten words. In some embodiments, the words in the custom language model may be duplicated to meet the minimum word count. For example, the ten words may be repeated two additional times in an associated document or file that defines the custom language model to generate a total of thirty words, thereby enabling the resulting custom language model to meet the minimum word requirement of twenty-five words. It is noted that if all of the words are replicated the same number of times, the word frequency table (e.g., how often each of the words in the custom language model is used), and the word tri-graph (e.g., indicative of other words that precede and followed a given word) of the custom language model should remain accurate. That is the frequencies and words that precede or follow a given word remain the same.
- In some embodiments, it may be desirable to automatically and systematically identifying some or all entities (e.g., dialogue and events) of a script that are of interest to production personnel who work with the script. For example, it may be desirable to identify people, places, and thing/noun entities contained in the script. In the usage chain of video content, such as a movie, users (e.g., marketing personnel, advertisers, and legal personnel) may be interested in identifying and locating when specific people, places, or things occur in the final production video or film to enable, for example, identifying prominent entities that occur in a scene in order to perform contextual advertising (e.g., an advertisement showing a certain type of car ad if the car appears in a crucial segment.) Thus, the processed script, extracted entities, and time-aligned dialogue/entity metadata may enable third-parties applications (e.g., contextual advertisers) to perform high relevancy ad placement.
- In some embodiments, a method for identifying and aligning some or all entities within a script includes receiving script data, processing the script data, receiving video content data (e.g., video and audio data), processing the video content data, and synchronizing the script data with the video content data to generate time-aligned script data, and categorizing each regular or proper noun entity within the time-aligned script data. In some embodiments, receiving and processing script data and receiving and processing video content data are performed in series or parallel prior to performing synchronizing the script data with the video content data which is flowed by categorizing each regular or proper noun entity within the time-aligned script data.
- Receiving script data may include processes similar to those above described with respect to
document extractor 108. For example, receiving script data may include accepting a Hollywood “Spec.” Movie Script or dramatic screenplay script document (e.g., document 104), converting this script into specific structured and tagged representation (e.g., document data 110) via systematically extracting and tagging all key script elements (e.g., Scene Headings, Action Descriptions, Dialogue Lines), and then storing these elements as objects in a specialized document object model (DOM) (e.g., a structured/tagged document) for subsequent processing. - Processing the script data may include extracting specific portions of the script. Extracted portions may include noun items. For example, for a given script DOM, processing script data may include processing the objects (e.g., entire sentences tagged by script section) within the script DOM using an NLP engine that identifies, extracts, and tags the noun items identified by the system for each sentence. The extracted and tagged noun elements are then recorded into a specialized metadata database.
- Receiving video content data may include processes similar to those described above with respect to
audio extractor 112. For example, receiving video content data may include receiving a video or audio file (e.g., video content 112) that contains spoken dialogue that closely but not necessarily exactly corresponds to the dialogue sections of the input script (e.g., document 104). The audio track in the provided video or audio file is then processed using a Speech-to-Text engine (e.g., audio extractor 112) to generate a transcription of the spoken dialogue (e.g., transcript 114). The transcription may include extremely accurate timecode information but potentially higher error rates due to noise and language model artifacts. All spoken words and timecode information of the transcript that indicates at exactly what point in time in the video or audio the words were spoken, is stored. - Synchronizing the script data with the video content data to generate time-aligned script data may include processes similar to those described above with respect to
synchronization module 102. For example, synchronizing the script data with the video content data to generate time-aligned script data may include analyzing and synchronizing the structured (but untimed) information in a tagged script document (e.g., document data 110) and the text resulting from the STT transcription stored in metadata repository (e.g., transcript 114) to generate a time-aligned script data (e.g., time aligned script data 116). The time-aligned script data is provided to a named Entity Recognition system to categorize each regular or proper noun entity contained within the time-aligned script data. -
FIGS. 8A and 8B are block diagrams that illustrates components of and dataflow in a document time-alignment technique in accordance with one or more embodiments of the present technique. Note, the dashed lines indicate potential communication paths between various portions of the two block diagrams.System 800 may include features similar to that of previously describedsystem 100. - In some embodiments, script data is provided to
system 800. Script document/data 802 may be similar todocument 104. For example, movie script documents, closed caption data, and source transcripts are presented as inputs to thesystem 100. Movie scripts may be represented using a semi-structured Hollywood “Spec.” or dramatic screenplay format which provides descriptions of all scene, action, and dialogue events within a movie. - In some embodiments,
script data 802 may be provided to ascript converter 804.Script converter 804 may be similar todocument extractor 108. For example, script elements may be systematically extracted and imported into a standard structured (e.g., XML, ASTX, etc.).Script converter 804 may enable all script elements (e.g., Scenes, Shots, Action, Characters, Dialogue, Parentheticals, and Camera transitions) to be accessible as metadata to applications (e.g., Adobe Story, Adobe OnLocation, and Adobe Premiere Pro) enabling indexing, searching, and organization of video by textual content.Script converter 804 may enable scripts to be captured from a wide variety of sources including: professional screenwriters using word processing or script writing tools, from fan-transcribed scripts of film and television content, and from legacy script archives captured by OCR.Script converter 804 may employ various techniques for extracting and transcribing audio data, such as those described in described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein. - In some embodiments, converted script data 805 (e.g., an ASTX format movie script) from
script converter 804 may be provided to ascript parser 806. In some embodiments, parser may be implemented as a portion ofdocument extractor 108. Spec. scripts captured and converted into a standard (e.g., Adobe) script format may be parsed byscript parser 806 to identify and tag specific script elements such as scenes, actions, camera transitions, dialogue, and parenthetical. The ability to capture, analyze, and generate structured movie scripts may be used in certain time-alignment workflows (e.g., Adobe Pro “Script Align” feature where dialogue text within a movie script is automatically synchronized to the audio dialogue portion of video content). - In some embodiments, parsed script data is processed by a natural language (processing) engine (NLP) 808. In some embodiments, a
filter 808 a analyzes dialogue and action text from the parsed script data. For example, the input text is normalized and then broken into individual sentences for further processing. Each sentence may form a basic information unit for lines of the script, such as lines of dialogue in the script, or descriptive sentences that describe the setting of a scene or the action within a scene. - In some embodiments, grammatical units of each sentence are tagged at a part-of-speech (POS)
tagger 808 b. For example, a specialized (POS)tagger 808 b is then used to parse, identify, and tag the grammatical units of each sentence with its POS tag (e.g., noun, verb, article, etc.).POS tagger 808 b may use a transformational grammar rules technique to first induce and learn a set of lexical and contextual grammar rules from an annotated and tagged reference corpus, and then apply the learned runs for performing the POS tagging step of submitted script sentences. - In some embodiments, tagged verb and noun phrases are submitted to a Named Entity Recognition (NER)
system 808 c.NER system 808 c may then identify and classify entities and actions within each verb or noun phrase.NER 808 c may employ one or more external world-knowledge ontologies (API's) to perform the final entity tagging and classification. - In some embodiments, some or all extracted entities from
NER system 808 c are then represented using a script Entity-Relationship (E-R)data model 810 that includes Scripts, Movie Sets, Scenes, Actions, Transitions, Characters, Parentheticals, Dialogue, and/or Entities. The instantiatedmodel 810 may be physically stored into arelational database 812. In some embodiments, the instantiatedmodel 810 may be mapped into an RDF-Triplestore 814 (seeFIG. 8B ). In some embodiments, a specialized relational database schema may be provided for certain application (e.g., for Adobe Story). For example, script metadata may be used to record all script metadata and entities and the interrelationships between all entities. - In some embodiments, a relational database to
RDF mapping processor 816 may then used automatically processes the relational database schema representation of theE-R model 810 to transfer all script entities in relational database table rows into the RDF-Triplestore 814. Mapping may include RDF mapping system and process techniques, such as those described in described in U.S. patent application Ser. No. 12/507,746 entitled “CONVERSION OF RELATIONAL DATABASES INTO TRIPLESTORES”, filed Jul. 22, 2009, which is hereby incorporated by reference as though fully set forth herein. - In some embodiments,
E-R model 810 may be saved torelational database 812.Relational database 812 may implementE-R model 810 though a set of specially defined tables and primary key/foreign key referential integrity constraints between tables. - In some embodiments, an RDF-
Triplestore 820 may be used to store to the mappedrelational database 812 using output of relational database toRDF mapping processor 816. RDF-Triplestore 820 may represent the relational information as a directed acyclic graph and may enable both sub-graph and inference chain queries needed by movie or script query applications that retrieve script metadata. Use of RDF-Triplestore 820 may allow video scene entities to be queried using an RDF query language such as SPARQL or a logic programming language, like Prolog. Use of the RDF-Triplestore enables certain kinds of limited machine reasoning and inferences on the script entities (e.g., finding prop objects common to specific movie sets, classifying a scene entity using its IS_A generalization chain for a particular prop, or determining the usage and ownership rights to specific cartoon characters within a movie, for example. Script dialogue data may be stored within RDF-Triplestore 820. - In some embodiments, an
application server 822 may be used to process incoming job requests and then communicate RDF-Triplestore data back to one ormore client applications 824, such as Adobe Story.Application server 822 may contain a workflow engine along with one or more optional web-servers. Script analysis requests or queries for video and script metadata may be processed byserver 822, and then dispatched to a workflow engine which invokes either theNLP analysis engine 808 or a multimodalvideo query engine 826.Application server 822 may include a Triad/Metasky web server. - In some embodiments,
client application 824 may be used to implement further processing. For example, Adobe Story is a product that a client may use to leverage outputs of the workflows described herein to allow script writers to edit and collaborate on movie scripts, to extract, index, and to tag script entities such as people, places, and objects mentioned in the dialogue and action sections of a script. Adobe story may include a script editing service. - The above described steps may describe certain aspects of text processing. The following described steps may describe certain aspects of video and audio processing.
- In some embodiments, video/
audio content 830 is input and accepted by theworkflow system 800. Video/audio content 830 may be similar to that ofvideo content 106. Video/audio content 830 may provide video footage and corresponding dialogue sound tracks. The audio data may be analyzed and transcribed into text using an STT engine, such as those described herein. A resulting generated STT transcript (e.g., similar to transcript 114) may be aligned with convertedtextual movie scripts 805. In the event scripts are not available for metadata and time-alignment, the STT transcript may be processed by the natural language analysis and entity extraction components for keyword searching of the video. Natural language analysis and entity extraction components for keyword searching of the video may use multimodal video search techniques, such as those described in U.S. patent application Ser. No. 12/618,353 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, which is hereby incorporated by reference as though fully set forth herein. - In some embodiments, audio content is provided. For example, input audio dialogue tracks may be directly provided by television or movie studios, or extracted from the provided video files using standard known extraction methods. For use with certain application (e.g., Adobe STT CLM and STT multicore application), the extracted audio may be converted to a mono channel format that uses 16-bit samples with a 16 kHz frequency response.
- In some embodiments, operation of an
STT engine 832 is modified by use of a custom language model (CLM). For example,STT engine 832 may employ transcription based at least partially or completely on a provided CLM. The CLM may be provided/built using certain methods, such as those described herein. In some embodiment,STT engine 832 includes a multicore STT engine. The multicore STT engine may segment the source audio data, may provide STT transcriptions using parallel processing. In some embodiments, speech-to-text (STT) technology may implement a custom language model and/or an enhanced multicore STT transcription engine such as those described in U.S. patent application Ser. No. 12/332,297 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, and/or U.S. patent application Ser. No. 12/332,309 entitled “MULTI-CORE PROCESSING FOR PARALLEL SPEECH-TO-TEXT PROCESSING”, filed Dec. 10, 2008, which are both hereby incorporated by reference as though fully set forth herein. - In some embodiments, a metadata
time synchronization service 834 aligns elements oftranscript 832 with corresponding portions ofscript data 802 to generate time-aligned script data. Metadatatime synchronization service 834 may be similar tosynchronization module 102. For example, in some embodiments, metadatatime synchronization service 834 implements a specialized STT/Script alignment component to provide time alignment of non-timecoded words in the script with timecoded words in the STT transcript using a hybrid two-level alignment process, such as that described herein with regard tosynchronization module 102. For example, in level one processing, smaller regions or partitions of text and STT transcription keywords are accurately identified and prepared for detailed alignment. In level two processing, known alignment methods based on Viterbi or dynamic programming techniques for edit distance can be used to align the words within each partition. However, in some embodiments, a modified Viterbi method and hybrid phonetic/text comparator may be implemented, as described below. As a result, each script word may be assigned an accurate video timecode. This facilitates keyword search and time-indexing of the video by client applications such as the multimodalvideo search engine 826, or other applications. - In some embodiments, a modified Viterbi and/or phonetic/text comparator is implemented by metadata
time synchronization service 834. Further, the alignment process may also implement special override rules to resolve alignment option ties. As described herein, a decision as to whether or not an alignment is made may not rely only on precise text matches between the transcribed STT word and the script word, but rather, may rely on how closely words sound to each other; this may be provided for using a specialize phonetic encoding of the STT words and script words. Such a technique may be applicable to supplement a wide variety of STT alignment applications. - In some embodiments, data relating to the user is provided a graphical display that presents source script dialogue, the resulting time aligned words, and/or video content in association with one another. For example, a GUI/visualization element of an application (e.g., CS5 Premiere Pro Script Align feature) may enable a user to see source script dialogue words time-aligned with video action.
- In some embodiments, a user may search a video based on the corresponding words in the time-aligned script data. For example, a multimodal video search engine may allow a user to search for specific segments of video based on provided query keywords. The search feature may implement various techniques, such as those described in U.S. patent application Ser. No. 12/618,353 entitled “ACCESSING MEDIA DATA USING METADATA REPOSITORY”, filed Nov. 13, 2009, which is hereby incorporated by reference as though fully set forth herein.
- In some embodiments, locations for the insertion of video descriptions can be located, video description content can be extracted from the script and automatically inserted into a time aligned script and/or audio track using time aligned script data (e.g., time aligned
script data 116 as described with respect toFIGS. 1 and 2 ) provided bysystem 100. Video descriptions may include an audio track in a movie or television program containing descriptions of the setting and action. Video description narrations fill in the story gaps by describing visual elements and provide a more complete description of what's happening in the program. This may be of particular value to the blind or visually impaired by helping to describe visual elements that they cannot view. The video description may be inserted into the natural pauses in dialogue or between critical sound elements, or the video and audio may be modified to enable insertion of video descriptions that may other wise be too long for the natural pauses. - Video description content may be generated by extracting descriptive information and narrative content from a script written for the project, syncing and editing it to the video program for playback. Video description content may be extracted directly from descriptive text embedded in the script. For example, location settings, actor movements, non-verbal events, etc. that may be provided in script elements (e.g., title, author name(s), scene headings, action elements, character names, parentheticals, transitions, shot elements, dialogue/narrations, and the like) may be extracted as the video description content, aligned to the correct portion of scenes (e.g., to pauses in dialogue) using time alignment data, and the video description content may be manually or automatically edited (if needed) to fit into the spaces available between dialogue segments.
- In some embodiments, the time aligned data acquired using
system 100 may be used to identify the location of pauses within the audio content for embedding narrative content (e.g., action elements). The locations of the pauses in the audio content may be provided to a user as locations for inserting video description content. Thus, a user may be able to quickly identify the location of pauses for adding video description content. In some embodiments, narrative content (e.g., action element descriptions embedded in the script) may be automatically inserted into corresponding pauses within the dialogue of the audio track to provide the corresponding video description content. The resulting video description content may be reviewable and editable by a user. A text version of the video description content can be used as a blueprint for recording by a human voiceover talent. Thus, a voicer may simply have to read the corresponding narration content as opposed to having to manually search through a program, manually identify breaks in the dialog, and derive/record narrations to describe the video. In some embodiments, the video description track can be created automatically using synthesized speech to read the video description content (e.g., without necessarily requiring any or at least a significant amount of human labor). - As noted above, a script may include a variety of script elements such as a scene heading, action, character, parenthetical, dialogue, transition, or other text that cannot be classified. Any or all of these and other script elements can be used to generate useful information for a video description track. A scene heading (also referred to as a “slugline”) includes a description of where the scene physically occurs. For example, a scene heading may indicate that the scene takes place indoors (e.g., INT.) or outdoors (e.g., EXT.), or possibly both indoors and outdoors (e.g., INT./EXT.) Typically, a location name follows the description of where the scene physically occurs. For example, “INT./EXT.” may be immediately followed by a more detailed description of where the scene occurs. (e.g., INT. KITCHEN, INT. LIVING ROOM, EXT. BASEBALL STADIUM, INT. AIRPLANE, etc.). The scene heading may also include the time of day (e.g., NIGHT, DAY, DAWN, EVENING, etc.). This information embedded in the script helps to “set the scene.” The scene type is typically designated as internal (INT.) or external (EXT.), and includes a period following the INT or EXT designation. A hyphen is typically used between other elements of the scene heading. For example, a complete scene heading may read, “INT. FERRY TERMINAL BAR—DAY” or “EXT. MAROON MOVIE STUDIO—DAY”.
- An action element (also referred to as a description element) typically describes the setting of the scene and introduces the characters in a scene. Action elements may also describe what will actually happen during the scene.
- A character name element may include an actual name (e.g., MS. SUTTER), description (e.g., BIG MAN) or occupation (e.g., BARTENDER) of a character. Sequence numbers are typically used to differentiate similar characters (e.g.,
COP # 1 and COP #2). A character name is almost always inserted prior to a character speaking (e.g., just before dialog element), to indicate that the character's dialogue follows. - A dialog element indicates what a character says when anyone on screen or off screen speaks. This may include conversation between characters, when a character speaks out loud to themselves, or when a character is off-screen and only their voice is heard (e.g., in a narration). Dialog elements may also include voice-overs or narration when the speaker is on screen but is not actively speaking on screen.
- A parenthetical typically includes a remark that indicates an attitude in dialog delivery, and/or specifies a verbal direction or action direction for the actor who is speaking the part of a character. Parentheticals are typically short, concise and descriptive statements located under the characters name.
- A transition typically includes a notation indicating an editing transition within the telling of a story. For example, “DISSOLVE TO:” means the action seems to blur and refocus into another scene, as generally used to denote a passage of time. Transitions almost always follow an action element and precede a scene heading. Common transitions include: “DISSOLVE TO:”, “CUT TO:”, “SMASH CUT:”, “QUICK CUT:”, “FADE IN:”, “FADE OUT:”, and “FADE TO:”.
- A shot element typically indicates what the camera sees. For example, a shot element that recites “TRACKING SHOT” generally indicates the camera should follow a character as he walks in a scene. “WIDE SHOT” generally indicates that every character appears in the scene. A SHOT tells the reader the focal point within a scene has changed. Example of shot elements include: “ANGLE ON . . . ”, “PAN TO . . . ”, “EXTREME CLOSE UP . . . ”, “FRANKIE′S POV . . . ”, and “REVERSE ANGLE . . . ”.
- In some embodiments, script elements may be identified and extracted as described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein. Moreover, the script elements may be time aligned to provide time-aligned
data 116 as described herein. The time aligned data may include dialogue as well as other script elements having corresponding timecodes that identify when each of the respective words/elements occur within the video/audio corresponding to the script. -
FIG. 9A illustrates anexemplary script document 900 in accordance with one or more embodiments of the present technique.Script document 900 depicts an exemplary layout of the above described script elements. For example,script document 900 includes atransition element 902, ascene heading element 904,action elements character name elements 908,dialog elements 910,parenthetical elements 912, and shotelement 914. - Script writers and describers often have closely aligned goals to describe onscreen actions succinctly, vividly and imaginatively. Often the action element text may be the most useful for creating video description content, as action elements typically provide the descriptions that clearly describe what has happened, is happening, or about to happen in a scene. Typically, long text passages in a script describing major changes in the setting or complex action sequences translate to longer spaces between dialogue in the recorded program (often filled with music and sound effects) and provide opportunities for including longer segments of video description content. For example, in the
script 900 ofFIG. 9A , the action described under the scene heading 904 andaction element 906 a is a wide establishing shot that follows the character out onto a busy studio lot. Since it describes a change of scene and establishes the new setting, there is a lot of descriptive text. The director filmed this shot on a crane, which swooped down from a high angle and followed the character through his action in this shot. Since there is a lot of information for the audience to take in during this lengthy transition shot, it begins without dialogue and continues for nearly half a minute. This gap in the dialogue provides a gap in which some or all of the descriptive action element text can be inserted. - Although some elements may be more useful than others, some or all of the script elements may be used to generate video description content. In some embodiments, a user may have control over which script elements to use in creating a video description. For example, a user may select to use only action elements and shot elements and to ignore other elements of the script. In some embodiments, the selection may be done before or after the video description is generated. For example, a user may allow the system to generate a video description using all or some of the script elements, and may subsequently pick-and-choose which elements to keep after the initial video description is generated.
-
FIG. 9B illustrates an exemplary portion of avideo description script 920 that corresponds to the portion ofscript 900 ofFIG. 9A .Video description script 920 includes avideo description track 922 broken into discrete segments (1-9) provided relative to gaps and dialogue of an audio track (e.g., main audio program recorded dialogue) 924 that corresponds to spoken words of dialogue content ofscript 920. In the illustrated embodiment, the content ofvideo description track 922 corresponds to action element text ofaction elements script 900 ofFIG. 9A . Each corresponding pause/gap in dialogue ofaudio track 922 is identified with a time of duration (e.g., “00:00:28:00 Gap” indicating a gap of twenty-eight seconds prior to the beginning of the script dialogue of segment 2). The corresponding content ofvideo description 922 is provided adjacent the gap/pause, and is identified with a time of duration for the video description content (e.g., “00:00:27:00” indicating twenty-seven seconds for the video description content to be spoken) where applicable. In some embodiments, the content ofvideo description 922 may be modified to fit within the corresponding gap. For example, in the illustrated embodiment, a portion of the first segment of video description content is removed to enable the resulting video description content to fit within the duration of the gap when spoken. In some embodiments, the entire video description content may be deleted or ignored where there is not a gap of sufficient length for the video description content. For example, the video description content ofsegment 3 was deleted/ignored as the corresponding pause in dialogue was only about twelve frames (or ½ a second) in duration—too short for the insertion of the corresponding video description content.Video description script 920 andvideo description content 922 can be used as a blueprint for recording by a human voiceover talent. Thus, a voicer may simply have to read the corresponding narration content as opposed to having to manually search through a program, manually identify breaks in the dialog, and derive/record narrations to describe the video. In some embodiments, the video description track can be created automatically using synthesized speech to read the video description content 922 (e.g., without necessarily requiring any or at least a significant amount of human labor). -
FIG. 9C is a flowchart that illustrates amethod 950 of generating a video description in accordance with one or more embodiments of the present technique.Method 950 may provide video description techniques using components and dataflow implemented atsystem 100.Method 950 generally includes identifying script elements, time aligning the script, identifying gaps/pauses in dialogue, aligning video description content to the gaps/pauses, generating a script with video description content, and generating a video description. -
Method 950 may include identifying script elements, as depicted atblock 952. Identifying script elements may include identifying some or all of the script elements contained within a script from which a video description is to be generated. For example, a script may be analyzed to provide script metadata that identifies a variety of script elements, such as scene headings, actions, characters, parentheticals, dialogue, transitions, or other text that cannot be classified. In some embodiments, script elements may be identified and extracted as described in U.S. patent application Ser. No. 12/713,008 entitled “METHOD AND APPARATUS FOR CAPTURING, ANALYZING, AND CONVERTING SCRIPTS”, filed Feb. 25, 2010, which is hereby incorporated by reference as though fully set forth herein. In some embodiments, the identification of the elements may not actually be performed but may simply be provided or retrieved for analysis. -
Method 950 may also include time aligning the script, as depicted atblock 954. Time aligning the script may include using techniques, such as those described herein with regard tosystem 100, to provide a timecode for some or all elements of the corresponding script. For example, a script may be processed to provide a timecode for some or all of the words within the script, including dialogue or other script elements. In some embodiments, the timecode information may provide stop and start time for various elements, including dialogue, which enables the identification of pauses between spoken words of dialogue. In some embodiments, the time alignment may not actually be performed but may simply be provided. For example, a system generating a video description may be provided or retrieve time alignedscript data 116. -
Method 950 may also include identifying gaps/pauses in dialogue, as depicted at block 956. In some embodiments, identifying gaps/pauses in dialogue may include assessing timecode information for each word of spoken dialogue to identify the beginning and end of spoken lines of dialogue, as well as any pauses in the spoken lines of dialogue that may provide gaps for the insertion of video description content. For example, invideo description script 920 ofFIG. 9B , a pause of twenty-eight seconds was identified atsegment 1, prior to the start of recorded dialogue ofsegment 2, a pause of 0.12 seconds was identified atsegment 3, and a pause of 4.06 seconds was identified atsegment 7. In some embodiments, a gap threshold may be used to identify what pauses are of sufficient length to constitute a gap that may be of sufficient length to be used for inserting video description content. For example, a gap threshold of three seconds may be set, thereby ignoring all pauses of less than three seconds and identifying only pauses equal to or greater than three-seconds as gaps of sufficient length to be used for inserting video description content. Such a technique may be useful to ignore normal pauses in speech (e.g., between spoken words) or short breaks between characters lines of dialogue that may be so short that it would be difficult to provide any substantive video description within the pause. In some embodiments, the gap threshold value may be user selectable. As depicted inFIG. 9B , the user may be provided with an indication that a gap is too short where there is a corresponding script element. For example,segment 3 of recordeddialogue 924 includes an inserted statement of “No gap available”, and the corresponding action text was deleted/ignored (as indicated by the strikethrough). Moreover, where there is no video description content (e.g., script elements) corresponding to a gap, the gap may be detected, but may be ignored. In some embodiments, the user may be alerted to the gap, thereby enabling them to readily identify gaps that could be used for the insertion of additional video description content. In some embodiments, video descriptions may be inserted into any available gaps, even out of sequence with their corresponding location in the script, according to rules or preferences provided by the user. For example, insegment 3, there was no available gap for the video description that would normally be inserted at that point according to the script. However, if there were another available gap within a prescribed number of seconds before or after that segment (e.g., segment 3), the video description could be inserted at that other location nearby within the prescribed number of seconds before or after that segment (e.g., segment 3). -
Method 950 may also include aligning video description content to gaps/pauses, as depicted at block 958. Aligning the video description content may include aligning the script elements with dialogue relative to where they occur within the script. InFIG. 9B , each of theaction elements FIG. 9B the script action elements have been aligned to the recorded dialog and the action element text from the script has been aligned with the available gaps when possible. Two gaps were identified atsegments - In some embodiments, a user may have control over the resulting video description. For example, a user may modify a video description at their choosing, or may be provided an opportunity to select how to truncate a video description that does not fit within a gap. For example, in the illustrated embodiment of
FIG. 9B , a user may select to remove the text of segment 1 (as indicated by the strikethrough) in an effort to make the video description fit within the corresponding gap. In some embodiments, video description content may be automatically modified to fit within a given gap. If a gap is too short to fit the corresponding video description content, the video description content may be automatically truncated using rules of grammar. For example, the last word(s) or entire last sentence(s) may be incrementally truncated/removed until the remaining video content description is short enough to fit within the gap. In the illustrated embodiment ofFIG. 9B , the last sentence “Maroon is leading an entourage of ASSISTANTS trying to keep up” may have been automatically removed, relieving the user of the need to manually modify the content. Of course, even in the event of automatic modification of the video description content, the user may have the opportunity to approve or modify the changes. In some embodiments, as the video description content is edited, the duration may be updated dynamically to indicate to the user whether the revised description will fit within an available gap. - In some embodiments, a gap in the recorded program may be created or the duration of a gap may be modified to provide for the insertion of video description content. For example, at
segment 3, the gap in the recorded audio may be increased (e.g., by inserting an additional amount of pause in the audio track between the end ofsegment 2 and the beginning of segment 4) to five seconds to enable the action element text to be fit within the resulting gap. Such a technique may be automatically applied at some or all instances where a gap is too short in duration to fit the corresponding video description content. Although such modifications of the dialogue may introduce delays or pauses within the corresponding video and, thus, may modify the video and dialogue of a traditional program, it may be particularly helpful in the context of audio-only programs. For example, for books-on-tape or similar audio tracks produced for the blind or visually impaired. - In some embodiments, video description content may be allowed to overlap certain portions of the audio track. For example, a user may have the option of modifying the video description content to overlap seemingly less important portions of the dialogue, music, sound effects, or the like. In some embodiments, the main audio recorded dialogue, music, sound effects, or the like may be dipped (e.g., reduced) in volume so that the video description may be heard more clearly. For example, the volume of music may be lowered while the video description content is being recited.
-
Method 950 may also include generating a script with video description content, as depicted atblock 960. Generating a script with video content may include generating a script document that includes video description content; script/recorded dialogue, and/or other script elements aligned with respect to one another.FIG. 9B illustrates an exemplaryvideo description script 920 that includesvideo description content 922 and recordeddialogue 924. In the illustrated embodiment, the modifications to the video description content are displayed. In some embodiments, a “clean” version of the video description script may be provided. For example, clean video description script may incorporate some or all of the modifications that are not visible. A text version of the video description content can be used as a blueprint for recording by a human voiceover talent. Thus, a voicer may simply have to read the corresponding narration content as opposed to having to manually search through a program, manually identify breaks in the dialog, compose appropriate video descriptions of correct lengths, and/or derive/record narrations to describe the program. -
Method 950 may also include generating a video description, as depicted atblock 962. Generating the video description may include recording a reading of the video description content. For example, a reading by a voicer and/or a synthesized reading of the video description content may be recorded to generate a video description track. In some embodiments, the video description track may be merged with the original audio of the program to generate a program containing both the original audio and the video description audio. - Various components of embodiments of a document time-alignment technique as described herein may be executed on one or more computer systems, which may interact with various other devices. One such computer system is illustrated by
FIG. 10 . In the illustrated embodiment,computer system 1000 includes one or more processors 1010 coupled to asystem memory 1020 via an input/output (I/O)interface 1030.Computer system 1000 further includes anetwork interface 1040 coupled to I/O interface 1030, and one or more input/output devices 1050, such ascursor control device 1060,keyboard 1070, audio device 1090, and display(s) 1080. In some embodiments, it is contemplated that embodiments may be implemented using a single instance ofcomputer system 1000, while in other embodiments multiple such systems, or multiple nodes making upcomputer system 1000, may be configured to host different portions or instances of embodiments. For example, in one embodiment some elements may be implemented via one or more nodes ofcomputer system 1000 that are distinct from those nodes implementing other elements. - In various embodiments,
computer system 1000 may be a uniprocessor system including one processor 1010, or a multiprocessor system including several processors 1010 (e.g., two, four, eight, or another suitable number). Processors 1010 may be any suitable processor capable of executing instructions. For example, in various embodiments, processors 1010 may be general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, PowerPC, SPARC, or MIPS ISAs, or any other suitable ISA. In multiprocessor systems, each of processors 1010 may commonly, but not necessarily, implement the same ISA. - In some embodiments, at least one processor 1010 may be a graphics processing unit. A graphics processing unit or GPU may be considered a dedicated graphics-rendering device for a personal computer, workstation, game console or other computer system. Modern GPUs may be very efficient at manipulating and displaying computer graphics and their highly parallel structure may make them more effective than typical CPUs for a range of complex graphical algorithms. For example, a graphics processor may implement a number of graphics primitive operations in a way that makes executing them much faster than drawing directly to the screen with a host central processing unit (CPU). In various embodiments, the methods disclosed herein for layout-preserved text generation may be implemented by program instructions configured for execution on one of, or parallel execution on two or more of, such GPUs. The GPU(s) may implement one or more application programmer interfaces (APIs) that permit programmers to invoke the functionality of the GPU(s). Suitable GPUs may be commercially available from vendors such as NVIDIA Corporation having headquarters in Santa Clara, Calif., ATI Technologies of AMD having headquarters in Sunnyvale, Calif., and others.
-
System memory 1020 may be configured to store program instructions and/or data accessible by processor 1010.System memory 1020 may include tangible a non-transitory storage medium for storing program instructions and other data thereon. In various embodiments,system memory 1020 may be implemented using any suitable memory technology, such as static random access memory (SRAM), synchronous dynamic RAM (SDRAM), nonvolatile/Flash-type memory, or any other type of memory. In the illustrated embodiment, program instructions and data implementing desired functions, such as those described above for time-alignment methods, are shown stored withinsystem memory 1020 asprogram instructions 1025 anddata storage 1035, respectively. In other embodiments, program instructions and/or data may be received, sent or stored upon different types of computer-accessible media or on similar media separate fromsystem memory 1020 orcomputer system 1000. Generally speaking, a computer-accessible medium may include storage media or memory media such as magnetic or optical media, e.g., disk or CD/DVD-ROM coupled tocomputer system 1000 via I/O interface 1030. Program instructions and data stored via a computer-accessible medium may be transmitted by transmission media or signals such as electrical, electromagnetic, or digital signals, which may be conveyed via a communication medium such as a network and/or a wireless link, such as may be implemented vianetwork interface 1040. - In one embodiment, I/
O interface 1030 may be configured to coordinate I/O traffic between processor 1010,system memory 1020, and any peripheral devices in the device, includingnetwork interface 1040 or other peripheral interfaces, such as input/output devices 1050. In some embodiments, I/O interface 1030 may perform any necessary protocol, timing or other data transformations to convert data signals from one component (e.g., system memory 1020) into a format suitable for use by another component (e.g., processor 1010). In some embodiments, I/O interface 1030 may include support for devices attached through various types of peripheral buses, such as a variant of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard, for example. In some embodiments, the function of I/O interface 1030 may be split into two or more separate components. In addition, in some embodiments some or all of the functionality of I/O interface 1030, such as an interface tosystem memory 1020, may be incorporated directly into processor 1010. -
Network interface 1040 may be configured to allow data to be exchanged betweencomputer system 1000 and other devices attached to a network, such as other computer systems, or between nodes ofcomputer system 1000. In various embodiments,network interface 1040 may support communication via wired or wireless general data networks, such as any suitable type of Ethernet network, for example; via telecommunications/telephony networks such as analog voice networks or digital fiber communications networks; via storage area networks such as Fibre Channel SANs, or via any other suitable type of network and/or protocol. - Input/
output devices 1050 may, in some embodiments, include one or more display terminals, keyboards, keypads, touchpads, scanning devices, voice or optical recognition devices, or any other devices suitable for entering or retrieving data by one ormore computer system 1000. Multiple input/output devices 1050 may be present incomputer system 1000 or may be distributed on various nodes ofcomputer system 1000. In some embodiments, similar input/output devices may be separate fromcomputer system 1000 and may interact with one or more nodes ofcomputer system 1000 through a wired or wireless connection, such as overnetwork interface 1040. - As shown in
FIG. 10 ,memory 1020 may includeprogram instructions 1025, configured to implement embodiments of a layout-preserved text generation method as described herein, anddata storage 1035, comprising various data accessible byprogram instructions 1025. In one embodiment,program instructions 1025 may include software elements of a layout-preserved text generation method illustrated in the above Figures.Data storage 1035 may include data that may be used in embodiments, for example input PDF documents or output layout-preserved text documents. In other embodiments, other or different software elements and/or data may be included. - Those skilled in the art will appreciate that
computer system 1000 is merely illustrative and is not intended to limit the scope of a layout-preserved text generation method as described herein. In particular, the computer system and devices may include any combination of hardware or software that can perform the indicated functions, including computers, network devices, internet appliances, PDAs, wireless phones, pagers, etc.Computer system 1000 may also be connected to other devices that are not illustrated, or instead may operate as a stand-alone system. In addition, the functionality provided by the illustrated components may in some embodiments be combined in fewer components or distributed in additional components. Similarly, in some embodiments, the functionality of some of the illustrated components may not be provided and/or other additional functionality may be available. - Those skilled in the art will also appreciate that, while various items are illustrated as being stored in memory or on storage while being used, these items or portions of them may be transferred between memory and other storage devices for purposes of memory management and data integrity. Alternatively, in other embodiments some or all of the software components may execute in memory on another device and communicate with the illustrated computer system via inter-computer communication. Some or all of the system components or data structures may also be stored (e.g., as instructions or structured data) on a computer-accessible medium or a portable article to be read by an appropriate drive, various examples of which are described above. In some embodiments, instructions stored on a computer-accessible medium separate from
computer system 1000 may be transmitted tocomputer system 1000 via transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as a network and/or a wireless link. Various embodiments may further include receiving, sending or storing instructions and/or data implemented in accordance with the foregoing description upon a computer-accessible medium. Accordingly, the present invention may be practiced with other computer system configurations. In some embodiments, portions of the techniques described herein (e.g., preprocessing of script and metadata may be hosted in a cloud computing infrastructure. - Various embodiments may further include receiving, sending or storing instructions and/or data implemented in accordance with the foregoing description upon a computer-accessible medium. Generally speaking, a computer-accessible storage medium may include a non-transitory storage media or memory media such as magnetic or optical media, e.g., disk or DVD/CD-ROM, volatile or non-volatile media such as RAM (e.g. SDRAM, DDR, RDRAM, SRAM, etc.), ROM, etc., as well as transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as network and/or a wireless link.
- Some portions of the detailed description provided herein are presented in terms of algorithms or symbolic representations of operations on binary digital signals stored within a memory of a specific apparatus or special purpose computing device or platform. In the context of this particular specification, the term specific apparatus or the like includes a general purpose computer once it is programmed to perform particular functions pursuant to instructions from program software. Algorithmic descriptions or symbolic representations are examples of techniques used by those of ordinary skill in the signal processing or related arts to convey the substance of their work to others skilled in the art. An algorithm is here, and is generally, considered to be a self-consistent sequence of operations or similar signal processing leading to a desired result. In this context, operations or processing involve physical manipulation of physical quantities. Typically, although not necessarily, such quantities may take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared or otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to such signals as bits, data, values, elements, symbols, characters, terms, numbers, numerals or the like. It should be understood, however, that all of these or similar terms are to be associated with appropriate physical quantities and are merely convenient labels. Unless specifically stated otherwise, as apparent from the discussion, it is appreciated that throughout this specification discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining” or the like refer to actions or processes of a specific apparatus, such as a special purpose computer or a similar special purpose electronic computing device. In the context of this specification, therefore, a special purpose computer or a similar special purpose electronic computing device is capable of manipulating or transforming signals, typically represented as physical electronic or magnetic quantities within memories, registers, or other information storage devices, transmission devices, or display devices of the special purpose computer or similar special purpose electronic computing device.
- Various methods as illustrated in the Figures and described herein represent examples of embodiments of methods. The methods may be implemented in software, hardware, or a combination thereof. The order of method may be changed, and various elements may be added, reordered, combined, omitted, modified, etc.
- Various modifications and changes may be to the above technique made as would be obvious to a person skilled in the art having the benefit of this disclosure. For example, although several embodiments are discussed with regard to dialogue/narrative elements of script documents, the techniques described herein may be applied to assess and determine data relating other elements of a script document. It is intended that the invention embrace all such modifications and changes and, accordingly, the above description to be regarded in an illustrative rather than a restrictive sense.
- Adobe and Adobe PDF are either registered trademarks or trademarks of Adobe Systems Incorporated in the United States and other countries.
Claims (21)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/789,708 US8447604B1 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for processing scripts and related data |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US32312110P | 2010-04-12 | 2010-04-12 | |
US12/789,708 US8447604B1 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for processing scripts and related data |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130124202A1 true US20130124202A1 (en) | 2013-05-16 |
US8447604B1 US8447604B1 (en) | 2013-05-21 |
Family
ID=48280305
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/789,785 Active 2031-04-04 US8825489B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for interpolating script data |
US12/789,791 Active 2031-01-31 US9066049B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for processing scripts |
US12/789,720 Active 2031-02-08 US8825488B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for time synchronized script metadata |
US12/789,708 Active 2031-03-17 US8447604B1 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for processing scripts and related data |
US12/789,760 Active 2031-10-25 US9191639B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for generating video descriptions |
US12/789,749 Abandoned US20130124984A1 (en) | 2010-04-12 | 2010-05-28 | Method and Apparatus for Providing Script Data |
Family Applications Before (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/789,785 Active 2031-04-04 US8825489B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for interpolating script data |
US12/789,791 Active 2031-01-31 US9066049B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for processing scripts |
US12/789,720 Active 2031-02-08 US8825488B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for time synchronized script metadata |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/789,760 Active 2031-10-25 US9191639B2 (en) | 2010-04-12 | 2010-05-28 | Method and apparatus for generating video descriptions |
US12/789,749 Abandoned US20130124984A1 (en) | 2010-04-12 | 2010-05-28 | Method and Apparatus for Providing Script Data |
Country Status (1)
Country | Link |
---|---|
US (6) | US8825489B2 (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100263005A1 (en) * | 2009-04-08 | 2010-10-14 | Eric Foster White | Method and system for egnaging interactive web content |
US20120143606A1 (en) * | 2010-12-01 | 2012-06-07 | At&T Intellectual Property I, L.P. | Method and system for testing closed caption content of video assets |
US20130080174A1 (en) * | 2011-09-22 | 2013-03-28 | Kabushiki Kaisha Toshiba | Retrieving device, retrieving method, and computer program product |
US20140013268A1 (en) * | 2012-07-09 | 2014-01-09 | Mobitude, LLC, a Delaware LLC | Method for creating a scripted exchange |
US20140122079A1 (en) * | 2012-10-25 | 2014-05-01 | Ivona Software Sp. Z.O.O. | Generating personalized audio programs from text content |
US20140180697A1 (en) * | 2012-12-20 | 2014-06-26 | Amazon Technologies, Inc. | Identification of utterance subjects |
US8825489B2 (en) | 2010-04-12 | 2014-09-02 | Adobe Systems Incorporated | Method and apparatus for interpolating script data |
US9002703B1 (en) * | 2011-09-28 | 2015-04-07 | Amazon Technologies, Inc. | Community audio narration generation |
EP4283490A1 (en) * | 2022-05-27 | 2023-11-29 | Tata Consultancy Services Limited | Systems and methods for rules-based mapping of answer scripts to markers |
Families Citing this family (249)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9141860B2 (en) | 2008-11-17 | 2015-09-22 | Liveclips Llc | Method and system for segmenting and transmitting on-demand live-action video in real-time |
US20130138637A1 (en) * | 2009-09-21 | 2013-05-30 | Walter Bachtiger | Systems and methods for ranking media files |
US20130311181A1 (en) * | 2009-09-21 | 2013-11-21 | Walter Bachtiger | Systems and methods for identifying concepts and keywords from spoken words in text, audio, and video content |
US20120029918A1 (en) * | 2009-09-21 | 2012-02-02 | Walter Bachtiger | Systems and methods for recording, searching, and sharing spoken content in media files |
US8355903B1 (en) | 2010-05-13 | 2013-01-15 | Northwestern University | System and method for using data and angles to automatically generate a narrative story |
US11989659B2 (en) | 2010-05-13 | 2024-05-21 | Salesforce, Inc. | Method and apparatus for triggering the automatic generation of narratives |
US9208147B1 (en) | 2011-01-07 | 2015-12-08 | Narrative Science Inc. | Method and apparatus for triggering the automatic generation of narratives |
WO2012006024A2 (en) * | 2010-06-28 | 2012-01-12 | Randall Lee Threewits | Interactive environment for performing arts scripts |
US20120092232A1 (en) * | 2010-10-14 | 2012-04-19 | Zebra Imaging, Inc. | Sending Video Data to Multiple Light Modulators |
US10095367B1 (en) * | 2010-10-15 | 2018-10-09 | Tivo Solutions Inc. | Time-based metadata management system for digital media |
US9443324B2 (en) * | 2010-12-22 | 2016-09-13 | Tata Consultancy Services Limited | Method and system for construction and rendering of annotations associated with an electronic image |
US9697197B1 (en) | 2011-01-07 | 2017-07-04 | Narrative Science Inc. | Automatic generation of narratives from data using communication goals and narrative analytics |
US10185477B1 (en) | 2013-03-15 | 2019-01-22 | Narrative Science Inc. | Method and system for configuring automatic generation of narratives from data |
US10657201B1 (en) | 2011-01-07 | 2020-05-19 | Narrative Science Inc. | Configurable and portable system for generating narratives |
US9720899B1 (en) | 2011-01-07 | 2017-08-01 | Narrative Science, Inc. | Automatic generation of narratives from data using communication goals and narrative analytics |
US9576009B1 (en) | 2011-01-07 | 2017-02-21 | Narrative Science Inc. | Automatic generation of narratives from data using communication goals and narrative analytics |
US9697178B1 (en) | 2011-01-07 | 2017-07-04 | Narrative Science Inc. | Use of tools and abstraction in a configurable and portable system for generating narratives |
GB2500356A (en) | 2011-01-20 | 2013-09-18 | Box Inc | Real time notification of activities that occur in a web-based collaboration environment |
US20120303643A1 (en) * | 2011-05-26 | 2012-11-29 | Raymond Lau | Alignment of Metadata |
RU2564681C2 (en) * | 2011-07-01 | 2015-10-10 | Долби Лабораторис Лайсэнзин Корпорейшн | Methods and systems of synchronisation and changeover for adaptive sound system |
EP2729877A4 (en) | 2011-07-08 | 2015-06-17 | Box Inc | Desktop application for access and interaction with workspaces in a cloud-based content management system and synchronization mechanisms thereof |
US20130035936A1 (en) * | 2011-08-02 | 2013-02-07 | Nexidia Inc. | Language transcription |
US9704111B1 (en) | 2011-09-27 | 2017-07-11 | 3Play Media, Inc. | Electronic transcription job market |
US9098474B2 (en) | 2011-10-26 | 2015-08-04 | Box, Inc. | Preview pre-generation based on heuristics and algorithmic prediction/assessment of predicted user behavior for enhancement of user experience |
US11210610B2 (en) | 2011-10-26 | 2021-12-28 | Box, Inc. | Enhanced multimedia content preview rendering in a cloud content management system |
US20130117012A1 (en) * | 2011-11-03 | 2013-05-09 | Microsoft Corporation | Knowledge based parsing |
US9069850B2 (en) * | 2011-11-08 | 2015-06-30 | Comcast Cable Communications, Llc | Content descriptor |
US9270718B2 (en) * | 2011-11-25 | 2016-02-23 | Harry E Emerson, III | Internet streaming and the presentation of dynamic content |
US9773051B2 (en) | 2011-11-29 | 2017-09-26 | Box, Inc. | Mobile platform file and folder selection functionalities for offline access and synchronization |
US9904435B2 (en) | 2012-01-06 | 2018-02-27 | Box, Inc. | System and method for actionable event generation for task delegation and management via a discussion forum in a web-based collaboration environment |
US9172983B2 (en) * | 2012-01-20 | 2015-10-27 | Gorilla Technology Inc. | Automatic media editing apparatus, editing method, broadcasting method and system for broadcasting the same |
US11232481B2 (en) | 2012-01-30 | 2022-01-25 | Box, Inc. | Extended applications of multimedia content previews in the cloud-based content management system |
US9471551B1 (en) * | 2012-02-03 | 2016-10-18 | Google Inc. | Promoting content |
US9378191B1 (en) | 2012-02-03 | 2016-06-28 | Google Inc. | Promoting content |
US9304985B1 (en) | 2012-02-03 | 2016-04-05 | Google Inc. | Promoting content |
CN103297389B (en) * | 2012-02-24 | 2018-09-07 | 腾讯科技(深圳)有限公司 | Interactive method and device |
US9965745B2 (en) | 2012-02-24 | 2018-05-08 | Box, Inc. | System and method for promoting enterprise adoption of a web-based collaboration environment |
US8918311B1 (en) * | 2012-03-21 | 2014-12-23 | 3Play Media, Inc. | Intelligent caption systems and methods |
JP6045175B2 (en) * | 2012-04-05 | 2016-12-14 | 任天堂株式会社 | Information processing program, information processing apparatus, information processing method, and information processing system |
US9575981B2 (en) | 2012-04-11 | 2017-02-21 | Box, Inc. | Cloud service enabled to handle a set of files depicted to a user as a single file in a native operating system |
US20130283143A1 (en) | 2012-04-24 | 2013-10-24 | Eric David Petajan | System for Annotating Media Content for Automatic Content Understanding |
US9367745B2 (en) | 2012-04-24 | 2016-06-14 | Liveclips Llc | System for annotating media content for automatic content understanding |
US9413587B2 (en) | 2012-05-02 | 2016-08-09 | Box, Inc. | System and method for a third-party application to access content within a cloud-based platform |
US9275636B2 (en) | 2012-05-03 | 2016-03-01 | International Business Machines Corporation | Automatic accuracy estimation for audio transcriptions |
US9396216B2 (en) | 2012-05-04 | 2016-07-19 | Box, Inc. | Repository redundancy implementation of a system which incrementally updates clients with events that occurred via a cloud-enabled platform |
US20130308922A1 (en) * | 2012-05-15 | 2013-11-21 | Microsoft Corporation | Enhanced video discovery and productivity through accessibility |
US9691051B2 (en) | 2012-05-21 | 2017-06-27 | Box, Inc. | Security enhancement through application access control |
US8914900B2 (en) | 2012-05-23 | 2014-12-16 | Box, Inc. | Methods, architectures and security mechanisms for a third-party application to access content in a cloud-based platform |
US9712510B2 (en) | 2012-07-06 | 2017-07-18 | Box, Inc. | Systems and methods for securely submitting comments among users via external messaging applications in a cloud-based platform |
GB2505072A (en) | 2012-07-06 | 2014-02-19 | Box Inc | Identifying users and collaborators as search results in a cloud-based system |
US9794256B2 (en) | 2012-07-30 | 2017-10-17 | Box, Inc. | System and method for advanced control tools for administrators in a cloud-based service |
US10109278B2 (en) * | 2012-08-02 | 2018-10-23 | Audible, Inc. | Aligning body matter across content formats |
US9558202B2 (en) | 2012-08-27 | 2017-01-31 | Box, Inc. | Server side techniques for reducing database workload in implementing selective subfolder synchronization in a cloud-based environment |
US20140067394A1 (en) * | 2012-08-28 | 2014-03-06 | King Abdulaziz City For Science And Technology | System and method for decoding speech |
US9135462B2 (en) | 2012-08-29 | 2015-09-15 | Box, Inc. | Upload and download streaming encryption to/from a cloud-based platform |
US9195519B2 (en) | 2012-09-06 | 2015-11-24 | Box, Inc. | Disabling the self-referential appearance of a mobile application in an intent via a background registration |
US9117087B2 (en) | 2012-09-06 | 2015-08-25 | Box, Inc. | System and method for creating a secure channel for inter-application communication based on intents |
US9292833B2 (en) | 2012-09-14 | 2016-03-22 | Box, Inc. | Batching notifications of activities that occur in a web-based collaboration environment |
US10915492B2 (en) * | 2012-09-19 | 2021-02-09 | Box, Inc. | Cloud-based platform enabled with media content indexed for text-based searches and/or metadata extraction |
US9959420B2 (en) | 2012-10-02 | 2018-05-01 | Box, Inc. | System and method for enhanced security and management mechanisms for enterprise administrators in a cloud-based environment |
US9495364B2 (en) | 2012-10-04 | 2016-11-15 | Box, Inc. | Enhanced quick search features, low-barrier commenting/interactive features in a collaboration platform |
US8965880B2 (en) * | 2012-10-05 | 2015-02-24 | Google Inc. | Transcoding and serving resources |
US9665349B2 (en) | 2012-10-05 | 2017-05-30 | Box, Inc. | System and method for generating embeddable widgets which enable access to a cloud-based collaboration platform |
US9251790B2 (en) * | 2012-10-22 | 2016-02-02 | Huseby, Inc. | Apparatus and method for inserting material into transcripts |
US20140142925A1 (en) * | 2012-11-16 | 2014-05-22 | Raytheon Bbn Technologies | Self-organizing unit recognition for speech and other data series |
KR102091003B1 (en) * | 2012-12-10 | 2020-03-19 | 삼성전자 주식회사 | Method and apparatus for providing context aware service using speech recognition |
US10235383B2 (en) | 2012-12-19 | 2019-03-19 | Box, Inc. | Method and apparatus for synchronization of items with read-only permissions in a cloud-based environment |
US9396245B2 (en) | 2013-01-02 | 2016-07-19 | Box, Inc. | Race condition handling in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform |
US9208784B2 (en) * | 2013-01-08 | 2015-12-08 | C21 Patents, Llc | Methododolgy for live text broadcasting |
US9953036B2 (en) | 2013-01-09 | 2018-04-24 | Box, Inc. | File system monitoring in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform |
US9507795B2 (en) | 2013-01-11 | 2016-11-29 | Box, Inc. | Functionalities, features, and user interface of a synchronization client to a cloud-based environment |
US20140201778A1 (en) * | 2013-01-15 | 2014-07-17 | Sap Ag | Method and system of interactive advertisement |
EP2757491A1 (en) | 2013-01-17 | 2014-07-23 | Box, Inc. | Conflict resolution, retry condition management, and handling of problem files for the synchronization client to a cloud-based platform |
US10339452B2 (en) | 2013-02-06 | 2019-07-02 | Verint Systems Ltd. | Automated ontology development |
US9378739B2 (en) * | 2013-03-13 | 2016-06-28 | Nuance Communications, Inc. | Identifying corresponding positions in different representations of a textual work |
US9804729B2 (en) | 2013-03-15 | 2017-10-31 | International Business Machines Corporation | Presenting key differences between related content from different mediums |
US9495365B2 (en) * | 2013-03-15 | 2016-11-15 | International Business Machines Corporation | Identifying key differences between related content from different mediums |
JP2014202848A (en) * | 2013-04-03 | 2014-10-27 | 株式会社東芝 | Text generation device, method and program |
JP6155821B2 (en) * | 2013-05-08 | 2017-07-05 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
US10725968B2 (en) | 2013-05-10 | 2020-07-28 | Box, Inc. | Top down delete or unsynchronization on delete of and depiction of item synchronization with a synchronization client to a cloud-based platform |
US10846074B2 (en) | 2013-05-10 | 2020-11-24 | Box, Inc. | Identification and handling of items to be ignored for synchronization with a cloud-based platform by a synchronization client |
WO2014191054A1 (en) * | 2013-05-31 | 2014-12-04 | Longsand Limited | Processing of audio data |
US9633037B2 (en) | 2013-06-13 | 2017-04-25 | Box, Inc | Systems and methods for synchronization event building and/or collapsing by a synchronization component of a cloud-based platform |
US20140379346A1 (en) * | 2013-06-21 | 2014-12-25 | Google Inc. | Video analysis based language model adaptation |
US9805050B2 (en) | 2013-06-21 | 2017-10-31 | Box, Inc. | Maintaining and updating file system shadows on a local device by a synchronization client of a cloud-based platform |
CN104252439B (en) * | 2013-06-26 | 2017-08-29 | 华为技术有限公司 | Diary generation method and device |
US8947596B2 (en) * | 2013-06-27 | 2015-02-03 | Intel Corporation | Alignment of closed captions |
US20150019206A1 (en) * | 2013-07-10 | 2015-01-15 | Datascription Llc | Metadata extraction of non-transcribed video and audio streams |
US9230547B2 (en) | 2013-07-10 | 2016-01-05 | Datascription Llc | Metadata extraction of non-transcribed video and audio streams |
US9535924B2 (en) | 2013-07-30 | 2017-01-03 | Box, Inc. | Scalability improvement in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform |
US20150058006A1 (en) * | 2013-08-23 | 2015-02-26 | Xerox Corporation | Phonetic alignment for user-agent dialogue recognition |
US20150066506A1 (en) | 2013-08-30 | 2015-03-05 | Verint Systems Ltd. | System and Method of Text Zoning |
KR101747873B1 (en) * | 2013-09-12 | 2017-06-27 | 한국전자통신연구원 | Apparatus and for building language model for speech recognition |
US10509527B2 (en) | 2013-09-13 | 2019-12-17 | Box, Inc. | Systems and methods for configuring event-based automation in cloud-based collaboration platforms |
US9535909B2 (en) | 2013-09-13 | 2017-01-03 | Box, Inc. | Configurable event-based automation architecture for cloud-based collaboration platforms |
US9477752B1 (en) * | 2013-09-30 | 2016-10-25 | Verint Systems Inc. | Ontology administration and application to enhance communication data analytics |
KR20150049700A (en) * | 2013-10-30 | 2015-05-08 | 삼성전자주식회사 | Method and apparautus for controlling input in portable device |
US9232063B2 (en) | 2013-10-31 | 2016-01-05 | Verint Systems Inc. | Call flow and discourse analysis |
US10078689B2 (en) | 2013-10-31 | 2018-09-18 | Verint Systems Ltd. | Labeling/naming of themes |
US9977830B2 (en) | 2014-01-31 | 2018-05-22 | Verint Systems Ltd. | Call summary |
US10037380B2 (en) | 2014-02-14 | 2018-07-31 | Microsoft Technology Licensing, Llc | Browsing videos via a segment list |
US9635219B2 (en) * | 2014-02-19 | 2017-04-25 | Nexidia Inc. | Supplementary media validation system |
US10304458B1 (en) * | 2014-03-06 | 2019-05-28 | Board of Trustees of the University of Alabama and the University of Alabama in Huntsville | Systems and methods for transcribing videos using speaker identification |
US9699404B2 (en) * | 2014-03-19 | 2017-07-04 | Microsoft Technology Licensing, Llc | Closed caption alignment |
US9451335B2 (en) | 2014-04-29 | 2016-09-20 | At&T Intellectual Property I, Lp | Method and apparatus for augmenting media content |
US10530854B2 (en) | 2014-05-30 | 2020-01-07 | Box, Inc. | Synchronization of permissioned content in cloud-based environments |
US10206014B2 (en) | 2014-06-20 | 2019-02-12 | Google Llc | Clarifying audible verbal information in video content |
US9946769B2 (en) | 2014-06-20 | 2018-04-17 | Google Llc | Displaying information related to spoken dialogue in content playing on a device |
US9805125B2 (en) | 2014-06-20 | 2017-10-31 | Google Inc. | Displaying a summary of media content items |
US9838759B2 (en) | 2014-06-20 | 2017-12-05 | Google Inc. | Displaying information related to content playing on a device |
KR102340251B1 (en) * | 2014-06-27 | 2021-12-16 | 삼성전자주식회사 | Method for managing data and an electronic device thereof |
US9575936B2 (en) * | 2014-07-17 | 2017-02-21 | Verint Systems Ltd. | Word cloud display |
US9478059B2 (en) * | 2014-07-28 | 2016-10-25 | PocketGems, Inc. | Animated audiovisual experiences driven by scripts |
US20160042765A1 (en) * | 2014-08-05 | 2016-02-11 | Avid Technology, Inc. | Media composition with timing blocks |
US20160042766A1 (en) * | 2014-08-06 | 2016-02-11 | Echostar Technologies L.L.C. | Custom video content |
US10038731B2 (en) | 2014-08-29 | 2018-07-31 | Box, Inc. | Managing flow-based interactions with cloud-based shared content |
US9894119B2 (en) | 2014-08-29 | 2018-02-13 | Box, Inc. | Configurable metadata-based automation and content classification architecture for cloud-based collaboration platforms |
US11922344B2 (en) | 2014-10-22 | 2024-03-05 | Narrative Science Llc | Automatic generation of narratives from data using communication goals and narrative analytics |
US11288328B2 (en) | 2014-10-22 | 2022-03-29 | Narrative Science Inc. | Interactive and conversational data exploration |
US10747823B1 (en) | 2014-10-22 | 2020-08-18 | Narrative Science Inc. | Interactive and conversational data exploration |
US11238090B1 (en) | 2015-11-02 | 2022-02-01 | Narrative Science Inc. | Applied artificial intelligence technology for using narrative analytics to automatically generate narratives from visualization data |
US10360925B2 (en) * | 2014-10-29 | 2019-07-23 | International Business Machines Corporation | Computerized tool for creating variable length presentations |
KR20160064337A (en) * | 2014-11-27 | 2016-06-08 | 삼성전자주식회사 | Content providing method and apparatus |
US9332221B1 (en) | 2014-11-28 | 2016-05-03 | International Business Machines Corporation | Enhancing awareness of video conference participant expertise |
US20160217127A1 (en) | 2015-01-27 | 2016-07-28 | Verint Systems Ltd. | Identification of significant phrases using multiple language models |
US9914218B2 (en) | 2015-01-30 | 2018-03-13 | Toyota Motor Engineering & Manufacturing North America, Inc. | Methods and apparatuses for responding to a detected event by a robot |
US10037712B2 (en) | 2015-01-30 | 2018-07-31 | Toyota Motor Engineering & Manufacturing North America, Inc. | Vision-assist devices and methods of detecting a classification of an object |
US10217379B2 (en) | 2015-01-30 | 2019-02-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Modifying vision-assist device parameters based on an environment classification |
US9904505B1 (en) * | 2015-04-10 | 2018-02-27 | Zaxcom, Inc. | Systems and methods for processing and recording audio with integrated script mode |
US9886423B2 (en) * | 2015-06-19 | 2018-02-06 | International Business Machines Corporation | Reconciliation of transcripts |
US12153618B2 (en) | 2015-11-02 | 2024-11-26 | Salesforce, Inc. | Applied artificial intelligence technology for automatically generating narratives from visualization data |
US11222184B1 (en) | 2015-11-02 | 2022-01-11 | Narrative Science Inc. | Applied artificial intelligence technology for using narrative analytics to automatically generate narratives from bar charts |
US11170038B1 (en) | 2015-11-02 | 2021-11-09 | Narrative Science Inc. | Applied artificial intelligence technology for using narrative analytics to automatically generate narratives from multiple visualizations |
US11232268B1 (en) | 2015-11-02 | 2022-01-25 | Narrative Science Inc. | Applied artificial intelligence technology for using narrative analytics to automatically generate narratives from line charts |
US10349141B2 (en) | 2015-11-19 | 2019-07-09 | Google Llc | Reminders of media content referenced in other media content |
CN106782627B (en) * | 2015-11-23 | 2019-08-27 | 广州酷狗计算机科技有限公司 | Audio file rerecords method and device |
EP3182297A1 (en) * | 2015-12-18 | 2017-06-21 | Thomson Licensing | Method for generating semantic description of textual content and apparatus performing the same |
US10034053B1 (en) | 2016-01-25 | 2018-07-24 | Google Llc | Polls for media program moments |
US10169033B2 (en) | 2016-02-12 | 2019-01-01 | International Business Machines Corporation | Assigning a computer to a group of computers in a group infrastructure |
BE1023431B1 (en) * | 2016-06-01 | 2017-03-17 | Limecraft Nv | AUTOMATIC IDENTIFICATION AND PROCESSING OF AUDIOVISUAL MEDIA |
US11409791B2 (en) | 2016-06-10 | 2022-08-09 | Disney Enterprises, Inc. | Joint heterogeneous language-vision embeddings for video tagging and search |
US10854190B1 (en) * | 2016-06-13 | 2020-12-01 | United Services Automobile Association (Usaa) | Transcription analysis platform |
US10489516B2 (en) * | 2016-07-13 | 2019-11-26 | Fujitsu Social Science Laboratory Limited | Speech recognition and translation terminal, method and non-transitory computer readable medium |
US9774911B1 (en) | 2016-07-29 | 2017-09-26 | Rovi Guides, Inc. | Methods and systems for automatically evaluating an audio description track of a media asset |
CN106331844A (en) * | 2016-08-17 | 2017-01-11 | 北京金山安全软件有限公司 | Method and device for generating subtitles of media file and electronic equipment |
US11042579B2 (en) | 2016-08-25 | 2021-06-22 | Lakeside Software, Llc | Method and apparatus for natural language query in a workspace analytics system |
US11144838B1 (en) | 2016-08-31 | 2021-10-12 | Narrative Science Inc. | Applied artificial intelligence technology for evaluating drivers of data presented in visualizations |
US10445052B2 (en) | 2016-10-04 | 2019-10-15 | Descript, Inc. | Platform for producing and delivering media content |
US10354008B2 (en) * | 2016-10-07 | 2019-07-16 | Productionpro Technologies Inc. | System and method for providing a visual scroll representation of production data |
WO2018084910A1 (en) * | 2016-11-07 | 2018-05-11 | Axon Enterprise, Inc. | Systems and methods for interrelating text transcript information with video and/or audio information |
US10546063B2 (en) * | 2016-12-13 | 2020-01-28 | International Business Machines Corporation | Processing of string inputs utilizing machine learning |
US10564817B2 (en) | 2016-12-15 | 2020-02-18 | Descript, Inc. | Techniques for creating and presenting media content |
EP3343537A1 (en) | 2016-12-29 | 2018-07-04 | Tata Consultancy Services Limited | Method and system for language skill assessment and development |
US20180203925A1 (en) * | 2017-01-17 | 2018-07-19 | Acoustic Protocol Inc. | Signature-based acoustic classification |
US10943069B1 (en) | 2017-02-17 | 2021-03-09 | Narrative Science Inc. | Applied artificial intelligence technology for narrative generation based on a conditional outcome framework |
US11568148B1 (en) | 2017-02-17 | 2023-01-31 | Narrative Science Inc. | Applied artificial intelligence technology for narrative generation based on explanation communication goals |
US10699079B1 (en) | 2017-02-17 | 2020-06-30 | Narrative Science Inc. | Applied artificial intelligence technology for narrative generation based on analysis communication goals |
US11954445B2 (en) | 2017-02-17 | 2024-04-09 | Narrative Science Llc | Applied artificial intelligence technology for narrative generation based on explanation communication goals |
US11068661B1 (en) | 2017-02-17 | 2021-07-20 | Narrative Science Inc. | Applied artificial intelligence technology for narrative generation based on smart attributes |
US10713442B1 (en) | 2017-02-17 | 2020-07-14 | Narrative Science Inc. | Applied artificial intelligence technology for interactive story editing to support natural language generation (NLG) |
CN106971749A (en) * | 2017-03-30 | 2017-07-21 | 联想(北京)有限公司 | Audio-frequency processing method and electronic equipment |
US9741337B1 (en) * | 2017-04-03 | 2017-08-22 | Green Key Technologies Llc | Adaptive self-trained computer engines with associated databases and methods of use thereof |
US10580457B2 (en) * | 2017-06-13 | 2020-03-03 | 3Play Media, Inc. | Efficient audio description systems and methods |
US11263489B2 (en) * | 2017-06-29 | 2022-03-01 | Intel Corporation | Techniques for dense video descriptions |
US10057537B1 (en) * | 2017-08-18 | 2018-08-21 | Prime Focus Technologies, Inc. | System and method for source script and video synchronization interface |
US11190855B2 (en) | 2017-08-30 | 2021-11-30 | Arris Enterprises Llc | Automatic generation of descriptive video service tracks |
US10083006B1 (en) * | 2017-09-12 | 2018-09-25 | Google Llc | Intercom-style communication using multiple computing devices |
US20190087870A1 (en) * | 2017-09-15 | 2019-03-21 | Oneva, Inc. | Personal video commercial studio system |
GB201715753D0 (en) * | 2017-09-28 | 2017-11-15 | Royal Nat Theatre | Caption delivery system |
US11397855B2 (en) * | 2017-12-12 | 2022-07-26 | International Business Machines Corporation | Data standardization rules generation |
US10225621B1 (en) | 2017-12-20 | 2019-03-05 | Dish Network L.L.C. | Eyes free entertainment |
US11055348B2 (en) * | 2017-12-29 | 2021-07-06 | Facebook, Inc. | Systems and methods for automatically generating stitched media content |
US11042709B1 (en) | 2018-01-02 | 2021-06-22 | Narrative Science Inc. | Context saliency-based deictic parser for natural language processing |
EP3714382A4 (en) * | 2018-01-11 | 2021-01-20 | End Cue, LLC | Script writing and content generation tools and improved operation of same |
US10896294B2 (en) | 2018-01-11 | 2021-01-19 | End Cue, Llc | Script writing and content generation tools and improved operation of same |
US11023689B1 (en) | 2018-01-17 | 2021-06-01 | Narrative Science Inc. | Applied artificial intelligence technology for narrative generation using an invocable analysis service with analysis libraries |
US10755046B1 (en) | 2018-02-19 | 2020-08-25 | Narrative Science Inc. | Applied artificial intelligence technology for conversational inferencing |
US10726838B2 (en) | 2018-06-14 | 2020-07-28 | Disney Enterprises, Inc. | System and method of generating effects during live recitations of stories |
US11232270B1 (en) | 2018-06-28 | 2022-01-25 | Narrative Science Inc. | Applied artificial intelligence technology for using natural language processing to train a natural language generation system with respect to numeric style features |
US10558761B2 (en) * | 2018-07-05 | 2020-02-11 | Disney Enterprises, Inc. | Alignment of video and textual sequences for metadata analysis |
CN110164420B (en) * | 2018-08-02 | 2022-07-19 | 腾讯科技(深圳)有限公司 | Voice recognition method, and method and device for sentence breaking by voice |
US11983183B2 (en) * | 2018-08-07 | 2024-05-14 | Disney Enterprises, Inc. | Techniques for training machine learning models using actor data |
CN109271495B (en) * | 2018-08-14 | 2023-02-17 | 创新先进技术有限公司 | Question-answer recognition effect detection method, device, equipment and readable storage medium |
US10885942B2 (en) | 2018-09-18 | 2021-01-05 | At&T Intellectual Property I, L.P. | Video-log production system |
US11094327B2 (en) * | 2018-09-28 | 2021-08-17 | Lenovo (Singapore) Pte. Ltd. | Audible input transcription |
WO2020086580A1 (en) | 2018-10-22 | 2020-04-30 | Carlson William D | Therapeutic combinations of tdfrps and additional agents and methods of use |
CN109547831B (en) * | 2018-11-19 | 2021-06-01 | 网宿科技股份有限公司 | Method and device for synchronizing white board and video, computing equipment and storage medium |
US11122099B2 (en) * | 2018-11-30 | 2021-09-14 | Motorola Solutions, Inc. | Device, system and method for providing audio summarization data from video |
CN109584882B (en) * | 2018-11-30 | 2022-12-27 | 南京天溯自动化控制系统有限公司 | Method and system for optimizing voice to text conversion aiming at specific scene |
US11729475B2 (en) * | 2018-12-21 | 2023-08-15 | Bce Inc. | System and method for providing descriptive video |
US10785385B2 (en) * | 2018-12-26 | 2020-09-22 | NBCUniversal Media, LLC. | Systems and methods for aligning text and multimedia content |
US11238886B1 (en) * | 2019-01-09 | 2022-02-01 | Audios Ventures Inc. | Generating video information representative of audio clips |
CN109840273B (en) * | 2019-01-18 | 2020-09-15 | 珠海天燕科技有限公司 | Method and device for generating file |
US11341330B1 (en) | 2019-01-28 | 2022-05-24 | Narrative Science Inc. | Applied artificial intelligence technology for adaptive natural language understanding with term discovery |
US11769012B2 (en) | 2019-03-27 | 2023-09-26 | Verint Americas Inc. | Automated system and method to prioritize language model and ontology expansion and pruning |
US10891489B2 (en) * | 2019-04-08 | 2021-01-12 | Nedelco, Incorporated | Identifying and tracking words in a video recording of captioning session |
US11245950B1 (en) * | 2019-04-24 | 2022-02-08 | Amazon Technologies, Inc. | Lyrics synchronization |
CN114788293B (en) | 2019-06-11 | 2023-07-14 | 唯众挚美影视技术公司 | System, method and medium for producing multimedia digital content including movies |
US11429789B2 (en) * | 2019-06-12 | 2022-08-30 | International Business Machines Corporation | Natural language processing and candidate response identification |
CN112231275B (en) | 2019-07-14 | 2024-02-27 | 阿里巴巴集团控股有限公司 | Method, system and equipment for classifying multimedia files, processing information and training models |
US20220414349A1 (en) * | 2019-07-22 | 2022-12-29 | wordly, Inc. | Systems, methods, and apparatus for determining an official transcription and speaker language from a plurality of transcripts of text in different languages |
US11276419B2 (en) * | 2019-07-30 | 2022-03-15 | International Business Machines Corporation | Synchronized sound generation from videos |
WO2021022499A1 (en) | 2019-08-07 | 2021-02-11 | WeMovie Technologies | Adaptive marketing in cloud-based content production |
WO2021039925A1 (en) * | 2019-08-29 | 2021-03-04 | 有限会社Bond | Program creating method, program creating device, and recording medium |
US11295084B2 (en) | 2019-09-16 | 2022-04-05 | International Business Machines Corporation | Cognitively generating information from videos |
CN110675896B (en) * | 2019-09-30 | 2021-10-22 | 北京字节跳动网络技术有限公司 | Character time alignment method, device and medium for audio and electronic equipment |
WO2021068105A1 (en) | 2019-10-08 | 2021-04-15 | WeMovie Technologies | Pre-production systems for making movies, tv shows and multimedia contents |
US11410658B1 (en) * | 2019-10-29 | 2022-08-09 | Dialpad, Inc. | Maintainable and scalable pipeline for automatic speech recognition language modeling |
EP4049270B1 (en) | 2019-11-04 | 2023-10-04 | Google LLC | Using video clips as dictionary usage examples |
US11430485B2 (en) * | 2019-11-19 | 2022-08-30 | Netflix, Inc. | Systems and methods for mixing synthetic voice with original audio tracks |
US11138970B1 (en) * | 2019-12-06 | 2021-10-05 | Asapp, Inc. | System, method, and computer program for creating a complete transcription of an audio recording from separately transcribed redacted and unredacted words |
US11562743B2 (en) * | 2020-01-29 | 2023-01-24 | Salesforce.Com, Inc. | Analysis of an automatically generated transcription |
US11570099B2 (en) | 2020-02-04 | 2023-01-31 | Bank Of America Corporation | System and method for autopartitioning and processing electronic resources |
US11360937B2 (en) | 2020-03-20 | 2022-06-14 | Bank Of America Corporation | System for natural language processing-based electronic file scanning for processing database queries |
CN111462775B (en) * | 2020-03-30 | 2023-11-03 | 腾讯科技(深圳)有限公司 | Audio similarity determination method, device, server and medium |
US10924636B1 (en) | 2020-04-30 | 2021-02-16 | Gopro, Inc. | Systems and methods for synchronizing information for videos |
WO2021225608A1 (en) | 2020-05-08 | 2021-11-11 | WeMovie Technologies | Fully automated post-production editing for movies, tv shows and multimedia contents |
CN114205665B (en) | 2020-06-09 | 2023-05-09 | 抖音视界有限公司 | Information processing method, device, electronic equipment and storage medium |
US11669295B2 (en) * | 2020-06-18 | 2023-06-06 | Sony Group Corporation | Multiple output control based on user input |
US11070888B1 (en) | 2020-08-27 | 2021-07-20 | WeMovie Technologies | Content structure aware multimedia streaming service for movies, TV shows and multimedia contents |
US11625928B1 (en) * | 2020-09-01 | 2023-04-11 | Amazon Technologies, Inc. | Language agnostic drift correction |
US11423941B2 (en) * | 2020-09-28 | 2022-08-23 | TCL Research America Inc. | Write-a-movie: unifying writing and shooting |
US11871138B2 (en) * | 2020-10-13 | 2024-01-09 | Grass Valley Canada | Virtualized production switcher and method for media production |
CN112256672B (en) * | 2020-10-22 | 2023-05-30 | 中国联合网络通信集团有限公司 | Database change approval method and device |
US11812121B2 (en) | 2020-10-28 | 2023-11-07 | WeMovie Technologies | Automated post-production editing for user-generated multimedia contents |
US11166086B1 (en) | 2020-10-28 | 2021-11-02 | WeMovie Technologies | Automated post-production editing for user-generated multimedia contents |
US11508411B2 (en) * | 2020-10-28 | 2022-11-22 | Meta Platforms Technologies, Llc | Text-driven editor for audio and video assembly |
US11481475B2 (en) | 2020-11-03 | 2022-10-25 | Capital One Services, Llc | Computer-based systems configured for automated computer script analysis and malware detection and methods thereof |
GB2600933B (en) * | 2020-11-11 | 2023-06-28 | Sony Interactive Entertainment Inc | Apparatus and method for analysis of audio recordings |
US11922943B1 (en) * | 2021-01-26 | 2024-03-05 | Wells Fargo Bank, N.A. | KPI-threshold selection for audio-transcription models |
US11729476B2 (en) * | 2021-02-08 | 2023-08-15 | Sony Group Corporation | Reproduction control of scene description |
US11887586B2 (en) * | 2021-03-03 | 2024-01-30 | Spotify Ab | Systems and methods for providing responses from media content |
EP4060519A1 (en) * | 2021-03-18 | 2022-09-21 | Prisma Analytics GmbH | Data transformation considering data integrity |
US11521639B1 (en) | 2021-04-02 | 2022-12-06 | Asapp, Inc. | Speech sentiment analysis using a speech sentiment classifier pretrained with pseudo sentiment labels |
WO2022239053A1 (en) * | 2021-05-10 | 2022-11-17 | ソニーグループ株式会社 | Information processing device, information processing method, and information processing program |
CN115514987B (en) * | 2021-06-23 | 2024-10-18 | 视见科技(杭州)有限公司 | System and method for automated narrative video production through the use of script annotations |
US20230027035A1 (en) * | 2021-07-09 | 2023-01-26 | Transitional Forms Inc. | Automated narrative production system and script production method with real-time interactive characters |
US12118986B2 (en) * | 2021-07-20 | 2024-10-15 | Conduent Business Services, Llc | System and method for automated processing of natural language using deep learning model encoding |
US11330154B1 (en) | 2021-07-23 | 2022-05-10 | WeMovie Technologies | Automated coordination in multimedia content production |
US11763803B1 (en) | 2021-07-28 | 2023-09-19 | Asapp, Inc. | System, method, and computer program for extracting utterances corresponding to a user problem statement in a conversation between a human agent and a user |
US11735186B2 (en) | 2021-09-07 | 2023-08-22 | 3Play Media, Inc. | Hybrid live captioning systems and methods |
US11769481B2 (en) * | 2021-10-07 | 2023-09-26 | Nvidia Corporation | Unsupervised alignment for text to speech synthesis using neural networks |
US11321639B1 (en) | 2021-12-13 | 2022-05-03 | WeMovie Technologies | Automated evaluation of acting performance using cloud services |
US20230237990A1 (en) * | 2022-01-27 | 2023-07-27 | Asapp, Inc. | Training speech processing models using pseudo tokens |
US12067363B1 (en) | 2022-02-24 | 2024-08-20 | Asapp, Inc. | System, method, and computer program for text sanitization |
US11763099B1 (en) | 2022-04-27 | 2023-09-19 | VoyagerX, Inc. | Providing translated subtitle for video content |
WO2025042388A1 (en) * | 2023-08-22 | 2025-02-27 | Google Llc | Video context aware editing agent |
CN117240983B (en) * | 2023-11-16 | 2024-01-26 | 湖南快乐阳光互动娱乐传媒有限公司 | Method and device for automatically generating sound drama |
JP7538574B1 (en) | 2024-04-19 | 2024-08-22 | 史睦 川口 | Video creation device, video creation method, video creation program, and video creation system |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6219642B1 (en) * | 1998-10-05 | 2001-04-17 | Legerity, Inc. | Quantization using frequency and mean compensated frequency input data for robust speech recognition |
US6473778B1 (en) * | 1998-12-24 | 2002-10-29 | At&T Corporation | Generating hypermedia documents from transcriptions of television programs using parallel text alignment |
US6728682B2 (en) * | 1998-01-16 | 2004-04-27 | Avid Technology, Inc. | Apparatus and method using speech recognition and scripts to capture, author and playback synchronized audio and video |
US7168953B1 (en) * | 2003-01-27 | 2007-01-30 | Massachusetts Institute Of Technology | Trainable videorealistic speech animation |
US20080243503A1 (en) * | 2007-03-30 | 2008-10-02 | Microsoft Corporation | Minimum divergence based discriminative training for pattern recognition |
US7672830B2 (en) * | 2005-02-22 | 2010-03-02 | Xerox Corporation | Apparatus and methods for aligning words in bilingual sentences |
US20100299131A1 (en) * | 2009-05-21 | 2010-11-25 | Nexidia Inc. | Transcript alignment |
US20100332225A1 (en) * | 2009-06-29 | 2010-12-30 | Nexidia Inc. | Transcript alignment |
US8249871B2 (en) * | 2005-11-18 | 2012-08-21 | Microsoft Corporation | Word clustering for input data |
US8301451B2 (en) * | 2008-09-03 | 2012-10-30 | Svox Ag | Speech synthesis with dynamic constraints |
Family Cites Families (48)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5677739A (en) | 1995-03-02 | 1997-10-14 | National Captioning Institute | System and method for providing described television services |
US5900908A (en) | 1995-03-02 | 1999-05-04 | National Captioning Insitute, Inc. | System and method for providing described television services |
US5801685A (en) | 1996-04-08 | 1998-09-01 | Tektronix, Inc. | Automatic editing of recorded video elements sychronized with a script text read or displayed |
US6172675B1 (en) * | 1996-12-05 | 2001-01-09 | Interval Research Corporation | Indirect manipulation of data using temporally related data, with particular application to manipulation of audio or audiovisual data |
JP3690043B2 (en) | 1997-03-03 | 2005-08-31 | ソニー株式会社 | Audio information transmission apparatus and method, and audio information recording apparatus |
US6370503B1 (en) | 1999-06-30 | 2002-04-09 | International Business Machines Corp. | Method and apparatus for improving speech recognition accuracy |
US6442518B1 (en) * | 1999-07-14 | 2002-08-27 | Compaq Information Technologies Group, L.P. | Method for refining time alignments of closed captions |
US6477493B1 (en) | 1999-07-15 | 2002-11-05 | International Business Machines Corporation | Off site voice enrollment on a transcription device for speech recognition |
GB0008537D0 (en) * | 2000-04-06 | 2000-05-24 | Ananova Ltd | Character animation |
US6505153B1 (en) * | 2000-05-22 | 2003-01-07 | Compaq Information Technologies Group, L.P. | Efficient method for producing off-line closed captions |
DE60142967D1 (en) * | 2000-06-09 | 2010-10-14 | British Broadcasting Corp | GENERATION OF SUBTITLES FOR MOVING PICTURES |
GB0023930D0 (en) * | 2000-09-29 | 2000-11-15 | Canon Kk | Database annotation and retrieval |
US6975985B2 (en) | 2000-11-29 | 2005-12-13 | International Business Machines Corporation | Method and system for the automatic amendment of speech recognition vocabularies |
US6925455B2 (en) * | 2000-12-12 | 2005-08-02 | Nec Corporation | Creating audio-centric, image-centric, and integrated audio-visual summaries |
JPWO2002103591A1 (en) | 2001-06-13 | 2004-10-07 | 富士通株式会社 | Agenda progress support device and agenda progress support program |
US7668718B2 (en) | 2001-07-17 | 2010-02-23 | Custom Speech Usa, Inc. | Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile |
EP2202979A1 (en) | 2002-04-12 | 2010-06-30 | Mitsubishi Denki Kabushiki Kaisha | Hint information describing method for manipulating metadata |
US7231351B1 (en) * | 2002-05-10 | 2007-06-12 | Nexidia, Inc. | Transcript alignment |
US7260738B2 (en) | 2002-06-17 | 2007-08-21 | Microsoft Corporation | System and method for splitting an image across multiple computer readable media |
US20040001106A1 (en) | 2002-06-26 | 2004-01-01 | John Deutscher | System and process for creating an interactive presentation employing multi-media components |
US7123696B2 (en) | 2002-10-04 | 2006-10-17 | Frederick Lowe | Method and apparatus for generating and distributing personalized media clips |
US20050120391A1 (en) * | 2003-12-02 | 2005-06-02 | Quadrock Communications, Inc. | System and method for generation of interactive TV content |
US8064752B1 (en) | 2003-12-09 | 2011-11-22 | Apple Inc. | Video encoding |
US20050228663A1 (en) | 2004-03-31 | 2005-10-13 | Robert Boman | Media production system using time alignment to scripts |
US7836389B2 (en) | 2004-04-16 | 2010-11-16 | Avid Technology, Inc. | Editing system for audiovisual works and corresponding text for television news |
US7356469B2 (en) | 2004-08-20 | 2008-04-08 | International Business Machines Corporation | Method and system for trimming audio files |
EP1851973A4 (en) | 2005-02-14 | 2010-07-07 | Teresis Media Man Inc | Multipurpose media players |
US7668869B2 (en) * | 2006-04-03 | 2010-02-23 | Digitalsmiths Corporation | Media access system |
US7623755B2 (en) | 2006-08-17 | 2009-11-24 | Adobe Systems Incorporated | Techniques for positioning audio and video clips |
US20080114603A1 (en) | 2006-11-15 | 2008-05-15 | Adacel, Inc. | Confirmation system for command or speech recognition using activation means |
US8345159B2 (en) | 2007-04-16 | 2013-01-01 | Caption Colorado L.L.C. | Captioning evaluation system |
US8170396B2 (en) | 2007-04-16 | 2012-05-01 | Adobe Systems Incorporated | Changing video playback rate |
KR101445869B1 (en) | 2007-07-11 | 2014-09-29 | 엘지전자 주식회사 | Media Interface |
US8990848B2 (en) * | 2008-07-22 | 2015-03-24 | At&T Intellectual Property I, L.P. | System and method for temporally adaptive media playback |
JP2010074823A (en) | 2008-08-22 | 2010-04-02 | Panasonic Corp | Video editing system |
US8219899B2 (en) | 2008-09-22 | 2012-07-10 | International Business Machines Corporation | Verbal description method and system |
US8131545B1 (en) * | 2008-09-25 | 2012-03-06 | Google Inc. | Aligning a transcript to audio data |
US9049477B2 (en) | 2008-11-13 | 2015-06-02 | At&T Intellectual Property I, Lp | Apparatus and method for managing media content |
US8497939B2 (en) | 2008-12-08 | 2013-07-30 | Home Box Office, Inc. | Method and process for text-based assistive program descriptions for television |
US8793282B2 (en) * | 2009-04-14 | 2014-07-29 | Disney Enterprises, Inc. | Real-time media presentation using metadata clips |
US20100260482A1 (en) | 2009-04-14 | 2010-10-14 | Yossi Zoor | Generating a Synchronized Audio-Textual Description of a Video Recording Event |
US8701007B2 (en) * | 2009-04-30 | 2014-04-15 | Apple Inc. | Edit visualizer for modifying and evaluating uncommitted media content |
US8843368B2 (en) | 2009-08-17 | 2014-09-23 | At&T Intellectual Property I, L.P. | Systems, computer-implemented methods, and tangible computer-readable storage media for transcription alignment |
US8799953B2 (en) | 2009-08-27 | 2014-08-05 | Verizon Patent And Licensing Inc. | Media content distribution systems and methods |
US8281231B2 (en) | 2009-09-11 | 2012-10-02 | Digitalsmiths, Inc. | Timeline alignment for closed-caption text using speech recognition transcripts |
US8302010B2 (en) | 2010-03-29 | 2012-10-30 | Avid Technology, Inc. | Transcript editor |
US8572488B2 (en) | 2010-03-29 | 2013-10-29 | Avid Technology, Inc. | Spot dialog editor |
US8825489B2 (en) | 2010-04-12 | 2014-09-02 | Adobe Systems Incorporated | Method and apparatus for interpolating script data |
-
2010
- 2010-05-28 US US12/789,785 patent/US8825489B2/en active Active
- 2010-05-28 US US12/789,791 patent/US9066049B2/en active Active
- 2010-05-28 US US12/789,720 patent/US8825488B2/en active Active
- 2010-05-28 US US12/789,708 patent/US8447604B1/en active Active
- 2010-05-28 US US12/789,760 patent/US9191639B2/en active Active
- 2010-05-28 US US12/789,749 patent/US20130124984A1/en not_active Abandoned
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6728682B2 (en) * | 1998-01-16 | 2004-04-27 | Avid Technology, Inc. | Apparatus and method using speech recognition and scripts to capture, author and playback synchronized audio and video |
US6219642B1 (en) * | 1998-10-05 | 2001-04-17 | Legerity, Inc. | Quantization using frequency and mean compensated frequency input data for robust speech recognition |
US6473778B1 (en) * | 1998-12-24 | 2002-10-29 | At&T Corporation | Generating hypermedia documents from transcriptions of television programs using parallel text alignment |
US7168953B1 (en) * | 2003-01-27 | 2007-01-30 | Massachusetts Institute Of Technology | Trainable videorealistic speech animation |
US7672830B2 (en) * | 2005-02-22 | 2010-03-02 | Xerox Corporation | Apparatus and methods for aligning words in bilingual sentences |
US8249871B2 (en) * | 2005-11-18 | 2012-08-21 | Microsoft Corporation | Word clustering for input data |
US20080243503A1 (en) * | 2007-03-30 | 2008-10-02 | Microsoft Corporation | Minimum divergence based discriminative training for pattern recognition |
US8301451B2 (en) * | 2008-09-03 | 2012-10-30 | Svox Ag | Speech synthesis with dynamic constraints |
US20100299131A1 (en) * | 2009-05-21 | 2010-11-25 | Nexidia Inc. | Transcript alignment |
US20100332225A1 (en) * | 2009-06-29 | 2010-12-30 | Nexidia Inc. | Transcript alignment |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100263005A1 (en) * | 2009-04-08 | 2010-10-14 | Eric Foster White | Method and system for egnaging interactive web content |
US9066049B2 (en) | 2010-04-12 | 2015-06-23 | Adobe Systems Incorporated | Method and apparatus for processing scripts |
US8825489B2 (en) | 2010-04-12 | 2014-09-02 | Adobe Systems Incorporated | Method and apparatus for interpolating script data |
US8825488B2 (en) | 2010-04-12 | 2014-09-02 | Adobe Systems Incorporated | Method and apparatus for time synchronized script metadata |
US9191639B2 (en) | 2010-04-12 | 2015-11-17 | Adobe Systems Incorporated | Method and apparatus for generating video descriptions |
US20120143606A1 (en) * | 2010-12-01 | 2012-06-07 | At&T Intellectual Property I, L.P. | Method and system for testing closed caption content of video assets |
US9620118B2 (en) | 2010-12-01 | 2017-04-11 | Nuance Communications, Inc. | Method and system for testing closed caption content of video assets |
US8826354B2 (en) * | 2010-12-01 | 2014-09-02 | At&T Intellectual Property I, L.P. | Method and system for testing closed caption content of video assets |
US20130080174A1 (en) * | 2011-09-22 | 2013-03-28 | Kabushiki Kaisha Toshiba | Retrieving device, retrieving method, and computer program product |
US9002703B1 (en) * | 2011-09-28 | 2015-04-07 | Amazon Technologies, Inc. | Community audio narration generation |
US20140013268A1 (en) * | 2012-07-09 | 2014-01-09 | Mobitude, LLC, a Delaware LLC | Method for creating a scripted exchange |
US20140122079A1 (en) * | 2012-10-25 | 2014-05-01 | Ivona Software Sp. Z.O.O. | Generating personalized audio programs from text content |
US9190049B2 (en) * | 2012-10-25 | 2015-11-17 | Ivona Software Sp. Z.O.O. | Generating personalized audio programs from text content |
US8977555B2 (en) * | 2012-12-20 | 2015-03-10 | Amazon Technologies, Inc. | Identification of utterance subjects |
US9240187B2 (en) | 2012-12-20 | 2016-01-19 | Amazon Technologies, Inc. | Identification of utterance subjects |
US20140180697A1 (en) * | 2012-12-20 | 2014-06-26 | Amazon Technologies, Inc. | Identification of utterance subjects |
EP4283490A1 (en) * | 2022-05-27 | 2023-11-29 | Tata Consultancy Services Limited | Systems and methods for rules-based mapping of answer scripts to markers |
Also Published As
Publication number | Publication date |
---|---|
US8447604B1 (en) | 2013-05-21 |
US20130120654A1 (en) | 2013-05-16 |
US8825489B2 (en) | 2014-09-02 |
US8825488B2 (en) | 2014-09-02 |
US9066049B2 (en) | 2015-06-23 |
US20130124203A1 (en) | 2013-05-16 |
US9191639B2 (en) | 2015-11-17 |
US20130124984A1 (en) | 2013-05-16 |
US20130124213A1 (en) | 2013-05-16 |
US20130124212A1 (en) | 2013-05-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8447604B1 (en) | Method and apparatus for processing scripts and related data | |
US12217756B2 (en) | Systems and methods for improved digital transcript creation using automated speech recognition | |
US11545156B2 (en) | Automated meeting minutes generation service | |
US20230205985A1 (en) | Automated meeting minutes generator | |
CN101382937B (en) | Speech recognition-based multimedia resource processing method and its online teaching system | |
US7693717B2 (en) | Session file modification with annotation using speech recognition or text to speech | |
US20200126583A1 (en) | Discovering highlights in transcribed source material for rapid multimedia production | |
US7668718B2 (en) | Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile | |
US20100299131A1 (en) | Transcript alignment | |
EP1692629B1 (en) | System & method for integrative analysis of intrinsic and extrinsic audio-visual data | |
US20200126559A1 (en) | Creating multi-media from transcript-aligned media recordings | |
US20070244700A1 (en) | Session File Modification with Selective Replacement of Session File Components | |
US20070011012A1 (en) | Method, system, and apparatus for facilitating captioning of multi-media content | |
US20020164151A1 (en) | Automatic content analysis and representation of multimedia presentations | |
US20110239107A1 (en) | Transcript editor | |
US11922944B2 (en) | Phrase alternatives representation for automatic speech recognition and methods of use | |
US8312379B2 (en) | Methods, systems, and computer program products for editing using an interface | |
US20130080384A1 (en) | Systems and methods for extracting and processing intelligent structured data from media files | |
CN113923479A (en) | Audio and video editing method and device | |
CN100538696C (en) | The system and method that is used for the analysis-by-synthesis of intrinsic and extrinsic audio-visual data | |
KR101783872B1 (en) | Video Search System and Method thereof | |
Dinkar et al. | From local hesitations to global impressions of a speaker’s feeling of knowing | |
KR20240041144A (en) | Automatic translation and dubbing method using AI | |
JP2025034460A (en) | Processing system, program and processing method | |
CN119807407A (en) | A conference summary generation method based on paralinguistic acoustic features |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ADOBE SYSTEMS INCORPORATED, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHANG, WALTER W.;REEL/FRAME:024454/0776 Effective date: 20100527 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: ADOBE INC., CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:ADOBE SYSTEMS INCORPORATED;REEL/FRAME:048867/0882 Effective date: 20181008 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |