CN110290425A - A kind of method for processing video frequency, device and storage medium - Google Patents
A kind of method for processing video frequency, device and storage medium Download PDFInfo
- Publication number
- CN110290425A CN110290425A CN201910691577.4A CN201910691577A CN110290425A CN 110290425 A CN110290425 A CN 110290425A CN 201910691577 A CN201910691577 A CN 201910691577A CN 110290425 A CN110290425 A CN 110290425A
- Authority
- CN
- China
- Prior art keywords
- video
- foreground
- background
- video frame
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 94
- 238000012545 processing Methods 0.000 title claims abstract description 66
- 238000003860 storage Methods 0.000 title claims abstract description 11
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 43
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 43
- 230000004044 response Effects 0.000 claims abstract description 33
- 238000005520 cutting process Methods 0.000 claims abstract description 23
- 239000011159 matrix material Substances 0.000 claims description 48
- 230000008569 process Effects 0.000 claims description 37
- 230000015654 memory Effects 0.000 claims description 18
- 238000010586 diagram Methods 0.000 description 20
- 239000000203 mixture Substances 0.000 description 17
- 230000011218 segmentation Effects 0.000 description 11
- 230000003068 static effect Effects 0.000 description 11
- 238000005516 engineering process Methods 0.000 description 10
- 238000013519 translation Methods 0.000 description 10
- 230000000694 effects Effects 0.000 description 6
- 230000006870 function Effects 0.000 description 6
- 230000002452 interceptive effect Effects 0.000 description 6
- 238000009877 rendering Methods 0.000 description 6
- 238000012549 training Methods 0.000 description 6
- 238000003062 neural network model Methods 0.000 description 5
- 238000004519 manufacturing process Methods 0.000 description 4
- 238000012512 characterization method Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 238000006073 displacement reaction Methods 0.000 description 3
- 238000005538 encapsulation Methods 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 230000009466 transformation Effects 0.000 description 3
- 238000010276 construction Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000013515 script Methods 0.000 description 2
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 238000000429 assembly Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000000155 melt Substances 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 208000001491 myopia Diseases 0.000 description 1
- 230000036961 partial effect Effects 0.000 description 1
- 238000004321 preservation Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000002829 reductive effect Effects 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/23424—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44016—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47205—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8547—Content authoring involving timestamps for synchronizing content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2628—Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/265—Mixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/272—Means for inserting a foreground image in a background image, i.e. inlay, outlay
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Business, Economics & Management (AREA)
- Marketing (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- Computer Security & Cryptography (AREA)
- Studio Circuits (AREA)
Abstract
The present invention provides a kind of method for processing video frequency, device and storage mediums;Method includes: acquisition target video;In response to the cutting operation for target object in the target video, obtain from the target video using the target object as the foreground video of prospect;The foreground video includes at least one foreground video frame;Obtain background video;The background video includes at least one background video frame;In response to being directed to the synthetic operation of the foreground video and the background video, the foreground video frame in the foreground video is overlapped with the background video frame in the background video, and the video frame that superposition obtains is encapsulated as synthetic video.By means of the invention it is possible to carry out the synthesis of dynamic video.
Description
Technical field
This application involves multimedia technology more particularly to a kind of method for processing video frequency, device and storage medium.
Background technique
With the continuous development of communication and mobile Internet, the epoch based on text and picture have become past, net
Network live streaming and short video traffic start to be skyrocketed through, the appearance of various video class application programs, significantly reduce people's production
The threshold of video, more and more users begin participating in video creation.
But video production scheme in the related technology, object that only can be static are synthesized in template video, can not carry out
The synthesis of dynamic video.
Summary of the invention
The embodiment of the present invention provides a kind of method for processing video frequency, device and storage medium, is able to carry out the conjunction of dynamic video
At.
The technical solution of the embodiment of the present invention is achieved in that
The embodiment of the present invention provides a kind of method for processing video frequency, comprising:
Obtain target video;
In response to the cutting operation for target object in the target video, obtain from the target video with described
Target object is the foreground video of prospect;The foreground video includes at least one foreground video frame;
Obtain background video;The background video includes at least one background video frame;
In response to being directed to the synthetic operation of the foreground video and the background video, by the prospect in the foreground video
Video frame is overlapped with the background video frame in the background video, and
The video frame that superposition obtains is encapsulated as synthetic video.
The embodiment of the present invention provides a kind of video process apparatus, comprising:
First acquisition unit, for obtaining target video;
Cutting unit, for being regarded from the target in response to the cutting operation for target object in the target video
It obtains in frequency using the target object as the foreground video of prospect;The foreground video includes at least one foreground video frame;
Second acquisition unit, for obtaining background video;The background video includes at least one background video frame;
Synthesis unit, for the synthetic operation in response to being directed to the foreground video and the background video, before described
Foreground video frame in scape video and the background video frame in the background video into being superimposed, and
The video frame that superposition obtains is encapsulated as synthetic video.
In the above scheme, the cutting unit, is also used to:
The batch splitting received at least two target videos operates;
It operates in response to the batch splitting, obtains from each target video using the target object as prospect
Video clip, and it is determined as corresponding foreground video.
In the above scheme, the synthesis unit, is also used to:
Receive the batch synthetic operation for being directed at least two foreground videos and the background video;
In response to the batch synthetic operation, the foreground video frame at least two foreground video is added to respectively
In background video frame in the background video.
In the above scheme, the second acquisition unit, is also used to:
Loaded and displayed has the video selection window of alternative background video;
The video selection received for the video selection window operates;
Obtain the selected background video of the video selection operation.
In the above scheme, described device further include: preview unit is used for:
In response to being directed to the preview operation of the foreground video and the background video, the foreground video frame and institute is presented
State the Overlay of background video frame.
In the above scheme, the cutting unit, is also used to:
Target area where identifying the target object in the video frame of the target video, and by the video frame
Described in areas transparentization processing except target area;
Video frame after transparency process is encapsulated as the foreground video.
In the above scheme, the cutting unit, is also used to:
Identify the target area where target object described in the video frame of the target video, and according to the target area
Domain obtains the corresponding image array of video frame of the target video;Element in described image matrix characterizes corresponding view respectively
The pixel of frequency frame belongs to the probability of the target area;
Described image matrix is subjected to mask process with corresponding video frame, the target area will be removed in the video frame
Areas transparent except domain.
In the above scheme, the synthesis unit, is also used to:
Obtain the timestamp alignment relation of the foreground video frame Yu background video frame;
The timestamp alignment relation pair will be met in foreground video frame and the background video in the foreground video
The background video frame answered is overlapped.
In the above scheme, the synthesis unit, is also used to:
In response to the edit operation for the foreground video and background video setting synthetic parameters, by the prospect
Video frame covers the background video frame, and overlay area of the foreground video frame in the background video frame meets setting
Synthetic parameters.
In the above scheme, the synthesis unit, is also used to:
Construct initial matrix identical with the foreground video frame sign;
The element in the initial matrix is adjusted according to the edit operation, obtains the synthetic parameters of characterization setting
Variable quantity objective matrix.
In the above scheme, the synthesis unit, is also used to:
The objective matrix is multiplied with the foreground video frame in the foreground video, the foreground video after being adjusted
Frame;
The foreground video frame adjusted is covered into the background video frame.
The embodiment of the present invention provides a kind of video process apparatus, comprising:
Memory, for storing executable instruction;
Processor when for executing the executable instruction stored in the memory, is realized provided in an embodiment of the present invention
Method for processing video frequency.
The embodiment of the present invention provides a kind of storage medium, is stored with executable instruction, real when for causing processor to execute
Existing method for processing video frequency provided in an embodiment of the present invention.
The prospect that will be split as foreground video using target object in target video, and will be divided from target video
The video frame that the foreground video frame of video is synthesized with the background video frame of background video is encapsulated as synthetic video, to be based on video
Content, new view is synthesized as background using the target object in target video as prospect and using the video frame of background video
Frequently, the dynamic video of image content coordination is obtained.
Detailed description of the invention
Fig. 1 is an optional structural schematic diagram of processing system for video framework provided in an embodiment of the present invention;
Fig. 2 is an optional structural schematic diagram of video process apparatus provided in an embodiment of the present invention;
Fig. 3 is an optional flow diagram of method for processing video frequency provided in an embodiment of the present invention;
Fig. 4 is that the embodiment of the present invention provides an optional display interface schematic diagram;
Fig. 5 A is that the embodiment of the present invention provides an optional Overlay schematic diagram;
Fig. 5 B is that the embodiment of the present invention provides an optional Overlay schematic diagram;
Fig. 6 is an optional flow diagram of method for processing video frequency provided in an embodiment of the present invention;
Fig. 7 is that the embodiment of the present invention provides an optional training sample schematic diagram;
Fig. 8 is that the embodiment of the present invention provides an optional editing interface schematic diagram;
Fig. 9 is that the embodiment of the present invention provides an optional editing interface schematic diagram;
Figure 10 is an optional encoding and decoding configuration diagram of video encoder provided in an embodiment of the present invention;
Figure 11 is an optional flow diagram of method for processing video frequency in the related technology;
Figure 12 is the synthetic effect schematic diagram of method for processing video frequency in the related technology;
Figure 13 is an optional flow diagram of method for processing video frequency in the related technology;
Figure 14 is an optional flow diagram of method for processing video frequency provided in an embodiment of the present invention;
Figure 15 is an optional flow diagram of method for processing video frequency provided in an embodiment of the present invention;
Figure 16 is one provided in an embodiment of the present invention optional display interface schematic diagram.
Specific embodiment
To make the objectives, technical solutions, and advantages of the present invention clearer, below in conjunction with attached drawing to the present invention make into
It is described in detail to one step, described embodiment is not construed as limitation of the present invention, and those of ordinary skill in the art are not having
All other embodiment obtained under the premise of creative work is made, shall fall within the protection scope of the present invention.
In the following description, it is related to " some embodiments ", which depict the subsets of all possible embodiments, but can
To understand, " some embodiments " can be the same subsets or different subsets of all possible embodiments, and can not conflict
In the case where be combined with each other.
Unless otherwise defined, all technical and scientific terms used herein and belong to technical field of the invention
The normally understood meaning of technical staff is identical.Term used herein is intended merely to the purpose of the description embodiment of the present invention,
It is not intended to limit the present invention.
Before the embodiment of the present invention is further elaborated, to noun involved in the embodiment of the present invention and term
It is illustrated, noun involved in the embodiment of the present invention and term are suitable for following explanation.
1) background, the scenery of main body behind, can show space-time environment locating for personage or event, example in the picture of video
The building, wall at such as personage rear, ground.
2) prospect is the main body that video shows compared with background closer to the content of camera lens in video pictures, such as stands and building
People before building object.
3) target video, carry out Video Composition when for extracting the video of prospect.
4) background video, carry out Video Composition when for extracting the video of background.
5) it is superimposed, using the partial region in one (or multiple images) as prospect, and using another image as background
It is synthesized, obtains new image.Such as: by image A some region and image B synthesize, obtain image C.This
In, image can be the video frame in video.
6) exposure mask (mask) is the image for being shielded to (some or all of pixel) in image to be processed
Matrix, so that the part in specific image highlights.Exposure mask can be two-dimensional matrix array, also use multivalue matrix function sometimes
According to.
7) mask process shields the processing of (such as transparence) based on exposure mask to some regions in image.In image
Binary number (also referred to as mask) progress and operation of same position in each pixel and exposure mask, such as 1&1=1;1&0=0.
8) it encapsulates, multiple video frames is converted to by video file based on certain frame per second and video format.Wherein, frame per second table
Show frame number per second, such as: 25 frames/per second (fps), 60fps etc..Video format can include: Matroska multimedia container
(Matroska Multimedia Container, MLV), Audio Video Interleaved format (Audio Video Interleaved,
AVI), the video file formats such as dynamic image expert group (Moving Picture Experts Group, MPEG) -4.
Analytic explanation is carried out to the technical solution for being directed to Video Composition in the related technology first.
Technical solution 1) still image and dynamic video synthesis
To still image carry out AI segmentation, be partitioned into the corresponding region of target object, using the region divided as with work
It is merged for the background video of background, the video after being synthesized.Here, the object of fusion is static image and dynamic
Video, the target object in video after synthesis are static, that is to say, that in video in post synthesis, in each video frame
Target object be static.
Technical solution 2) video pictures whole splicing
The video frame or so of two videos is stitched together, synthesize a bigger video, for the background to video into
Row processing, the picture of the video of synthesis are not merged in terms of content.
Above-mentioned several technical solutions there are aiming at the problem that, the embodiment of the present invention provides a kind of method for processing video frequency, by mesh
It is split from target video in mark video using target object as foreground video, and the prospect for the foreground video divided is regarded
The video frame that frequency frame is synthesized with the background video frame of background video is encapsulated as synthetic video, thus based on the content of video to dynamic
Video synthesized, obtain image content coordination dynamic video.
Illustrate the exemplary application for realizing the video process apparatus of the embodiment of the present invention below, it is provided in an embodiment of the present invention
Video process apparatus is desirably integrated into various forms of electronic equipments, and the electronic equipment that the present invention implements to provide may be embodied as
Various terminals, such as mobile phone (mobile phone), tablet computer, laptop etc. have the mobile terminal of wireless communication ability,
In another example desktop computer, desktop computer etc..In addition, electronic equipment also may be embodied as a server or by multiple servers
The server cluster of composition, is not limited herein.
It is an optional configuration diagram of processing system for video 100 provided in an embodiment of the present invention referring to Fig. 1, Fig. 1,
Both terminal 400 connects server 200 by network 300, and network 300 can be wide area network or local area network, or be
Combination realizes that data are transmitted using Radio Link.Operation has video processing applications program in terminal 400, and video processing is answered
It is provided with interface 410 with program, to receive the relevant operation of the synthetic video of user.
For being provided with video process apparatus provided in an embodiment of the present invention in server 200, exemplary answered at one
In, when terminal 400 needs synthetic video, target video and background video can be the video recorded using terminal, at this point,
Target video and background video can be sent to server by terminal 400, and request server 200 carries out Video Composition.At this point, clothes
Device 200 be engaged in after receiving target video and background video, using method for processing video frequency provided in an embodiment of the present invention, by target
Target object in video is split, and is prospect and using background video as background using the foreground video split, by prospect
Foreground video frame in video and the background video frame in background video are overlapped, and superimposed video frame is sealed
Dress, obtains synthetic video, the synthetic video after encapsulation is finally sent to terminal 400 again.
Such as: as shown in Figure 1, target video is video 101, background video is video 102, and terminal 400 is by 101 He of video
Video 102 is sent to server 200, and server 200 extracts the foreground video 104 with portrait 103 for prospect from video 101,
And by the foreground video frame 1041 of foreground video 104 (the background video frame including 1041-1 to 1041-n) and background video 102
1021 (including 1021-1 to 1021-n) is overlapped respectively, and obtaining the video frame 1051 of synthetic video 105, (including 1051-1 is extremely
1051-n), wherein n is the integer greater than 1.
To be provided with another exemplary application of video process apparatus provided in an embodiment of the present invention in server 200
In, when terminal 400 needs synthetic video, the identification information of target video and background video can be sent to server 200.Clothes
Business device 200 is based on the received identification information of institute and determines corresponding target video and background video, is mentioned using the embodiment of the present invention
The method for processing video frequency of confession splits the target object in target video, using the foreground video that splits as prospect simultaneously
Using background video as background, the foreground video frame in foreground video and the background video frame in background video are overlapped, and
Superimposed video frame is packaged, synthetic video is obtained, the video after encapsulation is finally sent to terminal 400 again.Terminal
400 can go out the video distribution after synthesis.
In an example using terminal 400 as electronic equipment, target video and background video be can be in terminal 400
The video file of encapsulation, by terminal 400 itself using the method provided in an embodiment of the present invention in video processing, by foreground video
In foreground video frame and background video in background video frame be overlapped, and superimposed video frame is packaged, is obtained
Video file after to synthesis.
It is carried out for being provided with video process apparatus provided in an embodiment of the present invention in server and terminal respectively above
Explanation, it is possible to understand that ground, video process apparatus provided in an embodiment of the present invention, which can be distributed, to be arranged in terminal and server, from
And method for processing video frequency provided in an embodiment of the present invention is completed by terminal and server collaboration.
It should be noted that in embodiments of the present invention, the type of target video and background video can be identical, it can also not
Together.Such as: target video and background video are encapsulated video file.For another example: target video is video flowing, background
Video is encapsulated video file.
Video process apparatus provided in an embodiment of the present invention may be embodied as the side of hardware, software or software and hardware combining
Formula.
As the example of software implementation, video process apparatus may include one or more software module, for independent
Or cooperative achievement method for processing video frequency provided in an embodiment of the present invention, software module can be using the various of various front ends or rear end
Programming language.
As the example that hardware is implemented, video process apparatus may include one or more hardware modules, and hardware module can
With using (ASIC, Application Specific Integrated Circuit), Complex Programmable Logic Devices (CPLD,
Complex Programmable Logic Device), field programmable gate array (F PGA, Field-Programmable
Gate Array) etc. hardware decoders, be programmed to realize video processing side provided in an embodiment of the present invention alone or synergistically
Method.
Below again by taking software and hardware combining as an example, illustrate the exemplary reality of video process apparatus provided in an embodiment of the present invention
It applies.
Referring to fig. 2, Fig. 2 is 20 1 optional structural schematic diagrams of video process apparatus provided in an embodiment of the present invention, root
According to Fig. 2 shows video process apparatus 20 structure, it is anticipated that other exemplary structures of video process apparatus 20, therefore
Structure as described herein is not construed as limiting, such as can be omitted members described below, alternatively, adding hereafter institute
The component that do not record is to adapt to the specific demands of certain applications.
Video process apparatus 20 shown in Fig. 2 includes: at least one processor 210, memory 240, at least one network
Interface 220 and user interface 230.Various components in video process apparatus 20 are coupled by bus system 250.It can manage
Solution, bus system 250 is for realizing the connection communication between these components.Bus system 250 is in addition to including data/address bus, also
Including power bus, control bus and status signal bus in addition.But for the sake of clear explanation, in Fig. 2 all by various buses
It is designated as bus system 250.
Memory 240 can be volatile memory or nonvolatile memory, may also comprise volatile and non-volatile
Both memories.Wherein, nonvolatile memory can be read-only memory (ROM, Read Only Memory), and volatibility is deposited
Reservoir can be random access memory (RAM, Random Access Memory).The memory of description of the embodiment of the present invention
240 are intended to include the memory of any suitable type.
Memory 240 in the embodiment of the present invention can storing data to support the operation of server 200.These data
Example includes: any computer program for operating on video process apparatus 20, such as operating system and application program.Its
In, operating system includes various system programs, such as ccf layer, core library layer, driving layer etc., for realizing various basic businesses
And the hardware based task of processing.Application program may include various application programs.
As the example that method for processing video frequency provided in an embodiment of the present invention uses software and hardware combining to implement, the present invention is implemented
Method provided by example can be embodied directly in be combined by the software module that processor 210 executes, and software module, which can be located at, deposits
In storage media, storage medium is located at memory 240, and processor 210 reads the executable finger that software module includes in memory 240
It enables, it is real to complete the present invention in conjunction with necessary hardware (e.g., including processor 210 and the other assemblies for being connected to bus 250)
The method for processing video frequency of example offer is provided.
The exemplary application and implementation of the video process apparatus above-mentioned for realizing the embodiment of the present invention will be combined, illustrates to realize
The method for processing video frequency of the embodiment of the present invention.It is to be appreciated that method for processing video frequency shown in Fig. 3 can be held by various electronic equipments
Row, such as executed by terminal or server, or, it is performed in unison with by terminal and server.
It is an optional flow diagram of method for processing video frequency provided in an embodiment of the present invention referring to Fig. 3, Fig. 3, it will
The step of showing in conjunction with Fig. 3 is illustrated.
Step S301 obtains target video.
Target video can be encapsulated video file.Target video can also be video flowing, such as: the stream of live video
Media data.
The quantity of target video can be one or more.
Operation has video processing applications program in terminal, and target video selection window is provided in video processing applications program
Mouthful, the identification information of alternative target video is provided in target video selection window, such as: video thumbnails, video name
Claim etc..Terminal receives the selection operation of user, and the corresponding video of identification information that selection operation is selected is as target video.Its
In, when electronic equipment is server, the video processing applications program run in terminal is the client of server.
For the target video of user's selection, target video can be presented at the terminal, so that user is to selected mesh
It marks video and carries out preview, whether be required target video with the selected target video of determination, if not needed for user
When the target video wanted, window can be selected to carry out reselecting for target video based on target video.
Illustratively, target video selection window can icon 401, icon as shown in window 401 in Fig. 4, in window 401
402, icon 403 is respectively the icon of alternative target video, when selection operation selectes icon 402, then the corresponding view of icon 402
Frequency is target video.In window 401 further include: the more multi-option 404 of the more alternative target videos of triggering, when more multi-option
404 when receiving the touch control operation of user, and the identification information of more alternative target videos can be presented.When acquisition target video
Afterwards, window 405 can be regard as the preview window, the picture of target video is presented in window 405.
Step S302, in response to the cutting operation for target object in the target video, from the target video
It obtains using the target object as the foreground video of prospect.
The segmentation entrance for receiving the cutting operation for being directed to target object can be loaded in terminal.Terminal can be based on receiving
Cutting operation generate instruction from target video obtain using the target object as the split order of the foreground video of prospect.
In one example, electronic equipment be terminal when, terminal be based on split order local from target video obtain before
Scape video.
In another example, when electronic equipment is server, split order is sent to server by terminal, and server is based on
The split order received obtains foreground video from target video.
Electronic equipment is based on split order, can call the interface of video encoder, be regarded target based on the interface called
Target video frame is decomposed into video frame by video encoder by frequency input video encoder.Electronic equipment to each video frame into
Row image recognition identifies target object from each video frame, obtains constituting foreground video based on the region where target object
Foreground video frame.Target object can be the object in the prospect in the video frames of target videos such as people, animal.Here, by structure
It is known as foreground video frame at the video frame of foreground video, the foreground video includes at least one foreground video frame.
It identifies from target video and is regarded by the video of prospect of target object, can be known at least one of in the following manner
Target object in the video frame of other target video:
Identification method one, calibration mode
Receive the proving operation for the video frame that user is directed in target video, pair that the proving operation of user is calibrated
As being determined as target object.Wherein, the object that the proving operation of user is calibrated can be specific object, for example, multiple
A people in people, is also possible to a class object, for example, male, women.
Identification method two, image recognition model automatic identification
Target object is used as by the prospect (such as people, animal) of image recognition model automatic identification video frame.
Step S303 obtains background video.
Background video can be encapsulated video file.Background video can also be video flowing.
The video processing applications program run in terminal can provide video selection window to receive user and select background view
The video selection of frequency operates, and the identification information of determining background video is operated based on the video selection of user.
Here, the video frame in background video is known as background video frame, the background video includes at least one background
Video frame.
It should be noted that in embodiments of the present invention, step S301, the execution sequence of step S302 and step S303 is not
Successively, step S301, step S302 can be first carried out, step S303 can also be first carried out.
Step S304 regards the prospect in response to being directed to the synthetic operation of the foreground video and the background video
Foreground video frame in frequency is overlapped with the background video frame in the background video, and the video frame that superposition is obtained encapsulates
For synthetic video.
The video processing applications program run in terminal can provide the interactive entrance of triggering Video Composition, to receive instruction
The synthetic operation that foreground video and background video are synthesized, and synthetic instruction is generated based on synthetic operation.
When electronic equipment is server, synthetic instruction is sent to server by terminal, server be based on synthetic instruction into
The superposition of foreground video frame in row foreground video and the background video frame in background video, to realize foreground video and background view
The synthesis of frequency.
As shown in Figure 5A, such as: when prospect video bag includes: video A', when background video is video D, by the view of video A'
Frequency frame is overlapped with the background video frame in background video, and Overlay can be as shown in Figure 5A, wherein background area 501 is
The picture of video D, object 502 are the corresponding region object a in the prospect in video A.
Electronic equipment carries out the background video frame of the foreground video frame of foreground video and background video according to synthetic parameters
Superposition.Here, can using relative position of the target object in target video and/or opposite imaging size as synthetic parameters,
The edit operation that user can be received based on edit page, so that user is adjusted synthetic parameters.
It should be noted that the behaviour of the user's operations such as selection operation, cutting operation, synthetic operation in the embodiment of the present invention
The mode of work can are as follows: touch-control, voice, gesture etc., the embodiment of the present invention is to the mode of operation of user without any restriction.
It will be the foreground video of prospect from target using target object in method for processing video frequency provided in an embodiment of the present invention
The view for being split in video, and the foreground video frame for the foreground video divided being superimposed with the background video frame of background video
Frequency frame is encapsulated as synthetic video, to synthesize based on the content of video to dynamic video, obtains image content coordination
Dynamic video.Here, it is based on image Segmentation Technology, target object is extracted from a video in real time, with another
Video is synthesized, and realizes two automatic syncretizing effects of video, the efficiency of energy significant increase user video production, and excite use
More interesting videos are created at family, allow ordinary user that can also produce the video of similar film special efficacy.
In some embodiments, when the quantity of target video is at least two, step S302 can be executed are as follows: reception is directed to
The batch splitting of at least described two target videos operates;It is operated in response to the batch splitting, from each target video
Middle acquisition is determined as corresponding foreground video using the target object as the video clip of prospect.
When the quantity of target video is multiple, cutting operation can operate for batch splitting.For multiple target videos, often
A target video can be identical as the target object of prospect, can not also be identical.Wherein, the corresponding target pair of different target videos
As that can be same class object.Here, the foreground video being partitioned into is the video clip that the target object in target video is constituted.
At this point, step S303 can be executed are as follows: receive at least two foreground videos and the background video
Batch synthetic operation;In response to the batch synthetic operation, the foreground video frame at least two foreground video is distinguished
It is overlapped with the background video frame in the background video.
Such as: when prospect video bag includes: video A', video B' and video C', background video be video D when, by video A',
The video frame of video B' and video C' are overlapped with the background video frame in background video jointly, and Overlay can be such as Fig. 5 B institute
Show, wherein background area 501 be video D picture, object 502, object 503 and object 504 be respectively video A, video B and
Object a, object b and the corresponding region object c in prospect in video C.
In some embodiments, when step S303 may be so performed that loaded and displayed has the video of alternative background video
Select window;The video selection received for the video selection window operates;Obtain the selected back of the video selection operation
Scape video.
It is provided with video selection window in the video processing applications program run in terminal, is shown in video selection window
The identification information of alternative background video.In video selection window the identification information of alternative background video can from local acquisition,
It can also be obtained from network side.Terminal is operated based on video selection window reception video selection, so that user is grasped based on video selection
Make the background video that the selection from alternative background video carries out Video Composition.
For the background video of user's selection, background video can be presented at the terminal, so that user is to selected back
Whether scape video carries out preview, be required background video with the selected background video of determination, if not needed for user
When the background video wanted, reselecting for background video can be carried out based on video selection window, to selected background video
It is replaced.Illustratively, video selection window can be as shown in window 401 in Fig. 4, here not to the selection course of background video
It is repeated again.
In some embodiments, in response to the preview operation for the foreground video and the background video, institute is presented
State the Overlay of foreground video frame Yu the background video frame.
The video processing applications program run in terminal can provide the interactive entrance for receiving preview operation, to receive instruction
The preview operation of the Overlay of preview foreground video and background video.
In some embodiments, as shown in fig. 6, after receiving cutting operation in step S302, following steps can be passed through
Foreground video is partitioned into from target video frame:
Step S3021, the target area where identifying the target object in the video frame of the target video, and will
Areas transparentization processing except target area described in the video frame;
The target of target object is identified from the video frame of target video by way of image recognition model or calibration
Region after identifying target area, keeps the pixel value for the pixel for belonging to target area constant, will belong to other than target area
The pixel value of pixel in region be set as 0, so that the areas transparentization except target area is handled, be partitioned into target video
Video frame in target object.
Video frame after transparency process is encapsulated as the foreground video by step S3022.
The foreground video frame of transparency process is encapsulated as foreground video based on Video Codec.
In some embodiments, step S3021 this can be implemented so that
Identify the target area where target object described in the video frame of the target video, and according to the target area
Domain obtains the corresponding image array of video frame of the target video;Element in described image matrix characterizes corresponding view respectively
The pixel of frequency frame belongs to the probability of the target area;Described image matrix is subjected to mask process with corresponding video frame, with
By the areas transparent in the video frame in addition to the target area.
Here, the target area where the target object in target video frame, image can be identified by image recognition model
Image array of the identification model based on the target area output binaryzation identified.Target can also be identified by the calibration of user
The target area where target object in video frame, and the image array of binaryzation is obtained according to determining target area.?
In image array, the corresponding element of pixel other than target area is 0, characterizes the pixel and is not belonging to target area, target area
The corresponding element of the pixel in domain is 1, characterizes the pixel and belongs to target area.By the video frame of image array and target video into
Line mask processing, the pixel value of the pixel of target area is constant, and the pixel value of the pixel in the region other than target area is
0, thus by the areas transparent in video frame in addition to the target area.
Here, image recognition model can be trained by carrying out the sample set of target object mark.Work as target object
When for portrait, the training sample in sample set can be as shown in fig. 7, be labeled portrait 702 in portrait picture 701.
In some embodiments, the background in the foreground video frame by the foreground video and the background video
Video frame is overlapped, comprising:
Obtain the timestamp alignment relation of the foreground video frame Yu background video frame;By the prospect in the foreground video
Video frame is overlapped with the background video frame for meeting the timestamp alignment relation in the background video.
Here, before by foreground video frame and the superposition of background video frame, each foreground video frame in foreground video is obtained
Timestamp, and the timestamp of each background video frame in background video is obtained, and according to the timestamp of acquisition, determine foreground video frame
With the timestamp alignment relation of background video frame, that is to say, that the pass between the period of foreground video and the period of background video
System, and by with timestamp alignment relation foreground video frame and background video frame be overlapped.Wherein, timestamp alignment relation
It can be and automatically determined according to the position of position and each background video frame on a timeline of each foreground video frame on a timeline,
The editting function that can also be provided based on video processing applications program is determined.Wherein, the editor that video processing applications program provides
Function can be existed based on the position of the timestamp of user adjustment operation adjustment foreground video frame on a timeline or foreground video frame
Position on time shaft.
Such as: background video when it is 2 minutes a length of, period on time shaft is 0 to 2 minute, foreground video when it is a length of
30 seconds, and its timestamp is aligned with the 1st point of 16 seconds to the 1st point 45 seconds this period of background video, then in foreground video frame
First frame and the 1st point of 16 seconds first frame of background video there is timestamp alignment relation, and correspond to frame by frame, by foreground video frame
In in each foreground video frame and background video the 1st point of each background video frame in 45 seconds 16 seconds to the 1st point be overlapped.Here, preceding
Scape video frame and the frame per second of background video frame can be identical.
For another example: ibid example is adjusted such as Fig. 8 institute the timestamp alignment relation of foreground video frame and background video frame
Show, before adjustment, the initial time of foreground video is aligned with the T1 of background video, wherein T1 be 1 point 16 seconds, foreground video
It is aligned with the 1st point of 16 seconds to the 1st point 45 seconds this period of background video.User is based on slidably control prolongs arrow shown in dotted line
Head shown in direction be adjusted, the initial time of foreground video is adjusted to the T2 of background video, wherein T2 be 1 point 06 second,
Then by time adjustment operation by initial position of the foreground video frame in the time shaft of background video by the 1st point adjust within 16 seconds to
1st point 06 second, the 1st point of the 1st point of 06- of the period of foreground video frame and background video frame is aligned for 35 seconds at this time, then is regarded prospect
The 1st point of each background video frame in 35 seconds 06 second to the 1st point is overlapped in each foreground video frame and background video in frequency frame.
In embodiments of the present invention, terminal can provide the slidably times such as control adjustment interface in user interface, so that
User passes through the starting that slidably times such as control adjustment interface selection foreground video is synthesized with background video on a user interface
Time, synthesis end time.It should be noted that the initial time of synthesis or the end time of synthesis are between background video
Between initial time and end time.When foreground video frame and background video frame are superimposed by electronic equipment, based on selected
It is background video frame that the initial time of synthesis, which starts decoding background video respectively, and based on the background video frame decomposed and prospect
Video frame is overlapped frame by frame, until the end time of synthesis.If between the initial time of synthesis and the end time of synthesis
Every the duration than foreground video, then it is subject to end time of foreground video.If the initial time of synthesis and the end of synthesis
The interval of time is shorter than the foreground video time, then be subject to selection synthesis end time, i.e., do not arrive also foreground video ending
Just terminate to synthesize.
In some embodiments, the background in the foreground video frame by the foreground video and the background video
Video frame is overlapped, comprising:
In response to the edit operation for the foreground video and background video setting synthetic parameters;By the prospect
Video frame covers the background video frame, and overlay area of the foreground video frame in the background video frame meets setting
Synthetic parameters.Synthetic parameters include at least one following parameter: position, size etc., are regarded with characterizing foreground video frame in background
The superposed positions such as relative position, relative size in frequency frame.
The video processing applications program run in terminal can provide edit page, can display foreground view in edit page
Background video frame in the foreground video frame and background video of frequency can show the prospect view with timestamp alignment relation here
Frequency frame and background video frame.
It is loaded with editor's interactive interface on editing interface, receives the edit operation of setting synthetic parameters, to set synthesis
Parameter.Wherein, edit operation can be the operations such as translation, rotation, scaling.
In practical applications, the editing interface for carrying out edit operation can be as shown in figure 9, provide one in editing interface 901
Rectangle frame 902 identical with the size of the target object in foreground video frame receives user to foreground video based on the rectangle frame
Edit operation.
When determine user complete edit operation after, can automatic trigger synthetic operation, may be based on user in the display interface
The interactive entrance of operation receives synthetic operation.Prospect is regarded based on the synthetic parameters of edit operation setting in response to synthetic operation
Frequency frame covers the background video frame, so that covering of the foreground video in background video frame meets the synthetic parameters of setting.
In some embodiments, synthetic parameters of the determination foreground video frame in the background video frame, packet
It includes: construction initial matrix identical with the foreground video frame sign;According to the edit operation in the initial matrix
Element is adjusted, and obtains the objective matrix for characterizing the variable quantity of the synthetic parameters.
Here, the matrix identical with the high width of target object in foreground video frame of construction, referred to as initial matrix.According to editor
Operation is adjusted initial matrix, obtains the objective matrix of the variable quantity of characterization synthetic parameters.When edit operation is translation,
The value of the corresponding element of pixel where the position of translation is then revised as displacement.When edit operation is scaling, then will
The value of the corresponding element of pixel where the position of scaling is revised as scaling.It, then will rotation when edit operation is rotation
Position where the corresponding element of pixel value be revised as scaling angle function.
Illustratively, when the high width of prospect video frame is 3, initial matrix can be the matrix of 3*3When translation
Objective matrix can beObjective matrix when translation can beObjective matrix when rotation can beWherein, tx、tyRespectively indicate the displacement along the translation of the direction x, y, sx、syRespectively indicate edge
X, the ratio of the direction y scaling, the q in sin (q)/cos (q) indicate the angle of rotation.Wherein, the ratio along the direction x, y scaling is
Indicate sx、syIndicate that 2-d spatial coordinate (x, y) scales s centered on (0,0) in the horizontal directionxTimes, it contracts in vertical direction
Put syTimes, that is to say, that the horizontal distance of coordinate position distance (0,0) becomes level of the former coordinate from place-centric point after transformation
The s of distancexTimes, vertical range becomes the s of vertical range of the former coordinate from place-centric pointyTimes.Wherein, 1,0 is without practical meaning
Justice is will to calculate the default parameters for being expressed as obtaining when math matrix.
In some embodiments, described that the foreground video frame is covered into the background video frame, and the foreground video
Overlay area of the frame in the background video frame meets the synthetic parameters of setting, comprising:
The objective matrix is multiplied with the foreground video frame in the foreground video, the foreground video after being adjusted
Frame;The foreground video frame adjusted is covered into the background video frame.
Here, objective matrix can be multiplied with the bitmap of foreground video frame, the bitmap of the foreground video frame after being adjusted.
Bitmap (Bitmap) is stored in a manner of the two-dimensional array of RGBA pixel.Coordinate position is p0 (x0, y0) in foreground video frame
Pixel when being converted, the parameters such as the displacement of transformation, scaling size are inputted into R-matrix, obtain corresponding objective matrix M
(x0, y0), then the coordinate position of the pixel is p1 (x1, y1) in the foreground video frame after adjustment, and the calculating of p1 (x1, y1) is public
Formula are as follows:
P1 (x1, y1)=p0 (x0, y0) * M (x0, y0);
Wherein, p0 (x0, y0) is with the transposition [x y] of matrix [x y]TIt is calculated.
Such as: when a space coordinate p0 (x0, y0) first prolongs the direction x translation tx, then prolong the direction y translation tyThe seat finally obtained
Mark p1 (x1, y1)=(x0+tx, y0+ty), it, can when being indicated with the form of matrix are as follows:
Each pixel can obtain a new coordinate position in foreground video frame, to obtain a new pixel two
Dimension group, the bitmap that can be reduced to by this two-dimensional array after new Bitmap, that is, adjustment.
Method for processing video frequency provided in an embodiment of the present invention is capable of providing edit page, and is received and used based on edit page
Edit operation of the family to foreground video frame, when adjustment foreground video frame is synthesized with background video frame, relative to background video frame
Relative position and imaging size.
Illustratively, by taking electronic equipment is using Android platform as an example, to video encoder involved in the embodiment of the present invention
It is illustrated, the encoding and decoding framework of video encoder is as shown in Figure 10:
Codec, which can handle, enters data to generate output data, and codec uses one group of input buffer and defeated
Buffer carrys out asynchronous process data out.An empty input block can be created by loader, to send after filling data
It is handled to codec.The input data that codec provides client is converted, be then output to one it is empty
Output buffer.Last client gets the data of output buffer, consumes the data of the inside, and the output of occupancy is buffered
Area is released back into codec.If subsequent, there are also data needs to continue with, and codec will repeat these operations.
The data type that codec can be handled can include: compressed data and original video data.Buffer area can be passed through
(ByteBuffers) these data are handled, screen buffer (Surface) is needed to open up original video data at this time
Show, can also improve the performance of encoding and decoding in this way.Surface can be used local screen buffer, this buffer area do not map or
Copy ByteBuffers to.Such mechanism allows the more efficient of codec.It, can not usually when using Surface
Original video data is accessed, but cis (ImageReader) can be used and access decoded original video frame.
In the following, electronic equipment is that terminal is actual application scenarios using target object as portrait, it will illustrate implementation of the present invention
Exemplary application of the example in actual application scenarios.
The relevant technologies, Video Composition scheme can be as shown in figure 11, comprising: selection background template 1101.Background template 1101
For a background video.The portrait picture 1102 that picture material includes portrait is chosen, selects portrait figure in such a way that user smears
The portrait area of piece 1102, based on user selection portrait area, divided by AI, by portrait picture 1102 be divided into portrait and
Background two parts show that portrait 1103 forms editor's image 1104, and right to pluck out portrait 1103 on background template 1101
Show that portrait 1103 and the position of background template 1101 are edited in editor's image 1104, after completing the editing, and based on volume
The synthetic parameters collected are merged to by portrait 1103 with background template 1101, obtain the synthetic video that picture material includes portrait
1105。
The synthetic effect of Video Composition scheme shown in Figure 11 is as shown in figure 12, by the people in static portrait picture 1102
As 1103 are synthesized in background template 1101, the display page 1105 of synthetic video 1104 is obtained.Video Composition shown in Figure 11
Scheme is to carry out portrait background segment to static picture, is then synthesized in video, limitation is larger, first is that must have
The good template background video of particular production, second is that can only be split for static images, the portrait plucked out is static, forfeiture
Many interests.In addition the portrait segmentation of picture is needed to carry out region manually to smear selection, there is multiframe for processing
The video efficiency of image is too low.
In the related technology, user is when watching some short-sighted frequency, can initiate video and be in step with function, two videos are carried out
Synthesis, is combined into a video with frame.The technic relization scheme that video is in step with is that two videos are directly carried out left and right splicing,
Two videos can seem more stiff because scene is different.The effect that video is in step with is as shown in figure 13, wherein picture 1301 is
The picture for one of video that video is in step with, picture 1302 are the pictures of another video.
Therefore, the be in step with scheme of function of video is that simply two videos or so are stitched together, and synthesis one is bigger
Video, the background of two videos is not handled, there are two scenes for the video of synthesis, it appears loftier.
In order to which the limitation for solving the above-mentioned Video Composition scheme for being only capable of being synthesized to static picture in video is fixed big, or will
The lofty technological deficiency of the Video Composition scheme scene that two videos are spliced, the embodiment of the present invention provide a kind of video processing
Method, comprising: video selection, video decoding, portrait segmentation, picture editting, Video Composition, as shown in figure 14, comprising:
Video selection is carried out from local video, obtains background video 1401 and portrait video 1402 i.e. target video.
Video decoding is carried out to background video 1401, obtains video frame 1403 i.e. background video frame.Video is carried out to portrait video 1402
Decoding, obtains video frame 1404.Video frame 1404 is inputted into neural network model 1405 and carries out portrait segmentation, exports portrait
Mask Figure 140 6 obtains portrait image 1407 i.e. prospect by mask Figure 140 6 of portrait and the mask process of video frame 1404
Video frame.
When receiving beginning edit operation, by background video frame 1402 and the display of portrait image 1407 in editing interface
On.Portrait image 1407 on editing interface receives the edit operation of user, based on the edit operation of user to portrait image
The relative position of 1407 relative video frames 1402 and relative size are adjusted, and obtain relativeness, and be based on relativeness people
As image is edited.Wherein, edit operation portrait image 1407 carried out can include: the processing such as translation, scaling and rotation.
When receiving preview operation, edited portrait image 1408 and video frame 1402 are rendered, output is used as portrait figure
As 1407 and the Overlay 1409 of video frame 1402.When receiving synthetic operation, again by rendering to edited people
As image 1408 and video frame 1402 are rendered, synthetic frame 1410 is obtained, again by Multimedia Encoder by synthetic frame
1410 are packaged into synthetic video 1411.Wherein, it after exporting Overlay 1409, can also continue to receive edit operation, to portrait figure
As the relative position of 1407 relative video frames 1402 and relative size are adjusted.
Terminal device can show video selection option by system photograph album or the customized photograph album page, based on shown choosing
Item selection background video 1401 and portrait video 1402.Terminal device is regarded background video 1401 and portrait by MediaCodec
Frequently 1402 it is decoded into multiple single-frame images respectively.Each frame image decoded to portrait video 1402 carries out portrait segmentation, obtains
To portrait image 1407.Terminal device passes through each frame portrait image 1407 of the target Matrix to segmentation of characterization synthetic parameters
Bitmap carry out matrixing, the Bitmap of the portrait image after being edited, then Bitmap after editor is passed through
OpenGL ES API uploads to the texture cell of graphics processing unit (Graphic Process Unit, GPU), terminal device
GPU by tinter by the corresponding texture of background video frame 702 and editor after portrait image texture carry out image mix
Closing operation obtains final synthetic frame, and synthetic frame is encoded into synthetic video by MediaCodec.
In the following, to being described with the next stage in method for processing video frequency provided in an embodiment of the present invention: portrait segmentation, figure
Piece editor, rendering, the decoding of video and synthesis.
1, portrait is divided
In the server, using include multiple portrait class pictures manually marked set as training set, to neural network
Model is trained, and is transplanted to terminal by the neural network model after training and preservation, and by the neural network model after training
In equipment.
Server collects portrait class picture, and is labeled by portrait class picture of the artificial mode to collection, will
The corresponding region of portrait is as prospect in portrait class picture, and using the region other than portrait as background, by foreground and background
Each pixel distinguishes.The portrait class picture manually marked can be as shown in fig. 7, carry out portrait 702 in portrait picture 701
Mark.
For dynamic target video, target video is decoded into real time by static state by Video Decoder (MediaCodec)
Frame, then the neural network model by static frames input training is handled, and is returned to the picture mask (binary map) of segmentation, is passed through
Original image in mask and target video frame carries out transparency blending, in the portrait i.e. foreground video after can cutting out segmentation
Foreground video frame.
2, picture editor
User it is divisible go out portrait picture translate, scale, rotate etc. and edit.User can be according to self-demand to volume
The position and size collected are controlled.
The portrait picture being partitioned into is stored in a manner of Bitmap in memory, can be by Matrix matrix to storage
Bitmap is converted.By constructing a rectangle frame equal with the wide height of portrait picture, then it is supplied on graphical interfaces
The interactive entrance of user dragging and rotation, available user edit the Matrix i.e. objective matrix of rectangle frame generation, will be original
Portrait picture pixels be multiplied with Matrix, the Bitmap after transformation such as available translation, scaling and rotation.
3, it renders
Portrait can carry out live preview after splitting in target video.In addition in editor's size and location information
It later, can also live preview.In live preview, OpenGL ES has been adopted in the terminal and has been rendered, by the image of each frame
RGB data uploads to the texture cell of GPU, is then rendered by the rendering pipeline of GPU, and final GPU can be by the figure of output
As data render to screen Frame Buffer in, to be shown on screen.
Have efficient parallel processing and rendering framework based on GPU, be very suitable to the processing and rendering of image, therefore, leads to
The API of OpenGL ES is crossed to utilize GPU to be rendered, may be implemented to achieve the purpose that real-time rendering special efficacy.
4, video decoding and synthesis
By the video frame being decoded to obtain to video in video, to handled frame by frame video.It is merging most
When the video at end, need using Video Composition technology, that is, Video coding.
By taking Android platform as an example, coding and decoding video can be carried out based on the MediaCodec module on Android.
Method for processing video frequency provided in an embodiment of the present invention can be in the video processing applications program of mobile video and live streaming class
In, the multiple videos of rapid synthesis are used for, entertaining video editing efficiency is promoted.As shown in figure 15, user when in use, can be successively
Select multiple videos, such as: video 151-1, video 151-2 ... video 151-m, wherein video 151-2 be template video (i.e.
Background video).User clicks the interactive interface that portrait is scratched for a key that video processing applications program provides, image application program
To other than place video 151-2 video 151-1 ... each video carries out portrait segmentation in video 151-m, corresponding to generate multiple portraits
Video: portrait video 152-1 ... portrait video 152-m.Wherein, in portrait video, the background area of non-portrait is transparent.User
It can be sequentially adjusted in size and relative position of the portrait video relative to template video (may include a background), and live preview melts
Close effect.When image, which is applied to figure, receives user's click synthesis button, Video Composition is carried out, by portrait video 152-1 ...
Portrait video 152-m is blended into jointly in video 151-2, obtains synthetic video 153, and can protect the video 153 that synthesis is completed
It is stored to local.
The video matting interface that video processing applications program provides can be currently in work such as 1601 in Figure 16, window 1602
That makees area needs to be scratched portrait list of videos, in addition there is the button 1603 of one " stingy portrait " and the button of " starting to edit "
1604.Portrait in the video currently selected can be extracted, and shown in real time by " stingy portrait " button 1603 when the user clicks
Show in preview area, clicks " starting to edit " button 1604, editing interface 1605 of entering.Editing interface 1605 is for editing portrait
In addition there are " replacement background video " button 1608 and one in the relative position and size of video 1606 and background video 1607
" starting to synthesize " button 1609, " replacement background video " button 1608 for replacing the background video currently selected, " start to close
At " button 1609 is for starting final Video Composition.
Method for processing video frequency provided in an embodiment of the present invention by carrying out automatic portrait background segment to dynamic video, this
The portrait that sample plucks out is movement and fresh and alive image, and user is allowed to select arbitrary background video, and portrait is synthesized to this
In video, to realize two sections even fusion of multistage video.For example, the video danced indoors using two sections of users,
Two portraits wherein respectively are plucked out, are synthesized in the scene video of another stage, to realize the cooperation of two people strange lands
The effect of performance.Due to having carried out portrait background segment, the video scene finally synthesized is unified, therefore this programme introduces more
Writing space, the imagination and creativity of user can be sufficiently excited, to promote the playability and interest of software entirety.
Illustrate the exemplary structure of software module below, in some embodiments, as shown in Fig. 2, in video process apparatus
Software module may include:
First acquisition unit 2401, for obtaining target video;
Cutting unit 2402, for the cutting operation in response to being directed to target object in the target video, from the mesh
It marks in video and obtains using the target object as the foreground video of prospect;The foreground video includes at least one foreground video
Frame;
Second acquisition unit 2403, for obtaining background video;The background video includes at least one background video frame;
Synthesis unit 2404, for the synthetic operation in response to being directed to the foreground video and the background video, by institute
The foreground video frame stated in foreground video is overlapped with the background video frame in the background video, and the view that superposition is obtained
Frequency frame is encapsulated as synthetic video.
In some embodiments, cutting unit 2402 are also used to:
The batch splitting received at least described two target videos operates;It is operated in response to the batch splitting, from
It obtains in each target video using the target object as the video clip of prospect, and is determined as corresponding foreground video.
In some embodiments, synthesis unit 2404 are also used to:
Receive the batch synthetic operation for being directed at least two foreground videos and the background video;In response to described batch
Measure synthetic operation, the background that the foreground video frame at least two foreground video is added to respectively in the background video
In video frame.
In some embodiments, second acquisition unit 2403 are also used to:
Loaded and displayed has the video selection window of alternative background video;Receive the video for being directed to the video selection window
Selection operation;Obtain the selected background video of the video selection operation.
In some embodiments, described device further include: preview unit is used for:
In response to being directed to the preview operation of the foreground video and the background video, the foreground video frame and institute is presented
State the Overlay of background video frame.
In some embodiments, cutting unit 2402 are also used to:
Target area where identifying the target object in the video frame of the target video, and by the video frame
Described in areas transparentization processing except target area;Video frame after transparency process is encapsulated as the foreground video.
In some embodiments, cutting unit 2402 are also used to:
Identify the target area where target object described in the video frame of the target video, and according to the target area
Domain obtains the corresponding image array of video frame of the target video;Element in described image matrix characterizes corresponding view respectively
The pixel of frequency frame belongs to the probability of the target area;Described image matrix is subjected to mask process with corresponding video frame, with
By the areas transparent in the video frame in addition to the target area.
In some embodiments, synthesis unit 2403 are also used to:
Obtain the timestamp alignment relation of the foreground video frame Yu background video frame;By the prospect in the foreground video
Video frame is overlapped with the background video frame for meeting the timestamp alignment relation in the background video.
In some embodiments, the synthesis unit 2403, is also used to:
In response to the edit operation for the foreground video and background video setting synthetic parameters, by the prospect
Video frame covers the background video frame, and overlay area of the foreground video frame in the background video frame meets setting
Synthetic parameters.
In some embodiments, synthesis unit 2403 are also used to:
Construct initial matrix identical with the foreground video frame sign;According to the edit operation to the initial matrix
In element be adjusted, obtain the objective matrix for characterizing the variable quantity of the synthetic parameters.
In some embodiments, synthesis unit 2403 are also used to:
The objective matrix is multiplied with the foreground video frame in the foreground video, the foreground video after being adjusted
Frame;The foreground video frame adjusted is covered into the background video frame.
The example of hardware implementation, method provided by the embodiment of the present invention are used as method provided in an embodiment of the present invention
The processor 410 of hardware decoding processor form can be directly used to execute completion, for example, dedicated by one or more application
Integrated circuit (ASIC, Application Specific Integrated Circuit), DSP, programmable logic device
(PLD, Programmable Logic Device), Complex Programmable Logic Devices (CPLD, Complex Programmable
Logic Device), field programmable gate array (FPGA, Field-Programmable Gate Array) or other electronics
Element, which executes, realizes method provided in an embodiment of the present invention.
The embodiment of the present invention provides a kind of storage medium for being stored with executable instruction, wherein it is stored with executable instruction,
When executable instruction is executed by processor, processor will be caused to execute method provided in an embodiment of the present invention, for example, such as Fig. 3
The method shown.
In some embodiments, executable instruction can use program, software, software module, the form of script or code,
By any form of programming language (including compiling or interpretative code, or declaratively or process programming language) write, and its
It can be disposed by arbitrary form, including be deployed as independent program or be deployed as module, component, subroutine or be suitble to
Calculate other units used in environment.
As an example, executable instruction can with but not necessarily correspond to the file in file system, can be stored in
A part of the file of other programs or data is saved, for example, being stored in hypertext markup language (HTML, Hyper Text
Markup Language) in one or more scripts in document, it is stored in the single file for being exclusively used in discussed program
In, alternatively, being stored in multiple coordinated files (for example, the file for storing one or more modules, subprogram or code section).
As an example, executable instruction can be deployed as executing in a calculating equipment, or it is being located at one place
Multiple calculating equipment on execute, or, be distributed in multiple places and by multiple calculating equipment of interconnection of telecommunication network
Upper execution.
It will using target object be foreground video from target video in target video in conclusion through the embodiment of the present invention
In split, and the foreground video frame of the foreground video to be divided is as prospect and using the background video frame of background video as back
Scape carries out the synthesis of video frame, and the video frame of synthesis is encapsulated as synthetic video, thus based on the content of video to dynamic view
Frequency is synthesized, and the dynamic video of image content coordination is obtained.A key operation based on display interface, to multiple target videos into
The dividing processing of row mass target object.Also, editing interface is provided a user, the edit operation based on user, to prospect
Video is edited relative to the position of background video and imaging size.
The above, only the embodiment of the present invention, are not intended to limit the scope of the present invention.It is all in this hair
Made any modifications, equivalent replacements, and improvements etc. within bright spirit and scope, be all contained in protection scope of the present invention it
It is interior.
Claims (14)
1. a kind of method for processing video frequency characterized by comprising
Obtain target video;
In response to the cutting operation for target object in the target video, obtain from the target video with the target
Object is the foreground video of prospect;The foreground video includes at least one foreground video frame;
Obtain background video;The background video includes at least one background video frame;
In response to being directed to the synthetic operation of the foreground video and the background video, by the foreground video in the foreground video
Frame is overlapped with the background video frame in the background video, and
The video frame that superposition obtains is encapsulated as synthetic video.
2. the method according to claim 1, wherein described in response to for target object in the target video
Cutting operation, from the target video obtain using the target object as the foreground video of prospect, comprising:
The batch splitting received at least two target videos operates;
It operates in response to the batch splitting, obtains from each target video using the target object as the video of prospect
Segment, and it is determined as corresponding foreground video.
3. the method according to claim 1, wherein described in response to being directed to the foreground video and the background
The synthetic operation of video folds the foreground video frame in the foreground video with the background video frame in the background video
Add, comprising:
Receive the batch synthetic operation for being directed at least two foreground videos and the background video;
In response to the batch synthetic operation, the foreground video frame at least two foreground video is added to respectively described
In background video frame in background video.
4. the method according to claim 1, wherein the acquisition background video, comprising:
Loaded and displayed has the video selection window of alternative background video;
The video selection received for the video selection window operates;
Obtain the selected background video of the video selection operation.
5. the method according to claim 1, wherein the method also includes:
In response to being directed to the preview operation of the foreground video and the background video, the foreground video frame and the back is presented
The Overlay of scape video frame.
6. the method according to claim 1, wherein described obtain from the target video with the target pair
As the foreground video for prospect, comprising:
Target area where identifying the target object in the video frame of the target video, and by institute in the video frame
State the areas transparentization processing except target area;
Video frame after transparency process is encapsulated as the foreground video.
7. according to the method described in claim 6, it is characterized in that, it is described from the video frame of the target video identification described in
Target area where target object, and the areas transparentization except target area described in the video frame is handled, comprising:
It identifies the target area where target object described in the video frame of the target video, and is obtained according to the target area
To the corresponding image array of video frame of the target video;Element in described image matrix characterizes corresponding video frame respectively
Pixel belong to the probability of the target area;
Described image matrix and corresponding video frame are subjected to mask process, will in the video frame except the target area it
Outer areas transparent.
8. method according to any one of claims 1 to 7, which is characterized in that the prospect by the foreground video
Video frame is overlapped with the background video frame in the background video, comprising:
Obtain the timestamp alignment relation of the foreground video frame Yu background video frame;
The background of the timestamp alignment relation will be met in foreground video frame and the background video in the foreground video
Video frame is overlapped.
9. method according to any one of claims 1 to 7, which is characterized in that the prospect by the foreground video
Video frame is overlapped with the background video frame in the background video, comprising:
In response to the edit operation for the foreground video and background video setting synthetic parameters, by the foreground video
Frame covers the background video frame, and overlay area of the foreground video frame in the background video frame meets the conjunction of setting
At parameter.
10. according to the method described in claim 9, it is characterized in that, the method also includes:
Construct initial matrix identical with the foreground video frame sign;
The element in the initial matrix is adjusted according to the edit operation, obtains the variation for characterizing the synthetic parameters
The objective matrix of amount.
11. according to the method described in claim 10, it is characterized in that, described cover the background view for the foreground video frame
Frequency frame, and overlay area of the foreground video frame in the background video frame meets the synthetic parameters of setting, comprising:
The objective matrix is multiplied with the foreground video frame in the foreground video, the foreground video frame after being adjusted;
The foreground video frame adjusted is covered into the background video frame.
12. a kind of video process apparatus characterized by comprising
First acquisition unit, for obtaining target video;
Cutting unit, for the cutting operation in response to being directed to target object in the target video, from the target video
It obtains using the target object as the foreground video of prospect;The foreground video includes at least one foreground video frame;
Second acquisition unit, for obtaining background video;The background video includes at least one background video frame;
Synthesis unit regards the prospect for the synthetic operation in response to being directed to the foreground video and the background video
Foreground video frame in frequency is overlapped with the background video frame in the background video, and the video frame that superposition is obtained encapsulates
For synthetic video.
13. a kind of video process apparatus characterized by comprising
Memory, for storing executable instruction;
Processor when for executing the executable instruction stored in the memory, is realized described in any one of claim 1 to 11
Method for processing video frequency.
14. a kind of storage medium, which is characterized in that being stored with executable instruction, when for causing processor to execute, realizing right
It is required that 1 to 11 described in any item method for processing video frequency.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910691577.4A CN110290425B (en) | 2019-07-29 | 2019-07-29 | Video processing method, device and storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910691577.4A CN110290425B (en) | 2019-07-29 | 2019-07-29 | Video processing method, device and storage medium |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN110290425A true CN110290425A (en) | 2019-09-27 |
| CN110290425B CN110290425B (en) | 2023-04-07 |
Family
ID=68024154
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201910691577.4A Active CN110290425B (en) | 2019-07-29 | 2019-07-29 | Video processing method, device and storage medium |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN110290425B (en) |
Cited By (43)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110891181A (en) * | 2019-12-03 | 2020-03-17 | 广州酷狗计算机科技有限公司 | Live broadcast picture display method and device, storage medium and terminal |
| CN111145192A (en) * | 2019-12-30 | 2020-05-12 | 维沃移动通信有限公司 | Image processing method and electronic device |
| CN111179386A (en) * | 2020-01-03 | 2020-05-19 | 广州虎牙科技有限公司 | Animation generation method, device, equipment and storage medium |
| CN111242962A (en) * | 2020-01-15 | 2020-06-05 | 中国平安人寿保险股份有限公司 | Method, device and equipment for generating remote training video and storage medium |
| CN111447389A (en) * | 2020-04-22 | 2020-07-24 | 广州酷狗计算机科技有限公司 | Video generation method, device, terminal and storage medium |
| CN111464761A (en) * | 2020-04-07 | 2020-07-28 | 北京字节跳动网络技术有限公司 | Video processing method and device, electronic equipment and computer readable storage medium |
| CN111464865A (en) * | 2020-06-18 | 2020-07-28 | 北京美摄网络科技有限公司 | Video generation method and device, electronic equipment and computer readable storage medium |
| CN111629151A (en) * | 2020-06-12 | 2020-09-04 | 北京字节跳动网络技术有限公司 | Video co-shooting method and device, electronic equipment and computer readable medium |
| CN111669638A (en) * | 2020-02-28 | 2020-09-15 | 海信视像科技股份有限公司 | Video rotation playing method and display equipment |
| CN111722902A (en) * | 2020-06-15 | 2020-09-29 | 朱利戈 | Method and system for realizing rich media interactive teaching based on window transparent processing |
| CN111726536A (en) * | 2020-07-03 | 2020-09-29 | 腾讯科技(深圳)有限公司 | Video generation method, device, storage medium and computer equipment |
| CN111783729A (en) * | 2020-07-17 | 2020-10-16 | 商汤集团有限公司 | Video classification method, device, equipment and storage medium |
| CN112037227A (en) * | 2020-09-09 | 2020-12-04 | 脸萌有限公司 | Video shooting method, device, equipment and storage medium |
| CN112087663A (en) * | 2020-09-10 | 2020-12-15 | 北京小糖科技有限责任公司 | Method for generating dance video with adaptive light and shade environment by mobile terminal |
| CN112087662A (en) * | 2020-09-10 | 2020-12-15 | 北京小糖科技有限责任公司 | Method for generating dance combination dance video by mobile terminal |
| CN112087664A (en) * | 2020-09-10 | 2020-12-15 | 北京小糖科技有限责任公司 | A method for obtaining a real-time dance video with a customized background on a mobile terminal |
| CN112118397A (en) * | 2020-09-23 | 2020-12-22 | 腾讯科技(深圳)有限公司 | Video synthesis method, related device, equipment and storage medium |
| CN112132931A (en) * | 2020-09-29 | 2020-12-25 | 新华智云科技有限公司 | Processing method, device and system for templated video synthesis |
| CN112235520A (en) * | 2020-12-07 | 2021-01-15 | 腾讯科技(深圳)有限公司 | Image processing method and device, electronic equipment and storage medium |
| CN112261313A (en) * | 2020-09-22 | 2021-01-22 | 网娱互动科技(北京)股份有限公司 | Method for making video with replaceable foreground |
| CN112653851A (en) * | 2020-12-22 | 2021-04-13 | 维沃移动通信有限公司 | Video processing method and device and electronic equipment |
| CN112752116A (en) * | 2020-12-30 | 2021-05-04 | 广州繁星互娱信息科技有限公司 | Display method, device, terminal and storage medium of live video picture |
| CN112752038A (en) * | 2020-12-28 | 2021-05-04 | 广州虎牙科技有限公司 | Background replacing method and device, electronic equipment and computer readable storage medium |
| CN112822542A (en) * | 2020-08-27 | 2021-05-18 | 腾讯科技(深圳)有限公司 | Video synthesis method and device, computer equipment and storage medium |
| CN112860209A (en) * | 2021-02-03 | 2021-05-28 | 合肥宏晶微电子科技股份有限公司 | Video overlapping method and device, electronic equipment and computer readable storage medium |
| CN113067983A (en) * | 2021-03-29 | 2021-07-02 | 维沃移动通信(杭州)有限公司 | Video processing method and device, electronic equipment and storage medium |
| CN113256499A (en) * | 2021-07-01 | 2021-08-13 | 北京世纪好未来教育科技有限公司 | Image splicing method, device and system |
| CN113810624A (en) * | 2021-09-18 | 2021-12-17 | 维沃移动通信有限公司 | Video generation method and device and electronic equipment |
| CN113873319A (en) * | 2021-09-27 | 2021-12-31 | 维沃移动通信有限公司 | Video processing method and device, electronic equipment and storage medium |
| CN114040129A (en) * | 2021-11-30 | 2022-02-11 | 北京字节跳动网络技术有限公司 | Video generation method, device, equipment and storage medium |
| CN114189635A (en) * | 2021-12-01 | 2022-03-15 | 惠州Tcl移动通信有限公司 | Video processing method and device, mobile terminal and storage medium |
| CN114339401A (en) * | 2021-12-30 | 2022-04-12 | 北京翼鸥教育科技有限公司 | A kind of video background processing method and device |
| CN114363697A (en) * | 2022-01-06 | 2022-04-15 | 上海哔哩哔哩科技有限公司 | Video file generation and playing method and device |
| CN114663633A (en) * | 2022-03-24 | 2022-06-24 | 航天宏图信息技术股份有限公司 | AR virtual live broadcast method and system |
| CN114900736A (en) * | 2022-03-28 | 2022-08-12 | 网易(杭州)网络有限公司 | Video generation method and device and electronic equipment |
| CN115100259A (en) * | 2022-05-27 | 2022-09-23 | 深圳市大头兄弟科技有限公司 | A video remapping method and related equipment |
| WO2022242497A1 (en) * | 2021-05-20 | 2022-11-24 | 北京字跳网络技术有限公司 | Video photographing method and apparatus, electronic device, and storage medium |
| CN115623244A (en) * | 2022-10-08 | 2023-01-17 | 苏州市广播电视总台 | A video synthesis method and system based on real-time video stream portrait matting |
| WO2023036160A1 (en) * | 2021-09-07 | 2023-03-16 | 上海商汤智能科技有限公司 | Video processing method and apparatus, computer-readable storage medium, and computer device |
| CN116229337A (en) * | 2023-05-10 | 2023-06-06 | 瀚博半导体(上海)有限公司 | Method, apparatus, system, device and medium for video processing |
| CN117372238A (en) * | 2023-09-21 | 2024-01-09 | 启朔(深圳)科技有限公司 | Registration function method, video frame processing method, device, equipment and storage medium |
| CN117714787A (en) * | 2024-02-05 | 2024-03-15 | 哈尔滨学院 | A video data processing method |
| WO2024088322A1 (en) * | 2022-10-28 | 2024-05-02 | 影石创新科技股份有限公司 | Video display method and apparatus, video display device, and storage medium |
Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090245571A1 (en) * | 2008-03-31 | 2009-10-01 | National Taiwan University | Digital video target moving object segmentation method and system |
| CN101588445A (en) * | 2009-06-09 | 2009-11-25 | 宁波大学 | Video area-of-interest exacting method based on depth |
| US20130170760A1 (en) * | 2011-12-29 | 2013-07-04 | Pelco, Inc. | Method and System for Video Composition |
| EP2627097A2 (en) * | 2012-02-13 | 2013-08-14 | Acer Incorporated | Video/audio switching in a computing device |
| CN103279494A (en) * | 2013-05-03 | 2013-09-04 | 吴军 | Dynamic video analysis moving target retrieval system |
| US20160198097A1 (en) * | 2015-01-05 | 2016-07-07 | GenMe, Inc. | System and method for inserting objects into an image or sequence of images |
| US20170094194A1 (en) * | 2015-09-28 | 2017-03-30 | Gopro, Inc. | Automatic composition of video with dynamic background and composite frames selected based on frame and foreground object criteria |
| CN106572385A (en) * | 2015-10-10 | 2017-04-19 | 北京佳讯飞鸿电气股份有限公司 | Image overlaying method for remote training video presentation |
| WO2018059206A1 (en) * | 2016-09-29 | 2018-04-05 | 努比亚技术有限公司 | Terminal, method of acquiring video, and data storage medium |
| CN108200359A (en) * | 2017-12-13 | 2018-06-22 | 苏州长风航空电子有限公司 | A kind of multi-standard video frequency superimposer for airborne indicator |
| CN108259781A (en) * | 2017-12-27 | 2018-07-06 | 努比亚技术有限公司 | image synthesizing method, terminal and computer readable storage medium |
| CN108694738A (en) * | 2017-04-01 | 2018-10-23 | 英特尔公司 | The multilayer of decoupling renders frequency |
| CN109146827A (en) * | 2018-08-24 | 2019-01-04 | 合肥景彰科技有限公司 | A kind of image processing method and device in video fusion |
| CN110050461A (en) * | 2016-09-02 | 2019-07-23 | 罗素·霍姆斯 | System and method for delivering real-time composite video from multi-source devices featuring augmented reality elements |
-
2019
- 2019-07-29 CN CN201910691577.4A patent/CN110290425B/en active Active
Patent Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090245571A1 (en) * | 2008-03-31 | 2009-10-01 | National Taiwan University | Digital video target moving object segmentation method and system |
| CN101588445A (en) * | 2009-06-09 | 2009-11-25 | 宁波大学 | Video area-of-interest exacting method based on depth |
| US20130170760A1 (en) * | 2011-12-29 | 2013-07-04 | Pelco, Inc. | Method and System for Video Composition |
| EP2627097A2 (en) * | 2012-02-13 | 2013-08-14 | Acer Incorporated | Video/audio switching in a computing device |
| CN103279494A (en) * | 2013-05-03 | 2013-09-04 | 吴军 | Dynamic video analysis moving target retrieval system |
| US20160198097A1 (en) * | 2015-01-05 | 2016-07-07 | GenMe, Inc. | System and method for inserting objects into an image or sequence of images |
| US20170094194A1 (en) * | 2015-09-28 | 2017-03-30 | Gopro, Inc. | Automatic composition of video with dynamic background and composite frames selected based on frame and foreground object criteria |
| CN106572385A (en) * | 2015-10-10 | 2017-04-19 | 北京佳讯飞鸿电气股份有限公司 | Image overlaying method for remote training video presentation |
| CN110050461A (en) * | 2016-09-02 | 2019-07-23 | 罗素·霍姆斯 | System and method for delivering real-time composite video from multi-source devices featuring augmented reality elements |
| WO2018059206A1 (en) * | 2016-09-29 | 2018-04-05 | 努比亚技术有限公司 | Terminal, method of acquiring video, and data storage medium |
| CN108694738A (en) * | 2017-04-01 | 2018-10-23 | 英特尔公司 | The multilayer of decoupling renders frequency |
| CN108200359A (en) * | 2017-12-13 | 2018-06-22 | 苏州长风航空电子有限公司 | A kind of multi-standard video frequency superimposer for airborne indicator |
| CN108259781A (en) * | 2017-12-27 | 2018-07-06 | 努比亚技术有限公司 | image synthesizing method, terminal and computer readable storage medium |
| CN109146827A (en) * | 2018-08-24 | 2019-01-04 | 合肥景彰科技有限公司 | A kind of image processing method and device in video fusion |
Non-Patent Citations (1)
| Title |
|---|
| 郎洪: ""复杂场景下的交通视频显著性前景目标提取"", 《中国图像图形学报》 * |
Cited By (76)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110891181A (en) * | 2019-12-03 | 2020-03-17 | 广州酷狗计算机科技有限公司 | Live broadcast picture display method and device, storage medium and terminal |
| CN110891181B (en) * | 2019-12-03 | 2022-02-18 | 广州酷狗计算机科技有限公司 | Live broadcast picture display method and device, storage medium and terminal |
| CN111145192A (en) * | 2019-12-30 | 2020-05-12 | 维沃移动通信有限公司 | Image processing method and electronic device |
| CN111179386A (en) * | 2020-01-03 | 2020-05-19 | 广州虎牙科技有限公司 | Animation generation method, device, equipment and storage medium |
| CN111242962A (en) * | 2020-01-15 | 2020-06-05 | 中国平安人寿保险股份有限公司 | Method, device and equipment for generating remote training video and storage medium |
| CN111242962B (en) * | 2020-01-15 | 2025-07-22 | 中国平安人寿保险股份有限公司 | Remote training video generation method, device, equipment and storage medium |
| CN111669638A (en) * | 2020-02-28 | 2020-09-15 | 海信视像科技股份有限公司 | Video rotation playing method and display equipment |
| CN111669638B (en) * | 2020-02-28 | 2022-07-15 | 海信视像科技股份有限公司 | Video rotation playing method and display device |
| US11962932B2 (en) * | 2020-04-07 | 2024-04-16 | Beijing Bytedance Network Technology Co., Ltd. | Video generation based on predetermined background |
| JP2023519625A (en) * | 2020-04-07 | 2023-05-11 | 北京字節跳動網絡技術有限公司 | Video processing method, apparatus, electronic equipment and non-transitory computer-readable storage medium |
| KR20220159453A (en) * | 2020-04-07 | 2022-12-02 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | Video processing method, device, electronic device and non-transitory computer readable storage medium |
| US20220377259A1 (en) * | 2020-04-07 | 2022-11-24 | Beijing Bytedance Network Technology Co., Ltd. | Video processing method and apparatus, electronic device, and non-transitory computer readable storage medium |
| WO2021203996A1 (en) * | 2020-04-07 | 2021-10-14 | 北京字节跳动网络技术有限公司 | Video processing method and apparatus, and electronic device, and non-transitory computer readable storage medium |
| JP7518189B2 (en) | 2020-04-07 | 2024-07-17 | 北京字節跳動網絡技術有限公司 | Video processing method, device, electronic device and non-transitory computer-readable storage medium |
| CN111464761A (en) * | 2020-04-07 | 2020-07-28 | 北京字节跳动网络技术有限公司 | Video processing method and device, electronic equipment and computer readable storage medium |
| KR102758914B1 (en) * | 2020-04-07 | 2025-01-23 | 두인 비전 컴퍼니 리미티드 | Method for processing video, apparatus, electronic device and non-transitory computer-readable storage medium |
| CN111447389B (en) * | 2020-04-22 | 2022-11-04 | 广州酷狗计算机科技有限公司 | Video generation method, device, terminal and storage medium |
| CN111447389A (en) * | 2020-04-22 | 2020-07-24 | 广州酷狗计算机科技有限公司 | Video generation method, device, terminal and storage medium |
| WO2021249428A1 (en) * | 2020-06-12 | 2021-12-16 | 北京字节跳动网络技术有限公司 | Method and apparatus for composite video filming, electronic device, and computer readable medium |
| CN111629151A (en) * | 2020-06-12 | 2020-09-04 | 北京字节跳动网络技术有限公司 | Video co-shooting method and device, electronic equipment and computer readable medium |
| US11875556B2 (en) | 2020-06-12 | 2024-01-16 | Beijing Bytedance Network Technology Co., Ltd. | Video co-shooting method, apparatus, electronic device and computer-readable medium |
| CN111722902A (en) * | 2020-06-15 | 2020-09-29 | 朱利戈 | Method and system for realizing rich media interactive teaching based on window transparent processing |
| CN111464865A (en) * | 2020-06-18 | 2020-07-28 | 北京美摄网络科技有限公司 | Video generation method and device, electronic equipment and computer readable storage medium |
| CN111726536A (en) * | 2020-07-03 | 2020-09-29 | 腾讯科技(深圳)有限公司 | Video generation method, device, storage medium and computer equipment |
| US12400438B2 (en) | 2020-07-03 | 2025-08-26 | Tencent Technology (Shenzhen) Company Limited | Video generation method and apparatus, storage medium, and computer device |
| CN111726536B (en) * | 2020-07-03 | 2024-01-05 | 腾讯科技(深圳)有限公司 | Video generation method, device, storage medium and computer equipment |
| CN111783729A (en) * | 2020-07-17 | 2020-10-16 | 商汤集团有限公司 | Video classification method, device, equipment and storage medium |
| CN112822542A (en) * | 2020-08-27 | 2021-05-18 | 腾讯科技(深圳)有限公司 | Video synthesis method and device, computer equipment and storage medium |
| JP7540073B2 (en) | 2020-09-09 | 2024-08-26 | レモン インコーポレイテッド | Video recording method, device, equipment and storage medium |
| CN112037227B (en) * | 2020-09-09 | 2024-02-20 | 脸萌有限公司 | Video shooting method, device, equipment and storage medium |
| CN112037227A (en) * | 2020-09-09 | 2020-12-04 | 脸萌有限公司 | Video shooting method, device, equipment and storage medium |
| US12041374B2 (en) | 2020-09-09 | 2024-07-16 | Lemon Inc. | Segmentation-based video capturing method, apparatus, device and storage medium |
| EP4170588A4 (en) * | 2020-09-09 | 2023-08-23 | Lemon Inc. | VIDEO PHOTOGRAPHY METHOD AND DEVICE, AS WELL AS DEVICE AND STORAGE MEDIUM |
| CN116055800B (en) * | 2020-09-10 | 2025-06-24 | 北京小糖科技有限责任公司 | A method for obtaining a real-time dance video with a customized background on a mobile terminal |
| CN112087662A (en) * | 2020-09-10 | 2020-12-15 | 北京小糖科技有限责任公司 | Method for generating dance combination dance video by mobile terminal |
| CN112087664A (en) * | 2020-09-10 | 2020-12-15 | 北京小糖科技有限责任公司 | A method for obtaining a real-time dance video with a customized background on a mobile terminal |
| CN112087663A (en) * | 2020-09-10 | 2020-12-15 | 北京小糖科技有限责任公司 | Method for generating dance video with adaptive light and shade environment by mobile terminal |
| CN112087663B (en) * | 2020-09-10 | 2021-09-28 | 北京小糖科技有限责任公司 | Method for generating dance video with adaptive light and shade environment by mobile terminal |
| CN116055800A (en) * | 2020-09-10 | 2023-05-02 | 北京小糖科技有限责任公司 | A method for mobile terminal to obtain real-time dance video with customized background |
| CN112261313A (en) * | 2020-09-22 | 2021-01-22 | 网娱互动科技(北京)股份有限公司 | Method for making video with replaceable foreground |
| CN112118397B (en) * | 2020-09-23 | 2021-06-22 | 腾讯科技(深圳)有限公司 | Video synthesis method, related device, equipment and storage medium |
| CN112118397A (en) * | 2020-09-23 | 2020-12-22 | 腾讯科技(深圳)有限公司 | Video synthesis method, related device, equipment and storage medium |
| CN112132931B (en) * | 2020-09-29 | 2023-12-19 | 新华智云科技有限公司 | Processing method, device and system for templated video synthesis |
| CN112132931A (en) * | 2020-09-29 | 2020-12-25 | 新华智云科技有限公司 | Processing method, device and system for templated video synthesis |
| CN112235520A (en) * | 2020-12-07 | 2021-01-15 | 腾讯科技(深圳)有限公司 | Image processing method and device, electronic equipment and storage medium |
| CN112653851A (en) * | 2020-12-22 | 2021-04-13 | 维沃移动通信有限公司 | Video processing method and device and electronic equipment |
| CN112752038B (en) * | 2020-12-28 | 2024-04-19 | 广州虎牙科技有限公司 | Background replacement method, device, electronic equipment and computer readable storage medium |
| CN112752038A (en) * | 2020-12-28 | 2021-05-04 | 广州虎牙科技有限公司 | Background replacing method and device, electronic equipment and computer readable storage medium |
| CN112752116A (en) * | 2020-12-30 | 2021-05-04 | 广州繁星互娱信息科技有限公司 | Display method, device, terminal and storage medium of live video picture |
| CN112860209B (en) * | 2021-02-03 | 2024-07-12 | 宏晶微电子科技股份有限公司 | Video overlapping method and device, electronic equipment and computer readable storage medium |
| CN112860209A (en) * | 2021-02-03 | 2021-05-28 | 合肥宏晶微电子科技股份有限公司 | Video overlapping method and device, electronic equipment and computer readable storage medium |
| CN113067983A (en) * | 2021-03-29 | 2021-07-02 | 维沃移动通信(杭州)有限公司 | Video processing method and device, electronic equipment and storage medium |
| US11895424B2 (en) | 2021-05-20 | 2024-02-06 | Beijing Zitiao Network Technology Co., Ltd. | Video shooting method and apparatus, electronic device and storage medium |
| US12368815B2 (en) | 2021-05-20 | 2025-07-22 | Beijing Zitiao Network Technology Co., Ltd. | Video shooting method and apparatus, electronic device and storage medium |
| WO2022242497A1 (en) * | 2021-05-20 | 2022-11-24 | 北京字跳网络技术有限公司 | Video photographing method and apparatus, electronic device, and storage medium |
| CN113256499A (en) * | 2021-07-01 | 2021-08-13 | 北京世纪好未来教育科技有限公司 | Image splicing method, device and system |
| WO2023036160A1 (en) * | 2021-09-07 | 2023-03-16 | 上海商汤智能科技有限公司 | Video processing method and apparatus, computer-readable storage medium, and computer device |
| CN113810624A (en) * | 2021-09-18 | 2021-12-17 | 维沃移动通信有限公司 | Video generation method and device and electronic equipment |
| CN113873319A (en) * | 2021-09-27 | 2021-12-31 | 维沃移动通信有限公司 | Video processing method and device, electronic equipment and storage medium |
| CN114040129B (en) * | 2021-11-30 | 2023-12-05 | 北京字节跳动网络技术有限公司 | Video generation method, device, equipment and storage medium |
| CN114040129A (en) * | 2021-11-30 | 2022-02-11 | 北京字节跳动网络技术有限公司 | Video generation method, device, equipment and storage medium |
| CN114189635A (en) * | 2021-12-01 | 2022-03-15 | 惠州Tcl移动通信有限公司 | Video processing method and device, mobile terminal and storage medium |
| CN114339401A (en) * | 2021-12-30 | 2022-04-12 | 北京翼鸥教育科技有限公司 | A kind of video background processing method and device |
| CN114363697B (en) * | 2022-01-06 | 2024-04-26 | 上海哔哩哔哩科技有限公司 | Video file generation and playing method and device |
| CN114363697A (en) * | 2022-01-06 | 2022-04-15 | 上海哔哩哔哩科技有限公司 | Video file generation and playing method and device |
| CN114663633A (en) * | 2022-03-24 | 2022-06-24 | 航天宏图信息技术股份有限公司 | AR virtual live broadcast method and system |
| CN114900736A (en) * | 2022-03-28 | 2022-08-12 | 网易(杭州)网络有限公司 | Video generation method and device and electronic equipment |
| CN115100259A (en) * | 2022-05-27 | 2022-09-23 | 深圳市大头兄弟科技有限公司 | A video remapping method and related equipment |
| CN115100259B (en) * | 2022-05-27 | 2025-07-25 | 深圳市闪剪智能科技有限公司 | Video remapping method and related equipment |
| CN115623244A (en) * | 2022-10-08 | 2023-01-17 | 苏州市广播电视总台 | A video synthesis method and system based on real-time video stream portrait matting |
| WO2024088322A1 (en) * | 2022-10-28 | 2024-05-02 | 影石创新科技股份有限公司 | Video display method and apparatus, video display device, and storage medium |
| CN116229337A (en) * | 2023-05-10 | 2023-06-06 | 瀚博半导体(上海)有限公司 | Method, apparatus, system, device and medium for video processing |
| CN116229337B (en) * | 2023-05-10 | 2023-09-26 | 瀚博半导体(上海)有限公司 | Methods, devices, systems, equipment and media for video processing |
| CN117372238A (en) * | 2023-09-21 | 2024-01-09 | 启朔(深圳)科技有限公司 | Registration function method, video frame processing method, device, equipment and storage medium |
| CN117714787B (en) * | 2024-02-05 | 2024-05-07 | 哈尔滨学院 | Video data processing method |
| CN117714787A (en) * | 2024-02-05 | 2024-03-15 | 哈尔滨学院 | A video data processing method |
Also Published As
| Publication number | Publication date |
|---|---|
| CN110290425B (en) | 2023-04-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN110290425A (en) | A kind of method for processing video frequency, device and storage medium | |
| CN108989830A (en) | A kind of live broadcasting method, device, electronic equipment and storage medium | |
| CN106713988A (en) | Beautifying method and system for virtual scene live | |
| US20080012988A1 (en) | System and method for virtual content placement | |
| CN113302622A (en) | System and method for providing personalized video | |
| CN113660528B (en) | Video synthesis method and device, electronic equipment and storage medium | |
| US11943489B2 (en) | Method and system for automatic real-time frame segmentation of high resolution video streams into constituent features and modifications of features in each frame to simultaneously create multiple different linear views from same video source | |
| CN118674839B (en) | Animation generation method, device, electronic equipment, storage medium and program product | |
| CN111598983A (en) | Animation system, animation method, storage medium, and program product | |
| CN113781660A (en) | Method and device for rendering and processing virtual scene on line in live broadcast room | |
| CA3139657C (en) | Apparatus for multi-angle screen coverage analysis | |
| CN111246122A (en) | Method and device for synthesizing video by multiple photos | |
| Willment et al. | What is virtual production? An explainer & research agenda | |
| CN119273591A (en) | Three-dimensional image generation method, device, storage medium and program product | |
| KR20100062822A (en) | System for cooperative digital image production | |
| CN113781618B (en) | Three-dimensional model light weight method, device, electronic equipment and storage medium | |
| Kuoppala | Real-time image generation for interactive art: developing an artwork platform with ComfyUI and TouchDesigner | |
| KR20190092922A (en) | Customized Motion Graphic Video Production System | |
| CN117201720A (en) | Providing real-time virtual background in video sessions | |
| Feiler et al. | Archiving the Memory of the Holocaust | |
| KR100448914B1 (en) | Method for manufacturing and supplying animation composed with real picture | |
| CN117409108A (en) | Image processing method, device, electronic equipment, medium and program product | |
| KR102859571B1 (en) | System for providing shooting service using double camera | |
| Zeng et al. | Design and Implementation of Virtual Real Fusion Metaverse Scene Based on Deep Learning | |
| US20240236388A1 (en) | Method and system for automatic real-time frame segmentation of high resolution video streams into constituent features and modifications of features in each frame to simultaneously create multiple different linear views from same video source |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |