+

WO2018176303A1 - Video transmitting and receiving method, system, and device, and unmanned aerial vehicle - Google Patents

Video transmitting and receiving method, system, and device, and unmanned aerial vehicle Download PDF

Info

Publication number
WO2018176303A1
WO2018176303A1 PCT/CN2017/078728 CN2017078728W WO2018176303A1 WO 2018176303 A1 WO2018176303 A1 WO 2018176303A1 CN 2017078728 W CN2017078728 W CN 2017078728W WO 2018176303 A1 WO2018176303 A1 WO 2018176303A1
Authority
WO
WIPO (PCT)
Prior art keywords
sub
video data
image
data units
encoded
Prior art date
Application number
PCT/CN2017/078728
Other languages
French (fr)
Chinese (zh)
Inventor
朱磊
崔浩
龚明
Original Assignee
深圳市大疆创新科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市大疆创新科技有限公司 filed Critical 深圳市大疆创新科技有限公司
Priority to CN201780005017.8A priority Critical patent/CN108496369A/en
Priority to PCT/CN2017/078728 priority patent/WO2018176303A1/en
Publication of WO2018176303A1 publication Critical patent/WO2018176303A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/164Feedback from the receiver or from the transmission channel
    • H04N19/166Feedback from the receiver or from the transmission channel concerning the amount of transmission errors, e.g. bit error rate [BER]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/238Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
    • H04N21/2385Channel allocation; Bandwidth allocation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display

Definitions

  • Embodiments of the present invention relate to the field of image processing, and in particular, to a video transmission and reception method, system, device, and unmanned aerial vehicle.
  • the unmanned aerial vehicle is equipped with a photographing device, and the photographing device can perform aerial photography, and the aerial video is transmitted to the ground receiving device such as a user terminal, a remote controller, and the like through a communication system of the unmanned aerial vehicle.
  • the code stream data size corresponding to each frame of image data will change in real time (ie, the source will change in real time); in addition, it will be subject to unmanned aerial vehicles and receiving equipment.
  • the distance between the distance, the relative position, the presence of occlusion, and the presence of electromagnetic interference, etc., the channel bandwidth between the UAV and the receiving device will also change in real time (ie, the channel will change in real time), and the source channel changes to each other.
  • the transmission and reception modes of frame-level image data are relatively fixed, which is difficult to adapt to real-time changing source channels, and lacks effective image transmission and image receiving methods, which may result in image transmission delay due to real-time source and channel mismatch problems. shake.
  • Embodiments of the present invention provide a video transmission and reception method, system, device, and an unmanned aerial vehicle to effectively reduce transmission delay jitter of video data.
  • An aspect of the embodiments of the present invention provides a video transmission method, including:
  • each sub-video data unit includes one or more sub-images
  • One or more encoded sub-video data units are selected and transmitted based on one or more characteristics of the channel, and one or more characteristics of the sub-video data unit.
  • Another aspect of the embodiments of the present invention provides a video receiving method, including:
  • the video data includes one or more image frames
  • the sub-video data unit includes a plurality of sub-decompositions obtained by decomposing each of the image frames At least one sub-image in the image.
  • One or more imaging devices configured to acquire video data
  • One or more processors on the movable object working alone or in concert, the processor being configured to:
  • each sub-video data unit includes one or more sub-images
  • One or more encoded sub-video data units are selected and transmitted based on one or more characteristics of the channel, and one or more characteristics of the sub-video data unit.
  • a receiving device including: a communication interface, one or more processors, working alone or in cooperation, and the communication interface is in communication with the processor;
  • the communication interface is configured to receive a plurality of encoded sub-video data units
  • the one or more processors are configured to: control a decoder to decode the plurality of encoded sub-video data units; reconstruct the video data according to the decoded sub-video data unit, wherein the video data includes one Or a plurality of image frames, the sub-video data unit including at least one of the plurality of sub-images obtained by decomposing each of the image frames.
  • a power system mounted to the fuselage for providing flight power
  • the video transmission and receiving method, the system, the device and the unmanned aerial vehicle provided by the embodiment are respectively decomposed into a plurality of sub-video data units by the video data, and respectively, and the plurality of sub-video data units are respectively Encoding, selecting one or more encoded sub-video data units and said selected sub-video data units according to channel characteristics and characteristics of sub-video data units, such that one or more encoded sub-elements are selected
  • the video data unit conforms to the channel characteristics, and when the selected one or more encoded sub-video data units are transmitted on the channel matched with the same, the problem of mismatch between the source and the channel can be effectively solved, and the video data can be effectively reduced. Transmission delay jitter caused by source channel mismatch problem.
  • 1 is a schematic diagram of transmission delay jitter according to an embodiment of the present invention.
  • FIG. 2 is a flowchart of a video transmission method according to an embodiment of the present invention.
  • FIG. 3 is a schematic diagram of a sub video data unit according to an embodiment of the present invention.
  • FIG. 4 is a schematic diagram of a sub video data unit according to another embodiment of the present invention.
  • FIG. 5 is a schematic diagram of a sub video data unit according to another embodiment of the present invention.
  • FIG. 6 is a flowchart of a video transmission method according to another embodiment of the present invention.
  • FIG. 7 is a schematic structural diagram of a frame image according to an embodiment of the present disclosure.
  • FIG. 8 is a coefficient image of a frame image after Hadamard transform according to an embodiment of the present invention.
  • FIG. 9 is a schematic diagram of spatial transformation decomposition according to an embodiment of the present invention.
  • FIG. 10 is a schematic diagram of spatial downsampling decomposition according to an embodiment of the present invention.
  • FIG. 11 is a flowchart of a video receiving method according to an embodiment of the present invention.
  • FIG. 12 is a schematic diagram of a decoded sub-image according to an embodiment of the present invention.
  • FIG. 13 is a schematic diagram of a decoded sub-image according to another embodiment of the present invention.
  • FIG. 14 is a schematic diagram of reconstructing an original image according to an embodiment of the present invention.
  • FIG. 15 is a schematic diagram of reconstructing an original image according to another embodiment of the present invention.
  • FIG. 16 is a structural diagram of a video transmission system according to an embodiment of the present invention.
  • FIG. 17 is a structural diagram of a receiving device according to an embodiment of the present invention.
  • FIG. 18 is a structural diagram of an unmanned aerial vehicle according to an embodiment of the present invention.
  • a component when referred to as being "fixed” to another component, it can be directly on the other component or the component can be present. When a component is considered to "connect” another component, it can be directly connected to another component or possibly a central component.
  • the stability of the transmission delay of image data is an important indicator to measure the performance of the image transmission system.
  • the image data transmission delay is the basic condition for ensuring the smooth display of the video image at the receiving end.
  • the source and channel are used.
  • the real-time change will cause the jitter of the transmission delay between frames and frames, and reduce the performance of the image transmission system.
  • the source data changes and channel changes are taken as examples to describe the frame-level image data transmission delay in detail. Jitter problem.
  • Figure 1 includes scenario 1 and scenario 2.
  • scenario 1 the bandwidth of the channel between the sender and the receiver remains stable.
  • the camera at the transmitting end suddenly moves, or the object within the camera shooting range suddenly moves rapidly, for example, at a certain moment, the camera's subject is a blue sky at the next moment.
  • the camera suddenly turns to shoot a colorful hot air balloon flying in the sky, causing the corresponding stream data size after frame 4 encoding to increase to twice the size of the corresponding stream data after the frame 3 encoding, that is, the sudden change of the source
  • the transmission delay of frame 4 will become twice the transmission delay of frame 3.
  • the code stream data corresponding to each frame image is basically stable, that is, the source remains stable.
  • the channel bandwidth corresponding to frame 4 suddenly drops to half of the channel bandwidth corresponding to frame 3.
  • the UAV suddenly approaches the nearby wireless communication base station.
  • the wireless communication base station will affect the transmission channel of the UAV, that is, the channel changes, and the bandwidth of the channel decreases to half of the original bandwidth.
  • the transmission delay of frame 4 will also become twice the transmission delay of frame 3.
  • the method for video transmission provided by the embodiment of the present invention is described in detail below with reference to FIG. 1 to effectively reduce the transmission delay jitter of the video data.
  • FIG. 2 is the basis A flowchart of a video transmission method provided by an embodiment of the invention. As shown in FIG. 2, the method in this embodiment may include:
  • Step S201 the video data is decomposed into a plurality of sub-video data units, wherein each sub-video data unit includes one or more sub-images.
  • the execution body of the embodiment may be a processor, a controller, or a general-purpose processor having an image processing function, and is not specifically limited herein.
  • an image processor is taken as an example to introduce the principle of a video transmission method.
  • the image processor acquires video data captured by a shooting device mounted on an unmanned aerial vehicle in real time, and the video data may include one frame image or continuous multi-frame.
  • the image processor may decompose the video data into a plurality of sub-video data units.
  • the embodiment does not limit the number of sub-video data units obtained by decomposing the video data, and each sub-video data unit includes one or more sub-images.
  • a feasible implementation manner of decomposing the video data into a plurality of sub-video data units is: decomposing each image frame included in the video data, that is, each frame image into a plurality of sub-images, and selecting at least one sub-image of each image frame.
  • the selected sub-image constitutes a sub-video data unit, that is, the sub-video data unit includes at least one of a plurality of sub-images obtained by decomposing each image frame in the video data.
  • This embodiment does not limit the number of image frames included in one video data.
  • the video data includes 6 image frames, that is, 6 frames, and further, in other embodiments, the video The number of image frames included in the data may also be other values.
  • the video data includes six image frames, which are frame 1, frame 2, frame 3, frame 4, frame 5, frame 6, and frame 1, frame 2, frame 3, frame 4, and frame 5.
  • the frame 6 is separately decomposed.
  • the number of sub-images into which each image frame is decomposed is not limited.
  • each image frame is decomposed into four sub-images, which is only schematically illustrated herein.
  • the number of sub-images obtained after each image frame is decomposed may also be other values.
  • Each of the sub-video data units includes at least one of the four sub-images corresponding to each of the image frames after each of the six image frames is decomposed.
  • each sub-video data unit includes one of the four sub-images corresponding to each of the six image frames.
  • the sub-video data unit 310 includes one sub-image 11 of the frame 1.
  • the sub-video data unit 410 includes two sub-images 11, 12 of the frame 1, and two sub-images 21, 22 of the frame 2.
  • sub-video data unit 420 includes one sub-image 13 of frame 1, frame 2
  • two sub-images 52, 53 of frame 5 two sub-images 62, 63 of frame 6
  • sub-video data unit 430 includes One sub-image of each of the six image frames.
  • each sub-video data unit do not overlap.
  • the manner in which at least one of the plurality of sub-images corresponding to each of the plurality of image frames is combined to form a sub-video data unit may also have other combinations, which are not enumerated here.
  • the video data may also include only one image frame, that is, one frame image.
  • 50 indicates an image frame included in the video data, and the image frame 50 is decomposed. This embodiment does not limit one image frame.
  • the number of sub-images obtained after the decomposition is selected, and the image frame 50 is optionally decomposed into four sub-images, such as the sub-image 11, the sub-image 12, the sub-image 13, and the sub-image 14 as shown in FIG.
  • Sub-picture 11, sub-image 12, sub-image 13, and sub-image 14 can be divided into the following achievable ways:
  • each sub-video data unit includes one sub-image, such as sub-video data unit 510, sub-video data unit 520, sub-video data unit 530, and sub-video data unit 540 as shown in FIG.
  • each sub-video data unit includes two sub-images. This embodiment does not limit the combination of two sub-images included in one sub-video data unit.
  • each sub-video data unit includes a different number of sub-images, such as sub-video data unit 570 and sub-video data unit 580 as shown in FIG. 5, wherein sub-video data unit 570 includes three sub-pictures.
  • Combination, optional, sub-video data unit 570 includes sub-image 11, sub-image 12, sub-picture
  • the sub-video data unit 580 includes a sub-image 14.
  • Step S202 encoding the plurality of sub-video data units separately.
  • the image processor separately encodes each of the plurality of sub-video data units by using each sub-video data unit as a coding unit, and obtains a plurality of code stream data after encoding, and optionally encodes one sub-video data unit. Then, a code stream data is obtained, where the coding includes source coding and/or channel coding, and the manner of source coding may include H.263, H.264, H.265, MPEG4, etc., and the channel coding method may include correcting
  • the error coding type may include an RS code, that is, a Reed-Solomon code, a convolutional code, a Turbo code, a Polar code, an interleaving code, a pseudo random sequence scrambling code, and the like.
  • Step S203 Select one or more encoded sub-video data units based on one or more characteristics of the channel and one or more characteristics of the sub-video data unit.
  • one or more characteristics of the channel include at least a bandwidth.
  • one or more characteristics of the channel include at least one of: noise, interference, signal to noise ratio, bit error rate, fading rate, bandwidth.
  • the one or more characteristics of the sub-video data unit include: a code stream data size encoded by the sub-video data unit, or an energy concentration of the sub-video data unit.
  • the image processor selects one or more of the plurality of encoded sub-video data units for transmission over the wireless channel based on one or more characteristics of the current wireless channel and one or more characteristics of the sub-video data unit, eg, transmitting
  • the receiving device can be a remote control smartphone, a tablet, a ground control station, a laptop, a watch, a wristband, and the like, and combinations thereof.
  • selecting one or more encoded sub-video data units can be implemented in the following feasible ways:
  • a first feasible way is to select one or more encoded sub-video data units such that the total code stream data size of the one or more encoded sub-video data units matches the channel bandwidth.
  • the image processor decomposes the video data to obtain four sub-video data units 310-340 as shown in FIG. 3, and encodes the four sub-video data units separately to obtain the size of the code stream data as S0, S1, S2, and S3.
  • the bandwidth of the current wireless channel is T
  • the image processor may select one or more code stream data from the four code stream data according to the bandwidth T of the wireless channel, and the selection may be based on: one or more selected codes.
  • the total size of the stream is as close as possible to the bandwidth T of the wireless channel.
  • one or more of the plurality of sub-video data units are selected for combination, so that the total code stream data size of the combined sub-video data unit is as close as possible to the bandwidth of the wireless channel, and the wireless channel can transmit the code stream with the current bandwidth.
  • the child video data unit is as large as possible.
  • the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, and the encoded sub-video data unit 310, the sub-video data unit 320, the sub-video data unit 330, and the sub-video data unit 340 may be selected for transmission.
  • the three largest code stream data may be selected from S0, S1, S2, and S3, and the three largest code stream data are S0, S1, and S2, if S0, If the sum of S1 and S2 is less than T, the image processor may select the encoded sub-video data unit 310, the sub-video data unit 320, and the sub-video data unit 330 for transmission, and so on, when the sum of S0, S1, S2 is greater than T, other combinations of the code stream data that can be transmitted with the bandwidth of the current wireless channel and the largest sub-video data unit can be selected.
  • the second feasible way is that a plurality of sub-video data units are prioritized according to the energy concentration.
  • the one or more encoded sub-video data units are selected based on a priority of the sub-video data unit and a channel bandwidth.
  • the energy concentration of each of the plurality of sub-video data units may be the same or similar, it may be different. Therefore, in other embodiments, if the energy concentration of each sub-video data unit is different, it may be based on each The energy concentration of the sub-video data units prioritizes the plurality of sub-video data units. Alternatively, the greater the energy concentration, the higher the priority. For example, the image processor decomposes the video data to obtain four sub-video data units as shown in FIG. 3, 4 or 5, and the four sub-video data units are respectively recorded as sub-video data unit A, sub-video data unit B, and sub-video.
  • the data unit C, the sub-video data unit D, and the priority of the four sub-video data units are successively decremented, the code stream data size encoded by the sub-video data unit A is S0, and the code stream data encoded by the sub-video data unit B is sequentially obtained.
  • the size of the code stream data encoded by the sub video data unit C is S2
  • the size of the code stream data encoded by the sub video data unit D is S3.
  • the image processor needs to select and transmit the sub-video data unit according to the code stream data size and channel bandwidth of the encoded sub-video data unit, and in some cases, according to the priority and channel of the sub-video data unit.
  • the bandwidth selects the sub-video data unit and transmits it.
  • one or more encoded sub-video data units selected from the above four encoded sub-video data units may be selected according to one or more selected sub-views selected.
  • the total code stream data size of the frequency data unit is smaller than the channel bandwidth, and the higher priority sub video data unit is preferentially selected, thereby ensuring that the high priority sub video data unit is preferentially transmitted.
  • the encoded sub-video data unit A, sub-video data unit B, sub-video data unit C, and sub-video data unit D may be selected for transmission.
  • the image processor transmits the encoded sub-video data unit A, the encoded sub-video data unit B, and the encoded sub-video data unit C; If the sum of S0, S1, and S2 is greater than T, the first two sub-video data units A and sub-video data units B having higher priority are determined, and so on, to ensure that at least the sub-video data unit A having the highest priority is transmitted. .
  • the video data is decomposed into a plurality of sub-video data units, and the plurality of sub-video data units are separately encoded, and one or more encoded sub-video data units are selected according to channel characteristics and characteristics of the sub-video data units.
  • the selected one or more encoded sub-video data units conform to the channel characteristics, which can effectively solve the problem of mismatch between the source and the channel, and can effectively reduce the transmission delay of the video data due to the source channel mismatch problem. shake.
  • FIG. 6 is a flowchart of a video transmission method according to another embodiment of the present invention.
  • the video data includes one or more image frames.
  • the method in this embodiment may include:
  • Step S601 Decompose each of the one or more image frames in the video data into a plurality of sub-images.
  • the video data may include one frame image, and may also include consecutive multi-frame images.
  • This embodiment does not limit the number of pixels included in one frame image, and does not limit the pixels of each pixel. value.
  • the image processor decomposes the video data, specifically, each frame image of one or more frames in the video data may be decomposed into a plurality of sub-images, and the sub-video data unit includes each of the image frames. At least one of the plurality of sub-images obtained after the decomposition is performed.
  • each of the one or more image frames in the video data is decomposed into a plurality of sub-images.
  • a frame image included in the video data is taken as an example to introduce the frame image.
  • the process of spatial decomposition can be achieved in several possible ways:
  • the Fourier correlation transform or orthogonal transform is selected from a Hadamard transform, a discrete cosine transform, a discrete Fourier transform, a Walsh-Hadamard transform, a Haar transform, or an oblique transform.
  • This embodiment specifically takes the Hadamard transform as an example to introduce a process of spatially decomposing the frame image.
  • other spatial transforms may also be used to spatially decompose the frame image.
  • a schematic diagram of a frame image is used.
  • This embodiment does not limit the number of pixels included in a frame image, and a frame image includes 16 pixels.
  • P1-P16 represents 16 pixels.
  • the pixel value is spatially transformed and decomposed into pixel values of every four adjacent pixel points of the 16 pixel points, and is decomposed into four sub-images.
  • the following is schematically illustrated by Hadamard transform, and the specific spatial transformation decomposition process includes The following steps:
  • Step 1 Perform a Hadamard transform by using four adjacent pixels of the 16 pixels as a unit.
  • the conversion coefficients obtained by the Hadamard transform of P1, P2, P3, and P4 are H1, H2, and H3. H4, wherein the relationship between P1, P2, P3, P4 and H1, H2, H3, H4 satisfies the formulas (1), (2), (3), (4):
  • H1 contains the average energy of 4 pixels
  • H2 contains the average gradient of 4 pixels in the vertical direction
  • H3 contains 4 pixels.
  • H4 contains a cross gradient of 4 pixels, ie texture information. It can be seen that the energy concentration of H1 is the highest, the energy concentration of H2 and H3 is second, and the energy concentration of H4 is the smallest.
  • P1, P2, P3, and P4 with H1, H2, H3, and H4, H1 is essential. H1 is the most important, followed by H2 and H3, and finally H4 is important.
  • H1 is the most important
  • H2 and H3 are the second most important
  • H4 is the least important, that is, H1, H2, H3, and H4 are important.
  • Sex is decremented in turn.
  • the same Hadamard transform is performed on P5-P8 to obtain H5-H8, and the energy concentration of H5, H6, H7 and H8 is successively decreased, and the importance of H5, H6, H7 and H8 is successively decreased, and P9-P12 is successively performed.
  • the same Hadamard transform gets H9-H12, H9, The energy concentration of H10, H11 and H12 decreases in turn, and the importance of H9, H10, H11 and H12 decreases successively.
  • Step 2 Decompose the conversion coefficients obtained by the Hadamard transform into different sub-images.
  • This embodiment does not limit the number of sub-images obtained by spatially transforming each frame of the image, for example, the sub-images obtained after the decomposition.
  • the number is four, which is only a schematic description. In other embodiments, the number of sub-images obtained by spatially transforming and decomposing each frame of image may be other values.
  • H1 is assigned to the first sub-image
  • H2 is assigned to the second sub-image
  • H3 is assigned to the third sub-image
  • H4 is assigned to the fourth sub-image.
  • H5-H8 is decomposed into 4 sub-images
  • H9-H12 is decomposed into 4 sub-images
  • H13-H16 is decomposed into 4 sub-images
  • the decomposition result shown in FIG. 9 is obtained, wherein the sub-image 1 Concentrates the conversion coefficient with the highest energy concentration
  • sub-image 2 and sub-image 3 concentrate the conversion coefficient of energy concentration second
  • sub-image 4 concentrates the conversion coefficient with the smallest energy concentration, so sub-image 1 is the most important, sub-image 2.
  • the importance of sub-image 3 is second, sub-image 4 is of the least importance.
  • the resolution of each of the four sub-images after the spatial transformation decomposition is one quarter of the original image before the decomposition.
  • This embodiment does not limit the number of sub-images obtained by down-sampling the image space of each frame.
  • the number of sub-images obtained after decomposition is four, which is only schematically illustrated herein.
  • the number of sub-images obtained by down-sampling the image space of each frame may also be other values.
  • spatially down-sampling the pixel values of every four adjacent pixel points of the 16 pixel points is performed.
  • the specific spatial down-sampling decomposition process is: decomposing four pixel points in one unit into different sub-images by using four adjacent pixel points of 16 pixels as one unit.
  • P1 is decomposed into the first sub-image
  • P2 is decomposed into the second sub-image
  • P3 is decomposed into the third sub-image
  • P4 is decomposed into the fourth sub-image
  • P5-P8 is decomposed into 4 sub-images.
  • P9-P12 is decomposed into 4 sub-images
  • P13-P16 is decomposed into 4 sub-images, and the decomposition result as shown in FIG. 10 is obtained.
  • the resolution of each of the four sub-images after spatial downsampling is one quarter of the original image before the decomposition. Without loss of generality, it is assumed that the size of the original image before decomposition is W*H. If the original image is decomposed into four sub-images, the row number or column number of the pixel matrix corresponding to the original image or sub-image is counted from 0, then the first The sub-images may include pixels in the original image with coordinates (2i, 2j), where The second sub-image may include a pixel with coordinates (2i+1, 2j) in the original image, and the third sub-image may include a pixel with coordinates (2i, 2j+1) in the original image, and the fourth sub-image. The pixel in the original image with coordinates (2i+1, 2j+1) may be included.
  • Each of the one or more image frames in the video data can be decomposed into a plurality of sub-images according to any of the above-described decomposition modes, that is, spatial transformation or spatial downsampling.
  • one image frame is one frame image
  • a plurality of image frames are multi-frame images.
  • the video data includes one or more image frames, the sub-video data unit including at least one of a plurality of sub-images obtained by decomposing each of the image frames.
  • the image frame is decomposed in the manner shown in FIG. 9 or FIG. 10, and the sub-video data unit may include at least one of the plurality of sub-images obtained by decomposing the image frame, for example, one.
  • the sub-video data unit includes a sub-image, and each sub-image obtained after the decomposition is compressed and encoded to obtain encoded code stream data.
  • each image frame that is, each frame image is decomposed as shown in FIG. 9 or FIG. 10, for example, the video data includes 24 image frames, and if each image frame is decomposed For 4 sub-images, 24 consecutive image frames are decomposed to obtain 24*4 sub-images, and each sub-video data unit may include multiple sub-images of 24*4 sub-images, specifically, the sub-video data unit At least one of the plurality of sub-images corresponding to each image frame obtained by decomposing each of the image frames, the sub-video data unit may be four, wherein the sub-video data unit Each of them includes a sub-image of each of the 24 image frames.
  • the encoded code stream data is obtained by compressing and encoding each sub-video data unit.
  • each sub-image includes a portion of an image frame.
  • each sub-image includes one or more pixels of an image frame.
  • each sub-image includes one or more conversion coefficients of the image frame.
  • the receiving device reconstructs a frame of image as described in FIG. 7, in FIG. 10, the importance of the sub-image 1, the sub-image 2, the sub-image 3, and the sub-image 4 are the same.
  • the sub-image 1 is the most important, and the sub-image 2 and the sub-image 3 are the second most important, and the sub-image 4 has the lowest importance.
  • the plurality of sub-video data units are prioritized according to the energy concentration. Sort by level.
  • sub-image 1 concentrates the conversion coefficient with the highest energy concentration
  • sub-picture The image 2 and the sub-image 3 concentrate the conversion coefficient of the energy concentration second
  • sub-image 4 concentrates the conversion coefficient with the smallest energy concentration. If a sub-video data unit includes a sub-image, the sub-video data unit including the sub-image 1
  • the priority of the sub-video data unit including sub-image 2 or sub-image 3 is the highest, and the sub-video data unit including sub-image 4 has the lowest priority.
  • Step S602 encoding the plurality of sub-video data units separately.
  • a first possible way multiple sub-video data units are encoded by a plurality of separate encoders.
  • the plurality of sub-video data units are encoded in parallel by using the plurality of separate encoders; or the plurality of sub-video data units are encoded by using different video encoding rules; or, the same video encoding rule is adopted.
  • the plurality of sub-video data units are encoded.
  • a second possible way two or more of the plurality of sub-video data units are encoded by the same encoder.
  • a third possible way encoding at least one of the plurality of sub-video data units based on a motion compensated video compression standard.
  • a fourth feasible way compressing the plurality of sub-video data units according to different compression ratios.
  • the compression ratio is determined based on one or more characteristics of the sub-video data unit.
  • Step S603 Select one or more encoded sub-video data units based on one or more characteristics of the channel and one or more characteristics of the sub-video data unit.
  • one sub-video data unit includes one sub-image, as shown in FIG. 9, the sub-image 1 is the most important, the sub-image 2, the sub-image 3 are of the second most important, and the sub-image 4 has the lowest importance. Then, the sub-video data unit including the sub-image 1 has the highest priority, and the sub-video data unit including the sub-image 2 or the sub-image 3 has the lower priority, and the sub-video data unit including the sub-image 4 has the lowest priority.
  • the code stream data size encoded by the sub video data unit A including the sub image 1 is S0
  • the code stream data size encoded by the sub video data unit B including the sub image 2 is S1
  • the sub video including the sub image 3 is used.
  • the code stream data size encoded by the data unit C is S2
  • the code stream data size encoded by the sub video data unit D including the sub image 4 is S3.
  • the image processor can receive and transmit error status according to historical data, priority of the current sub-video data unit,
  • the size of the code stream data encoded by the current sub-video data unit, and the estimated value of the real-time channel, such as the channel bandwidth, are combined and transmitted to the four sub-video data units to achieve real-time matching between the source and the channel. For example, when the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, the encoded sub-video data unit A, sub-video data unit B, sub-video data unit C, and sub-video data unit D may be selected for transmission.
  • the image processor transmits the encoded sub-video data unit A, the encoded sub-video data unit B, and the encoded sub-video data unit C; If the sum of S0, S1, S2 is greater than T, then the first two higher priority sub-video data units A, sub-video data units B are determined, and so on, ensuring that at least the highest priority sub-video data unit A is transmitted.
  • sub-video data unit includes one sub-image, as shown in FIG. 10, sub-image 1, sub-image 2, sub-image 3, sub-image 4 have the same importance, or in some cases The importance of the image 1, the sub-image 2, the sub-image 3, and the sub-image 4 cannot be determined, that is, the sub-video data unit A including the sub-image 1, the sub-video data unit B including the sub-image 2, and the sub-video including the sub-image 3.
  • the data unit C, the sub-video data unit D including the sub-image 4 have the same priority or the priority cannot be determined, and the code stream data size encoded by the sub-video data unit A is S0, and the code encoded by the sub-video data unit B
  • the stream data size is S1
  • the code stream data size encoded by the sub video data unit C is S2
  • the code stream data size encoded by the sub video data unit D is S3.
  • the image processor can jointly transmit the four sub-video data units according to the transmission and error status of the historical data, the current stream data size of the sub-video data unit, and the estimated value of the real-time channel, such as the channel bandwidth, to achieve real-time source and channel. match.
  • the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, and the encoded sub-video data unit A, the sub-video data unit B, the sub-video data unit C, and the sub-video data unit D may be selected for transmission.
  • the three largest code stream data can be selected from S0, S1, S2, and S3, and the three largest code stream data are S0, S1, S2, if S0, S1 If the sum of S2 is less than T, the image processor may select the encoded sub-video data unit A, the sub-video data unit B, and the sub-video data unit C to transmit, and so on, when the sum of S0, S1, S2 is greater than T. It is possible to select other combinations of the code stream data that can be transmitted with the bandwidth of the current wireless channel and the largest sub-picture data unit.
  • the combined code stream data size may fall within the range of [T-D, T+D].
  • the specific spatial decomposition includes spatial transformation decomposition and spatial down sampling decomposition, so that the sub-image obtained after the decomposition includes the image frame.
  • One or more pixels, or one or more conversion coefficients including an image frame the sub-images obtained by the decomposition are combined into a sub-video data unit, and when the encoded sub-video data unit is transmitted, the sub-video data unit is encoded
  • the code stream data size matches the characteristics of the channel (such as bandwidth), or the priority of the sub-video data unit is matched to the characteristics of the channel (such as bandwidth), which enables source and channel matching and reduces the frame level.
  • the transmission delay jitter caused by the video channel mismatch problem is a sub-video data unit.
  • FIG. 11 is a flowchart of a video receiving method according to an embodiment of the present invention. As shown in FIG. 11, the method in this embodiment may include:
  • Step S1101 receiving a plurality of encoded sub-video data units.
  • the video data includes one or more image frames
  • the sub-video data unit includes at least one of a plurality of sub-images obtained by decomposing each of the image frames.
  • the decomposition of the video data may be as shown in FIG. 3, FIG. 4 or FIG. 5, and the sub-video data unit is obtained; wherein each frame of the video data is performed.
  • the decomposition may be as shown in FIG. 9 or FIG. 10, and the specific decomposition process is consistent with the foregoing embodiment, and details are not described herein again.
  • the receiving device receives a plurality of encoded sub-video data units transmitted by the communication system of the unmanned aerial vehicle, wherein the video data may include one or more image frames, and the unmanned aerial vehicle transmits the video data before transmitting the video data.
  • Each image frame included in the data is decomposed into a plurality of sub-images.
  • the sub-video data unit includes at least one of a plurality of sub-images corresponding to each image frame decomposed by each image frame.
  • each sub-video data unit includes at least one sub-image of each image frame to reduce the correlation between the sub-images in each sub-video data unit, to avoid loss or distortion due to the sub-video data unit during transmission.
  • the resulting sub-images with high correlation are lost or distorted. Because the correlation between the sub-images obtained after the same frame image is decomposed is higher, and the sub-images with higher correlation are lost. Or when it is distorted, it will be difficult to recover an image frame composed of a highly correlated sub-image.
  • the receiving device receives 4 sub-video data units, and each sub-video data unit includes a sub-image obtained as shown in FIG. 9 or FIG.
  • Step S1102 Decode the plurality of encoded sub-video data units.
  • the plurality of encoded sub-video data units are separately decoded.
  • the receiving device separately decodes the four encoded sub-video data units, that is, separately performs code stream data corresponding to the four sub-images. Decode to obtain the decoded sub-video data unit.
  • the sub-video data unit is transmitted in the wireless channel, the sub-video data unit obtained by the receiving device and the encoded system actually transmitted by the communication system of the UAV may be caused due to noise interference, multipath effect, fading, and the like.
  • the video data units are different, causing the receiving device to receive an error.
  • the four sub-images obtained by the receiving device after decoding the code stream data are as shown in FIG. 12, wherein if the sub-image 1 is transmitted correctly, H1 Same as h1, H2 and h2 are the same, H3 and h3 are the same, H4 and h4 are the same, and if sub-image 1 is transmitted incorrectly, at least one of H1 and h1, H2 and h2, H3 and h3, H4 and h4 are different. Similarly, other sub-pictures are transmitted correctly or transmitted incorrectly, and the conversion coefficients before transmission and the conversion coefficients after transmission also have the same relationship.
  • the four sub-images obtained by the receiving device after decoding the code stream data are as shown in FIG. 13, wherein if the sub-image 1 is transmitted correctly, P1 The same as p1, P2 and p2 are the same, P3 and p3 are the same, and P4 and p4 are the same. If sub-image 1 is transmitted incorrectly, at least one of P1 and p1, P2 and p2, P3 and p3, P4 and p4 are different. Similarly, other sub-images are transmitted correctly or transmitted incorrectly, and the pixels before transmission and the pixels after transmission also have the same relationship.
  • Step S1103 Reconstruct the video data according to the decoded sub video data unit.
  • a transmission error of one or more sub-images of the sub-video data unit is detected, and the video data is reconstructed according to receiving the correct sub-image.
  • the receiving device decodes the code stream data to obtain 4 sub-images, it detects that each sub-image is transmitted correctly or transmits an error, and reconstructs the original image according to receiving the correct sub-image.
  • the sub-image transmitted by the communication system is the sub-image shown in FIG. 9, and the sub-image received by the receiving device is as shown in FIG. 12.
  • FIG. 12 it is assumed that the sub-image 2 receives an error, and the sub-image 1, The sub-image 3 and the sub-image 4 are all received correctly, and the receiving device reconstructs the original image according to the sub-image 1, the sub-image 3, and the sub-image 4 shown in FIG. 12, and reconstructs the original image, reconstructs the video data by using an inverse transform, and gives
  • the sub-picture in which the error is transmitted in the sub-picture data unit is assigned a value.
  • a possible implementation is to assign a value of 0 to the sub-picture that transmits the error in the sub-video data unit.
  • h2, h6, h10, and h14 of the sub-image 2 are both set to 0.
  • H1, H2, H3, and H4 are obtained based on P1, P2, P3, and P4.
  • spatial transformation is used to decompose the image frame, other spatial transformations are used.
  • the receiving device reconstructs the original image, the corresponding spatial transformation is adopted.
  • H1 and h1 are the same, H3 and h3 are the same, and H4 and h4 are the same. Therefore, p1, p2, p3, p4 obtained by inverse Hadamard transform and pixel values P1, P2, P3 in the original image, P4 may be different, but reconstructing the original image based on receiving the correct sub-image ensures that the reconstructed image is close to the original image.
  • h5, h6, h7, h8 are inversely transformed by Hadamard to obtain p5, p6, p7, p8, and h9, h10, h11, h12 are inversely transformed by Hadamama to obtain p9, p10, p11, p12, h13, H14, h15, h16 perform inverse Hadamard transform to obtain p13, p14, p15, p16, wherein h6, h10, h14 are all 0, and reconstruct the original image according to p1-p16 obtained by inverse Hadamard transform, as shown in Fig. 14. .
  • the sub-image transmitted by the communication system is the sub-image shown in FIG. 10, and the sub-image received by the receiving device is as shown in FIG. 13, and in FIG. 13, it is assumed that the sub-image 3 receives the error, and the sub-image 1, The sub-image 2 and the sub-image 4 are all received correctly, and the receiving device reconstructs the original image according to the sub-image 1, the sub-image 2, the sub-image 3, and the sub-image 4 shown in FIG. 13, and when the original image is reconstructed, the sub-video data unit is given.
  • the sub-image of the transmission error is assigned a value.
  • the value assigned to the sub-picture transmitted in the sub-video data unit is determined by interpolation, specifically, for transmission in the sub-video data unit.
  • the value of the wrong sub-image assignment is based on the transmission
  • the correct sub-image is determined, and the sub-image transmitted incorrectly and the sub-image transmitted correctly are from the same image frame.
  • the sub-image 3 receives an error, and the sub-image 1, the sub-image 2, and the sub-image 4 are all received correctly, and the sub-image 3 does not participate in the reconstruction process, that is, the receiving device only according to the sub-image 1, the sub-image 2,
  • the sub-image 4 reconstructs the original image.
  • the specific process is as follows: since the original image includes 16 pixels, the sub-image 1, the sub-image 2, and the sub-image 4 have a total of 12 pixels. According to FIG. 10, each of the 16 pixels in the original image is known. Four adjacent pixels are decomposed into four different sub-images. Therefore, when the original image is reconstructed from the sub-image 1, sub-image 2, and sub-image 4, the first pixel p1 of the sub-image 1 and the sub-image 2 The first pixel p2 and the first pixel p4 of the sub-image 4 are respectively three pixels P1, P2, and P4 of the first four adjacent pixels of the original image. Similarly, p5, p6, and p8 are original.
  • the three pixels P5, P6, P8, p9, p10, and p12 in the image P5-P8 are the three pixels P9, P10, and P12 in the original image P9-P12, respectively, and the p13, p14, and p16 are the original images P13-P16, respectively.
  • the three pixels P 13 , P 14 , P 16 in accordance with p1, p2, p4, p5, p6, p8, p9, p10, p12, p13, p14 p16 obtained image A shown in FIG.
  • the pixel image A spare reception error that is included in the sub-image pixel 3 due to reception error sub-image 3, and therefore, the sub-image 3 can not participate in the process of reconstruction of the original image. Since p1, p2, p4, p5, p6, p8, p9, p10, p12, p13, p14, and p16 are all correctly received, ie, p1, p2, p4, p5, p6, p8, p9, p10, p12, p13 , p14, p16 are the same as the pixel points at the same position of the original image. In this embodiment, the pixel value of the image A can be determined according to the interpolation method.
  • a feasible interpolation method is: p3 is equal to the arithmetic average of p1, p2, and p4 The value, p7 is equal to the arithmetic mean of p5, p6, p8, p11 is equal to the arithmetic mean of p9, p10, p12, and p15 is equal to the arithmetic mean of p13, p14, p16, resulting in the reconstructed original image B.
  • p3 is equal to the arithmetic average of p1, p2, and p4
  • the value, p7 is equal to the arithmetic mean of p5
  • p6, p8 p11 is equal to the arithmetic mean of p9, p10, p12
  • p15 is equal to the arithmetic mean of p13, p14, p16, resulting in the reconstructed original image B.
  • the receiving device separately decodes the plurality of sub-video data units to obtain the decoded sub-video data unit, and reconstructs the original image according to the decoded sub-video data unit, and specifically, receives the correct sub-image reconstruction according to the decoding.
  • the original image can make the reconstructed image be close to the original image to the greatest extent when the partial sub-image is received incorrectly, which improves the fault tolerance of the receiving device in reconstructing the image and enhances the robustness of the system.
  • FIG. 16 is a structural diagram of a video transmission system according to an embodiment of the present invention. As shown in FIG. 16, the video transmission system 1600 includes one or more imaging devices 1601, and one or more processors 1602 on a movable object.
  • the plurality of imaging devices 1601 are configured to acquire video data; the one or more processors 1602 work alone or in concert, configured to: decompose the video data into a plurality of sub-video data units, wherein each sub-video data unit includes one or Multiple sub-images; encoding the plurality of sub-video data units separately; and selecting one or more encoded sub-videos based on one or more characteristics of the channel and one or more characteristics of the sub-video data unit Data unit and transfer.
  • one or more characteristics of the channel include at least a bandwidth.
  • the one or more characteristics of the channel include at least one of: noise, interference, signal to noise ratio, bit error rate, fading rate, bandwidth.
  • the one or more characteristics of the sub video data unit include: a code stream data size encoded by the sub video data unit, or an energy concentration of the sub video data unit.
  • the total code stream data size of the one or more encoded sub-video data units matches the channel bandwidth.
  • the plurality of sub-video data units are prioritized according to the energy concentration.
  • the processor 1602, when selecting one or more encoded sub-video data units is configured to select the one or more encoded sub-video data units based on a priority and a channel bandwidth of the sub-video data units.
  • the video data is decomposed into a plurality of sub-video data units, and the plurality of sub-video data units are separately encoded, and one or more encoded sub-video data units are selected according to channel characteristics and characteristics of the sub-video data units.
  • the selected one or more encoded sub-video data units are made to conform to channel characteristics, and when the selected one or more encoded sub-video data units are transmitted on the channel matched thereto, the source and the source can be effectively solved.
  • the channel mismatch problem can effectively reduce the transmission delay jitter of video data due to the source channel mismatch problem.
  • Embodiments of the present invention provide a video transmission system.
  • the video data includes one or more image frames.
  • Processor 1602 will When the video data is decomposed into a plurality of sub-video data units, configured to: decompose each of the one or more image frames in the video data into a plurality of sub-images, wherein each of the sub-video data units At least one of the plurality of sub-images obtained by each of the image frames is included.
  • Each sub-image includes a portion of the image frame. Specifically, each sub-image includes one or more pixels of the image frame. Alternatively, each sub-image includes one or more conversion coefficients of the image frame.
  • the processor 1602 when the processor 1602 decomposes each of the one or more image frames in the video data into a plurality of sub-images, the processor 1602 is configured to: each of the one or more image frames in the video data A space is decomposed into multiple sub-images.
  • the processor 1602 When the processor 1602 decomposes each of the one or more image frames in the video data into a plurality of sub-images, the processor 1602 is configured to: use one of the video data by Fourier correlation transform or orthogonal transform Or each of the plurality of image frames is decomposed into a plurality of sub-images.
  • the Fourier correlation transform or orthogonal transform is selected from a Hadamard transform, a discrete cosine transform, a discrete Fourier transform, a Walsh-Hadamard transform, a Haar transform or a skew transform.
  • the processor 1602 when the processor 1602 decomposes each of the one or more image frames in the video data into a plurality of sub-images, the processor 1602 is configured to: spatially downsample one or more images in the video data Each space in the frame is decomposed into multiple sub-images.
  • processor 1602 separately encodes the plurality of sub video data units
  • the following may be implemented in several feasible manners:
  • the processor 1602 controls a plurality of encoders to encode the plurality of sub-video data units. Specifically, the processor 1602 controls the plurality of encoders to perform parallel encoding on the plurality of sub video data units; or the processor 1602 controls the plurality of encoders to respectively use the different video encoding rules to the plurality of sub video data. The unit performs encoding; or alternatively, the processor 1602 controls the plurality of encoders to encode the plurality of sub-video data units using the same video encoding rule.
  • the processor 1602 controls the encoder to encode two or more of the plurality of sub-video data units.
  • the processor 1602 controls the encoder to encode at least one of the plurality of sub-video data units based on a motion compensated video compression standard.
  • the processor 1602 is configured to compress the plurality of sub-video data units according to different compression ratios. Wherein the compression ratio is determined according to one or more characteristics of the sub-video data unit.
  • the encoder may be a hardware entity that is independent of the processor 1602 and is electrically connected to the processor 1602, or may be software for implementing the encoding function in the processor 1602.
  • the movable object is an unmanned aerial vehicle.
  • One or more imaging devices 1601 are coupled to the movable object by a carrier, which may be a multi-axis universal joint.
  • the specific spatial decomposition includes spatial transformation decomposition and spatial down sampling decomposition, so that the sub-image obtained after the decomposition includes the image frame.
  • One or more pixels, or one or more conversion coefficients including an image frame, for a frame image a plurality of sub-images obtained by the decomposition thereof may be combined and transmitted, so that the combined code stream data size and channel
  • the bandwidth matching, or the combined code stream data further includes code stream data corresponding to the sub-images with higher importance, so that each frame image can be matched on the wireless channel to achieve source and channel matching, and the frame is reduced.
  • FIG. 17 is a structural diagram of a receiving device according to an embodiment of the present invention. As shown in FIG. 17, the receiving device 1700 includes a communication interface 1701, one or more processors 1702, and one or more processors 1702 work alone or in cooperation.
  • the interface 1701 is communicatively coupled to the processor 1702; the communication interface 1701 is configured to receive a plurality of encoded sub-video data units, wherein the video data includes one or more image frames, and the sub-video data unit includes the At least one of the plurality of sub-images obtained after each of the image frames is decomposed; the one or more processors 1702 are configured to: control the decoder to decode the plurality of encoded sub-video data units; The subsequent sub-video data unit reconstructs the video data.
  • the processor 1702 when the processor 1702 controls the decoder to decode the plurality of encoded sub-video data units, the processor 1702 is specifically configured to: control the decoder to separately decode the plurality of encoded sub-video data units.
  • the decoder may be independent of the processor 1702 and with the processor
  • a hardware entity that is electrically connected to 1702 may also be software for implementing a decoding function in the processor 1702.
  • the processor 1702 is further configured to: detect a transmission error of one or more sub-images of the sub-video data unit; when the processor 1702 reconstructs the video data according to the decoded sub-video data unit, specifically, according to the receiving The correct sub-image reconstructs the video data.
  • the processor 1702 is further configured to assign a value to the sub-image in which the error is transmitted in the sub-video data unit.
  • the value assigned to the sub-image in which the error is transmitted in the sub-video data unit is 0.
  • the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined by interpolation.
  • the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined based on the transmission of the correct sub-image, and the sub-image of the transmission error and the sub-image of the transmission are from the same image frame.
  • the processor 1702 When the processor 1702 reconstructs the video data according to the decoded sub video data unit, the processor 1702 is specifically configured to: reconstruct the video data by using an inverse transform.
  • the receiving device may be a remote controller, a smart phone, a tablet computer, a ground control station, a laptop computer, a watch, a wristband, and the like, and combinations thereof.
  • the receiving device separately decodes the plurality of sub-video data units to obtain the decoded sub-video data unit, and reconstructs the original image according to the decoded sub-video data unit, and specifically, receives the correct sub-image reconstruction according to the decoding.
  • the original image can make the reconstructed image be close to the original image to the greatest extent when the partial sub-image is received incorrectly, which improves the fault tolerance of the receiving device in reconstructing the image and enhances the robustness of the system.
  • An embodiment of the present invention provides a control terminal, where the control terminal includes the receiving device described in the foregoing embodiment.
  • the control terminal can be a remote control, a smart phone, a tablet, a ground control station, a laptop, a watch, a wristband, etc., and combinations thereof, and can also control the unmanned aerial vehicle on the ground.
  • FIG. 18 is a schematic diagram of an embodiment of the present invention
  • the unmanned aerial vehicle 1800 includes: a fuselage, a power system, and a video transmission system, the power system including at least one of the following: a motor 1801, a propeller 1802, and an electronic governor 1803.
  • a power system is mounted to the airframe for providing flight power;
  • a flight controller 1804 is in communication with the power system for controlling the unmanned flight 1806, wherein the flight controller 1804 includes an inertial measurement unit And gyroscopes.
  • the inertial measurement unit and the gyroscope are configured to detect an acceleration, a pitch angle, a roll angle, a yaw angle, and the like of the unmanned aerial vehicle.
  • the video transmission system includes one or more imaging devices 1805, and one or more processors 1806 disposed on the movable object.
  • the imaging device 1805 is coupled to the body through the support device 1807, and the processor 1806 and the imaging device 1805 are communicatively coupled.
  • the supporting device 1807 may specifically be a pan/tilt. The principle and implementation of the video transmission system are similar to the above embodiments, and are not described herein again.
  • the disclosed apparatus and method may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of hardware plus software functional units.
  • the above-described integrated unit implemented in the form of a software functional unit can be stored in a computer readable storage medium.
  • the above software functional unit is stored in a storage medium and includes a plurality of instructions for causing a computer device (which may be a personal computer, a server, or A network device or the like) or a processor performs part of the steps of the method described in various embodiments of the present invention.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like, which can store program codes. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Embodiments of the present invention provide a video transmitting and receiving method, system, and device, and an unmanned aerial vehicle. The method comprises: decomposing video data into a plurality of sub-video data units; separately encoding the plurality of sub-video data units; and selecting one or more encoded sub-video data units on the basis of one or more characteristics of a channel and one or more characteristics of the sub-video data units, and transmitting the one or more encoded sub-video data units. In the embodiments of the present invention, video data is decomposed into a plurality of sub-video data units, and the plurality of sub-video data units are separately encoded, and one or more encoded sub-video data units are selected according to a channel characteristic and a characteristic of the sub-video data units, such that the selected one or more encoded sub-video data units match the channel characteristic. In this way, the problem of mismatch between a signal source and a channel can be effectively resolved, and transmission delay and jitter of video data caused by signal source-channel mismatch can be effectively reduced.

Description

视频传输、接收方法、系统、设备及无人飞行器Video transmission, receiving method, system, device and unmanned aerial vehicle 技术领域Technical field
本发明实施例涉及图像处理领域,尤其涉及一种视频传输、接收方法、系统、设备及无人飞行器。Embodiments of the present invention relate to the field of image processing, and in particular, to a video transmission and reception method, system, device, and unmanned aerial vehicle.
背景技术Background technique
现有技术中无人飞行器搭载有拍摄设备,拍摄设备可进行航拍,并通过无人飞行器的通信系统将航拍的视频通过无线的方式发送给地面的接收设备例如用户终端、遥控器等。In the prior art, the unmanned aerial vehicle is equipped with a photographing device, and the photographing device can perform aerial photography, and the aerial video is transmitted to the ground receiving device such as a user terminal, a remote controller, and the like through a communication system of the unmanned aerial vehicle.
无人飞行器的拍摄设备在不同时间拍摄的场景或物体可能不同,因此,每帧图像数据对应的码流数据大小会实时变化(即信源会实时变化);此外,受到无人飞行器与接收设备之间的距离、相对位置、是否存在遮挡、是否存在电磁干扰等因素的影响,无人飞行器与接收设备之间的信道带宽也会实时变化(即信道会实时变化),信源信道的变化相互独立,难以预测。目前帧级别图像数据的传输和接收方式相对固定,难以适应实时变化的信源信道,缺乏有效的图像传输和图像接收方法,可能导致由于实时信源和信道失配问题而形成的图像传输延时抖动。The scenes or objects captured by the UAV's shooting equipment may be different at different times. Therefore, the code stream data size corresponding to each frame of image data will change in real time (ie, the source will change in real time); in addition, it will be subject to unmanned aerial vehicles and receiving equipment. The distance between the distance, the relative position, the presence of occlusion, and the presence of electromagnetic interference, etc., the channel bandwidth between the UAV and the receiving device will also change in real time (ie, the channel will change in real time), and the source channel changes to each other. Independent and unpredictable. At present, the transmission and reception modes of frame-level image data are relatively fixed, which is difficult to adapt to real-time changing source channels, and lacks effective image transmission and image receiving methods, which may result in image transmission delay due to real-time source and channel mismatch problems. shake.
发明内容Summary of the invention
本发明实施例提供一种视频传输、接收方法、系统、设备及无人飞行器,以有效降低视频数据的传输延时抖动。Embodiments of the present invention provide a video transmission and reception method, system, device, and an unmanned aerial vehicle to effectively reduce transmission delay jitter of video data.
本发明实施例的一个方面是提供一种视频传输方法,包括:An aspect of the embodiments of the present invention provides a video transmission method, including:
将视频数据分解为多个子视频数据单元,其中,每个子视频数据单元包括一个或多个子图像;Decomposing video data into a plurality of sub-video data units, wherein each sub-video data unit includes one or more sub-images;
对所述多个子视频数据单元分别进行编码;以及Encoding the plurality of sub-video data units separately;
基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。One or more encoded sub-video data units are selected and transmitted based on one or more characteristics of the channel, and one or more characteristics of the sub-video data unit.
本发明实施例的另一个方面是提供一种视频接收方法,包括: Another aspect of the embodiments of the present invention provides a video receiving method, including:
接收多个编码后的子视频数据单元;Receiving a plurality of encoded sub-video data units;
对所述多个编码后的子视频数据单元进行解码;Decoding the plurality of encoded sub-video data units;
根据解码后的子视频数据单元重建所述视频数据,其中,所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。Reconstructing the video data according to the decoded sub-video data unit, wherein the video data includes one or more image frames, and the sub-video data unit includes a plurality of sub-decompositions obtained by decomposing each of the image frames At least one sub-image in the image.
本发明实施例的另一个方面是提供一种视频传输系统,包括:Another aspect of the embodiments of the present invention provides a video transmission system, including:
一个或多个成像装置,被配置为采集视频数据;以及One or more imaging devices configured to acquire video data;
所述可移动物体上的一个或多个处理器,单独或协同工作,所述处理器被配置为:One or more processors on the movable object, working alone or in concert, the processor being configured to:
将视频数据分解为多个子视频数据单元,其中,每个子视频数据单元包括一个或多个子图像;Decomposing video data into a plurality of sub-video data units, wherein each sub-video data unit includes one or more sub-images;
对所述多个子视频数据单元分别进行编码;以及Encoding the plurality of sub-video data units separately;
基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。One or more encoded sub-video data units are selected and transmitted based on one or more characteristics of the channel, and one or more characteristics of the sub-video data unit.
本发明实施例的另一个方面是提供一种接收设备,包括:通讯接口、一个或多个处理器,单独或协同工作,所述通讯接口和所述处理器通讯连接;Another aspect of the embodiments of the present invention provides a receiving device, including: a communication interface, one or more processors, working alone or in cooperation, and the communication interface is in communication with the processor;
所述通讯接口用于接收多个编码后的子视频数据单元;The communication interface is configured to receive a plurality of encoded sub-video data units;
所述一个或多个处理器用于:控制解码器对所述多个编码后的子视频数据单元进行解码;根据解码后的子视频数据单元重建所述视频数据,其中,所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。The one or more processors are configured to: control a decoder to decode the plurality of encoded sub-video data units; reconstruct the video data according to the decoded sub-video data unit, wherein the video data includes one Or a plurality of image frames, the sub-video data unit including at least one of the plurality of sub-images obtained by decomposing each of the image frames.
本发明实施例的另一个方面是提供一种无人飞行器,包括:Another aspect of an embodiment of the present invention provides an unmanned aerial vehicle comprising:
机身;body;
动力系统,安装在所述机身,用于提供飞行动力;a power system mounted to the fuselage for providing flight power;
以及上述的视频传输系统。And the above video transmission system.
本实施例提供的视频传输、接收方法、系统、设备及无人飞行器,通过将视频数据分解为多个子视频数据单元,并对多个子视频数据单元分别 进行编码,根据信道特性和子视频数据单元的特性,选择一个或多个编码后的子视频数据单元并所述所述选中的子视频数据单元,以使选择出的一个或多个编码后的子视频数据单元符合信道特性,当选择出的一个或多个编码后的子视频数据单元在与其匹配的信道上传输时,可以有效地解决信源与信道的失配问题,可有效降低视频数据由于信源信道失配问题而造成的传输延时抖动。The video transmission and receiving method, the system, the device and the unmanned aerial vehicle provided by the embodiment are respectively decomposed into a plurality of sub-video data units by the video data, and respectively, and the plurality of sub-video data units are respectively Encoding, selecting one or more encoded sub-video data units and said selected sub-video data units according to channel characteristics and characteristics of sub-video data units, such that one or more encoded sub-elements are selected The video data unit conforms to the channel characteristics, and when the selected one or more encoded sub-video data units are transmitted on the channel matched with the same, the problem of mismatch between the source and the channel can be effectively solved, and the video data can be effectively reduced. Transmission delay jitter caused by source channel mismatch problem.
附图说明DRAWINGS
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly described below. It is obvious that the drawings in the following description are some embodiments of the present invention. Other drawings may also be obtained from those of ordinary skill in the art in view of the drawings.
图1为本发明实施例提供的传输延时抖动的示意图;1 is a schematic diagram of transmission delay jitter according to an embodiment of the present invention;
图2为本发明实施例提供的视频传输方法的流程图;2 is a flowchart of a video transmission method according to an embodiment of the present invention;
图3为本发明实施例提供的构成子视频数据单元的示意图;FIG. 3 is a schematic diagram of a sub video data unit according to an embodiment of the present invention; FIG.
图4为本发明另一实施例提供的构成子视频数据单元的示意图;FIG. 4 is a schematic diagram of a sub video data unit according to another embodiment of the present invention; FIG.
图5为本发明另一实施例提供的构成子视频数据单元的示意图;FIG. 5 is a schematic diagram of a sub video data unit according to another embodiment of the present invention; FIG.
图6为本发明另一实施例提供的视频传输方法的流程图;FIG. 6 is a flowchart of a video transmission method according to another embodiment of the present invention;
图7为本发明实施例提供的一帧图像的结构示意图;FIG. 7 is a schematic structural diagram of a frame image according to an embodiment of the present disclosure;
图8为本发明实施例提供的一帧图像经过哈达玛变换后的系数图像;FIG. 8 is a coefficient image of a frame image after Hadamard transform according to an embodiment of the present invention;
图9为本发明实施例提供的空间变换分解的示意图;FIG. 9 is a schematic diagram of spatial transformation decomposition according to an embodiment of the present invention; FIG.
图10为本发明实施例提供的空间下采样分解的示意图;FIG. 10 is a schematic diagram of spatial downsampling decomposition according to an embodiment of the present invention; FIG.
图11为本发明实施例提供的视频接收方法的流程图;FIG. 11 is a flowchart of a video receiving method according to an embodiment of the present invention;
图12为本发明实施例提供的解码后的子图像的示意图;FIG. 12 is a schematic diagram of a decoded sub-image according to an embodiment of the present invention;
图13为本发明另一实施例提供的解码后的子图像的示意图;FIG. 13 is a schematic diagram of a decoded sub-image according to another embodiment of the present invention; FIG.
图14为本发明实施例提供的重建原始图像的示意图;FIG. 14 is a schematic diagram of reconstructing an original image according to an embodiment of the present invention;
图15为本发明另一实施例提供的重建原始图像的示意图;FIG. 15 is a schematic diagram of reconstructing an original image according to another embodiment of the present invention; FIG.
图16为本发明实施例提供的视频传输系统的结构图;FIG. 16 is a structural diagram of a video transmission system according to an embodiment of the present invention;
图17为本发明实施例提供的接收设备的结构图;FIG. 17 is a structural diagram of a receiving device according to an embodiment of the present invention;
图18为本发明实施例提供的无人飞行器的结构图。 FIG. 18 is a structural diagram of an unmanned aerial vehicle according to an embodiment of the present invention.
附图标记:Reference mark:
11-子图像 12-子图像 13-子图像 14-子图像11-Sub Image 12-Sub Image 13-Sub Image 14-Sub Image
21-子图像 22-子图像 23-子图像 24-子图像21-Sub Image 22-Sub Image 23-Sub Image 24-Sub Image
31-子图像 32-子图像 33-子图像 34-子图像31-Sub Image 32-Sub Image 33-Sub Image 34-Sub Image
41-子图像 42-子图像 43-子图像 44-子图像41-Sub Image 42-Sub Image 43-Sub Image 44-Sub Image
51-子图像 52-子图像 53-子图像 54-子图像51-sub image 52-sub image 53-sub image 54-sub image
61-子图像 62-子图像 63-子图像 64-子图像61-sub image 62-sub image 63-sub image 64-sub image
310-子视频数据单元 320-子视频数据单元 330-子视频数据单元310-Sub Video Data Unit 320-Sub Video Data Unit 330-Sub Video Data Unit
340-子视频数据单元 410-子视频数据单元 420-子视频数据单元340-sub video data unit 410-sub video data unit 420-sub video data unit
430-子视频数据单元 50-图像帧 510-子视频数据单元430-Sub Video Data Unit 50-Image Frame 510-Sub Video Data Unit
520-子视频数据单元 530-子视频数据单元 540-子视频数据单元520-sub video data unit 530-sub video data unit 540-sub video data unit
550-子视频数据单元 560-子视频数据单元 570-子视频数据单元550-sub video data unit 560-sub video data unit 570-sub video data unit
580-子视频数据单元 1600-视频传输系统 1601-成像装置580-sub video data unit 1600-video transmission system 1601-imaging device
1602-处理器 1700-接收设备 1701-通讯接口 1702-处理器1602-processor 1700-receiving device 1701-communication interface 1702-processor
1800-无人飞行器 1801-电机 1802-螺旋桨1800-Unmanned aerial vehicle 1801-motor 1802-propeller
1803-电子调速器 1804-飞行控制器1803-Electronic governor 1804-Flight controller
1806-处理器 1807-支撑设备 1805-成像装置1806-processor 1807-support device 1805-imaging device
具体实施方式detailed description
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly described with reference to the accompanying drawings in the embodiments of the present invention. It is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments obtained by those skilled in the art based on the embodiments of the present invention without creative efforts are within the scope of the present invention.
需要说明的是,当组件被称为“固定于”另一个组件,它可以直接在另一个组件上或者也可以存在居中的组件。当一个组件被认为是“连接”另一个组件,它可以是直接连接到另一个组件或者可能同时存在居中组件。It should be noted that when a component is referred to as being "fixed" to another component, it can be directly on the other component or the component can be present. When a component is considered to "connect" another component, it can be directly connected to another component or possibly a central component.
除非另有定义,本文所使用的所有的技术和科学术语与属于本发明的技术领域的技术人员通常理解的含义相同。本文中在本发明的说明书中所使用的术语只是为了描述具体的实施例的目的,不是旨在于限制本 发明。本文所使用的术语“及/或”包括一个或多个相关的所列项目的任意的和所有的组合。All technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs, unless otherwise defined. The terminology used in the description of the present invention is for the purpose of describing particular embodiments, and is not intended to invention. The term "and/or" used herein includes any and all combinations of one or more of the associated listed items.
下面结合附图,对本发明的一些实施方式作详细说明。在不冲突的情况下,下述的实施例及实施例中的特征可以相互组合。Some embodiments of the present invention are described in detail below with reference to the accompanying drawings. The features of the embodiments and examples described below can be combined with each other without conflict.
图像数据的传输延时的稳定性是衡量图传系统性能的重要指标,保证图像数据传输延时是保证接收端的视频图像流畅显示的基本条件,然而,在图像传输的过程中,信源、信道的实时变化会引起帧与帧之间传输延时的抖动,降低图传系统的性能,下面先结合图1,分别以信源变化和信道变化为例,详细描述帧级别的图像数据传输延时的抖动问题。The stability of the transmission delay of image data is an important indicator to measure the performance of the image transmission system. The image data transmission delay is the basic condition for ensuring the smooth display of the video image at the receiving end. However, in the process of image transmission, the source and channel are used. The real-time change will cause the jitter of the transmission delay between frames and frames, and reduce the performance of the image transmission system. First, with reference to Figure 1, the source data changes and channel changes are taken as examples to describe the frame-level image data transmission delay in detail. Jitter problem.
图1包括场景一和场景二,在场景一中,发送端和接收端之间的信道的带宽保持稳定。在利用该信道传输图像数据的过程中,假设发送端的相机突然运动,或者相机拍摄范围内的物体突然快速移动,例如,在某一时刻,相机的拍摄对象是蓝色的天空,在下一时刻时,相机突然转向去拍摄在天空中飞行的五颜六色的热气球,导致帧4编码后对应的码流数据大小增加至帧3编码后对应的码流数据大小的两倍,即信源发生突然的变化,此时,帧4的传输延时会变为帧3的传输延时的两倍。Figure 1 includes scenario 1 and scenario 2. In scenario 1, the bandwidth of the channel between the sender and the receiver remains stable. In the process of transmitting image data using the channel, it is assumed that the camera at the transmitting end suddenly moves, or the object within the camera shooting range suddenly moves rapidly, for example, at a certain moment, the camera's subject is a blue sky at the next moment. The camera suddenly turns to shoot a colorful hot air balloon flying in the sky, causing the corresponding stream data size after frame 4 encoding to increase to twice the size of the corresponding stream data after the frame 3 encoding, that is, the sudden change of the source At this time, the transmission delay of frame 4 will become twice the transmission delay of frame 3.
在场景二中,每帧图像对应的码流数据大小基本保持稳定,即信源保持稳定。在传输图像数据的过程中,假设帧4对应的信道带宽突然降为帧3对应的信道带宽的一半,例如携带拍摄设备的无人飞行器在对拍摄对象进行拍摄时,拍摄对象基本不变,然而在飞行的过程中,无人飞行器突然靠近了附近的无线通讯基站,此时无线通讯基站会对无人飞行器的传输信道产生影响,即导致信道发生变化,信道的带宽下降为原带宽的一半,同样地,帧4的传输延时也会变为帧3的传输延时的两倍。In scenario 2, the code stream data corresponding to each frame image is basically stable, that is, the source remains stable. In the process of transmitting image data, it is assumed that the channel bandwidth corresponding to frame 4 suddenly drops to half of the channel bandwidth corresponding to frame 3. For example, when an unmanned aerial vehicle carrying a photographing device shoots a subject, the subject is substantially unchanged. During the flight, the UAV suddenly approaches the nearby wireless communication base station. At this time, the wireless communication base station will affect the transmission channel of the UAV, that is, the channel changes, and the bandwidth of the channel decreases to half of the original bandwidth. Similarly, the transmission delay of frame 4 will also become twice the transmission delay of frame 3.
从图1的描述可以看出,无论是信源变化还是信道变化,都会引起帧级别的图像数据传输延时的抖动,另外信源的变化和信道的变化相互独立,难以预测,目前对每帧图像数据的编码方式相对固定,无法适应实时变化的信源信道。It can be seen from the description of FIG. 1 that whether the source change or the channel change causes frame-level image data transmission delay jitter, and the source change and the channel change are independent of each other, and it is difficult to predict, currently for each frame. Image data is encoded in a relatively fixed manner and cannot be adapted to real-time changing source channels.
下面结合图1,详细描述本发明实施例提供的用于视频传输的方法,以有效降低视频数据的传输延时抖动。The method for video transmission provided by the embodiment of the present invention is described in detail below with reference to FIG. 1 to effectively reduce the transmission delay jitter of the video data.
为了解决上述问题,本发明实施例提供一种视频传输方法。图2为本 发明实施例提供的视频传输方法的流程图。如图2所示,本实施例中的方法,可以包括:In order to solve the above problem, an embodiment of the present invention provides a video transmission method. Figure 2 is the basis A flowchart of a video transmission method provided by an embodiment of the invention. As shown in FIG. 2, the method in this embodiment may include:
步骤S201、将视频数据分解为多个子视频数据单元,其中,每个子视频数据单元包括一个或多个子图像。Step S201, the video data is decomposed into a plurality of sub-video data units, wherein each sub-video data unit includes one or more sub-images.
本实施例的执行主体可以是具有图像处理功能的处理器、控制器,也可能是通用处理器,在这里不作具体的限定。本实施例以图像处理器为例介绍视频传输方法的原理,图像处理器实时获取无人飞行器上搭载的拍摄设备拍摄的视频数据,该视频数据可以包括一帧图像,也可以包括连续的多帧图像,图像处理器可将视频数据分解为多个子视频数据单元,本实施例不限定对视频数据进行分解后得到的子视频数据单元的个数,每个子视频数据单元包括一个或多个子图像。其中,将视频数据分解为多个子视频数据单元的一种可行的实现方式是:将视频数据包括的每个图像帧即每帧图像分解为多个子图像,选中每个图像帧的至少一个子图像,被选中的子图像构成子视频数据单元,即子视频数据单元包括视频数据中每一个图像帧分解得到的多个子图像中的至少一个。本实施例不限定一个视频数据包括的图像帧的个数,为了示意性说明视频数据的分解过程,假设该视频数据包括6个图像帧,即6帧图像,此外,在其他实施例中,视频数据包括的图像帧的个数还可以是其他值。The execution body of the embodiment may be a processor, a controller, or a general-purpose processor having an image processing function, and is not specifically limited herein. In this embodiment, an image processor is taken as an example to introduce the principle of a video transmission method. The image processor acquires video data captured by a shooting device mounted on an unmanned aerial vehicle in real time, and the video data may include one frame image or continuous multi-frame. The image processor may decompose the video data into a plurality of sub-video data units. The embodiment does not limit the number of sub-video data units obtained by decomposing the video data, and each sub-video data unit includes one or more sub-images. Wherein, a feasible implementation manner of decomposing the video data into a plurality of sub-video data units is: decomposing each image frame included in the video data, that is, each frame image into a plurality of sub-images, and selecting at least one sub-image of each image frame. The selected sub-image constitutes a sub-video data unit, that is, the sub-video data unit includes at least one of a plurality of sub-images obtained by decomposing each image frame in the video data. This embodiment does not limit the number of image frames included in one video data. To schematically illustrate the decomposition process of the video data, it is assumed that the video data includes 6 image frames, that is, 6 frames, and further, in other embodiments, the video The number of image frames included in the data may also be other values.
如图3所示,该视频数据包括6个图像帧,分别为帧1、帧2、帧3、帧4、帧5、帧6,对帧1、帧2、帧3、帧4、帧5、帧6分别进行分解,本实施例不限定每个图像帧分解成的子图像的个数,如图3所示,每个图像帧被分解为4个子图像,此处只是示意性说明,对每个图像帧分解后得到的子图像的个数还可以是其他值。每个子视频数据单元包括对6个图像帧中每一个进行分解后,每一个图像帧对应的4个子图像中的至少一个子图像。可选的,每个子视频数据单元包括6个图像帧中每一个图像帧对应的4个子图像中的一个子图像,如图3所示,子视频数据单元310包括帧1的一个子图像11、帧2的一个子图像21、帧3的一个子图像31、帧4的一个子图像41、帧5的一个子图像51、帧6的一个子图像61;同理,子视频数据单元320、子视频数据单元330、子视频数据单元340分别包括6个图像帧中每个图像帧的一个子图像。 As shown in FIG. 3, the video data includes six image frames, which are frame 1, frame 2, frame 3, frame 4, frame 5, frame 6, and frame 1, frame 2, frame 3, frame 4, and frame 5. The frame 6 is separately decomposed. In this embodiment, the number of sub-images into which each image frame is decomposed is not limited. As shown in FIG. 3, each image frame is decomposed into four sub-images, which is only schematically illustrated herein. The number of sub-images obtained after each image frame is decomposed may also be other values. Each of the sub-video data units includes at least one of the four sub-images corresponding to each of the image frames after each of the six image frames is decomposed. Optionally, each sub-video data unit includes one of the four sub-images corresponding to each of the six image frames. As shown in FIG. 3, the sub-video data unit 310 includes one sub-image 11 of the frame 1. One sub-image 21 of frame 2, one sub-image 31 of frame 3, one sub-image 41 of frame 4, one sub-image 51 of frame 5, and one sub-image 61 of frame 6; similarly, sub-video data unit 320, sub- The video data unit 330 and the sub video data unit 340 respectively include one sub image of each of the six image frames.
另外,不同子视频数据单元包括的子图像的个数还可以不同,如图4所示,子视频数据单元410包括帧1的两个子图像11、12,帧2的两个子图像21、22,帧3的两个子图像31、32,帧4的一个子图像41、帧5的一个子图像51、帧6的一个子图像61;子视频数据单元420包括帧1的一个子图像13,帧2的一个子图像23,帧3的一个子图像33,帧4的两个子图像42、43,帧5的两个子图像52、53,帧6的两个子图像62、63;子视频数据单元430包括6个图像帧中每个图像帧的一个子图像。In addition, the number of sub-images included in different sub-video data units may also be different. As shown in FIG. 4, the sub-video data unit 410 includes two sub-images 11, 12 of the frame 1, and two sub-images 21, 22 of the frame 2. Two sub-images 31, 32 of frame 3, one sub-image 41 of frame 4, one sub-image 51 of frame 5, one sub-image 61 of frame 6; sub-video data unit 420 includes one sub-image 13 of frame 1, frame 2 One sub-image 23, one sub-image 33 of frame 3, two sub-images 42, 43 of frame 4, two sub-images 52, 53 of frame 5, two sub-images 62, 63 of frame 6, sub-video data unit 430 includes One sub-image of each of the six image frames.
可选的,每个子视频数据单元包括的子图像不重叠。多个图像帧中每个图像帧对应的多个子图像中的至少一个子图像组合构成子视频数据单元的方式还可以有其他的组合方式,此处不再一一列举。Optionally, the sub-images included in each sub-video data unit do not overlap. The manner in which at least one of the plurality of sub-images corresponding to each of the plurality of image frames is combined to form a sub-video data unit may also have other combinations, which are not enumerated here.
此外,视频数据还可能只包括一个图像帧即一帧图像,如图5所示,50表示视频数据包括的一个图像帧,对该图像帧50进行分解,本实施例不限定对一个图像帧进行分解后得到的子图像的个数,可选的,将该图像帧50分解为4个子图像,如图5所示的子图像11、子图像12、子图像13、子图像14。用子图像11、子图像12、子图像13、子图像14构成子视频数据单元可分为如下几种可实现的方式:In addition, the video data may also include only one image frame, that is, one frame image. As shown in FIG. 5, 50 indicates an image frame included in the video data, and the image frame 50 is decomposed. This embodiment does not limit one image frame. The number of sub-images obtained after the decomposition is selected, and the image frame 50 is optionally decomposed into four sub-images, such as the sub-image 11, the sub-image 12, the sub-image 13, and the sub-image 14 as shown in FIG. Sub-picture 11, sub-image 12, sub-image 13, and sub-image 14 can be divided into the following achievable ways:
一种可实现的方式是:每个子视频数据单元包括一个子图像,如图5所示的子视频数据单元510、子视频数据单元520、子视频数据单元530、子视频数据单元540。One achievable manner is that each sub-video data unit includes one sub-image, such as sub-video data unit 510, sub-video data unit 520, sub-video data unit 530, and sub-video data unit 540 as shown in FIG.
另一种可实现的方式是:每个子视频数据单元包括两个子图像,本实施例不限定一个子视频数据单元中包括的两个子图像的组合方式,可选的,如图5所示的子视频数据单元550和子视频数据单元560,其中,子视频数据单元550包括子图像11和子图像12,子视频数据单元560包括子图像13和子图像14。Another achievable manner is that each sub-video data unit includes two sub-images. This embodiment does not limit the combination of two sub-images included in one sub-video data unit. Alternatively, the sub-image shown in FIG. Video data unit 550 and sub-video data unit 560, wherein sub-video data unit 550 includes sub-image 11 and sub-image 12, sub-video data unit 560 includes sub-image 13 and sub-image 14.
再一种可实现的方式是:每个子视频数据单元包括的子图像的个数不同,如图5所示的子视频数据单元570和子视频数据单元580,其中,子视频数据单元570包括3个子图像,子视频数据单元580包括1个子图像,或者,子视频数据单元570包括1个子图像,子视频数据单元580包括3个子图像,本实施例不限定,构成一个子视频数据单元的3个子图像的组合方式,可选的,子视频数据单元570包括子图像11、子图像12、子图 像13,子视频数据单元580包括子图像14。Yet another achievable manner is that each sub-video data unit includes a different number of sub-images, such as sub-video data unit 570 and sub-video data unit 580 as shown in FIG. 5, wherein sub-video data unit 570 includes three sub-pictures. The image, the sub-video data unit 580 includes one sub-image, or the sub-video data unit 570 includes one sub-image, and the sub-video data unit 580 includes three sub-images, which are not limited in this embodiment, and constitute three sub-images of one sub-video data unit. Combination, optional, sub-video data unit 570 includes sub-image 11, sub-image 12, sub-picture Like 13, the sub-video data unit 580 includes a sub-image 14.
步骤S202、对所述多个子视频数据单元分别进行编码。Step S202: encoding the plurality of sub-video data units separately.
图像处理器以每个子视频数据单元为编码单位,对多个子视频数据单元中的每个子视频数据单元分别进行编码,编码后得到多个码流数据,可选的,对一个子视频数据单元编码后得到一个码流数据,此处的编码包括信源编码和/或信道编码,信源编码的方式可以包括H.263,H.264,H.265,MPEG4等,信道编码的方式可以包括纠错编码,纠错码的类型可以包括RS码即里德-所罗门码、卷积码、Turbo码、Polar码、交织码、伪随机序列扰码等。The image processor separately encodes each of the plurality of sub-video data units by using each sub-video data unit as a coding unit, and obtains a plurality of code stream data after encoding, and optionally encodes one sub-video data unit. Then, a code stream data is obtained, where the coding includes source coding and/or channel coding, and the manner of source coding may include H.263, H.264, H.265, MPEG4, etc., and the channel coding method may include correcting The error coding type may include an RS code, that is, a Reed-Solomon code, a convolutional code, a Turbo code, a Polar code, an interleaving code, a pseudo random sequence scrambling code, and the like.
步骤S203、基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。Step S203: Select one or more encoded sub-video data units based on one or more characteristics of the channel and one or more characteristics of the sub-video data unit.
在本实施例中,信道的一个或多个特性至少包括带宽。或者,信道的一个或多个特性包括如下至少一种:噪声、干扰、信噪比、误比特率、衰落速率、带宽。In this embodiment, one or more characteristics of the channel include at least a bandwidth. Alternatively, one or more characteristics of the channel include at least one of: noise, interference, signal to noise ratio, bit error rate, fading rate, bandwidth.
子视频数据单元的一个或多个特性包括:所述子视频数据单元编码后的码流数据大小,或者所述子视频数据单元的能量集中度。The one or more characteristics of the sub-video data unit include: a code stream data size encoded by the sub-video data unit, or an energy concentration of the sub-video data unit.
图像处理器基于当前无线信道的一个或多个特性,以及子视频数据单元的一个或多个特性,从多个编码后的子视频数据单元中选择一个或多个通过无线信道进行传输,例如发送给接收设备,该接收设备可以是遥控器智能手机、平板电脑、地面控制站、膝上型电脑、手表、手环等及其组合。其中,选择一个或多个编码后的子视频数据单元可通过如下几种可行的方式实现:The image processor selects one or more of the plurality of encoded sub-video data units for transmission over the wireless channel based on one or more characteristics of the current wireless channel and one or more characteristics of the sub-video data unit, eg, transmitting To the receiving device, the receiving device can be a remote control smartphone, a tablet, a ground control station, a laptop, a watch, a wristband, and the like, and combinations thereof. Wherein, selecting one or more encoded sub-video data units can be implemented in the following feasible ways:
第一种可行的方式为:选择一个或多个编码后的子视频数据单元,以使所述一个或多个编码后的子视频数据单元的总的码流数据大小与信道带宽匹配。A first feasible way is to select one or more encoded sub-video data units such that the total code stream data size of the one or more encoded sub-video data units matches the channel bandwidth.
例如,图像处理器对视频数据分解后得到如图3所示的4个子视频数据单元310-340,对4个子视频数据单元分别编码后得到码流数据的大小为S0、S1、S2、S3,当前无线信道的带宽为T,则图像处理器可根据无线信道的带宽T,从4个码流数据中选择一个或多个码流数据,选择的依据可以是:被选择的一个或多个码流总的大小尽量接近于无线信道的带宽T, 即从多个子视频数据单元中选取一个或多个来进行组合,使得组合后子视频数据单元的总的码流数据大小尽可能接近无线信道的带宽,保证无线信道能够以当前的带宽发送码流数据尽可能大的子视频数据单元。例如,S0、S1、S2、S3的和小于或等于信道的当前带宽,可以选择编码后的子视频数据单元310、子视频数据单元320、子视频数据单元330、子视频数据单元340来传输,当S0、S1、S2、S3的和大于T,可以从S0、S1、S2、S3中选取码流数据最大的三个,假设码流数据最大的三个为S0、S1、S2,若S0、S1、S2的和小于T,则图像处理器可选择编码后的子视频数据单元310、子视频数据单元320、子视频数据单元330来传输,以此类推,当S0、S1、S2的和大于T,可以选取其他以当前的无线信道的带宽可以发送的码流数据的和最大的子视频数据单元的组合。For example, the image processor decomposes the video data to obtain four sub-video data units 310-340 as shown in FIG. 3, and encodes the four sub-video data units separately to obtain the size of the code stream data as S0, S1, S2, and S3. The bandwidth of the current wireless channel is T, and the image processor may select one or more code stream data from the four code stream data according to the bandwidth T of the wireless channel, and the selection may be based on: one or more selected codes. The total size of the stream is as close as possible to the bandwidth T of the wireless channel. That is, one or more of the plurality of sub-video data units are selected for combination, so that the total code stream data size of the combined sub-video data unit is as close as possible to the bandwidth of the wireless channel, and the wireless channel can transmit the code stream with the current bandwidth. The child video data unit is as large as possible. For example, the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, and the encoded sub-video data unit 310, the sub-video data unit 320, the sub-video data unit 330, and the sub-video data unit 340 may be selected for transmission. When the sum of S0, S1, S2, and S3 is greater than T, the three largest code stream data may be selected from S0, S1, S2, and S3, and the three largest code stream data are S0, S1, and S2, if S0, If the sum of S1 and S2 is less than T, the image processor may select the encoded sub-video data unit 310, the sub-video data unit 320, and the sub-video data unit 330 for transmission, and so on, when the sum of S0, S1, S2 is greater than T, other combinations of the code stream data that can be transmitted with the bandwidth of the current wireless channel and the largest sub-video data unit can be selected.
第二种可行的方式为:多个子视频数据单元依据所述能量集中度进行优先级排序。根据所述子视频数据单元的优先级和信道带宽,选择所述一个或多个编码后的子视频数据单元。The second feasible way is that a plurality of sub-video data units are prioritized according to the energy concentration. The one or more encoded sub-video data units are selected based on a priority of the sub-video data unit and a channel bandwidth.
由于多个子视频数据单元中的每个子视频数据单元的能量集中度可以相同或相近,也可以不同,因此,在其他实施例中,如果每个子视频数据单元的能量集中度不同,还可以基于每个子视频数据单元的能量集中度对该多个子视频数据单元进行优先级排序,可选的,能量集中度越大,优先级越高。例如,图像处理器对视频数据分解后得到如图3、4或5所示的4个子视频数据单元,该4个子视频数据单元分别记为子视频数据单元A、子视频数据单元B、子视频数据单元C、子视频数据单元D,且4个子视频数据单元的优先级依次递减,对子视频数据单元A编码后的码流数据大小为S0,对子视频数据单元B编码后的码流数据大小为S1,对子视频数据单元C编码后的码流数据大小为S2,对子视频数据单元D编码后的码流数据大小为S3。Since the energy concentration of each of the plurality of sub-video data units may be the same or similar, it may be different. Therefore, in other embodiments, if the energy concentration of each sub-video data unit is different, it may be based on each The energy concentration of the sub-video data units prioritizes the plurality of sub-video data units. Alternatively, the greater the energy concentration, the higher the priority. For example, the image processor decomposes the video data to obtain four sub-video data units as shown in FIG. 3, 4 or 5, and the four sub-video data units are respectively recorded as sub-video data unit A, sub-video data unit B, and sub-video. The data unit C, the sub-video data unit D, and the priority of the four sub-video data units are successively decremented, the code stream data size encoded by the sub-video data unit A is S0, and the code stream data encoded by the sub-video data unit B is sequentially obtained. The size of the code stream data encoded by the sub video data unit C is S2, and the size of the code stream data encoded by the sub video data unit D is S3.
在某些情况下,图像处理器需要根据编码后子视频数据单元的码流数据大小和信道带宽来选择子视频数据单元并传输,在某些情况下需要根据子视频数据单元的优先级和信道的带宽选择子视频数据单元并传输。例如,从上述4个编码后的子视频数据单元中选择出的一个或多个编码后的子视频数据单元,选择的依据可以是:被选择的一个或多个编码后的子视 频数据单元的总的码流数据大小小于信道带宽,且优先选择优先级较高的子视频数据单元,从而保证高优先级的子视频数据单元优先被发送。例如,当S0、S1、S2、S3的和小于或等于信道的当前带宽,可以选择编码后的子视频数据单元A、子视频数据单元B、子视频数据单元C、子视频数据单元D来传输,当S0、S1、S2、S3的和大于T,则确定前3个优先级较高的码流数据即编码后的子视频数据单元A、编码后的子视频数据单元B和编码后的子视频数据单元C,若S0、S1、S2的和小于T,则图像处理器对编码后的子视频数据单元A、编码后的子视频数据单元B和编码后的子视频数据单元C进行传输;若S0、S1、S2的和大于T,则确定前两个优先级较高的子视频数据单元A、子视频数据单元B,以此类推,保证至少将优先级最高的子视频数据单元A传输。In some cases, the image processor needs to select and transmit the sub-video data unit according to the code stream data size and channel bandwidth of the encoded sub-video data unit, and in some cases, according to the priority and channel of the sub-video data unit. The bandwidth selects the sub-video data unit and transmits it. For example, one or more encoded sub-video data units selected from the above four encoded sub-video data units may be selected according to one or more selected sub-views selected. The total code stream data size of the frequency data unit is smaller than the channel bandwidth, and the higher priority sub video data unit is preferentially selected, thereby ensuring that the high priority sub video data unit is preferentially transmitted. For example, when the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, the encoded sub-video data unit A, sub-video data unit B, sub-video data unit C, and sub-video data unit D may be selected for transmission. When the sum of S0, S1, S2, and S3 is greater than T, determining the first three higher priority code stream data, that is, the encoded sub video data unit A, the encoded sub video data unit B, and the encoded sub- The video data unit C, if the sum of S0, S1, S2 is less than T, the image processor transmits the encoded sub-video data unit A, the encoded sub-video data unit B, and the encoded sub-video data unit C; If the sum of S0, S1, and S2 is greater than T, the first two sub-video data units A and sub-video data units B having higher priority are determined, and so on, to ensure that at least the sub-video data unit A having the highest priority is transmitted. .
本实施例通过将视频数据分解为多个子视频数据单元,并对多个子视频数据单元分别进行编码,根据信道特性和子视频数据单元的特性,选择一个或多个编码后的子视频数据单元,以使选择出的一个或多个编码后的子视频数据单元符合信道特性,可以有效地解决信源与信道的失配问题,可有效降低视频数据由于信源信道失配问题而造成的传输延时抖动。In this embodiment, the video data is decomposed into a plurality of sub-video data units, and the plurality of sub-video data units are separately encoded, and one or more encoded sub-video data units are selected according to channel characteristics and characteristics of the sub-video data units. The selected one or more encoded sub-video data units conform to the channel characteristics, which can effectively solve the problem of mismatch between the source and the channel, and can effectively reduce the transmission delay of the video data due to the source channel mismatch problem. shake.
本发明实施例提供一种视频传输方法。图6为本发明另一实施例提供的视频传输方法的流程图。在图2所示实施例的基础上,视频数据包括一个或多个图像帧,如图6所示,本实施例中的方法,可以包括:Embodiments of the present invention provide a video transmission method. FIG. 6 is a flowchart of a video transmission method according to another embodiment of the present invention. On the basis of the embodiment shown in FIG. 2, the video data includes one or more image frames. As shown in FIG. 6, the method in this embodiment may include:
步骤S601、将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像。Step S601: Decompose each of the one or more image frames in the video data into a plurality of sub-images.
在上述实施例的基础上,视频数据可以包括一帧图像,也可以包括连续的多帧图像,本实施例不限定一帧图像包括的像素点的个数,也不限定每个像素点的像素值。图像处理器对视频数据进行分解时,具体可以将该视频数据中的一帧或多帧图像中的每一帧图像分解为多个子图像,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。On the basis of the above embodiments, the video data may include one frame image, and may also include consecutive multi-frame images. This embodiment does not limit the number of pixels included in one frame image, and does not limit the pixels of each pixel. value. When the image processor decomposes the video data, specifically, each frame image of one or more frames in the video data may be decomposed into a plurality of sub-images, and the sub-video data unit includes each of the image frames. At least one of the plurality of sub-images obtained after the decomposition is performed.
具体的,将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像,本实施例以视频数据包括的一帧图像为例,介绍对该帧图 像进行空间分解的过程,具体可通过如下几种可行的方式实现:Specifically, each of the one or more image frames in the video data is decomposed into a plurality of sub-images. In this embodiment, a frame image included in the video data is taken as an example to introduce the frame image. The process of spatial decomposition can be achieved in several possible ways:
第一种可行的方式:利用傅里叶相关变换或正交变换将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。A first possible way to decompose each of the one or more image frames in the video data into a plurality of sub-images using a Fourier correlation transform or an orthogonal transform.
所述傅里叶相关变换或正交变换是从哈达玛变换、离散余弦变换、离散傅里叶变换、沃尔什-哈达玛变换、哈尔变换或斜变换中选择的。本实施例具体以哈达玛变换为例,介绍对该帧图像进行空间分解的过程。在其他实施例中,还可以采用其他的空间变换对该帧图像进行空间分解。The Fourier correlation transform or orthogonal transform is selected from a Hadamard transform, a discrete cosine transform, a discrete Fourier transform, a Walsh-Hadamard transform, a Haar transform, or an oblique transform. This embodiment specifically takes the Hadamard transform as an example to introduce a process of spatially decomposing the frame image. In other embodiments, other spatial transforms may also be used to spatially decompose the frame image.
如图7所示为一帧图像的示意图,本实施例不限定一帧图像包括的像素点的个数,以一帧图像包括16个像素点为例,其中,P1-P16表示16个像素点的像素值,对于16个像素点中每4个相邻的像素点的像素值进行空间变换分解,分解成4个子图像,下面以哈达玛变换来进行示意性说明,具体的空间变换分解过程包括如下步骤:As shown in FIG. 7 , a schematic diagram of a frame image is used. This embodiment does not limit the number of pixels included in a frame image, and a frame image includes 16 pixels. For example, P1-P16 represents 16 pixels. The pixel value is spatially transformed and decomposed into pixel values of every four adjacent pixel points of the 16 pixel points, and is decomposed into four sub-images. The following is schematically illustrated by Hadamard transform, and the specific spatial transformation decomposition process includes The following steps:
步骤1、以16个像素点中每4个相邻的像素点为一个单位进行哈达玛变换,例如,P1、P2、P3、P4经过哈达玛变换后得到的转换系数为H1、H2、H3、H4,其中,P1、P2、P3、P4和H1、H2、H3、H4之间的关系满足公式(1)、(2)、(3)、(4): Step 1. Perform a Hadamard transform by using four adjacent pixels of the 16 pixels as a unit. For example, the conversion coefficients obtained by the Hadamard transform of P1, P2, P3, and P4 are H1, H2, and H3. H4, wherein the relationship between P1, P2, P3, P4 and H1, H2, H3, H4 satisfies the formulas (1), (2), (3), (4):
H1=(P1+P2+P3+P4+1)>>1       (1)H1=(P1+P2+P3+P4+1)>>1 (1)
H2=(P1+P2–P3–P4+1)>>1        (2)H2=(P1+P2–P3–P4+1)>>1 (2)
H3=(P1+P3–P2–P4+1)>>1       (3)H3=(P1+P3–P2–P4+1)>>1 (3)
H4=(P1+P4–P2–P3+1)>>1       (4)H4=(P1+P4–P2–P3+1)>>1 (4)
根据公式(1)、(2)、(3)、(4)可知H1包含有4个像素点的平均能量,H2包含有4个像素点在垂直方向的平均梯度,H3包含有4个像素点在水平方向的平均梯度,H4包含有4个像素点的交叉梯度即纹理信息。可见,H1的能量集中度最高,H2和H3的能量集中度次之,H4的能量集中度最小,在用H1、H2、H3和H4恢复P1、P2、P3、P4时,H1是必不可少的,即H1最重要,其次是H2、H3重要,最后是H4重要,因此,H1最重要、H2、H3的重要性次之、H4的重要性最小,即H1、H2、H3、H4的重要性依次递减。同理,对P5-P8进行同样的哈达玛变换得到H5-H8,H5、H6、H7、H8的能量集中度依次递减,H5、H6、H7、H8的重要性依次递减,对P9-P12进行同样的哈达玛变换得到H9-H12,H9、 H10、H11、H12的能量集中度依次递减,H9、H10、H11、H12的重要性依次递减,对P13-P16进行同样的哈达玛变换得到H13-H16,H13、H14、H15、H16的能量集中度依次递减,H13、H14、H15、H16的重要性依次递减,得到如图8所示的系数图像。According to formulas (1), (2), (3), (4), H1 contains the average energy of 4 pixels, H2 contains the average gradient of 4 pixels in the vertical direction, and H3 contains 4 pixels. In the horizontal gradient, H4 contains a cross gradient of 4 pixels, ie texture information. It can be seen that the energy concentration of H1 is the highest, the energy concentration of H2 and H3 is second, and the energy concentration of H4 is the smallest. When using P1, P2, P3, and P4 with H1, H2, H3, and H4, H1 is essential. H1 is the most important, followed by H2 and H3, and finally H4 is important. Therefore, H1 is the most important, H2 and H3 are the second most important, and H4 is the least important, that is, H1, H2, H3, and H4 are important. Sex is decremented in turn. Similarly, the same Hadamard transform is performed on P5-P8 to obtain H5-H8, and the energy concentration of H5, H6, H7 and H8 is successively decreased, and the importance of H5, H6, H7 and H8 is successively decreased, and P9-P12 is successively performed. The same Hadamard transform gets H9-H12, H9, The energy concentration of H10, H11 and H12 decreases in turn, and the importance of H9, H10, H11 and H12 decreases successively. The same Hadamard transformation is performed on P13-P16 to obtain the energy concentration of H13-H16, H13, H14, H15 and H16. The degrees are successively decremented, and the importance of H13, H14, H15, and H16 is successively decreased, and a coefficient image as shown in FIG. 8 is obtained.
步骤2、将哈达玛变换后得到的转换系数分解到不同的子图像中,本实施例不限定对每帧图像空间变换分解后得到的子图像的个数,例如分解后得到的子图像的个数是4个,此处只是示意性说明,在其他实施例中,对每帧图像空间变换分解后得到的子图像的个数还可以是其他值。可选的,将H1分到第一个子图像、将H2分到第二个子图像、将H3分到第三个子图像、将H4分到第四个子图像。同理,将H5-H8分解到4个子图像中,将H9-H12分解到4个子图像中,将H13-H16分解到4个子图像中,得到如图9所示的分解结果,其中,子图像1集中了能量集中度最高的转换系数,子图像2和子图像3集中了能量集中度次之的转换系数,子图像4集中了能量集中度最小的转换系数,因此子图像1最重要,子图像2、子图像3的重要性次之,子图像4的重要性最小。Step 2: Decompose the conversion coefficients obtained by the Hadamard transform into different sub-images. This embodiment does not limit the number of sub-images obtained by spatially transforming each frame of the image, for example, the sub-images obtained after the decomposition. The number is four, which is only a schematic description. In other embodiments, the number of sub-images obtained by spatially transforming and decomposing each frame of image may be other values. Optionally, H1 is assigned to the first sub-image, H2 is assigned to the second sub-image, H3 is assigned to the third sub-image, and H4 is assigned to the fourth sub-image. Similarly, H5-H8 is decomposed into 4 sub-images, H9-H12 is decomposed into 4 sub-images, and H13-H16 is decomposed into 4 sub-images, and the decomposition result shown in FIG. 9 is obtained, wherein the sub-image 1 Concentrates the conversion coefficient with the highest energy concentration, sub-image 2 and sub-image 3 concentrate the conversion coefficient of energy concentration second, and sub-image 4 concentrates the conversion coefficient with the smallest energy concentration, so sub-image 1 is the most important, sub-image 2. The importance of sub-image 3 is second, sub-image 4 is of the least importance.
根据图9可知,空间变换分解后的4个子图像中每个子图像的分辨率是分解前原始图像的四分之一。As can be seen from FIG. 9, the resolution of each of the four sub-images after the spatial transformation decomposition is one quarter of the original image before the decomposition.
第二种可行的方式:利用空间下采样将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。A second possible way to use spatial downsampling to decompose each of the one or more image frames in the video data into a plurality of sub-images.
本实施例不限定对每帧图像空间下采样分解后得到的子图像的个数,例如分解后得到的子图像的个数是4个,此处只是示意性说明,在其他实施例中,对每帧图像空间下采样分解后得到的子图像的个数还可以是其他值,在图7的基础上,对于16个像素点中每4个相邻的像素点的像素值进行空间下采样,分解成4个子图像,具体的空间下采样分解过程为:以16个像素点中每4个相邻的像素点为一个单位,将一个单位中的4个像素点分解到不同的子图像中,例如,将P1分解到第一个子图像,将P2分解到第二个子图像,将P3分解到第三个子图像,将P4分解到第四个子图像,同理,将P5-P8分解到4个子图像,将P9-P12分解到4个子图像,将P13-P16分解到4个子图像,得到如图10所示的分解结果。This embodiment does not limit the number of sub-images obtained by down-sampling the image space of each frame. For example, the number of sub-images obtained after decomposition is four, which is only schematically illustrated herein. In other embodiments, The number of sub-images obtained by down-sampling the image space of each frame may also be other values. On the basis of FIG. 7, spatially down-sampling the pixel values of every four adjacent pixel points of the 16 pixel points is performed. Decomposed into four sub-images, the specific spatial down-sampling decomposition process is: decomposing four pixel points in one unit into different sub-images by using four adjacent pixel points of 16 pixels as one unit. For example, P1 is decomposed into the first sub-image, P2 is decomposed into the second sub-image, P3 is decomposed into the third sub-image, and P4 is decomposed into the fourth sub-image. Similarly, P5-P8 is decomposed into 4 sub-images. For the image, P9-P12 is decomposed into 4 sub-images, and P13-P16 is decomposed into 4 sub-images, and the decomposition result as shown in FIG. 10 is obtained.
根据图10可知,空间下采样分解后的4个子图像中每个子图像的分辨率是分解前原始图像的四分之一。不失一般性,假设分解前原始图像的 大小为W*H,若将原始图像分解为4个子图像,原始图像或子图像对应的像素矩阵的行号或列号从0开始计数,则第一个子图像中可包含原始图像中坐标为(2i,2j)的像素点,其中,
Figure PCTCN2017078728-appb-000001
第二个子图像中可包含原始图像中坐标为(2i+1,2j)的像素点,第三个子图像中可包含原始图像中坐标为(2i,2j+1)的像素点,第四个子图像中可包含原始图像中坐标为(2i+1,2j+1)的像素点。
As can be seen from FIG. 10, the resolution of each of the four sub-images after spatial downsampling is one quarter of the original image before the decomposition. Without loss of generality, it is assumed that the size of the original image before decomposition is W*H. If the original image is decomposed into four sub-images, the row number or column number of the pixel matrix corresponding to the original image or sub-image is counted from 0, then the first The sub-images may include pixels in the original image with coordinates (2i, 2j), where
Figure PCTCN2017078728-appb-000001
The second sub-image may include a pixel with coordinates (2i+1, 2j) in the original image, and the third sub-image may include a pixel with coordinates (2i, 2j+1) in the original image, and the fourth sub-image. The pixel in the original image with coordinates (2i+1, 2j+1) may be included.
根据上述分解方式中的任一一种即空间变换或空间下采样方式即可将视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。在本实施例中,一个图像帧即为一帧图像,多个图像帧即为多帧图像。所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。Each of the one or more image frames in the video data can be decomposed into a plurality of sub-images according to any of the above-described decomposition modes, that is, spatial transformation or spatial downsampling. In this embodiment, one image frame is one frame image, and a plurality of image frames are multi-frame images. The video data includes one or more image frames, the sub-video data unit including at least one of a plurality of sub-images obtained by decomposing each of the image frames.
若视频数据中有一个图像帧,则该图像帧的分解方式如图9或图10所示,子视频数据单元可以包括该图像帧分解后得到的多个子图像中的至少一个子图像,例如一个子视频数据单元包括一个子图像,对分解后得到的每个子图像进行压缩、编码即可得到编码后的码流数据。If there is an image frame in the video data, the image frame is decomposed in the manner shown in FIG. 9 or FIG. 10, and the sub-video data unit may include at least one of the plurality of sub-images obtained by decomposing the image frame, for example, one. The sub-video data unit includes a sub-image, and each sub-image obtained after the decomposition is compressed and encoded to obtain encoded code stream data.
若视频数据中有连续的多个图像帧,则每一个图像帧即每一帧图像的分解方式如图9或图10所示,例如视频数据包括24个图像帧,若每一个图像帧被分解为4个子图像,则连续的24个图像帧被分解后将得到24*4个子图像,每个子视频数据单元可包括24*4个子图像中的多个子图像,具体地,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的每个图像帧对应的多个子图像中的至少一个子图像,所述子视频数据单元可以为4个,其中,所述子视频数据单元中的每一个都包括24个图像帧中每一个图像帧的一个子图像。对每个子视频数据单元进行压缩、编码即可得到编码后的码流数据。If there are consecutive image frames in the video data, each image frame, that is, each frame image is decomposed as shown in FIG. 9 or FIG. 10, for example, the video data includes 24 image frames, and if each image frame is decomposed For 4 sub-images, 24 consecutive image frames are decomposed to obtain 24*4 sub-images, and each sub-video data unit may include multiple sub-images of 24*4 sub-images, specifically, the sub-video data unit At least one of the plurality of sub-images corresponding to each image frame obtained by decomposing each of the image frames, the sub-video data unit may be four, wherein the sub-video data unit Each of them includes a sub-image of each of the 24 image frames. The encoded code stream data is obtained by compressing and encoding each sub-video data unit.
根据图9或图10可知,每个子图像包括图像帧的一部分。如图10所示,每个子图像包括图像帧的一个或多个像素。如图9所示,每个子图像包括图像帧的一个或多个转换系数。接收设备在重建如图7所述的一帧图像时,图10中,子图像1、子图像2、子图像3、子图像4的重要性相同。图9中,子图像1最重要,子图像2、子图像3的重要性次之,子图像4的重要性最低,在本实施例中,多个子视频数据单元依据所述能量集中度进行优先级排序。由于子图像1集中了能量集中度最高的转换系数,子图 像2和子图像3集中了能量集中度次之的转换系数,子图像4集中了能量集中度最小的转换系数,若一个子视频数据单元包括一个子图像,则包含子图像1的子视频数据单元的优先级最高,包含子图像2或子图像3的子视频数据单元的优先级次之,包含子图像4的子视频数据单元的优先级最低。As can be seen from FIG. 9 or FIG. 10, each sub-image includes a portion of an image frame. As shown in FIG. 10, each sub-image includes one or more pixels of an image frame. As shown in Figure 9, each sub-image includes one or more conversion coefficients of the image frame. When the receiving device reconstructs a frame of image as described in FIG. 7, in FIG. 10, the importance of the sub-image 1, the sub-image 2, the sub-image 3, and the sub-image 4 are the same. In FIG. 9, the sub-image 1 is the most important, and the sub-image 2 and the sub-image 3 are the second most important, and the sub-image 4 has the lowest importance. In this embodiment, the plurality of sub-video data units are prioritized according to the energy concentration. Sort by level. Since sub-image 1 concentrates the conversion coefficient with the highest energy concentration, the sub-picture The image 2 and the sub-image 3 concentrate the conversion coefficient of the energy concentration second, and the sub-image 4 concentrates the conversion coefficient with the smallest energy concentration. If a sub-video data unit includes a sub-image, the sub-video data unit including the sub-image 1 The priority of the sub-video data unit including sub-image 2 or sub-image 3 is the highest, and the sub-video data unit including sub-image 4 has the lowest priority.
步骤S602、对所述多个子视频数据单元分别进行编码。Step S602, encoding the plurality of sub-video data units separately.
对所述多个子视频数据单元分别进行编码可通过如下几种可行的方式实现:The encoding of the plurality of sub-video data units separately can be implemented in the following feasible ways:
第一种可行的方式:多个子视频数据单元由多个单独的编码器进行编码。A first possible way: multiple sub-video data units are encoded by a plurality of separate encoders.
具体的,采用所述多个单独的编码器对所述多个子视频数据单元并行编码;或者,采用不同的视频编码规则对所述多个子视频数据单元进行编码;或者,采用相同的视频编码规则对所述多个子视频数据单元进行编码。Specifically, the plurality of sub-video data units are encoded in parallel by using the plurality of separate encoders; or the plurality of sub-video data units are encoded by using different video encoding rules; or, the same video encoding rule is adopted. The plurality of sub-video data units are encoded.
第二种可行的方式:多个子视频数据单元中的两个或多个由同一个编码器进行编码。A second possible way: two or more of the plurality of sub-video data units are encoded by the same encoder.
第三种可行的方式:基于运动补偿的视频压缩标准对所述多个子视频数据单元中的至少一个进行编码。A third possible way: encoding at least one of the plurality of sub-video data units based on a motion compensated video compression standard.
第四种可行的方式:依据不同的压缩率对所述多个子视频数据单元进行压缩。该压缩率是根据所述子视频数据单元的一个或多个特性确定的。A fourth feasible way: compressing the plurality of sub-video data units according to different compression ratios. The compression ratio is determined based on one or more characteristics of the sub-video data unit.
步骤S603、基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。Step S603: Select one or more encoded sub-video data units based on one or more characteristics of the channel and one or more characteristics of the sub-video data unit.
在本实施例中,若一个子视频数据单元包括一个子图像,如图9所示,子图像1最重要,子图像2、子图像3的重要性次之,子图像4的重要性最低,则包含子图像1的子视频数据单元的优先级最高,包含子图像2或子图像3的子视频数据单元的优先级次之,包含子图像4的子视频数据单元的优先级最低。对包含子图像1的子视频数据单元A编码后的码流数据大小为S0,对包含子图像2的子视频数据单元B编码后的码流数据大小为S1,对包含子图像3的子视频数据单元C编码后的码流数据大小为S2,对包含子图像4的子视频数据单元D编码后的码流数据大小为S3。图像处理器可根据历史数据的收发错误状态、当前子视频数据单元的优先级、 当前子视频数据单元编码后的码流数据大小,以及实时信道的估计值例如信道带宽,对4个子视频数据单元进行组合发送,以达到信源和信道实时匹配。例如,当S0、S1、S2、S3的和小于或等于信道的当前带宽,可以选择编码后的子视频数据单元A、子视频数据单元B、子视频数据单元C、子视频数据单元D来传输,当S0、S1、S2、S3的和大于T,则确定前3个优先级较高的码流数据即编码后的子视频数据单元A、编码后的子视频数据单元B和编码后的子视频数据单元C,若S0、S1、S2的和小于T,则图像处理器对编码后的子视频数据单元A、编码后的子视频数据单元B和编码后的子视频数据单元C进行传输;若S0、S1、S2的和大于T,则确定前两个优先级较高子视频数据单元A、子视频数据单元B,以此类推,保证至少将优先级最高的子视频数据单元A传输。In this embodiment, if one sub-video data unit includes one sub-image, as shown in FIG. 9, the sub-image 1 is the most important, the sub-image 2, the sub-image 3 are of the second most important, and the sub-image 4 has the lowest importance. Then, the sub-video data unit including the sub-image 1 has the highest priority, and the sub-video data unit including the sub-image 2 or the sub-image 3 has the lower priority, and the sub-video data unit including the sub-image 4 has the lowest priority. The code stream data size encoded by the sub video data unit A including the sub image 1 is S0, the code stream data size encoded by the sub video data unit B including the sub image 2 is S1, and the sub video including the sub image 3 is used. The code stream data size encoded by the data unit C is S2, and the code stream data size encoded by the sub video data unit D including the sub image 4 is S3. The image processor can receive and transmit error status according to historical data, priority of the current sub-video data unit, The size of the code stream data encoded by the current sub-video data unit, and the estimated value of the real-time channel, such as the channel bandwidth, are combined and transmitted to the four sub-video data units to achieve real-time matching between the source and the channel. For example, when the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, the encoded sub-video data unit A, sub-video data unit B, sub-video data unit C, and sub-video data unit D may be selected for transmission. When the sum of S0, S1, S2, and S3 is greater than T, determining the first three higher priority code stream data, that is, the encoded sub video data unit A, the encoded sub video data unit B, and the encoded sub- The video data unit C, if the sum of S0, S1, S2 is less than T, the image processor transmits the encoded sub-video data unit A, the encoded sub-video data unit B, and the encoded sub-video data unit C; If the sum of S0, S1, S2 is greater than T, then the first two higher priority sub-video data units A, sub-video data units B are determined, and so on, ensuring that at least the highest priority sub-video data unit A is transmitted.
在某些实施例中,若一个子视频数据单元包括一个子图像,如图10所示,子图像1、子图像2、子图像3、子图像4的重要性相同,或者在某些情况下子图像1、子图像2、子图像3、子图像4的重要性不能确定,即包含子图像1的子视频数据单元A、包含子图像2的子视频数据单元B、包含子图像3的子视频数据单元C、包含子图像4的子视频数据单元D的优先级相同或者优先级无法确定,对子视频数据单元A编码后的码流数据大小为S0,对子视频数据单元B编码后的码流数据大小为S1,对子视频数据单元C编码后的码流数据大小为S2,对子视频数据单元D编码后的码流数据大小为S3。图像处理器可根据历史数据的收发错误状态、当前子视频数据单元的码流数据大小,以及实时信道的估计值例如信道带宽,对4个子视频数据单元进行组合发送,以达到信源和信道实时匹配。例如,S0、S1、S2、S3的和小于或等于信道的当前带宽,可以选择编码后的子视频数据单元A、子视频数据单元B、子视频数据单元C、子视频数据单元D来传输,当S0、S1、S2、S3的和大于T,可以从S0、S1、S2、S3选取码流数据最大的三个,假设码流数据最大的三个为S0、S1、S2,若S0、S1、S2的和小于T,则图像处理器可选择编码后的子视频数据单元A、子视频数据单元B、子视频数据单元C来传输,以此类推,当S0、S1、S2的和大于T,可以选取其他以当前的无线信道的带宽可以发送的码流数据的和最大的子视频数据单元的组合。 In some embodiments, if one sub-video data unit includes one sub-image, as shown in FIG. 10, sub-image 1, sub-image 2, sub-image 3, sub-image 4 have the same importance, or in some cases The importance of the image 1, the sub-image 2, the sub-image 3, and the sub-image 4 cannot be determined, that is, the sub-video data unit A including the sub-image 1, the sub-video data unit B including the sub-image 2, and the sub-video including the sub-image 3. The data unit C, the sub-video data unit D including the sub-image 4 have the same priority or the priority cannot be determined, and the code stream data size encoded by the sub-video data unit A is S0, and the code encoded by the sub-video data unit B The stream data size is S1, the code stream data size encoded by the sub video data unit C is S2, and the code stream data size encoded by the sub video data unit D is S3. The image processor can jointly transmit the four sub-video data units according to the transmission and error status of the historical data, the current stream data size of the sub-video data unit, and the estimated value of the real-time channel, such as the channel bandwidth, to achieve real-time source and channel. match. For example, the sum of S0, S1, S2, and S3 is less than or equal to the current bandwidth of the channel, and the encoded sub-video data unit A, the sub-video data unit B, the sub-video data unit C, and the sub-video data unit D may be selected for transmission. When the sum of S0, S1, S2, and S3 is greater than T, the three largest code stream data can be selected from S0, S1, S2, and S3, and the three largest code stream data are S0, S1, S2, if S0, S1 If the sum of S2 is less than T, the image processor may select the encoded sub-video data unit A, the sub-video data unit B, and the sub-video data unit C to transmit, and so on, when the sum of S0, S1, S2 is greater than T. It is possible to select other combinations of the code stream data that can be transmitted with the bandwidth of the current wireless channel and the largest sub-picture data unit.
此外,若无线信道允许的延时抖动对应的发送数据量为D,则组合后的码流数据大小可落在[T-D,T+D]的范围内。In addition, if the amount of transmitted data corresponding to the delay jitter allowed by the wireless channel is D, the combined code stream data size may fall within the range of [T-D, T+D].
本实施例通过将视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像,具体的空间分解包括空间变换分解和空间下采样分解,使得分解后得到的子图像包括图像帧的一个或多个像素,或者包括图像帧的一个或多个转换系数,由分解后得到的子图像组合成子视频数据单元,在发送编码后的子视频数据单元时,使得子视频数据单元编码后的码流数据大小与信道的特性(例如带宽)相匹配,或者使得子视频数据单元的优先级与信道的特性(例如带宽)相匹配,这样能实现信源和信道匹配,减小了帧级别的视频数据由于信源信道失配问题而造成的传输延时抖动。In this embodiment, by decomposing each of the one or more image frames in the video data into a plurality of sub-images, the specific spatial decomposition includes spatial transformation decomposition and spatial down sampling decomposition, so that the sub-image obtained after the decomposition includes the image frame. One or more pixels, or one or more conversion coefficients including an image frame, the sub-images obtained by the decomposition are combined into a sub-video data unit, and when the encoded sub-video data unit is transmitted, the sub-video data unit is encoded The code stream data size matches the characteristics of the channel (such as bandwidth), or the priority of the sub-video data unit is matched to the characteristics of the channel (such as bandwidth), which enables source and channel matching and reduces the frame level. The transmission delay jitter caused by the video channel mismatch problem.
本发明实施例提供一种视频接收方法。图11为本发明实施例提供的视频接收方法的流程图。如图11所示,本实施例中的方法,可以包括:Embodiments of the present invention provide a video receiving method. FIG. 11 is a flowchart of a video receiving method according to an embodiment of the present invention. As shown in FIG. 11, the method in this embodiment may include:
步骤S1101、接收多个编码后的子视频数据单元。Step S1101, receiving a plurality of encoded sub-video data units.
,其中,所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。And wherein the video data includes one or more image frames, and the sub-video data unit includes at least one of a plurality of sub-images obtained by decomposing each of the image frames.
具体的,当视频数据包括一帧或多帧图像时,对视频数据的分解可以如图3、图4或图5所示,得到子视频数据单元;其中,对视频数据的每一帧图像进行分解可以如图9或图10所示,具体分解过程和上述实施例一致,此处不再赘述。Specifically, when the video data includes one or more frames of images, the decomposition of the video data may be as shown in FIG. 3, FIG. 4 or FIG. 5, and the sub-video data unit is obtained; wherein each frame of the video data is performed. The decomposition may be as shown in FIG. 9 or FIG. 10, and the specific decomposition process is consistent with the foregoing embodiment, and details are not described herein again.
在本实施例中,接收设备接收无人飞行器的通信系统发送的多个编码后的子视频数据单元,其中视频数据可以包括一个或多个图像帧,无人飞行器在发送视频数据前,将视频数据中包括的每一个图像帧都分解成多个子图像,具体的分解方法请参见本文前述部分,子视频数据单元包括每一个图像帧分解得到的每个图像帧对应的多个子图像中的至少一个,即每一个子视频数据单元中都包括每一个图像帧的至少一个子图像,以降低每个子视频数据单元中子图像之间的相关性,避免由于子视频数据单元在传输过程中丢失或失真而导致相关性较高的子图像丢失或失真。因为,同一帧图像分解后得到的子图像之间的相关性较高,而相关性较高的子图像丢失 或失真时,将很难恢复由相关性较高的子图像构成的图像帧。In this embodiment, the receiving device receives a plurality of encoded sub-video data units transmitted by the communication system of the unmanned aerial vehicle, wherein the video data may include one or more image frames, and the unmanned aerial vehicle transmits the video data before transmitting the video data. Each image frame included in the data is decomposed into a plurality of sub-images. For the specific decomposition method, refer to the foregoing section, where the sub-video data unit includes at least one of a plurality of sub-images corresponding to each image frame decomposed by each image frame. , that is, each sub-video data unit includes at least one sub-image of each image frame to reduce the correlation between the sub-images in each sub-video data unit, to avoid loss or distortion due to the sub-video data unit during transmission. The resulting sub-images with high correlation are lost or distorted. Because the correlation between the sub-images obtained after the same frame image is decomposed is higher, and the sub-images with higher correlation are lost. Or when it is distorted, it will be difficult to recover an image frame composed of a highly correlated sub-image.
例如,接收设备接收到4个子视频数据单元,每个子视频数据单元包括一个如图9或图10分解得到的子图像。For example, the receiving device receives 4 sub-video data units, and each sub-video data unit includes a sub-image obtained as shown in FIG. 9 or FIG.
步骤S1102、对所述多个编码后的子视频数据单元进行解码。Step S1102: Decode the plurality of encoded sub-video data units.
具体的,对所述多个编码后的子视频数据单元分别进行解码,例如,接收设备对该4个编码后的子视频数据单元分别进行解码,即对4个子图像对应的码流数据分别进行解码,得到解码后的子视频数据单元。当子视频数据单元在无线信道中传输时,由于噪声干扰、多径效应、衰落等原因,可能导致接收设备解码后得到的子视频数据单元与无人飞行器的通信系统实际发送的编码后的子视频数据单元不同,从而导致接收设备接收错误。Specifically, the plurality of encoded sub-video data units are separately decoded. For example, the receiving device separately decodes the four encoded sub-video data units, that is, separately performs code stream data corresponding to the four sub-images. Decode to obtain the decoded sub-video data unit. When the sub-video data unit is transmitted in the wireless channel, the sub-video data unit obtained by the receiving device and the encoded system actually transmitted by the communication system of the UAV may be caused due to noise interference, multipath effect, fading, and the like. The video data units are different, causing the receiving device to receive an error.
若通信系统发送的是如图9所示的4个编码后的子图像,则接收设备对码流数据解码后得到的4个子图像如图12所示,其中,若子图像1传输正确,则H1和h1相同、H2和h2相同、H3和h3相同、H4和h4相同,若子图像1传输错误,则H1和h1、H2和h2、H3和h3、H4和h4中至少有一组不相同。同理,其他子图像传输正确或传输错误,传输之前的转换系数和传输之后的转换系数也存在同样的关系。If the communication system transmits the four encoded sub-images as shown in FIG. 9, the four sub-images obtained by the receiving device after decoding the code stream data are as shown in FIG. 12, wherein if the sub-image 1 is transmitted correctly, H1 Same as h1, H2 and h2 are the same, H3 and h3 are the same, H4 and h4 are the same, and if sub-image 1 is transmitted incorrectly, at least one of H1 and h1, H2 and h2, H3 and h3, H4 and h4 are different. Similarly, other sub-pictures are transmitted correctly or transmitted incorrectly, and the conversion coefficients before transmission and the conversion coefficients after transmission also have the same relationship.
若通信系统发送的是如图10所示的4个编码后的子图像,则接收设备对码流数据解码后得到的4个子图像如图13所示,其中,若子图像1传输正确,则P1和p1相同、P2和p2相同、P3和p3相同、P4和p4相同,若子图像1传输错误,则P1和p1、P2和p2、P3和p3、P4和p4中至少有一组不相同。同理,其他子图像传输正确或传输错误,传输之前的像素和传输之后的像素也存在同样的关系。If the communication system transmits the four encoded sub-images as shown in FIG. 10, the four sub-images obtained by the receiving device after decoding the code stream data are as shown in FIG. 13, wherein if the sub-image 1 is transmitted correctly, P1 The same as p1, P2 and p2 are the same, P3 and p3 are the same, and P4 and p4 are the same. If sub-image 1 is transmitted incorrectly, at least one of P1 and p1, P2 and p2, P3 and p3, P4 and p4 are different. Similarly, other sub-images are transmitted correctly or transmitted incorrectly, and the pixels before transmission and the pixels after transmission also have the same relationship.
步骤S1103、根据解码后的子视频数据单元重建所述视频数据。Step S1103: Reconstruct the video data according to the decoded sub video data unit.
具体的,检测所述子视频数据单元的一个或多个子图像的传输错误,并根据接收正确的子图像重建所述视频数据。例如,接收设备对码流数据解码得到4个子图像后,检测每个子图像传输正确或传输错误,并根据接收正确的子图像重建原始图像。Specifically, a transmission error of one or more sub-images of the sub-video data unit is detected, and the video data is reconstructed according to receiving the correct sub-image. For example, after the receiving device decodes the code stream data to obtain 4 sub-images, it detects that each sub-image is transmitted correctly or transmits an error, and reconstructs the original image according to receiving the correct sub-image.
在本实施例中,通信系统发送的子图像为图9所示的子图像,接收设备接收到的子图像如图12所示,在图12中,假设子图像2接收错误,子图像1、 子图像3、子图像4均接收正确,则接收设备根据图12所示的子图像1、子图像3、子图像4重建原始图像,在重建原始图像时,采用逆变换重建视频数据,且给子视频数据单元中传输错误的子图像分配一个值,一种可行的实现方式是:给子视频数据单元中传输错误的子图像分配的值为0。例如,在图12中,将子图像2的h2、h6、h10、h14均置为0,通过上述实施例可知,H1、H2、H3、H4是根据P1、P2、P3、P4得到的,因此,在重建原始图像时,需要对h1、h2、h3、h4进行哈达玛逆变换,若空间变换分解图像帧时采用的是其他的空间变换,则接收设备重建原始图像时,采用相应的空间变换的逆变换,本实施例对h1、h2、h3、h4进行哈达玛逆变换后,得到p1、p2、p3、p4,其中,h1、h2、h3、h4和p1、p2、p3、p4之间的关系满足公式(5)、(6)、(7)、(8):In the present embodiment, the sub-image transmitted by the communication system is the sub-image shown in FIG. 9, and the sub-image received by the receiving device is as shown in FIG. 12. In FIG. 12, it is assumed that the sub-image 2 receives an error, and the sub-image 1, The sub-image 3 and the sub-image 4 are all received correctly, and the receiving device reconstructs the original image according to the sub-image 1, the sub-image 3, and the sub-image 4 shown in FIG. 12, and reconstructs the original image, reconstructs the video data by using an inverse transform, and gives The sub-picture in which the error is transmitted in the sub-picture data unit is assigned a value. A possible implementation is to assign a value of 0 to the sub-picture that transmits the error in the sub-video data unit. For example, in FIG. 12, h2, h6, h10, and h14 of the sub-image 2 are both set to 0. As can be seen from the above embodiment, H1, H2, H3, and H4 are obtained based on P1, P2, P3, and P4, When reconstructing the original image, it is necessary to perform inverse Hadamard transform on h1, h2, h3, and h4. If spatial transformation is used to decompose the image frame, other spatial transformations are used. When the receiving device reconstructs the original image, the corresponding spatial transformation is adopted. In the inverse transformation of this embodiment, after hdama inverse transformation of h1, h2, h3, h4, p1, p2, p3, p4 are obtained, wherein h1, h2, h3, h4 and p1, p2, p3, p4 The relationship satisfies the formulas (5), (6), (7), (8):
p1=(h1+h2+h3+h4+1)>>1       (5)P1=(h1+h2+h3+h4+1)>>1 (5)
p2=(h1+h2-h3-h4+1)>>1        (6)P2=(h1+h2-h3-h4+1)>>1 (6)
p3=(h1+h3-h2-h4+1)>>1        (7)P3=(h1+h3-h2-h4+1)>>1 (7)
p4=(h1+h4-h2-h3+1)>>1        (8)P4=(h1+h4-h2-h3+1)>>1 (8)
其中,h2为0,H1和h1相同、H3和h3相同、H4和h4相同,因此,经过哈达玛逆变换得到的p1、p2、p3、p4和原始图像中的像素值P1、P2、P3、P4可能不同,但是根据接收正确的子图像重建原始图像,能够保证重建的图像接近原始图像。同理,对h5、h6、h7、h8进行哈达玛逆变换得到p5、p6、p7、p8,对h9、h10、h11、h12进行哈达玛逆变换得到p9、p10、p11、p12,对h13、h14、h15、h16进行哈达玛逆变换得到p13、p14、p15、p16,其中,h6、h10、h14均为0,再根据哈达玛逆变换得到的p1-p16重建原始图像,如图14所示。Where h2 is 0, H1 and h1 are the same, H3 and h3 are the same, and H4 and h4 are the same. Therefore, p1, p2, p3, p4 obtained by inverse Hadamard transform and pixel values P1, P2, P3 in the original image, P4 may be different, but reconstructing the original image based on receiving the correct sub-image ensures that the reconstructed image is close to the original image. Similarly, h5, h6, h7, h8 are inversely transformed by Hadamard to obtain p5, p6, p7, p8, and h9, h10, h11, h12 are inversely transformed by Hadamama to obtain p9, p10, p11, p12, h13, H14, h15, h16 perform inverse Hadamard transform to obtain p13, p14, p15, p16, wherein h6, h10, h14 are all 0, and reconstruct the original image according to p1-p16 obtained by inverse Hadamard transform, as shown in Fig. 14. .
在其他实施例中,通信系统发送的子图像为图10所示的子图像,接收设备接收到的子图像如图13所示,在图13中,假设子图像3接收错误,子图像1、子图像2、子图像4均接收正确,则接收设备根据图13所示的子图像1、子图像2、子图像3、子图像4重建原始图像,在重建原始图像时,给子视频数据单元中传输错误的子图像分配一个值,另一种可行的实现方式是:给子视频数据单元中传输错误的子图像分配的值是采用插值法确定的,具体的,给子视频数据单元中传输错误的子图像分配的值是根据传输 正确的子图像确定的,且传输错误的子图像和传输正确的子图像来自同一个图像帧。例如,在图13中,子图像3接收错误,子图像1、子图像2、子图像4均接收正确,则子图像3不参与重建过程,即接收设备只根据子图像1、子图像2、子图像4重建原始图像,具体过程如下:由于原始图像包括16个像素,子图像1、子图像2、子图像4一共有12个像素,根据图10可知,原始图像中16个像素点中每4个相邻的像素点被分解到4个不同的子图像中,因此,根据子图像1、子图像2、子图像4重建原始图像时,子图像1的第一个像素p1、子图像2的第一个像素p2、子图像4的第一个像素p4分别是原始图像第一组相邻4个像素中的3个像素P1、P2、P4,同理,p5、p6、p8分别是原始图像P5-P8中的3个像素P5、P6、P8,p9、p10、p12分别是原始图像P9-P12中的3个像素P9、P10、P12,p13、p14、p16分别是原始图像P13-P16中的3个像素P 13、P 14、P 16,根据p1、p2、p4、p5、p6、p8、p9、p10、p12、p13、p14、p16可得到如图15所示的图像A,图像A中空余的像素即是接收错误的子图像3包括的像素,由于子图像3接收错误,因此,子图像3不能参与重建原始图像的过程。由于p1、p2、p4、p5、p6、p8、p9、p10、p12、p13、p14、p16都是正确接收的,即p1、p2、p4、p5、p6、p8、p9、p10、p12、p13、p14、p16分别与原始图像相同位置上的像素点相同,本实施例可根据插值法确定图像A中空余的像素值,一种可行的插值法是:p3等于p1、p2、p4的算数平均值,p7等于p5、p6、p8的算数平均值,p11等于p9、p10、p12的算数平均值,p15等于p13、p14、p16的算数平均值,从而得到重建的原始图像B。其中,这里只是给出了插值确定图像A中空余的像素值的一种方法,本领域技术人员可以采用其他的插值方式确定图像A中空余像素值,在这里不做具体限定。In other embodiments, the sub-image transmitted by the communication system is the sub-image shown in FIG. 10, and the sub-image received by the receiving device is as shown in FIG. 13, and in FIG. 13, it is assumed that the sub-image 3 receives the error, and the sub-image 1, The sub-image 2 and the sub-image 4 are all received correctly, and the receiving device reconstructs the original image according to the sub-image 1, the sub-image 2, the sub-image 3, and the sub-image 4 shown in FIG. 13, and when the original image is reconstructed, the sub-video data unit is given. The sub-image of the transmission error is assigned a value. Another feasible implementation is that the value assigned to the sub-picture transmitted in the sub-video data unit is determined by interpolation, specifically, for transmission in the sub-video data unit. The value of the wrong sub-image assignment is based on the transmission The correct sub-image is determined, and the sub-image transmitted incorrectly and the sub-image transmitted correctly are from the same image frame. For example, in FIG. 13, the sub-image 3 receives an error, and the sub-image 1, the sub-image 2, and the sub-image 4 are all received correctly, and the sub-image 3 does not participate in the reconstruction process, that is, the receiving device only according to the sub-image 1, the sub-image 2, The sub-image 4 reconstructs the original image. The specific process is as follows: since the original image includes 16 pixels, the sub-image 1, the sub-image 2, and the sub-image 4 have a total of 12 pixels. According to FIG. 10, each of the 16 pixels in the original image is known. Four adjacent pixels are decomposed into four different sub-images. Therefore, when the original image is reconstructed from the sub-image 1, sub-image 2, and sub-image 4, the first pixel p1 of the sub-image 1 and the sub-image 2 The first pixel p2 and the first pixel p4 of the sub-image 4 are respectively three pixels P1, P2, and P4 of the first four adjacent pixels of the original image. Similarly, p5, p6, and p8 are original. The three pixels P5, P6, P8, p9, p10, and p12 in the image P5-P8 are the three pixels P9, P10, and P12 in the original image P9-P12, respectively, and the p13, p14, and p16 are the original images P13-P16, respectively. The three pixels P 13 , P 14 , P 16 in accordance with p1, p2, p4, p5, p6, p8, p9, p10, p12, p13, p14 p16 obtained image A shown in FIG. 15, the pixel image A spare reception error that is included in the sub-image pixel 3 due to reception error sub-image 3, and therefore, the sub-image 3 can not participate in the process of reconstruction of the original image. Since p1, p2, p4, p5, p6, p8, p9, p10, p12, p13, p14, and p16 are all correctly received, ie, p1, p2, p4, p5, p6, p8, p9, p10, p12, p13 , p14, p16 are the same as the pixel points at the same position of the original image. In this embodiment, the pixel value of the image A can be determined according to the interpolation method. A feasible interpolation method is: p3 is equal to the arithmetic average of p1, p2, and p4 The value, p7 is equal to the arithmetic mean of p5, p6, p8, p11 is equal to the arithmetic mean of p9, p10, p12, and p15 is equal to the arithmetic mean of p13, p14, p16, resulting in the reconstructed original image B. Here, only one method of interpolating the pixel value of the image A is provided, and those skilled in the art may use other interpolation methods to determine the residual pixel value of the image A, which is not specifically limited herein.
本实施例通过接收设备对多个子视频数据单元分别进行解码,得到解码后的子视频数据单元,并根据解码后的子视频数据单元重建原始图像,具体的,根据解码后接收正确的子图像重建原始图像,可以在部分子图像接收错误的情况下,使得重建的图像最大程度的接近原始图像,提高了接收设备在重建图像过程中的容错能力,增强了系统的鲁棒性。 In this embodiment, the receiving device separately decodes the plurality of sub-video data units to obtain the decoded sub-video data unit, and reconstructs the original image according to the decoded sub-video data unit, and specifically, receives the correct sub-image reconstruction according to the decoding. The original image can make the reconstructed image be close to the original image to the greatest extent when the partial sub-image is received incorrectly, which improves the fault tolerance of the receiving device in reconstructing the image and enhances the robustness of the system.
本发明实施例提供一种视频传输系统。图16为本发明实施例提供的视频传输系统的结构图,如图16所示,视频传输系统1600包括一个或多个成像装置1601、以及可移动物体上的一个或多个处理器1602,一个或多个成像装置1601被配置为采集视频数据;一个或多个处理器1602单独或协同工作,被配置为:将视频数据分解为多个子视频数据单元,其中,每个子视频数据单元包括一个或多个子图像;对所述多个子视频数据单元分别进行编码;以及基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。Embodiments of the present invention provide a video transmission system. FIG. 16 is a structural diagram of a video transmission system according to an embodiment of the present invention. As shown in FIG. 16, the video transmission system 1600 includes one or more imaging devices 1601, and one or more processors 1602 on a movable object. Or the plurality of imaging devices 1601 are configured to acquire video data; the one or more processors 1602 work alone or in concert, configured to: decompose the video data into a plurality of sub-video data units, wherein each sub-video data unit includes one or Multiple sub-images; encoding the plurality of sub-video data units separately; and selecting one or more encoded sub-videos based on one or more characteristics of the channel and one or more characteristics of the sub-video data unit Data unit and transfer.
其中,所述信道的一个或多个特性至少包括带宽。或者,所述信道的一个或多个特性包括如下至少一种:噪声、干扰、信噪比、误比特率、衰落速率、带宽。Wherein one or more characteristics of the channel include at least a bandwidth. Alternatively, the one or more characteristics of the channel include at least one of: noise, interference, signal to noise ratio, bit error rate, fading rate, bandwidth.
其中,所述子视频数据单元的一个或多个特性包括:所述子视频数据单元编码后的码流数据大小,或者所述子视频数据单元的能量集中度。The one or more characteristics of the sub video data unit include: a code stream data size encoded by the sub video data unit, or an energy concentration of the sub video data unit.
可选的,所述一个或多个编码后的子视频数据单元的总的码流数据大小与信道带宽匹配。或者,所述多个子视频数据单元依据所述能量集中度进行优先级排序。处理器1602选择一个或多个编码后的子视频数据单元时,被配置为:根据所述子视频数据单元的优先级和信道带宽,选择所述一个或多个编码后的子视频数据单元。Optionally, the total code stream data size of the one or more encoded sub-video data units matches the channel bandwidth. Alternatively, the plurality of sub-video data units are prioritized according to the energy concentration. The processor 1602, when selecting one or more encoded sub-video data units, is configured to select the one or more encoded sub-video data units based on a priority and a channel bandwidth of the sub-video data units.
本发明实施例提供的视频传输系统的具体原理和实现方式均与图2所示实施例类似,此处不再赘述。The specific principles and implementation manners of the video transmission system provided by the embodiment of the present invention are similar to the embodiment shown in FIG. 2, and details are not described herein again.
本实施例通过将视频数据分解为多个子视频数据单元,并对多个子视频数据单元分别进行编码,根据信道特性和子视频数据单元的特性,选择一个或多个编码后的子视频数据单元,以使选择出的一个或多个编码后的子视频数据单元符合信道特性,当选择出的一个或多个编码后的子视频数据单元在与其匹配的信道上传输时,可以有效地解决信源与信道的失配问题,可有效降低视频数据由于信源信道失配问题而造成的传输延时抖动。In this embodiment, the video data is decomposed into a plurality of sub-video data units, and the plurality of sub-video data units are separately encoded, and one or more encoded sub-video data units are selected according to channel characteristics and characteristics of the sub-video data units. The selected one or more encoded sub-video data units are made to conform to channel characteristics, and when the selected one or more encoded sub-video data units are transmitted on the channel matched thereto, the source and the source can be effectively solved. The channel mismatch problem can effectively reduce the transmission delay jitter of video data due to the source channel mismatch problem.
本发明实施例提供一种视频传输系统。在图16所示实施例提供的技术方案的基础上,所述视频数据包括一个或多个图像帧。处理器1602将 视频数据分解为多个子视频数据单元时,被配置为:将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像,其中,所述子视频数据单元中的每一个包括所述图像帧中每一个分解得到的多个子图像中的至少一个子图像。每个子图像包括所述图像帧的一部分。具体的,每个子图像包括所述图像帧的一个或多个像素。或者,每个子图像包括所述图像帧的一个或多个转换系数。Embodiments of the present invention provide a video transmission system. Based on the technical solution provided by the embodiment shown in FIG. 16, the video data includes one or more image frames. Processor 1602 will When the video data is decomposed into a plurality of sub-video data units, configured to: decompose each of the one or more image frames in the video data into a plurality of sub-images, wherein each of the sub-video data units At least one of the plurality of sub-images obtained by each of the image frames is included. Each sub-image includes a portion of the image frame. Specifically, each sub-image includes one or more pixels of the image frame. Alternatively, each sub-image includes one or more conversion coefficients of the image frame.
具体的,处理器1602将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像时,被配置为:将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Specifically, when the processor 1602 decomposes each of the one or more image frames in the video data into a plurality of sub-images, the processor 1602 is configured to: each of the one or more image frames in the video data A space is decomposed into multiple sub-images.
处理器1602将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像时,被配置为:利用傅里叶相关变换或正交变换将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。其中,所述傅里叶相关变换或正交变换是从哈达玛变换、离散余弦变换、离散傅里叶变换、沃尔什-哈达玛变换、哈尔变换或斜变换中选择的。When the processor 1602 decomposes each of the one or more image frames in the video data into a plurality of sub-images, the processor 1602 is configured to: use one of the video data by Fourier correlation transform or orthogonal transform Or each of the plurality of image frames is decomposed into a plurality of sub-images. Wherein, the Fourier correlation transform or orthogonal transform is selected from a Hadamard transform, a discrete cosine transform, a discrete Fourier transform, a Walsh-Hadamard transform, a Haar transform or a skew transform.
或者,处理器1602将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像时,被配置为:利用空间下采样将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Alternatively, when the processor 1602 decomposes each of the one or more image frames in the video data into a plurality of sub-images, the processor 1602 is configured to: spatially downsample one or more images in the video data Each space in the frame is decomposed into multiple sub-images.
另外,处理器1602对所述多个子视频数据单元分别进行编码时可通过如下几种可行的方式实现:In addition, when the processor 1602 separately encodes the plurality of sub video data units, the following may be implemented in several feasible manners:
第一种可行的方式:The first possible way:
处理器1602控制多个编码器对所述多个子视频数据单元进行编码。具体的,处理器1602控制所述多个编码器对所述多个子视频数据单元并行编码;或者,处理器1602控制所述多个编码器分别采用不同的视频编码规则对所述多个子视频数据单元进行编码;再或者,处理器1602控制所述多个编码器采用相同的视频编码规则对所述多个子视频数据单元进行编码。The processor 1602 controls a plurality of encoders to encode the plurality of sub-video data units. Specifically, the processor 1602 controls the plurality of encoders to perform parallel encoding on the plurality of sub video data units; or the processor 1602 controls the plurality of encoders to respectively use the different video encoding rules to the plurality of sub video data. The unit performs encoding; or alternatively, the processor 1602 controls the plurality of encoders to encode the plurality of sub-video data units using the same video encoding rule.
第二种可行的方式:处理器1602控制编码器对所述多个子视频数据单元中的两个或多个进行编码。A second possible way: the processor 1602 controls the encoder to encode two or more of the plurality of sub-video data units.
第三种可行的方式:处理器1602控制编码器基于运动补偿的视频压缩标准对所述多个子视频数据单元中的至少一个进行编码。 A third possible way: the processor 1602 controls the encoder to encode at least one of the plurality of sub-video data units based on a motion compensated video compression standard.
第四种可行的方式:处理器1602被配置为:依据不同的压缩率对所述多个子视频数据单元进行压缩。其中,压缩率是根据所述子视频数据单元的一个或多个特性确定的。A fourth possible way: the processor 1602 is configured to compress the plurality of sub-video data units according to different compression ratios. Wherein the compression ratio is determined according to one or more characteristics of the sub-video data unit.
其中,编码器可以是独立于处理器1602且与处理器1602电连接的一个硬件实体,也可以是处理器1602内用于实现编码功能的软件。The encoder may be a hardware entity that is independent of the processor 1602 and is electrically connected to the processor 1602, or may be software for implementing the encoding function in the processor 1602.
在本实施例中,可移动物体为无人飞行器。一个或多个成像装置1601通过载体与所述可移动物体连接,该载体可以是一种多轴万向节。In this embodiment, the movable object is an unmanned aerial vehicle. One or more imaging devices 1601 are coupled to the movable object by a carrier, which may be a multi-axis universal joint.
本发明实施例提供的视频传输系统的具体原理和实现方式均与图2-10所示实施例类似,此处不再赘述。The specific principles and implementations of the video transmission system provided by the embodiment of the present invention are similar to the embodiment shown in FIG. 2-10, and details are not described herein again.
本实施例通过将视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像,具体的空间分解包括空间变换分解和空间下采样分解,使得分解后得到的子图像包括图像帧的一个或多个像素,或者包括图像帧的一个或多个转换系数,对于一帧图像,可对其分解后得到的多个子图像进行组合后发送,以使组合后的码流数据大小与信道带宽匹配,或者组合后的码流数据还包括重要性较高的子图像对应的码流数据,使得每一帧图像在无线信道上传输时,均能实现信源和信道匹配,减小了帧级别的视频数据由于信源信道失配问题而造成的传输延时抖动。In this embodiment, by decomposing each of the one or more image frames in the video data into a plurality of sub-images, the specific spatial decomposition includes spatial transformation decomposition and spatial down sampling decomposition, so that the sub-image obtained after the decomposition includes the image frame. One or more pixels, or one or more conversion coefficients including an image frame, for a frame image, a plurality of sub-images obtained by the decomposition thereof may be combined and transmitted, so that the combined code stream data size and channel The bandwidth matching, or the combined code stream data further includes code stream data corresponding to the sub-images with higher importance, so that each frame image can be matched on the wireless channel to achieve source and channel matching, and the frame is reduced. The transmission delay jitter caused by the level channel video data due to the source channel mismatch problem.
本发明实施例提供一种接收设备。图17为本发明实施例提供的接收设备的结构图,如图17所示,接收设备1700包括通讯接口1701、一个或多个处理器1702,一个或多个处理器1702单独或协同工作,通讯接口1701和处理器1702通讯连接;通讯接口1701用于接收多个编码后的子视频数据单元,,其中,所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像;一个或多个处理器1702用于:控制解码器对所述多个编码后的子视频数据单元进行解码;根据解码后的子视频数据单元重建所述视频数据。The embodiment of the invention provides a receiving device. FIG. 17 is a structural diagram of a receiving device according to an embodiment of the present invention. As shown in FIG. 17, the receiving device 1700 includes a communication interface 1701, one or more processors 1702, and one or more processors 1702 work alone or in cooperation. The interface 1701 is communicatively coupled to the processor 1702; the communication interface 1701 is configured to receive a plurality of encoded sub-video data units, wherein the video data includes one or more image frames, and the sub-video data unit includes the At least one of the plurality of sub-images obtained after each of the image frames is decomposed; the one or more processors 1702 are configured to: control the decoder to decode the plurality of encoded sub-video data units; The subsequent sub-video data unit reconstructs the video data.
可选的,处理器1702控制解码器对所述多个编码后的子视频数据单元进行解码时,具体用于:控制解码器对所述多个编码后的子视频数据单元分别进行解码。其中,解码器可以是独立于处理器1702且与处理器 1702电连接的一个硬件实体,也可以是处理器1702内用于实现解码功能的软件。Optionally, when the processor 1702 controls the decoder to decode the plurality of encoded sub-video data units, the processor 1702 is specifically configured to: control the decoder to separately decode the plurality of encoded sub-video data units. Wherein, the decoder may be independent of the processor 1702 and with the processor A hardware entity that is electrically connected to 1702 may also be software for implementing a decoding function in the processor 1702.
另外,处理器1702还用于:检测所述子视频数据单元的一个或多个子图像的传输错误;处理器1702根据解码后的子视频数据单元重建所述视频数据时,具体用于:根据接收正确的子图像重建所述视频数据。In addition, the processor 1702 is further configured to: detect a transmission error of one or more sub-images of the sub-video data unit; when the processor 1702 reconstructs the video data according to the decoded sub-video data unit, specifically, according to the receiving The correct sub-image reconstructs the video data.
此外,处理器1702还用于给所述子视频数据单元中传输错误的子图像分配一个值。In addition, the processor 1702 is further configured to assign a value to the sub-image in which the error is transmitted in the sub-video data unit.
其中,给所述子视频数据单元中传输错误的子图像分配的值为0。Wherein, the value assigned to the sub-image in which the error is transmitted in the sub-video data unit is 0.
或者,给所述子视频数据单元中传输错误的子图像分配的值是采用插值法确定的。给所述子视频数据单元中传输错误的子图像分配的值是根据传输正确的子图像确定的,且所述传输错误的子图像和所述传输正确的子图像来自同一个图像帧。Alternatively, the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined by interpolation. The value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined based on the transmission of the correct sub-image, and the sub-image of the transmission error and the sub-image of the transmission are from the same image frame.
处理器1702根据解码后的子视频数据单元重建所述视频数据时,具体用于:采用逆变换重建所述视频数据。When the processor 1702 reconstructs the video data according to the decoded sub video data unit, the processor 1702 is specifically configured to: reconstruct the video data by using an inverse transform.
其中,所述接收设备可以是遥控器、智能手机、平板电脑、地面控制站、膝上型电脑、手表、手环等及其组合。The receiving device may be a remote controller, a smart phone, a tablet computer, a ground control station, a laptop computer, a watch, a wristband, and the like, and combinations thereof.
本发明实施例提供的接收设备的具体原理和实现方式均与图11-15所示实施例类似,此处不再赘述。The specific principles and implementations of the receiving device provided by the embodiment of the present invention are similar to the embodiment shown in FIG. 11-15, and details are not described herein again.
本实施例通过接收设备对多个子视频数据单元分别进行解码,得到解码后的子视频数据单元,并根据解码后的子视频数据单元重建原始图像,具体的,根据解码后接收正确的子图像重建原始图像,可以在部分子图像接收错误的情况下,使得重建的图像最大程度的接近原始图像,提高了接收设备在重建图像过程中的容错能力,增强了系统的鲁棒性。In this embodiment, the receiving device separately decodes the plurality of sub-video data units to obtain the decoded sub-video data unit, and reconstructs the original image according to the decoded sub-video data unit, and specifically, receives the correct sub-image reconstruction according to the decoding. The original image can make the reconstructed image be close to the original image to the greatest extent when the partial sub-image is received incorrectly, which improves the fault tolerance of the receiving device in reconstructing the image and enhances the robustness of the system.
本发明实施例提供一种控制终端,该控制终端包括上述实施例所述的接收设备。例如,该控制终端可以是遥控器、智能手机、平板电脑、地面控制站、膝上型电脑、手表、手环等及其组合,同时还可以在地面控制无人飞行器。An embodiment of the present invention provides a control terminal, where the control terminal includes the receiving device described in the foregoing embodiment. For example, the control terminal can be a remote control, a smart phone, a tablet, a ground control station, a laptop, a watch, a wristband, etc., and combinations thereof, and can also control the unmanned aerial vehicle on the ground.
本发明实施例提供一种无人飞行器。图18为本发明实施例提供的无 人飞行器的结构图,如图18所示,无人飞行器1800包括:机身、动力系统和视频传输系统,所述动力系统包括如下至少一种:电机1801、螺旋桨1802和电子调速器1803,动力系统安装在所述机身,用于提供飞行动力;飞行控制器1804与所述动力系统通讯连接,用于控制所述无人飞1806行器飞行;其中,飞行控制器1804包括惯性测量单元及陀螺仪。所述惯性测量单元及所述陀螺仪用于检测所述无人飞行器的加速度、俯仰角、横滚角及偏航角等。Embodiments of the present invention provide an unmanned aerial vehicle. FIG. 18 is a schematic diagram of an embodiment of the present invention A structural diagram of a human aircraft, as shown in FIG. 18, the unmanned aerial vehicle 1800 includes: a fuselage, a power system, and a video transmission system, the power system including at least one of the following: a motor 1801, a propeller 1802, and an electronic governor 1803. A power system is mounted to the airframe for providing flight power; a flight controller 1804 is in communication with the power system for controlling the unmanned flight 1806, wherein the flight controller 1804 includes an inertial measurement unit And gyroscopes. The inertial measurement unit and the gyroscope are configured to detect an acceleration, a pitch angle, a roll angle, a yaw angle, and the like of the unmanned aerial vehicle.
视频传输系统包括一个或多个成像装置1805、以及配置在可移动物体上的一个或多个处理器1806,成像装置1805通过支撑设备1807连接到机身,且处理器1806和成像装置1805通讯连接,支撑设备1807具体可以是云台。视频传输系统的原理和实现方式均与上述实施例类似,此处不再赘述。The video transmission system includes one or more imaging devices 1805, and one or more processors 1806 disposed on the movable object. The imaging device 1805 is coupled to the body through the support device 1807, and the processor 1806 and the imaging device 1805 are communicatively coupled. The supporting device 1807 may specifically be a pan/tilt. The principle and implementation of the video transmission system are similar to the above embodiments, and are not described herein again.
在本发明所提供的几个实施例中,应该理解到,所揭露的装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided by the present invention, it should be understood that the disclosed apparatus and method may be implemented in other manners. For example, the device embodiments described above are merely illustrative. For example, the division of the unit is only a logical function division. In actual implementation, there may be another division manner, for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed. In addition, the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or in the form of hardware plus software functional units.
上述以软件功能单元的形式实现的集成的单元,可以存储在一个计算机可读取存储介质中。上述软件功能单元存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者 网络设备等)或处理器(processor)执行本发明各个实施例所述方法的部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The above-described integrated unit implemented in the form of a software functional unit can be stored in a computer readable storage medium. The above software functional unit is stored in a storage medium and includes a plurality of instructions for causing a computer device (which may be a personal computer, a server, or A network device or the like) or a processor performs part of the steps of the method described in various embodiments of the present invention. The foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like, which can store program codes. .
本领域技术人员可以清楚地了解到,为描述的方便和简洁,仅以上述各功能模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能模块完成,即将装置的内部结构划分成不同的功能模块,以完成以上描述的全部或者部分功能。上述描述的装置的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。A person skilled in the art can clearly understand that for the convenience and brevity of the description, only the division of each functional module described above is exemplified. In practical applications, the above function assignment can be completed by different functional modules as needed, that is, the device is installed. The internal structure is divided into different functional modules to perform all or part of the functions described above. For the specific working process of the device described above, refer to the corresponding process in the foregoing method embodiment, and details are not described herein again.
最后应说明的是:以上各实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述各实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分或者全部技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的范围。 Finally, it should be noted that the above embodiments are merely illustrative of the technical solutions of the present invention, and are not intended to be limiting; although the present invention has been described in detail with reference to the foregoing embodiments, those skilled in the art will understand that The technical solutions described in the foregoing embodiments may be modified, or some or all of the technical features may be equivalently replaced; and the modifications or substitutions do not deviate from the technical solutions of the embodiments of the present invention. range.

Claims (67)

  1. 一种视频传输方法,其特征在于,包括:A video transmission method, comprising:
    将视频数据分解为多个子视频数据单元,其中,每个子视频数据单元包括一个或多个子图像;Decomposing video data into a plurality of sub-video data units, wherein each sub-video data unit includes one or more sub-images;
    对所述多个子视频数据单元分别进行编码;以及Encoding the plurality of sub-video data units separately;
    基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。One or more encoded sub-video data units are selected and transmitted based on one or more characteristics of the channel, and one or more characteristics of the sub-video data unit.
  2. 根据权利要求1所述的方法,其特征在于,所述视频数据包括一个或多个图像帧;The method of claim 1 wherein said video data comprises one or more image frames;
    所述将视频数据分解为多个子视频数据单元,包括:Decomposing the video data into a plurality of sub-video data units, including:
    将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像,其中,所述子视频数据单元中的每一个包括所述图像帧中每一个分解得到的多个子图像中的至少一个子图像。Decomposing each of the one or more image frames in the video data into a plurality of sub-images, wherein each of the sub-video data units includes a plurality of sub-images obtained by each of the image frames At least one sub-image.
  3. 根据权利要求2所述的方法,其特征在于,每个子图像包括所述图像帧的一部分。The method of claim 2 wherein each sub-image comprises a portion of said image frame.
  4. 根据权利要求3所述的方法,其特征在于,所述每个子图像包括所述图像帧的一个或多个像素。The method of claim 3 wherein each of said sub-images comprises one or more pixels of said image frame.
  5. 根据权利要求3所述的方法,其特征在于,所述每个子图像包括所述图像帧的一个或多个转换系数。The method of claim 3 wherein each of said sub-images comprises one or more conversion coefficients of said image frame.
  6. 根据权利要求2-5任一项所述的方法,其特征在于,所述将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像,包括:The method according to any one of claims 2 to 5, wherein the decomposing each of the one or more image frames in the video data into a plurality of sub-images comprises:
    将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Decomposing each of the one or more image frames in the video data into a plurality of sub-images.
  7. 根据权利要求6所述的方法,其特征在于,所述将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像,包括:The method according to claim 6, wherein the decomposing each of the one or more image frames in the video data into a plurality of sub-images comprises:
    利用傅里叶相关变换或正交变换将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Each of the one or more image frames in the video data is decomposed into a plurality of sub-images using a Fourier correlation transform or an orthogonal transform.
  8. 根据权利要求7所述的方法,其特征在于,所述傅里叶相关变换或正交变换是从哈达玛变换、离散余弦变换、离散傅里叶变换、沃尔什-哈达玛变换、哈尔变换或斜变换中选择的。 The method according to claim 7, wherein said Fourier correlation transform or orthogonal transform is from Hadamard transform, discrete cosine transform, discrete Fourier transform, Walsh-Hadamard transform, Hal Selected in transform or skew transform.
  9. 根据权利要求7所述的方法,其特征在于,所述将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像,包括:The method according to claim 7, wherein the decomposing each of the one or more image frames in the video data into a plurality of sub-images comprises:
    利用空间下采样将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Each of the one or more image frames in the video data is decomposed into a plurality of sub-images using spatial downsampling.
  10. 根据权利要求1-9任一项所述的方法,其特征在于,所述信道的一个或多个特性至少包括带宽。The method of any of claims 1-9, wherein the one or more characteristics of the channel comprise at least a bandwidth.
  11. 根据权利要求1-9任一项所述的方法,其特征在于,所述信道的一个或多个特性包括如下至少一种:The method of any of claims 1-9, wherein the one or more characteristics of the channel comprise at least one of the following:
    噪声、干扰、信噪比、误比特率、衰落速率、带宽。Noise, interference, signal to noise ratio, bit error rate, fading rate, bandwidth.
  12. 根据权利要求1-11任一项所述的方法,其特征在于,所述子视频数据单元的一个或多个特性包括:The method of any of claims 1-11, wherein one or more characteristics of the sub-video data unit comprises:
    所述子视频数据单元编码后的码流数据大小,或者所述子视频数据单元的能量集中度。The size of the code stream data encoded by the sub video data unit or the energy concentration of the sub video data unit.
  13. 根据权利要求12所述的方法,其特征在于,所述选择一个或多个编码后的子视频数据单元,包括:The method according to claim 12, wherein said selecting one or more encoded sub-video data units comprises:
    选择一个或多个编码后的子视频数据单元,以使所述一个或多个编码后的子视频数据单元的总的码流数据大小与信道带宽匹配。One or more encoded sub-video data units are selected such that the total code stream data size of the one or more encoded sub-video data units matches the channel bandwidth.
  14. 根据权利要求12所述的方法,其特征在于,所述多个子视频数据单元依据所述能量集中度进行优先级排序。The method according to claim 12, wherein said plurality of sub-video data units are prioritized according to said energy concentration.
  15. 根据权利要求14所述的方法,其特征在于,所述选择一个或多个编码后的子视频数据单元,包括:The method according to claim 14, wherein said selecting one or more encoded sub-video data units comprises:
    根据所述子视频数据单元的优先级和信道带宽,选择所述一个或多个编码后的子视频数据单元。The one or more encoded sub-video data units are selected based on a priority of the sub-video data unit and a channel bandwidth.
  16. 根据权利要求1-15任一项所述的方法,其特征在于,所述对所述多个子视频数据单元分别进行编码,包括:The method according to any one of claims 1 to 15, wherein the encoding the plurality of sub-video data units separately comprises:
    所述多个子视频数据单元由多个单独的编码器进行编码。The plurality of sub-video data units are encoded by a plurality of separate encoders.
  17. 根据权利要求16所述的方法,其特征在于,所述多个子视频数据单元由多个单独的编码器进行编码,包括:The method of claim 16 wherein said plurality of sub-video data units are encoded by a plurality of separate encoders, comprising:
    采用所述多个单独的编码器对所述多个子视频数据单元并行编码。The plurality of sub-video data units are encoded in parallel using the plurality of separate encoders.
  18. 根据权利要求16所述的方法,其特征在于,所述多个子视频数 据单元由多个单独的编码器进行编码,包括:The method of claim 16 wherein said plurality of sub-videos The unit is encoded by a number of separate encoders, including:
    采用不同的视频编码规则对所述多个子视频数据单元进行编码。The plurality of sub-video data units are encoded using different video coding rules.
  19. 根据权利要求16所述的方法,其特征在于,所述多个子视频数据单元由多个单独的编码器进行编码,包括:The method of claim 16 wherein said plurality of sub-video data units are encoded by a plurality of separate encoders, comprising:
    采用相同的视频编码规则对所述多个子视频数据单元进行编码。The plurality of sub-video data units are encoded using the same video encoding rules.
  20. 根据权利要求1-15任一项所述的方法,其特征在于,所述对所述多个子视频数据单元分别进行编码,包括:The method according to any one of claims 1 to 15, wherein the encoding the plurality of sub-video data units separately comprises:
    所述多个子视频数据单元中的两个或多个由同一个编码器进行编码。Two or more of the plurality of sub-video data units are encoded by the same encoder.
  21. 根据权利要求1-15任一项所述的方法,其特征在于,所述对所述多个子视频数据单元分别进行编码,包括:The method according to any one of claims 1 to 15, wherein the encoding the plurality of sub-video data units separately comprises:
    基于运动补偿的视频压缩标准对所述多个子视频数据单元中的至少一个进行编码。At least one of the plurality of sub-video data units is encoded based on a motion compensated video compression standard.
  22. 根据权利要求1-15任一项所述的方法,其特征在于,所述对所述多个子视频数据单元分别进行编码,包括:The method according to any one of claims 1 to 15, wherein the encoding the plurality of sub-video data units separately comprises:
    依据不同的压缩率对所述多个子视频数据单元进行压缩。The plurality of sub-video data units are compressed according to different compression ratios.
  23. 根据权利要求22所述的方法,其特征在于,所述压缩率是根据所述子视频数据单元的一个或多个特性确定的。The method of claim 22 wherein said compression ratio is determined based on one or more characteristics of said sub-video data unit.
  24. 一种视频接收方法,其特征在于,包括:A video receiving method, comprising:
    接收多个编码后的子视频数据单元;Receiving a plurality of encoded sub-video data units;
    对所述多个编码后的子视频数据单元进行解码;Decoding the plurality of encoded sub-video data units;
    根据解码后的子视频数据单元重建所述视频数据,其中,所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。Reconstructing the video data according to the decoded sub-video data unit, wherein the video data includes one or more image frames, and the sub-video data unit includes a plurality of sub-decompositions obtained by decomposing each of the image frames At least one sub-image in the image.
  25. 根据权利要求24所述的方法,其特征在于,所述对所述多个编码后的子视频数据单元进行解码,包括:The method according to claim 24, wherein the decoding the plurality of encoded sub-video data units comprises:
    对所述多个编码后的子视频数据单元分别进行解码。Decoding the plurality of encoded sub-video data units separately.
  26. 根据权利要求24或25所述的方法,其特征在于,还包括:The method of claim 24 or 25, further comprising:
    检测所述子视频数据单元的一个或多个子图像的传输错误;Detecting a transmission error of one or more sub-images of the sub-video data unit;
    所述根据解码后的子视频数据单元重建所述视频数据,包括: Reconstructing the video data according to the decoded sub video data unit includes:
    根据接收正确的子图像重建所述视频数据。The video data is reconstructed based on receiving the correct sub-image.
  27. 根据权利要求26所述的方法,其特征在于,还包括:The method of claim 26, further comprising:
    给所述子视频数据单元中传输错误的子图像分配一个值。A value is assigned to the sub-image in which the error is transmitted in the sub-video data unit.
  28. 根据权利要求27所述的方法,其特征在于,给所述子视频数据单元中传输错误的子图像分配的值为0。The method according to claim 27, wherein a value of 0 is assigned to the sub-picture in which the error is transmitted in the sub-video data unit.
  29. 根据权利要求27所述的方法,其特征在于,给所述子视频数据单元中传输错误的子图像分配的值是采用插值法确定的。The method according to claim 27, wherein the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined by interpolation.
  30. 根据权利要求29所述的方法,其特征在于,给所述子视频数据单元中传输错误的子图像分配的值是根据传输正确的子图像确定的,且所述传输错误的子图像和所述传输正确的子图像来自同一个图像帧。The method according to claim 29, wherein the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined based on the transmission of the correct sub-image, and the sub-image of the transmission error and the The correct sub-image is transmitted from the same image frame.
  31. 根据权利要求28或29所述的方法,其特征在于,所述根据解码后的子视频数据单元重建所述视频数据,包括:The method according to claim 28 or 29, wherein the reconstructing the video data according to the decoded sub-video data unit comprises:
    采用逆变换重建所述视频数据。The video data is reconstructed using an inverse transform.
  32. 一种视频传输系统,其特征在于,包括:A video transmission system, comprising:
    一个或多个成像装置,被配置为采集视频数据;以及One or more imaging devices configured to acquire video data;
    所述可移动物体上的一个或多个处理器,单独或协同工作,所述处理器被配置为:One or more processors on the movable object, working alone or in concert, the processor being configured to:
    将视频数据分解为多个子视频数据单元,其中,每个子视频数据单元包括一个或多个子图像;Decomposing video data into a plurality of sub-video data units, wherein each sub-video data unit includes one or more sub-images;
    对所述多个子视频数据单元分别进行编码;以及Encoding the plurality of sub-video data units separately;
    基于信道的一个或多个特性,以及所述子视频数据单元的一个或多个特性,选择一个或多个编码后的子视频数据单元并传输。One or more encoded sub-video data units are selected and transmitted based on one or more characteristics of the channel, and one or more characteristics of the sub-video data unit.
  33. 根据权利要求32所述的系统,其特征在于,所述视频数据包括一个或多个图像帧;The system of claim 32 wherein said video data comprises one or more image frames;
    所述处理器将视频数据分解为多个子视频数据单元时,被配置为:When the processor decomposes the video data into a plurality of sub-video data units, it is configured to:
    将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像,其中,所述子视频数据单元中的每一个包括所述图像帧中每一个分解得到的多个子图像中的至少一个子图像。Decomposing each of the one or more image frames in the video data into a plurality of sub-images, wherein each of the sub-video data units includes a plurality of sub-images obtained by each of the image frames At least one sub-image.
  34. 根据权利要求33所述的系统,其特征在于,每个子图像包括所述图像帧的一部分。 The system of claim 33 wherein each sub-image comprises a portion of said image frame.
  35. 根据权利要求34所述的系统,其特征在于,所述每个子图像包括所述图像帧的一个或多个像素。The system of claim 34 wherein each of said sub-images comprises one or more pixels of said image frame.
  36. 根据权利要求34所述的系统,其特征在于,所述每个子图像包括所述图像帧的一个或多个转换系数。The system of claim 34 wherein each of said sub-images comprises one or more conversion coefficients of said image frame.
  37. 根据权利要求33-36任一项所述的系统,其特征在于,所述处理器将所述视频数据中的一个或多个图像帧中的每一个分解成多个子图像时,被配置为:A system according to any of claims 33-36, wherein the processor, when decomposing each of the one or more image frames in the video data into a plurality of sub-images, is configured to:
    将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Decomposing each of the one or more image frames in the video data into a plurality of sub-images.
  38. 根据权利要求37所述的系统,其特征在于,所述处理器将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像时,被配置为:The system of claim 37, wherein the processor, when decomposing each of the one or more image frames in the video data into a plurality of sub-images, is configured to:
    利用傅里叶相关变换或正交变换将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Each of the one or more image frames in the video data is decomposed into a plurality of sub-images using a Fourier correlation transform or an orthogonal transform.
  39. 根据权利要求38所述的系统,其特征在于,所述傅里叶相关变换或正交变换是从哈达玛变换、离散余弦变换、离散傅里叶变换、沃尔什-哈达玛变换、哈尔变换或斜变换中选择的。The system according to claim 38, wherein said Fourier correlation transform or orthogonal transform is from Hadamard transform, discrete cosine transform, discrete Fourier transform, Walsh-Hadamard transform, Hal Selected in transform or skew transform.
  40. 根据权利要求37所述的系统,其特征在于,所述处理器将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像时,被配置为:The system of claim 37, wherein the processor, when decomposing each of the one or more image frames in the video data into a plurality of sub-images, is configured to:
    利用空间下采样将所述视频数据中的一个或多个图像帧中的每一个空间分解成多个子图像。Each of the one or more image frames in the video data is decomposed into a plurality of sub-images using spatial downsampling.
  41. 根据权利要求32-40任一项所述的系统,其特征在于,所述信道的一个或多个特性至少包括带宽。A system according to any of claims 32-40, wherein one or more characteristics of the channel comprise at least a bandwidth.
  42. 根据权利要求32-40任一项所述的系统,其特征在于,所述信道的一个或多个特性包括如下至少一种:A system according to any one of claims 32 to 40, wherein one or more characteristics of the channel comprise at least one of the following:
    噪声、干扰、信噪比、误比特率、衰落速率、带宽。Noise, interference, signal to noise ratio, bit error rate, fading rate, bandwidth.
  43. 根据权利要求32-42任一项所述的系统,其特征在于,所述子视频数据单元的一个或多个特性包括:A system according to any of claims 32-42, wherein one or more characteristics of said sub-video data unit comprise:
    所述子视频数据单元编码后的码流数据大小,或者所述子视频数据 单元的能量集中度。The size of the code stream data encoded by the sub video data unit, or the sub video data The energy concentration of the unit.
  44. 根据权利要求43所述的系统,其特征在于,所述处理器选择一个或多个编码后的子视频数据单元时,被配置为:The system of claim 43 wherein said processor, when selecting one or more encoded sub-video data units, is configured to:
    选择一个或多个编码后的子视频数据单元,以使所述一个或多个编码后的子视频数据单元的总的码流数据大小与信道带宽匹配。One or more encoded sub-video data units are selected such that the total code stream data size of the one or more encoded sub-video data units matches the channel bandwidth.
  45. 根据权利要求35所述的系统,其特征在于,所述多个子视频数据单元依据所述能量集中度进行优先级排序。The system of claim 35 wherein said plurality of sub-video data units are prioritized in accordance with said energy concentration.
  46. 根据权利要求45所述的系统,其特征在于,所述处理器选择一个或多个编码后的子视频数据单元时,被配置为:The system of claim 45, wherein the processor, when selecting one or more encoded sub-video data units, is configured to:
    根据所述子视频数据单元的优先级和信道带宽,选择所述一个或多个编码后的子视频数据单元。The one or more encoded sub-video data units are selected based on a priority of the sub-video data unit and a channel bandwidth.
  47. 根据权利要求32-46任一项所述的系统,其特征在于,A system according to any of claims 32-46, wherein
    所述处理器还用于控制多个编码器对所述多个子视频数据单元进行编码。The processor is further configured to control a plurality of encoders to encode the plurality of sub-video data units.
  48. 根据权利要求47所述的系统,其特征在于,所述处理器具体用于控制所述多个编码器对所述多个子视频数据单元并行编码。The system according to claim 47, wherein said processor is specifically configured to control said plurality of encoders to encode said plurality of sub-video data units in parallel.
  49. 根据权利要求47所述的系统,其特征在于,所述处理器具体用于控制所述多个编码器分别采用不同的视频编码规则对所述多个子视频数据单元进行编码。The system according to claim 47, wherein the processor is specifically configured to control the plurality of encoders to respectively encode the plurality of sub-video data units by using different video encoding rules.
  50. 根据权利要求47所述的系统,其特征在于,所述处理器具体用于控制所述多个编码器采用相同的视频编码规则对所述多个子视频数据单元进行编码。The system according to claim 47, wherein said processor is specifically configured to control said plurality of encoders to encode said plurality of sub-video data units using the same video encoding rule.
  51. 根据权利要求32-46任一项所述的系统,其特征在于,所述处理器还用于控制编码器对所述多个子视频数据单元中的两个或多个进行编码。The system of any of claims 32-46, wherein the processor is further configured to control an encoder to encode two or more of the plurality of sub-video data units.
  52. 根据权利要求32-46任一项所述的系统,其特征在于,所述处理器还用于控制编码器基于运动补偿的视频压缩标准对所述多个子视频数据单元中的至少一个进行编码。A system according to any of claims 32-46, wherein the processor is further for controlling the encoder to encode at least one of the plurality of sub-video data units based on a motion compensated video compression standard.
  53. 根据权利要求32-46任一项所述的系统,其特征在于,所述处理器对所述多个子视频数据单元分别进行编码时,被配置为: The system according to any one of claims 32-46, wherein when the processor separately encodes the plurality of sub-video data units, the processor is configured to:
    依据不同的压缩率对所述多个子视频数据单元进行压缩。The plurality of sub-video data units are compressed according to different compression ratios.
  54. 根据权利要求53所述的系统,其特征在于,所述压缩率是根据所述子视频数据单元的一个或多个特性确定的。The system of claim 53 wherein said compression ratio is determined based on one or more characteristics of said sub-video data unit.
  55. 根据权利要求32-54任一项所述的系统,其特征在于,所述可移动物体为无人飞行器。A system according to any of claims 32-54, wherein the movable object is an unmanned aerial vehicle.
  56. 根据权利要求32-54任一项所述的系统,其特征在于,所述一个或多个成像装置通过载体与所述可移动物体连接。A system according to any of claims 32-54, wherein said one or more imaging devices are coupled to said movable object by a carrier.
  57. 根据权利要求56所述的系统,其特征在于,所述载体是一种多轴万向节。The system of claim 56 wherein said carrier is a multi-axis joint.
  58. 一种接收设备,其特征在于,包括:通讯接口、一个或多个处理器,单独或协同工作,所述通讯接口和所述处理器通讯连接;A receiving device, comprising: a communication interface, one or more processors, working alone or in cooperation, wherein the communication interface is in communication with the processor;
    所述通讯接口用于接收多个编码后的子视频数据单元;The communication interface is configured to receive a plurality of encoded sub-video data units;
    所述一个或多个处理器用于:控制解码器对所述多个编码后的子视频数据单元进行解码;根据解码后的子视频数据单元重建所述视频数据,其中,所述视频数据包括一个或多个图像帧,所述子视频数据单元包括对所述图像帧中每一个进行分解后得到的多个子图像中的至少一个子图像。The one or more processors are configured to: control a decoder to decode the plurality of encoded sub-video data units; reconstruct the video data according to the decoded sub-video data unit, wherein the video data includes one Or a plurality of image frames, the sub-video data unit including at least one of the plurality of sub-images obtained by decomposing each of the image frames.
  59. 根据权利要求58所述的接收设备,其特征在于,所述处理器控制解码器对所述多个编码后的子视频数据单元进行解码时,具体用于:The receiving device according to claim 58, wherein the processor controls the decoder to decode the plurality of encoded sub-video data units, specifically for:
    控制解码器对所述多个编码后的子视频数据单元分别进行解码。The control decoder separately decodes the plurality of encoded sub-video data units.
  60. 根据权利要求58或59所述的接收设备,其特征在于,所述处理器还用于:检测所述子视频数据单元的一个或多个子图像的传输错误;The receiving device according to claim 58 or 59, wherein the processor is further configured to: detect a transmission error of one or more sub-images of the sub-video data unit;
    所述处理器根据解码后的子视频数据单元重建所述视频数据时,具体用于:根据接收正确的子图像重建所述视频数据。When the processor reconstructs the video data according to the decoded sub video data unit, the processor is specifically configured to: reconstruct the video data according to receiving the correct sub image.
  61. 根据权利要求60所述的接收设备,其特征在于,所述处理器还用于:给所述子视频数据单元中传输错误的子图像分配一个值。The receiving device according to claim 60, wherein the processor is further configured to: assign a value to the sub-image in which the error is transmitted in the sub-video data unit.
  62. 根据权利要求61所述的接收设备,其特征在于,给所述子视频数据单元中传输错误的子图像分配的值为0。The receiving device according to claim 61, wherein a value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is 0.
  63. 根据权利要求61所述的接收设备,其特征在于,给所述子视频数据单元中传输错误的子图像分配的值是采用插值法确定的。 The receiving device according to claim 61, wherein the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined by interpolation.
  64. 根据权利要求63所述的接收设备,其特征在于,给所述子视频数据单元中传输错误的子图像分配的值是根据传输正确的子图像确定的,且所述传输错误的子图像和所述传输正确的子图像来自同一个图像帧。The receiving apparatus according to claim 63, wherein the value assigned to the sub-picture in which the error is transmitted in the sub-video data unit is determined based on the transmission of the correct sub-image, and the sub-image and the The correct sub-images are transmitted from the same image frame.
  65. 根据权利要求62或63所述的接收设备,其特征在于,所述处理器根据解码后的子视频数据单元重建所述视频数据时,具体用于:The receiving device according to claim 62 or 63, wherein when the processor reconstructs the video data according to the decoded sub-video data unit, the processor is specifically configured to:
    采用逆变换重建所述视频数据。The video data is reconstructed using an inverse transform.
  66. 一种控制终端,其特征在于,包括:如权利要求58-65任一项所述的接收设备。A control terminal, comprising: the receiving device according to any one of claims 58-65.
  67. 一种无人飞行器,其特征在于,包括:An unmanned aerial vehicle, comprising:
    机身;body;
    动力系统,安装在所述机身,用于提供飞行动力;a power system mounted to the fuselage for providing flight power;
    以及如权利要求32-57任一项所述的视频传输系统。 And a video transmission system according to any of claims 32-57.
PCT/CN2017/078728 2017-03-30 2017-03-30 Video transmitting and receiving method, system, and device, and unmanned aerial vehicle WO2018176303A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201780005017.8A CN108496369A (en) 2017-03-30 2017-03-30 Transmission of video, method of reseptance, system, equipment and unmanned vehicle
PCT/CN2017/078728 WO2018176303A1 (en) 2017-03-30 2017-03-30 Video transmitting and receiving method, system, and device, and unmanned aerial vehicle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/078728 WO2018176303A1 (en) 2017-03-30 2017-03-30 Video transmitting and receiving method, system, and device, and unmanned aerial vehicle

Publications (1)

Publication Number Publication Date
WO2018176303A1 true WO2018176303A1 (en) 2018-10-04

Family

ID=63344689

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/078728 WO2018176303A1 (en) 2017-03-30 2017-03-30 Video transmitting and receiving method, system, and device, and unmanned aerial vehicle

Country Status (2)

Country Link
CN (1) CN108496369A (en)
WO (1) WO2018176303A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021179304A1 (en) * 2020-03-13 2021-09-16 深圳市大疆创新科技有限公司 Method for automatically adjusting video live broadcast bit rate, video transmission device, and server

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020258241A1 (en) * 2019-06-28 2020-12-30 深圳市大疆创新科技有限公司 Image processing method and apparatus for movable platform, movable platform and medium
WO2021056575A1 (en) * 2019-09-29 2021-04-01 华为技术有限公司 Low-delay joint source-channel coding method, and related device
CN111565291B (en) * 2020-07-14 2020-11-10 深圳市慧明捷科技有限公司 Multi-base-station image transmission method for unmanned aerial vehicle
CN114040226B (en) * 2022-01-10 2022-03-11 北京小鸟科技股份有限公司 Data transmission method, system and equipment for low-bandwidth high-resolution video transmission

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102907096A (en) * 2010-05-10 2013-01-30 三星电子株式会社 Method and device for sending and receiving layered encoded video
WO2013128010A2 (en) * 2012-03-02 2013-09-06 Canon Kabushiki Kaisha Method and devices for encoding a sequence of images into a scalable video bit-stream, and decoding a corresponding scalable video bit-stream
CN103945165A (en) * 2014-05-05 2014-07-23 北京领通科技有限公司 Method and device for processing remote video transmission of terminal device
CN104349142A (en) * 2014-11-03 2015-02-11 南京航空航天大学 Layered representation-based unmanned plane video adaptive transmission method

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7480252B2 (en) * 2002-10-04 2009-01-20 Koniklijke Philips Electronics N.V. Method and system for improving transmission efficiency using multiple-description layered encoding
JP4187746B2 (en) * 2005-01-26 2008-11-26 三洋電機株式会社 Video data transmission device
US8189621B2 (en) * 2006-05-12 2012-05-29 Microsoft Corporation Stack signaling to application with lack of requested bandwidth
WO2008041954A1 (en) * 2006-10-06 2008-04-10 Agency For Science, Technology And Research Method for encoding, method for decoding, encoder, decoder and computer program products
CN101848499B (en) * 2009-03-25 2013-05-08 上海贝尔股份有限公司 Method for improving classified service transmission in wireless system, network element and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102907096A (en) * 2010-05-10 2013-01-30 三星电子株式会社 Method and device for sending and receiving layered encoded video
WO2013128010A2 (en) * 2012-03-02 2013-09-06 Canon Kabushiki Kaisha Method and devices for encoding a sequence of images into a scalable video bit-stream, and decoding a corresponding scalable video bit-stream
CN103945165A (en) * 2014-05-05 2014-07-23 北京领通科技有限公司 Method and device for processing remote video transmission of terminal device
CN104349142A (en) * 2014-11-03 2015-02-11 南京航空航天大学 Layered representation-based unmanned plane video adaptive transmission method

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021179304A1 (en) * 2020-03-13 2021-09-16 深圳市大疆创新科技有限公司 Method for automatically adjusting video live broadcast bit rate, video transmission device, and server

Also Published As

Publication number Publication date
CN108496369A (en) 2018-09-04

Similar Documents

Publication Publication Date Title
US12003770B2 (en) Face discontinuity filtering for 360-degree video coding
US10931851B2 (en) Image stitching with electronic rolling shutter correction
US11388438B2 (en) 360-degree video coding using geometry projection
WO2018176303A1 (en) Video transmitting and receiving method, system, and device, and unmanned aerial vehicle
US11064110B2 (en) Warp processing for image capture
US20230412839A1 (en) Geometry Conversion for 360-degree Video Coding
US11057646B2 (en) Image processor and image processing method
US11212496B2 (en) Geometry conversion and frame packing associated with 360-degree videos
EP3643063A1 (en) Adaptive quantization for 360-degree video coding
CA3059870A1 (en) 360-degree video coding using face continuities
WO2018045108A1 (en) Method and system for signaling of 360-degree video information
WO2018176340A1 (en) Video transmission method, reception method, system, and unmanned aerial vehicle
US11908111B2 (en) Image processing including noise reduction
WO2019143551A1 (en) Adaptive frame packing for 360-degree video coding
WO2017205597A1 (en) Image signal processing-based encoding hints for motion estimation
CN114339259A (en) Video compression method and device
WO2018017599A1 (en) Quality evaluation system and method for 360-degree video
WO2018176341A1 (en) Video transmission method, reception method, system, and unmanned aerial vehicle
KR20250053051A (en) Deep distribution-aware point feature extractor for AI-based point cloud compression
CN116033155A (en) Binocular image compression method, device and readable storage medium
JP2019083405A (en) Decoding device, transmission device, decoding method, control method for transmission device, and program
EP3091742A1 (en) Device and method for encoding a first image of a scene using a second image having a lower resolution and captured at the same instant

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17904344

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17904344

Country of ref document: EP

Kind code of ref document: A1

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载