+

CN108881886A - A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end - Google Patents

A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end Download PDF

Info

Publication number
CN108881886A
CN108881886A CN201810550078.9A CN201810550078A CN108881886A CN 108881886 A CN108881886 A CN 108881886A CN 201810550078 A CN201810550078 A CN 201810550078A CN 108881886 A CN108881886 A CN 108881886A
Authority
CN
China
Prior art keywords
camera
data
perspective images
display end
target object
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810550078.9A
Other languages
Chinese (zh)
Inventor
陈政权
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Medium Rope Technology Co Ltd
Original Assignee
Chengdu Medium Rope Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Medium Rope Technology Co Ltd filed Critical Chengdu Medium Rope Technology Co Ltd
Priority to CN201810550078.9A priority Critical patent/CN108881886A/en
Publication of CN108881886A publication Critical patent/CN108881886A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/131Protocols for games, networked simulations or virtual reality

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The present invention relates to technical field of computer vision, disclose a kind of method realized based on camera Matrix Technology and carry out the lossless interactive application of big data in display end.It creates through the invention, it can provide a kind of method that losslessization interactive application is carried out to magnanimity byte data to foreground display end from background data base, the 3D hologram image data of target object is placed in background data base, it is then based on specified viewpoint, image data losslessization at corresponding visual angle is sent to foreground display end and carries out output displaying, due to not needing to cut down in the whole process to collected initial data, the initial data of characterization physical details (comprising concave-convex surface and texture etc.) information can be retained in display data, ensure to export the authenticity or accuracy of vision.

Description

One kind realizing that carrying out the lossless interaction of big data in display end answers based on camera Matrix Technology Method
Technical field
The invention belongs to technical field of computer vision, and in particular to one kind is realized based on camera Matrix Technology in display end The method for carrying out the lossless interactive application of big data.
Background technique
Computer vision technique (i.e. Computer Vision, abbreviation CV) is current most popular one of research, refers to use Video camera and computer replace human eye the machine vision such as to be identified, tracked and measured to target, and further do graphics process, make Computer is treated as the image for being more suitable for eye-observation or sending instrument detection to.It is the research of a multi-crossed disciplines, is contained Lid computer science (graphics, algorithm, theoretical research etc.), mathematics (information retrieval, machine learning), engineering (robot, NLP Deng), biology (neuroscience) and psychology (cognitive science) etc..
In the current 3D Model Results way of output, most company also generallys use WebGL (i.e. Web Graphics Library is a kind of 3D drawing agreement, and this drawing technique standard allows 2.0 knot of JavaScript and OpenGL ES It is combined, a JavaScript by increasing OpenGL ES 2.0 is bound, and can provide hardware for HTML5Canvas 3D accelerates rendering, such web developer can come in browser to show more glibly by system video card 3D scene and Model, moreover it is possible to create complicated navigation and data and visualize) technology, although eliminating the fiber crops of the exploitation dedicated Rendering Plug of webpage It is tired, it can be used for the Website page that creation has complexity 3D structure, it might even be possible to for designing 3D web game etc., but meeting Detail section loss of data and the problem serious to model bevel are caused, simultaneously because recessed with material diffusing reflection, bloom, normal The shade that the modes such as convex textures and opacity show replaces physical details, has certain vision duplicity, causes to make Product precision is only game level, can not be applied to holographic interaction scenarios, especially cannot be used for the digitlization to historical relic or historic site etc. Protectiveness engineering field.
Summary of the invention
In order to solve the above problems existing in the present technology, it is an object of that present invention to provide one kind to be based on camera Matrix Technology Realize the method for carrying out the lossless interactive application of big data in display end.
The technical scheme adopted by the invention is as follows:
A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end, including is as follows Step:
S101. the camera matrix perspective data of target object is stored in advance in background data base, wherein the camera square Battle array perspective data includes that several camera spherical coordinates of spherical surface matrix arrangement are carried out using target object center as spherical coordinate system origin With the first perspective images data of each camera spherical coordinate of correspondence, and/or comprising using target object center as cylindrical-coordinate system Origin carries out several camera cylindrical coordinates of cylinder matrix arrangement and the second perspective images number of corresponding each camera cylindrical coordinates According to the sight of the first perspective images data is directed toward spherical coordinate system origin, second perspective images by camera spherical coordinate The sight of data is directed toward cylindrical-coordinate system central axis by camera cylindrical coordinates;
S102. it after receiving specified camera three-dimensional coordinate, is searched and the specified camera three-dimensional coordinate from background data base Corresponding perspective images data, wherein the specified camera three-dimensional coordinate is camera spherical coordinate or camera cylindrical coordinates;
S103. by the perspective images data transmission found to foreground display end;
S104. output displaying is carried out to the perspective images data received by foreground display end.
Optimization, the first perspective images data are the virtual reality figure that S201~204 is obtained in accordance with the following steps As data:
S201. using target object center as spherical coordinate system origin, if matrix arrangement arranges cadre's camera lens direction on spherical surface The ultra high-definition digital camera of spherical coordinate system origin, and sat the spherical coordinate of each portion's ultra high-definition digital camera as camera spherical surface Mark;
S202. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains and each portion's ultra high-definition digital camera Corresponding first multi-view image data;
S203. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the first physics letter of target object Cease data, wherein the first physical message data include point cloud information, material information, color information and/or shadow parameter letter Breath;
S204. the first multi-view image data matching is carried out with the first physical message data to merge, obtain with The consistent first perspective images data of first multi-view image clear data degree;
Alternatively, the first perspective images data are the virtual reality image that S205~207 is obtained in accordance with the following steps Data:
S205. the threedimensional model of target object is generated using 3 d modeling software;
S206. using the threedimensional model center of target object as spherical coordinate system origin, matrix arrangement setting is several on spherical surface A sight is directed toward the camera view of spherical coordinate system origin, and using the spherical coordinate of each camera view as camera spherical coordinate;
S207. corresponding first visual angle is obtained in each camera view using the virtual camera module in 3-D graphic software Image data.
Optimization, the second perspective images data are the virtual reality figure that S301~304 is obtained in accordance with the following steps As data:
S301. using target object center as cylindrical-coordinate system origin, if matrix arrangement arranges cadre's camera lens court on cylinder To the ultra high-definition digital camera of cylindrical-coordinate system central axis, and using the cylindrical coordinates of each portion's ultra high-definition digital camera as phase column Areal coordinate;
S302. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains and each portion's ultra high-definition digital camera Corresponding second multi-view image data;
S303. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the second physics letter of target object Cease data, wherein the second physical message data include point cloud information, material information, color information and/or shadow parameter letter Breath;
S304. the second multi-view image data matching is carried out with the second physical message data to merge, obtain with The consistent second perspective images data of second multi-view image clear data degree;
Alternatively, the second perspective images data are the virtual reality image that S305~307 is obtained in accordance with the following steps Data:
S305. the threedimensional model of target object is generated using 3 d modeling software;
S306. using the threedimensional model center of target object as cylindrical-coordinate system origin, if matrix arrangement is arranged on cylinder Dry sight is directed toward the camera view of cylindrical-coordinate system central axis, and using the cylindrical coordinates of each camera view as camera cylinder Coordinate;
S307. corresponding second visual angle is obtained in each camera view using the virtual camera module in 3-D graphic software Image data.
It advanced optimizes, when perspective images data are for virtual reality image data and in Image Acquisition, record is adopted Collect the time axial coordinate of multi-view image data, then includes together with corresponding obtained perspective images data by the time axial coordinate In camera matrix perspective data.
It advanced optimizes, when perspective images data are virtual reality image data and after acquisition, in conjunction with three-dimensional The physical attribute parameter of model, to perspective images data carry out computer graphical rendering processing, later period outdoor scene incorporate processing and/or Adjust colors countenance, wherein the physical attribute parameter includes material attribute parameter, environment attribute parameter and/or light source attributes Parameter.
Specifically, the model Phase One IQ3100MP Trichromatic of the ultra high-definition digital camera, square Battle array number of permutations is no less than 180.
Specifically, the three-dimensional scanning device is the spatial digitizer based on laser, radar and/or infrared ray.
Optimization, coordinate parameters control shaft or page operation of the specified camera three-dimensional coordinate from human-computer interaction interface Instruction, wherein the page operation instruction includes the instruction of visual angle moving operation and/or three-dimensional coordinate input instruction.
Optimization, further include following steps before the step S103:The perspective images data found are converted to Picture format file or WebGL formatted file.
Specifically, the foreground display end is LED display, LCD display, holographic film, pseudo- line holographic projections equipment, 3D electricity Fan, AR equipment, VR equipment, projector touch interactive device or Three-dimensional holographic image documentation equipment.
Beneficial effects of the present invention are:
(1) the invention provides a kind of lossless to the progress of magnanimity byte data from background data base to foreground display end The 3D hologram image data of target object, i.e., be placed in background data base, be then based on specified by the method for changing interactive application Image data losslessization at corresponding visual angle is sent to foreground display end and carries out output displaying, due in whole process by viewpoint In do not need to cut down collected initial data, can display data in retain characterization physical details (comprising surface it is recessed Convex and texture etc.) information initial data, it is ensured that export the authenticity or accuracy of vision;
(2) during acquiring virtual reality image data, it can not only accomplish the comprehensive data acquisition to the original. It can also accomplish to correspond to the accurate match of the physical messages such as the room and time of each point in Point Cloud Processing link, into And it can avoid carrying out the work such as image compensation in the later period;
(3) can make acquisition and treated three-dimensional holographic data (including room and time data etc.) can stable delivery, Ensure in the following various data processing circumstances, is not in the calculating such as data degradation, disorder, or even collapse error problem, i.e., After one-pass molding, the following various holographic applications environment can be coped with, realize the working effect of " once and for all ", be the various of future Holographic applications save a large amount of model development cost;
(4) it can be achieved visually to carry out material object the purpose that complete and lossless digitlization reduction reproduces, with fidelity Researching value, the digitlization particularly suitable for historical relic, historic site, spot, important security place and traditional architecture etc. are protected The application fields such as shield property engineering field and electronics/3D hologram interaction books field;
(5) when the hologram interactive application in transfer complex environment, without setting up processing large capacity near scene Low bi-dimensional end (the benefit matched can be used in the ancillary hardware server apparatus of data, such as user under current 4G communication environment In smart phone) hologram of ZB (i.e. 10,000,000,000,000 hundred million bytes) rank is browsed, it thus can look forward to logical in following 5G Interrogate under environment, can also long-range real-time delivery naked eye grade hologram interactive application data;
(6) output element is being applied, the simple interactivity hologram that both may be output as Entertainment or browse level is answered With, can be also the corresponding output format of complex interactions hologram application and development such as scientific research or engineering, promote convenient for practical and Using.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this Some embodiments of invention for those of ordinary skill in the art without creative efforts, can be with It obtains other drawings based on these drawings.
Fig. 1 is that provided by the invention realized based on camera Matrix Technology carries out the lossless interactive application of big data in display end Method flow schematic diagram.
Fig. 2 is the camera matrix schematic diagram provided by the invention on spherical surface.
Fig. 3 is transversal spherical surface camera matrix schematic diagram provided by the invention.
Fig. 4 is transversal cylinder camera matrix schematic diagram provided by the invention.
In above-mentioned attached drawing:1- target object;2- ultra high-definition digital camera.
Specific embodiment
With reference to the accompanying drawing and specific embodiment the present invention is further elaborated.It should be noted that for this The explanation of a little way of example is used to help understand the present invention, but and does not constitute a limitation of the invention.
The terms "and/or", only a kind of incidence relation for describing affiliated partner, indicates that there may be three kinds of passes System, for example, A and/or B, can indicate:Individualism A, individualism B exist simultaneously tri- kinds of situations of A and B, the terms "/and " it is to describe another affiliated partner relationship, indicate may exist two kinds of relationships, for example, A/ and B, can indicate:Individually deposit In A, two kinds of situations of individualism A and B, in addition, character "/" herein, typicallying represent forward-backward correlation object is a kind of "or" pass System.
Embodiment one
As illustrated in fig. 1 and 2, described realized based on camera Matrix Technology provided in this embodiment carries out big data in display end The method of lossless interactive application, includes the following steps.
S101. the camera matrix perspective data of target object is stored in advance in background data base, wherein the camera square Battle array perspective data includes that several camera spherical coordinates of spherical surface matrix arrangement are carried out using target object center as spherical coordinate system origin With the first perspective images data of each camera spherical coordinate of correspondence, and/or comprising using target object center as cylindrical-coordinate system Origin carries out several camera cylindrical coordinates of cylinder matrix arrangement and the second perspective images number of corresponding each camera cylindrical coordinates According to the sight of the first perspective images data is directed toward spherical coordinate system origin, second perspective images by camera spherical coordinate The sight of data is directed toward cylindrical-coordinate system central axis by camera cylindrical coordinates.
The crosspoint of spherical surface camera matrix as shown in Figure 2, each spherical surface warp and spherical surface weft is camera spherical surface Coordinate points.The crosspoint of cylinder camera matrix as shown in Figure 3, each cylinder loop wire and cylinder vertical line is that camera cylinder is sat Punctuate.The target object can be real object, such as the historical relic present in real world or engineering goods, or empty Quasi- body, such as the Virtual Building or the art work that are generated in a computer by three dimensional design.Thus first perspective images Data or the second perspective images data can be (to use number based on virtual reality image data obtained by real object acquisition The image data of word expression real object information), or based on virtual reality image obtained from virtual body three-dimensional models Data (i.e. with the image data of the quasi- real-world object information of number).
In the step S101, the first perspective images data can obtain for S201~204 in accordance with the following steps Virtual reality image data:S201. using target object center as spherical coordinate system origin, if matrix arrangement is arranged on spherical surface Cadre's camera lens towards spherical coordinate system origin ultra high-definition digital camera, and using the spherical coordinate of each portion's ultra high-definition digital camera as Camera spherical coordinate;S202. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains digital with each portion's ultra high-definition The corresponding first multi-view image data of camera;
S203. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the first physics letter of target object Cease data, wherein the first physical message data include point cloud information, material information, color information and/or shadow parameter letter Breath;S204. the first multi-view image data are carried out matching with the first physical message data to merge, is obtained and described The one consistent first perspective images data of multi-view image clear data degree.
Similarly, in the step S101, the second perspective images data may be in accordance with the following steps S301~ The 304 virtual reality image datas obtained:S301. using target object center as cylindrical-coordinate system origin, the matrix on cylinder If arranged cadre camera lens is towards the ultra high-definition digital camera of cylindrical-coordinate system central axis, and by each portion's ultra high-definition digital camera Cylindrical coordinates as camera cylindrical coordinates;S302. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtain with The corresponding second multi-view image data of each portion's ultra high-definition digital camera;S303. target object is carried out using three-dimensional scanning device 3-D scanning obtains the second physical message data of target object, wherein the second physical message data include point cloud letter Breath, material information, color information and/or shadow parameter information;S304. by the second multi-view image data and second object Reason information data carries out matching fusion, obtains and the consistent second perspective images number of the second multi-view image clear data degree According to.
By as described in abovementioned steps S201~S204 or step S301~S304 based on camera and with 3-D scanning Image acquisition scheme supplemented by equipment, it can be ensured that material picture contents OK acquisition (refer to realize maximal accuracy, it is most accurate and Most comprehensive acquisition mode), the acquisition data guaranteed will be long-term useful for future.And existing most of In type of industry, the image acquisition scheme based on laser acquisition data and supplemented by camera is generallyd use, no material is first used The laser of information, texture information and colouring information etc. acquires the physical message on surface in kind, then uses the collected material of camera Information is attached on the collected physical data of laser, this process will lead to collected data precision fusion and mismatch, must It so will cause certain information errors, such error message is worthless absolutely for historical relic and paleovolcanic structure etc..Such as Method such as/FARO Singapore, Tian Bao/Trimble, Bentley, Agisoft, Topcon, Polyworks, Geomagic, The companies such as Creaform, which have, is absorbed in this field, possesses to realize the relevant technologies such as outdoor scene modeling and 3-D scanning, although he Object scanning accuracy can achieve army grade, but do not accomplish also in texture and material precision ultimate attainment, do not do true 4D timing (space is recorded with multiframe model in real time) in positive meaning, including the information such as shadow, material texture, not because of object With and acquisition mode is different, such as height is saturating and bloom.In addition, employed in the step S204 or the step S304 It is the prior art with fusion method, is referred to《Mapping notification》Disclosed in article《Three-dimensional laser point cloud and full-view image Match fusion method》(publication date:Routinely displacement design 2016-06-30) is carried out to realize.
It advanced optimizes, when carrying out Image Acquisition in the step S202 or the step S303, record acquisition view The time axial coordinate of angle image data (the i.e. described first multi-view image data or the second multi-view image data), then should Time axial coordinate and corresponding obtained perspective images data (the i.e. described first perspective images data or the second perspective images number According to) be included in camera matrix perspective data together.So during acquiring virtual reality image data, can not only it do To the comprehensive data acquisition to the original, it can also accomplish the room and time etc. to each point in Point Cloud Processing link The accurate match of physical message is corresponding, and then can avoid carrying out the work such as image compensation in the later period, and for the number of other companies According to acquisition method, due to generally existing space and the deviation of time data, the later period needs to carry out textures compensation etc. and makes up work Make.
It advanced optimizes, in order to further ensure that the clarity and losslessization of Image Acquisition, the ultra high-definition number phase The model of machine is preferably Phase One IQ3100MP Trichromatic, and matrix arrangement number is no less than 180.Due to Built-in 100MP cmos sensor is integrated with newest Bayer filter dye technology, and sufficiently benefit in aforementioned model digital camera With the achievement that works closely of Phase One and Sony, the color of finding of naked eye can be accurately captured.This brand new technical is Phase One institute is exclusive, can promote the expressive force of digital color, unprecedented realistic colour is presented, with 1.01 hundred million pixels Creativity is played unrestrainedly, so that the color that collected color more infinite approach human eye is seen.It can so borrow Sensor most outstanding so far is helped, the Nature masterpiece of spectators' short distance appreciation riot of color and the ancient times older generations is allowed to stay Rarity.In addition, the three-dimensional scanning device can be, but not limited to as the spatial digitizer based on laser, radar and/or infrared ray.
In the step S101, the first perspective images data can also obtain for S205~207 in accordance with the following steps The virtual reality image data taken:S205. the threedimensional model of target object is generated using 3 d modeling software;S206. with mesh The threedimensional model center for marking object is spherical coordinate system origin, several sights direction spherical coordinate system is arranged in matrix arrangement on spherical surface The camera view of origin, and using the spherical coordinate of each camera view as camera spherical coordinate;S207. soft using 3-D graphic Virtual camera module in part obtains corresponding first perspective images data in each camera view.
Similarly, in the step S101, the second perspective images data can also for S305 in accordance with the following steps~ The 307 virtual reality image datas obtained:S305. the threedimensional model of target object is generated using 3 d modeling software;S306. Using the threedimensional model center of target object as cylindrical-coordinate system origin, several sights direction circle is arranged in matrix arrangement on cylinder The camera view of cylindrical coordinate central axis, and using the cylindrical coordinates of each camera view as camera cylindrical coordinates;S307. it applies Virtual camera module in 3-D graphic software obtains corresponding second perspective images data in each camera view.
In the step S205 or S305, the generating mode of the threedimensional model of the target object can be, but not limited to as Under:Target object and each dimension two dimensional image is first drawn, all two dimensional images of target object are then based on, using three Tie up the threedimensional model that modeling function module generates target object.In addition, the 3 d modeling software and the 3-D graphic software Can be, but not limited to for the three-dimensional applications such as CAD, SolidWorks, SketchUp, 3Ds Max, Rhino, Maya and ZBrush it is soft Part.
It advanced optimizes, it is right herein in connection with the physical attribute parameter of threedimensional model after the step S207 or S307 Perspective images data (the i.e. described first perspective images data or the second perspective images data) carry out computer graphical rendering Processing, later period outdoor scene incorporate processing and/or adjustment colors countenance, wherein the physical attribute parameter can be, but not limited to include Material attribute parameter, environment attribute parameter and/or light source attributes parameter etc..Aforementioned computer figure rendering processing, later period outdoor scene Incorporating processing and adjustment colors countenance is existing processing technique, so can carry out scientifically complex die to final image effect It is quasi-, intend the physical attribute of true real world completely, makes uninteresting data materialization.Such as in real world, for copper object, Its surface can be carried out scientific simulation in parameters such as the reflections and refraction at each visual angle;Or in real world, very much Object has different degrees of " Fresnel effect ":When so that sight is perpendicular to surface, reflection is weaker, and works as the non-perpendicular table of sight When face, angle is smaller, and reflection is more obvious (for seeing to a ball, the reflection at ball center is weaker, stronger by proximal edge).
S102. it after receiving specified camera three-dimensional coordinate, is searched and the specified camera three-dimensional coordinate from background data base Corresponding perspective images data, wherein the specified camera three-dimensional coordinate is camera spherical coordinate or camera cylindrical coordinates.
In the step S102, the specified camera three-dimensional coordinate can be, but not limited to the seat from human-computer interaction interface Mark state modulator axis or page operation instruction etc., wherein the page operation instruction can be, but not limited to comprising the mobile behaviour in visual angle Make instruction and/or three-dimensional coordinate input instruction etc..It can not only so see from each visual angle and simulate the true of real world completely Effect, moreover it is possible to adjust a certain item parameter and be reacted to generate a series of linkage to result.In addition, if also receiving specified time axis Coordinate, then the perspective images data found must also be corresponding with specified time axial coordinate, to realize the big number of time dimension According to interaction, wherein the specified time axial coordinate equally can be, but not limited to the control of the coordinate parameters from human-computer interaction interface Axis or page operation instruction etc..
S103. by the perspective images data transmission found to foreground display end.
It further include following steps before the step S103:The perspective images data found are converted into picture format File or WebGL formatted file etc..In the step S103, transmitting path based on RJ45 cable or RS232 either passed The cable network of the transmission such as defeated line, can also be based on WiFi, 3G, 4G and the mobile radio network of even future 5G technology.This Outside, the foreground display end can be, but not limited to set for existing LED display, LCD display, holographic film, pseudo- line holographic projections Standby, 3D electric fan, AR equipment, VR equipment, projector touch interactive device or Three-dimensional holographic image documentation equipment etc..
S104. output displaying is carried out to the perspective images data received by foreground display end.
In the step 104, optimization, the output exhibition method of the perspective images data can be specific as follows:According to The customized visual field sizes range of user selects the image data of corresponding regional area to carry out defeated from the perspective images data It shows out.It, can be to ZB grades of (i.e. 10,000,000,000,000 hundred million byte levels from there through the lossless information transfer scheme of abovementioned steps S101~S104 Data volume not) is browsed low with display end, without data degradation because the data seen of orientation in office all with The data seen on background data base are same data, are not only the same formats, whole process is not to original number According to any information reduction of progress.I.e. in other words, the information content of 100TB or so can be browsed on the picture of 145KB.
To sum up, it is realized based on camera Matrix Technology in the display end progress lossless friendship of big data using provided by the present embodiment The method mutually applied, has the following technical effect that:
(1) it present embodiments provides one kind and losslessization is carried out to magnanimity byte data from background data base to foreground display end The 3D hologram image data of target object is placed in background data base, is then based on specified view by the method for interactive application Image data losslessization at corresponding visual angle is sent to foreground display end and carries out output displaying, due in the whole process by point It does not need to cut down collected initial data, it (includes concave-convex surface that characterization physical details can be retained in display data With texture etc.) initial data of information, it is ensured that export the authenticity or accuracy of vision;
(2) during acquiring virtual reality image data, it can not only accomplish the comprehensive data acquisition to the original. It can also accomplish to correspond to the accurate match of the physical messages such as the room and time of each point in Point Cloud Processing link, into And it can avoid carrying out the work such as image compensation in the later period;
(3) can make acquisition and treated three-dimensional holographic data (including room and time data etc.) can stable delivery, Ensure in the following various data processing circumstances, is not in the calculating such as data degradation, disorder, or even collapse error problem, i.e., After one-pass molding, the following various holographic applications environment can be coped with, realize the working effect of " once and for all ", be the various of future Holographic applications save a large amount of model development cost;
(4) it can be achieved visually to carry out material object the purpose that complete and lossless digitlization reduction reproduces, with fidelity Researching value, the digitlization particularly suitable for historical relic, historic site, spot, important security place and traditional architecture etc. are protected The application fields such as shield property engineering field and electronics/3D hologram interaction books field;
(5) when the hologram interactive application in transfer complex environment, without setting up processing large capacity near scene Low bi-dimensional end (the benefit matched can be used in the ancillary hardware server apparatus of data, such as user under current 4G communication environment In smart phone) hologram of ZB (i.e. 10,000,000,000,000 hundred million bytes) rank is browsed, it thus can look forward to logical in following 5G Interrogate under environment, can also long-range real-time delivery naked eye grade hologram interactive application data;
(6) output element is being applied, the simple interactivity hologram that both may be output as Entertainment or browse level is answered With, can be also the corresponding output format of complex interactions hologram application and development such as scientific research or engineering, promote convenient for practical and Using.
The present invention is not limited to above-mentioned optional embodiment, anyone can show that other are each under the inspiration of the present invention The product of kind form.Above-mentioned specific embodiment should not be understood the limitation of pairs of protection scope of the present invention, protection of the invention Range should be subject to be defined in claims, and specification can be used for interpreting the claims.

Claims (10)

1. a kind of realize that, in the method that display end carries out the lossless interactive application of big data, feature exists based on camera Matrix Technology In including the following steps:
S101. the camera matrix perspective data of target object is stored in advance in background data base, wherein the camera matrix view It is spherical coordinate system origin progress several camera spherical coordinates of spherical surface matrix arrangement and right that angular data, which includes using target object center, The first perspective images data of each camera spherical coordinate are answered, and/or comprising using target object center as cylindrical-coordinate system origin Carry out several camera cylindrical coordinates of cylinder matrix arrangement and the second perspective images data of corresponding each camera cylindrical coordinates, institute The sight for stating the first perspective images data is directed toward spherical coordinate system origin by camera spherical coordinate, the second perspective images data Sight is directed toward cylindrical-coordinate system central axis by camera cylindrical coordinates;
S102. it after receiving specified camera three-dimensional coordinate, is searched from background data base corresponding with the specified camera three-dimensional coordinate Perspective images data, wherein the specified camera three-dimensional coordinate be camera spherical coordinate or camera cylindrical coordinates;
S103. by the perspective images data transmission found to foreground display end;
S104. output displaying is carried out to the perspective images data received by foreground display end.
2. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end Method, it is characterised in that:
The first perspective images data are the virtual reality image data that S201~204 is obtained in accordance with the following steps:
S201. using target object center as spherical coordinate system origin, if matrix arrangement arranges that cadre's camera lens is sat towards ball on spherical surface The ultra high-definition digital camera of mark system origin, and using the spherical coordinate of each portion's ultra high-definition digital camera as camera spherical coordinate;
S202. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains corresponding with each portion's ultra high-definition digital camera The first multi-view image data;
S203. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the first physical message number of target object According to, wherein the first physical message data include point cloud information, material information, color information and/or shadow parameter information;
S204. the first multi-view image data matching is carried out with the first physical message data to merge, obtain with it is described The first consistent first perspective images data of multi-view image clear data degree;
Alternatively, the first perspective images data are the virtual reality image data that S205~207 is obtained in accordance with the following steps:
S205. the threedimensional model of target object is generated using 3 d modeling software;
S206. using the threedimensional model center of target object as spherical coordinate system origin, several views are arranged in matrix arrangement on spherical surface Line is directed toward the camera view of spherical coordinate system origin, and using the spherical coordinate of each camera view as camera spherical coordinate;
S207. corresponding first perspective images are obtained in each camera view using the virtual camera module in 3-D graphic software Data.
3. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end Method, it is characterised in that:
The second perspective images data are the virtual reality image data that S301~304 is obtained in accordance with the following steps:
S301. using target object center as cylindrical-coordinate system origin, if matrix arrangement arrangement cadre's camera lens direction circle on cylinder The ultra high-definition digital camera of cylindrical coordinate central axis, and sat the cylindrical coordinates of each portion's ultra high-definition digital camera as camera cylinder Mark;
S302. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains corresponding with each portion's ultra high-definition digital camera The second multi-view image data;
S303. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the second physical message number of target object According to, wherein the second physical message data include point cloud information, material information, color information and/or shadow parameter information;
S304. the second multi-view image data matching is carried out with the second physical message data to merge, obtain with it is described The second consistent second perspective images data of multi-view image clear data degree;
Alternatively, the second perspective images data are the virtual reality image data that S305~307 is obtained in accordance with the following steps:
S305. the threedimensional model of target object is generated using 3 d modeling software;
S306. using the threedimensional model center of target object as cylindrical-coordinate system origin, several are arranged in matrix arrangement on cylinder Sight is directed toward the camera view of cylindrical-coordinate system central axis, and sits the cylindrical coordinates of each camera view as camera cylinder Mark;
S307. corresponding second perspective images are obtained in each camera view using the virtual camera module in 3-D graphic software Data.
4. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3 The method of application, it is characterised in that:
When perspective images data are virtual reality image data and in Image Acquisition, record acquisition multi-view image data when Between axial coordinate, the time axial coordinate is then included in camera matrix perspective data together with corresponding obtained perspective images data In.
5. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3 The method of application, it is characterised in that:
When perspective images data are virtual reality image data and after acquisition, join in conjunction with the physical attribute of threedimensional model Number carries out computer graphical rendering processing, later period outdoor scene incorporates processing and/or adjusts colors countenance to perspective images data, In, the physical attribute parameter includes material attribute parameter, environment attribute parameter and/or light source attributes parameter.
6. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3 The method of application, it is characterised in that:
The model Phase One IQ3100MP Trichromatic of the ultra high-definition digital camera, matrix arrangement number No less than 180.
7. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3 The method of application, it is characterised in that:
The three-dimensional scanning device is the spatial digitizer based on laser, radar and/or infrared ray.
8. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end Method, it is characterised in that:
The specified camera three-dimensional coordinate is instructed from the coordinate parameters control shaft or page operation of human-computer interaction interface, wherein The page operation instruction includes the instruction of visual angle moving operation and/or three-dimensional coordinate input instruction.
9. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end Method, it is characterised in that:
It further include following steps before the step S103:The perspective images data found are converted into picture format file Or WebGL formatted file.
10. a kind of realized based on camera Matrix Technology as described in claim 1 is answered in the display end progress lossless interaction of big data Method, it is characterised in that:
The foreground display end is that LED display, LCD display, holographic film, pseudo- line holographic projections equipment, 3D electric fan, AR are set Standby, VR equipment, projector touch interactive device or Three-dimensional holographic image documentation equipment.
CN201810550078.9A 2018-05-31 2018-05-31 A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end Pending CN108881886A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810550078.9A CN108881886A (en) 2018-05-31 2018-05-31 A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810550078.9A CN108881886A (en) 2018-05-31 2018-05-31 A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end

Publications (1)

Publication Number Publication Date
CN108881886A true CN108881886A (en) 2018-11-23

Family

ID=64336275

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810550078.9A Pending CN108881886A (en) 2018-05-31 2018-05-31 A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end

Country Status (1)

Country Link
CN (1) CN108881886A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110992474A (en) * 2019-12-13 2020-04-10 四川中绳矩阵技术发展有限公司 Method for realizing time domain technology
CN111063034A (en) * 2019-12-13 2020-04-24 四川中绳矩阵技术发展有限公司 Time domain interaction method
CN111815683A (en) * 2019-04-12 2020-10-23 北京京东尚科信息技术有限公司 Target positioning method and device, electronic equipment and computer readable medium
CN114648615A (en) * 2022-05-24 2022-06-21 四川中绳矩阵技术发展有限公司 Method, device and equipment for controlling interactive reproduction of target object and storage medium
CN115482325A (en) * 2022-09-29 2022-12-16 北京百度网讯科技有限公司 Image rendering method, device, system, equipment and medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101072366A (en) * 2007-05-24 2007-11-14 上海大学 Free stereo display system and method based on light field and binocular vision technology
CN103607578A (en) * 2013-12-03 2014-02-26 西安电子科技大学 Full-view angle picture browsing system based on video encoding
CN203708345U (en) * 2013-11-15 2014-07-09 中国科学院嘉兴光电工程中心 Instant three dimension shooting system based on camera array
WO2015021381A1 (en) * 2013-08-08 2015-02-12 University Of Florida Research Foundation, Incorporated Real-time reconstruction of the human body and automated avatar synthesis
CN105678701A (en) * 2016-01-28 2016-06-15 鲁鹏 An archaeological excavation unit sequence three-dimensional visualization method based on multi-view images and laser scanning
CN105847778A (en) * 2016-03-29 2016-08-10 朱光喜 360-degree multi-viewpoint 3D holographic video acquisition method, acquisition device, and realization method
CN106097433A (en) * 2016-05-30 2016-11-09 广州汉阈数据处理技术有限公司 Object industry and the stacking method of Image model and system
CN106358036A (en) * 2016-08-31 2017-01-25 杭州当虹科技有限公司 Method for watching virtual reality video at preset view angle
CN106504339A (en) * 2016-11-09 2017-03-15 四川长虹电器股份有限公司 Historical relic 3D methods of exhibiting based on virtual reality
CN107093387A (en) * 2017-06-26 2017-08-25 广州天翌云信息科技有限公司 An interactive holographic three-dimensional display system
CN107172442A (en) * 2017-06-11 2017-09-15 成都吱吖科技有限公司 A kind of interactive panoramic video storage method and device based on virtual reality

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101072366A (en) * 2007-05-24 2007-11-14 上海大学 Free stereo display system and method based on light field and binocular vision technology
WO2015021381A1 (en) * 2013-08-08 2015-02-12 University Of Florida Research Foundation, Incorporated Real-time reconstruction of the human body and automated avatar synthesis
CN203708345U (en) * 2013-11-15 2014-07-09 中国科学院嘉兴光电工程中心 Instant three dimension shooting system based on camera array
CN103607578A (en) * 2013-12-03 2014-02-26 西安电子科技大学 Full-view angle picture browsing system based on video encoding
CN105678701A (en) * 2016-01-28 2016-06-15 鲁鹏 An archaeological excavation unit sequence three-dimensional visualization method based on multi-view images and laser scanning
CN105847778A (en) * 2016-03-29 2016-08-10 朱光喜 360-degree multi-viewpoint 3D holographic video acquisition method, acquisition device, and realization method
CN106097433A (en) * 2016-05-30 2016-11-09 广州汉阈数据处理技术有限公司 Object industry and the stacking method of Image model and system
CN106358036A (en) * 2016-08-31 2017-01-25 杭州当虹科技有限公司 Method for watching virtual reality video at preset view angle
CN106504339A (en) * 2016-11-09 2017-03-15 四川长虹电器股份有限公司 Historical relic 3D methods of exhibiting based on virtual reality
CN107172442A (en) * 2017-06-11 2017-09-15 成都吱吖科技有限公司 A kind of interactive panoramic video storage method and device based on virtual reality
CN107093387A (en) * 2017-06-26 2017-08-25 广州天翌云信息科技有限公司 An interactive holographic three-dimensional display system

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111815683A (en) * 2019-04-12 2020-10-23 北京京东尚科信息技术有限公司 Target positioning method and device, electronic equipment and computer readable medium
CN111815683B (en) * 2019-04-12 2024-05-17 北京京东乾石科技有限公司 Target positioning method and device, electronic device and computer readable medium
CN110992474A (en) * 2019-12-13 2020-04-10 四川中绳矩阵技术发展有限公司 Method for realizing time domain technology
CN111063034A (en) * 2019-12-13 2020-04-24 四川中绳矩阵技术发展有限公司 Time domain interaction method
CN111063034B (en) * 2019-12-13 2023-08-04 四川中绳矩阵技术发展有限公司 Time domain interaction method
CN110992474B (en) * 2019-12-13 2023-10-31 四川中绳矩阵技术发展有限公司 Method for realizing time domain technology
CN114648615A (en) * 2022-05-24 2022-06-21 四川中绳矩阵技术发展有限公司 Method, device and equipment for controlling interactive reproduction of target object and storage medium
CN114648615B (en) * 2022-05-24 2022-07-29 四川中绳矩阵技术发展有限公司 Method, device and equipment for controlling interactive reproduction of target object and storage medium
CN115482325A (en) * 2022-09-29 2022-12-16 北京百度网讯科技有限公司 Image rendering method, device, system, equipment and medium
CN115482325B (en) * 2022-09-29 2023-10-31 北京百度网讯科技有限公司 Picture rendering method, device, system, equipment and medium

Similar Documents

Publication Publication Date Title
CN103337095B (en) The tridimensional virtual display methods of the three-dimensional geographical entity of a kind of real space
US10560687B2 (en) LED-based integral imaging display system as well as its control method and device
CN108881886A (en) A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end
KR100953931B1 (en) Mixed Reality Implementation System and Method
CN102568026B (en) A 3D augmented reality method for autostereoscopic display of multi-viewpoints
CN108919944A (en) A kind of virtual roaming method carrying out data lossless interaction in display end based on digital city model realization
CN107945282A (en) The synthesis of quick multi-view angle three-dimensional and methods of exhibiting and device based on confrontation network
US20070291035A1 (en) Horizontal Perspective Representation
US20050264558A1 (en) Multi-plane horizontal perspective hands-on simulator
CN107038745B (en) A 3D tourist landscape roaming interaction method and device
CN106101689A (en) Utilize the method that mobile phone monocular cam carries out augmented reality to virtual reality glasses
CN105869216A (en) Method and apparatus for presenting object target
US20050219240A1 (en) Horizontal perspective hands-on simulator
CN103959340A (en) Graphics rendering technique for autostereoscopic three dimensional display
EP1740998A2 (en) Horizontal perspective hand-on simulator
CN103077552A (en) Three-dimensional displaying method based on multi-view video
CN108043027A (en) Storage medium, electronic device, the display methods of game picture and device
CN118918293A (en) CAVE rescue scene construction method and system adopting digital twin and holographic technology
US20170104982A1 (en) Presentation of a virtual reality scene from a series of images
US20050248566A1 (en) Horizontal perspective hands-on simulator
JP3951362B2 (en) Image processing apparatus, game apparatus, method thereof, and recording medium
CN103871094A (en) Swept-volume-based three-dimensional display system data source generating method
CN108932055A (en) A kind of method and apparatus of augmented reality content
CN117218266A (en) 3D white-mode texture map generation method, device, equipment and medium
WO2023177773A1 (en) Stereoscopic features in virtual reality

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20181123

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载