CN108881886A - A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end - Google Patents
A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end Download PDFInfo
- Publication number
- CN108881886A CN108881886A CN201810550078.9A CN201810550078A CN108881886A CN 108881886 A CN108881886 A CN 108881886A CN 201810550078 A CN201810550078 A CN 201810550078A CN 108881886 A CN108881886 A CN 108881886A
- Authority
- CN
- China
- Prior art keywords
- camera
- data
- perspective images
- display end
- target object
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 239000011159 matrix material Substances 0.000 title claims abstract description 51
- 238000000034 method Methods 0.000 title claims abstract description 35
- 238000005516 engineering process Methods 0.000 title claims abstract description 25
- 230000002452 interceptive effect Effects 0.000 title claims abstract description 22
- 230000000007 visual effect Effects 0.000 claims abstract description 11
- 239000000463 material Substances 0.000 claims description 17
- 230000003993 interaction Effects 0.000 claims description 16
- 238000012545 processing Methods 0.000 claims description 16
- 238000009877 rendering Methods 0.000 claims description 6
- 230000005540 biological transmission Effects 0.000 claims description 4
- 239000003086 colorant Substances 0.000 claims description 4
- 230000008569 process Effects 0.000 abstract description 6
- 238000012512 characterization method Methods 0.000 abstract description 3
- 230000000717 retained effect Effects 0.000 abstract description 2
- 238000005457 optimization Methods 0.000 description 5
- 238000011160 research Methods 0.000 description 5
- 238000011161 development Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000015556 catabolic process Effects 0.000 description 3
- 238000006731 degradation reaction Methods 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 241001269238 Data Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 238000000465 moulding Methods 0.000 description 2
- 238000007500 overflow downdraw method Methods 0.000 description 2
- 230000003245 working effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 241000350052 Daniellia ogea Species 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000001149 cognitive effect Effects 0.000 description 1
- 238000004040 coloring Methods 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000009738 saturating Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/131—Protocols for games, networked simulations or virtual reality
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Processing Or Creating Images (AREA)
Abstract
The present invention relates to technical field of computer vision, disclose a kind of method realized based on camera Matrix Technology and carry out the lossless interactive application of big data in display end.It creates through the invention, it can provide a kind of method that losslessization interactive application is carried out to magnanimity byte data to foreground display end from background data base, the 3D hologram image data of target object is placed in background data base, it is then based on specified viewpoint, image data losslessization at corresponding visual angle is sent to foreground display end and carries out output displaying, due to not needing to cut down in the whole process to collected initial data, the initial data of characterization physical details (comprising concave-convex surface and texture etc.) information can be retained in display data, ensure to export the authenticity or accuracy of vision.
Description
Technical field
The invention belongs to technical field of computer vision, and in particular to one kind is realized based on camera Matrix Technology in display end
The method for carrying out the lossless interactive application of big data.
Background technique
Computer vision technique (i.e. Computer Vision, abbreviation CV) is current most popular one of research, refers to use
Video camera and computer replace human eye the machine vision such as to be identified, tracked and measured to target, and further do graphics process, make
Computer is treated as the image for being more suitable for eye-observation or sending instrument detection to.It is the research of a multi-crossed disciplines, is contained
Lid computer science (graphics, algorithm, theoretical research etc.), mathematics (information retrieval, machine learning), engineering (robot, NLP
Deng), biology (neuroscience) and psychology (cognitive science) etc..
In the current 3D Model Results way of output, most company also generallys use WebGL (i.e. Web Graphics
Library is a kind of 3D drawing agreement, and this drawing technique standard allows 2.0 knot of JavaScript and OpenGL ES
It is combined, a JavaScript by increasing OpenGL ES 2.0 is bound, and can provide hardware for HTML5Canvas
3D accelerates rendering, such web developer can come in browser to show more glibly by system video card 3D scene and
Model, moreover it is possible to create complicated navigation and data and visualize) technology, although eliminating the fiber crops of the exploitation dedicated Rendering Plug of webpage
It is tired, it can be used for the Website page that creation has complexity 3D structure, it might even be possible to for designing 3D web game etc., but meeting
Detail section loss of data and the problem serious to model bevel are caused, simultaneously because recessed with material diffusing reflection, bloom, normal
The shade that the modes such as convex textures and opacity show replaces physical details, has certain vision duplicity, causes to make
Product precision is only game level, can not be applied to holographic interaction scenarios, especially cannot be used for the digitlization to historical relic or historic site etc.
Protectiveness engineering field.
Summary of the invention
In order to solve the above problems existing in the present technology, it is an object of that present invention to provide one kind to be based on camera Matrix Technology
Realize the method for carrying out the lossless interactive application of big data in display end.
The technical scheme adopted by the invention is as follows:
A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end, including is as follows
Step:
S101. the camera matrix perspective data of target object is stored in advance in background data base, wherein the camera square
Battle array perspective data includes that several camera spherical coordinates of spherical surface matrix arrangement are carried out using target object center as spherical coordinate system origin
With the first perspective images data of each camera spherical coordinate of correspondence, and/or comprising using target object center as cylindrical-coordinate system
Origin carries out several camera cylindrical coordinates of cylinder matrix arrangement and the second perspective images number of corresponding each camera cylindrical coordinates
According to the sight of the first perspective images data is directed toward spherical coordinate system origin, second perspective images by camera spherical coordinate
The sight of data is directed toward cylindrical-coordinate system central axis by camera cylindrical coordinates;
S102. it after receiving specified camera three-dimensional coordinate, is searched and the specified camera three-dimensional coordinate from background data base
Corresponding perspective images data, wherein the specified camera three-dimensional coordinate is camera spherical coordinate or camera cylindrical coordinates;
S103. by the perspective images data transmission found to foreground display end;
S104. output displaying is carried out to the perspective images data received by foreground display end.
Optimization, the first perspective images data are the virtual reality figure that S201~204 is obtained in accordance with the following steps
As data:
S201. using target object center as spherical coordinate system origin, if matrix arrangement arranges cadre's camera lens direction on spherical surface
The ultra high-definition digital camera of spherical coordinate system origin, and sat the spherical coordinate of each portion's ultra high-definition digital camera as camera spherical surface
Mark;
S202. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains and each portion's ultra high-definition digital camera
Corresponding first multi-view image data;
S203. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the first physics letter of target object
Cease data, wherein the first physical message data include point cloud information, material information, color information and/or shadow parameter letter
Breath;
S204. the first multi-view image data matching is carried out with the first physical message data to merge, obtain with
The consistent first perspective images data of first multi-view image clear data degree;
Alternatively, the first perspective images data are the virtual reality image that S205~207 is obtained in accordance with the following steps
Data:
S205. the threedimensional model of target object is generated using 3 d modeling software;
S206. using the threedimensional model center of target object as spherical coordinate system origin, matrix arrangement setting is several on spherical surface
A sight is directed toward the camera view of spherical coordinate system origin, and using the spherical coordinate of each camera view as camera spherical coordinate;
S207. corresponding first visual angle is obtained in each camera view using the virtual camera module in 3-D graphic software
Image data.
Optimization, the second perspective images data are the virtual reality figure that S301~304 is obtained in accordance with the following steps
As data:
S301. using target object center as cylindrical-coordinate system origin, if matrix arrangement arranges cadre's camera lens court on cylinder
To the ultra high-definition digital camera of cylindrical-coordinate system central axis, and using the cylindrical coordinates of each portion's ultra high-definition digital camera as phase column
Areal coordinate;
S302. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains and each portion's ultra high-definition digital camera
Corresponding second multi-view image data;
S303. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the second physics letter of target object
Cease data, wherein the second physical message data include point cloud information, material information, color information and/or shadow parameter letter
Breath;
S304. the second multi-view image data matching is carried out with the second physical message data to merge, obtain with
The consistent second perspective images data of second multi-view image clear data degree;
Alternatively, the second perspective images data are the virtual reality image that S305~307 is obtained in accordance with the following steps
Data:
S305. the threedimensional model of target object is generated using 3 d modeling software;
S306. using the threedimensional model center of target object as cylindrical-coordinate system origin, if matrix arrangement is arranged on cylinder
Dry sight is directed toward the camera view of cylindrical-coordinate system central axis, and using the cylindrical coordinates of each camera view as camera cylinder
Coordinate;
S307. corresponding second visual angle is obtained in each camera view using the virtual camera module in 3-D graphic software
Image data.
It advanced optimizes, when perspective images data are for virtual reality image data and in Image Acquisition, record is adopted
Collect the time axial coordinate of multi-view image data, then includes together with corresponding obtained perspective images data by the time axial coordinate
In camera matrix perspective data.
It advanced optimizes, when perspective images data are virtual reality image data and after acquisition, in conjunction with three-dimensional
The physical attribute parameter of model, to perspective images data carry out computer graphical rendering processing, later period outdoor scene incorporate processing and/or
Adjust colors countenance, wherein the physical attribute parameter includes material attribute parameter, environment attribute parameter and/or light source attributes
Parameter.
Specifically, the model Phase One IQ3100MP Trichromatic of the ultra high-definition digital camera, square
Battle array number of permutations is no less than 180.
Specifically, the three-dimensional scanning device is the spatial digitizer based on laser, radar and/or infrared ray.
Optimization, coordinate parameters control shaft or page operation of the specified camera three-dimensional coordinate from human-computer interaction interface
Instruction, wherein the page operation instruction includes the instruction of visual angle moving operation and/or three-dimensional coordinate input instruction.
Optimization, further include following steps before the step S103:The perspective images data found are converted to
Picture format file or WebGL formatted file.
Specifically, the foreground display end is LED display, LCD display, holographic film, pseudo- line holographic projections equipment, 3D electricity
Fan, AR equipment, VR equipment, projector touch interactive device or Three-dimensional holographic image documentation equipment.
Beneficial effects of the present invention are:
(1) the invention provides a kind of lossless to the progress of magnanimity byte data from background data base to foreground display end
The 3D hologram image data of target object, i.e., be placed in background data base, be then based on specified by the method for changing interactive application
Image data losslessization at corresponding visual angle is sent to foreground display end and carries out output displaying, due in whole process by viewpoint
In do not need to cut down collected initial data, can display data in retain characterization physical details (comprising surface it is recessed
Convex and texture etc.) information initial data, it is ensured that export the authenticity or accuracy of vision;
(2) during acquiring virtual reality image data, it can not only accomplish the comprehensive data acquisition to the original.
It can also accomplish to correspond to the accurate match of the physical messages such as the room and time of each point in Point Cloud Processing link, into
And it can avoid carrying out the work such as image compensation in the later period;
(3) can make acquisition and treated three-dimensional holographic data (including room and time data etc.) can stable delivery,
Ensure in the following various data processing circumstances, is not in the calculating such as data degradation, disorder, or even collapse error problem, i.e.,
After one-pass molding, the following various holographic applications environment can be coped with, realize the working effect of " once and for all ", be the various of future
Holographic applications save a large amount of model development cost;
(4) it can be achieved visually to carry out material object the purpose that complete and lossless digitlization reduction reproduces, with fidelity
Researching value, the digitlization particularly suitable for historical relic, historic site, spot, important security place and traditional architecture etc. are protected
The application fields such as shield property engineering field and electronics/3D hologram interaction books field;
(5) when the hologram interactive application in transfer complex environment, without setting up processing large capacity near scene
Low bi-dimensional end (the benefit matched can be used in the ancillary hardware server apparatus of data, such as user under current 4G communication environment
In smart phone) hologram of ZB (i.e. 10,000,000,000,000 hundred million bytes) rank is browsed, it thus can look forward to logical in following 5G
Interrogate under environment, can also long-range real-time delivery naked eye grade hologram interactive application data;
(6) output element is being applied, the simple interactivity hologram that both may be output as Entertainment or browse level is answered
With, can be also the corresponding output format of complex interactions hologram application and development such as scientific research or engineering, promote convenient for practical and
Using.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below
There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this
Some embodiments of invention for those of ordinary skill in the art without creative efforts, can be with
It obtains other drawings based on these drawings.
Fig. 1 is that provided by the invention realized based on camera Matrix Technology carries out the lossless interactive application of big data in display end
Method flow schematic diagram.
Fig. 2 is the camera matrix schematic diagram provided by the invention on spherical surface.
Fig. 3 is transversal spherical surface camera matrix schematic diagram provided by the invention.
Fig. 4 is transversal cylinder camera matrix schematic diagram provided by the invention.
In above-mentioned attached drawing:1- target object;2- ultra high-definition digital camera.
Specific embodiment
With reference to the accompanying drawing and specific embodiment the present invention is further elaborated.It should be noted that for this
The explanation of a little way of example is used to help understand the present invention, but and does not constitute a limitation of the invention.
The terms "and/or", only a kind of incidence relation for describing affiliated partner, indicates that there may be three kinds of passes
System, for example, A and/or B, can indicate:Individualism A, individualism B exist simultaneously tri- kinds of situations of A and B, the terms
"/and " it is to describe another affiliated partner relationship, indicate may exist two kinds of relationships, for example, A/ and B, can indicate:Individually deposit
In A, two kinds of situations of individualism A and B, in addition, character "/" herein, typicallying represent forward-backward correlation object is a kind of "or" pass
System.
Embodiment one
As illustrated in fig. 1 and 2, described realized based on camera Matrix Technology provided in this embodiment carries out big data in display end
The method of lossless interactive application, includes the following steps.
S101. the camera matrix perspective data of target object is stored in advance in background data base, wherein the camera square
Battle array perspective data includes that several camera spherical coordinates of spherical surface matrix arrangement are carried out using target object center as spherical coordinate system origin
With the first perspective images data of each camera spherical coordinate of correspondence, and/or comprising using target object center as cylindrical-coordinate system
Origin carries out several camera cylindrical coordinates of cylinder matrix arrangement and the second perspective images number of corresponding each camera cylindrical coordinates
According to the sight of the first perspective images data is directed toward spherical coordinate system origin, second perspective images by camera spherical coordinate
The sight of data is directed toward cylindrical-coordinate system central axis by camera cylindrical coordinates.
The crosspoint of spherical surface camera matrix as shown in Figure 2, each spherical surface warp and spherical surface weft is camera spherical surface
Coordinate points.The crosspoint of cylinder camera matrix as shown in Figure 3, each cylinder loop wire and cylinder vertical line is that camera cylinder is sat
Punctuate.The target object can be real object, such as the historical relic present in real world or engineering goods, or empty
Quasi- body, such as the Virtual Building or the art work that are generated in a computer by three dimensional design.Thus first perspective images
Data or the second perspective images data can be (to use number based on virtual reality image data obtained by real object acquisition
The image data of word expression real object information), or based on virtual reality image obtained from virtual body three-dimensional models
Data (i.e. with the image data of the quasi- real-world object information of number).
In the step S101, the first perspective images data can obtain for S201~204 in accordance with the following steps
Virtual reality image data:S201. using target object center as spherical coordinate system origin, if matrix arrangement is arranged on spherical surface
Cadre's camera lens towards spherical coordinate system origin ultra high-definition digital camera, and using the spherical coordinate of each portion's ultra high-definition digital camera as
Camera spherical coordinate;S202. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains digital with each portion's ultra high-definition
The corresponding first multi-view image data of camera;
S203. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the first physics letter of target object
Cease data, wherein the first physical message data include point cloud information, material information, color information and/or shadow parameter letter
Breath;S204. the first multi-view image data are carried out matching with the first physical message data to merge, is obtained and described
The one consistent first perspective images data of multi-view image clear data degree.
Similarly, in the step S101, the second perspective images data may be in accordance with the following steps S301~
The 304 virtual reality image datas obtained:S301. using target object center as cylindrical-coordinate system origin, the matrix on cylinder
If arranged cadre camera lens is towards the ultra high-definition digital camera of cylindrical-coordinate system central axis, and by each portion's ultra high-definition digital camera
Cylindrical coordinates as camera cylindrical coordinates;S302. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtain with
The corresponding second multi-view image data of each portion's ultra high-definition digital camera;S303. target object is carried out using three-dimensional scanning device
3-D scanning obtains the second physical message data of target object, wherein the second physical message data include point cloud letter
Breath, material information, color information and/or shadow parameter information;S304. by the second multi-view image data and second object
Reason information data carries out matching fusion, obtains and the consistent second perspective images number of the second multi-view image clear data degree
According to.
By as described in abovementioned steps S201~S204 or step S301~S304 based on camera and with 3-D scanning
Image acquisition scheme supplemented by equipment, it can be ensured that material picture contents OK acquisition (refer to realize maximal accuracy, it is most accurate and
Most comprehensive acquisition mode), the acquisition data guaranteed will be long-term useful for future.And existing most of
In type of industry, the image acquisition scheme based on laser acquisition data and supplemented by camera is generallyd use, no material is first used
The laser of information, texture information and colouring information etc. acquires the physical message on surface in kind, then uses the collected material of camera
Information is attached on the collected physical data of laser, this process will lead to collected data precision fusion and mismatch, must
It so will cause certain information errors, such error message is worthless absolutely for historical relic and paleovolcanic structure etc..Such as
Method such as/FARO Singapore, Tian Bao/Trimble, Bentley, Agisoft, Topcon, Polyworks, Geomagic,
The companies such as Creaform, which have, is absorbed in this field, possesses to realize the relevant technologies such as outdoor scene modeling and 3-D scanning, although he
Object scanning accuracy can achieve army grade, but do not accomplish also in texture and material precision ultimate attainment, do not do true
4D timing (space is recorded with multiframe model in real time) in positive meaning, including the information such as shadow, material texture, not because of object
With and acquisition mode is different, such as height is saturating and bloom.In addition, employed in the step S204 or the step S304
It is the prior art with fusion method, is referred to《Mapping notification》Disclosed in article《Three-dimensional laser point cloud and full-view image
Match fusion method》(publication date:Routinely displacement design 2016-06-30) is carried out to realize.
It advanced optimizes, when carrying out Image Acquisition in the step S202 or the step S303, record acquisition view
The time axial coordinate of angle image data (the i.e. described first multi-view image data or the second multi-view image data), then should
Time axial coordinate and corresponding obtained perspective images data (the i.e. described first perspective images data or the second perspective images number
According to) be included in camera matrix perspective data together.So during acquiring virtual reality image data, can not only it do
To the comprehensive data acquisition to the original, it can also accomplish the room and time etc. to each point in Point Cloud Processing link
The accurate match of physical message is corresponding, and then can avoid carrying out the work such as image compensation in the later period, and for the number of other companies
According to acquisition method, due to generally existing space and the deviation of time data, the later period needs to carry out textures compensation etc. and makes up work
Make.
It advanced optimizes, in order to further ensure that the clarity and losslessization of Image Acquisition, the ultra high-definition number phase
The model of machine is preferably Phase One IQ3100MP Trichromatic, and matrix arrangement number is no less than 180.Due to
Built-in 100MP cmos sensor is integrated with newest Bayer filter dye technology, and sufficiently benefit in aforementioned model digital camera
With the achievement that works closely of Phase One and Sony, the color of finding of naked eye can be accurately captured.This brand new technical is
Phase One institute is exclusive, can promote the expressive force of digital color, unprecedented realistic colour is presented, with 1.01 hundred million pixels
Creativity is played unrestrainedly, so that the color that collected color more infinite approach human eye is seen.It can so borrow
Sensor most outstanding so far is helped, the Nature masterpiece of spectators' short distance appreciation riot of color and the ancient times older generations is allowed to stay
Rarity.In addition, the three-dimensional scanning device can be, but not limited to as the spatial digitizer based on laser, radar and/or infrared ray.
In the step S101, the first perspective images data can also obtain for S205~207 in accordance with the following steps
The virtual reality image data taken:S205. the threedimensional model of target object is generated using 3 d modeling software;S206. with mesh
The threedimensional model center for marking object is spherical coordinate system origin, several sights direction spherical coordinate system is arranged in matrix arrangement on spherical surface
The camera view of origin, and using the spherical coordinate of each camera view as camera spherical coordinate;S207. soft using 3-D graphic
Virtual camera module in part obtains corresponding first perspective images data in each camera view.
Similarly, in the step S101, the second perspective images data can also for S305 in accordance with the following steps~
The 307 virtual reality image datas obtained:S305. the threedimensional model of target object is generated using 3 d modeling software;S306.
Using the threedimensional model center of target object as cylindrical-coordinate system origin, several sights direction circle is arranged in matrix arrangement on cylinder
The camera view of cylindrical coordinate central axis, and using the cylindrical coordinates of each camera view as camera cylindrical coordinates;S307. it applies
Virtual camera module in 3-D graphic software obtains corresponding second perspective images data in each camera view.
In the step S205 or S305, the generating mode of the threedimensional model of the target object can be, but not limited to as
Under:Target object and each dimension two dimensional image is first drawn, all two dimensional images of target object are then based on, using three
Tie up the threedimensional model that modeling function module generates target object.In addition, the 3 d modeling software and the 3-D graphic software
Can be, but not limited to for the three-dimensional applications such as CAD, SolidWorks, SketchUp, 3Ds Max, Rhino, Maya and ZBrush it is soft
Part.
It advanced optimizes, it is right herein in connection with the physical attribute parameter of threedimensional model after the step S207 or S307
Perspective images data (the i.e. described first perspective images data or the second perspective images data) carry out computer graphical rendering
Processing, later period outdoor scene incorporate processing and/or adjustment colors countenance, wherein the physical attribute parameter can be, but not limited to include
Material attribute parameter, environment attribute parameter and/or light source attributes parameter etc..Aforementioned computer figure rendering processing, later period outdoor scene
Incorporating processing and adjustment colors countenance is existing processing technique, so can carry out scientifically complex die to final image effect
It is quasi-, intend the physical attribute of true real world completely, makes uninteresting data materialization.Such as in real world, for copper object,
Its surface can be carried out scientific simulation in parameters such as the reflections and refraction at each visual angle;Or in real world, very much
Object has different degrees of " Fresnel effect ":When so that sight is perpendicular to surface, reflection is weaker, and works as the non-perpendicular table of sight
When face, angle is smaller, and reflection is more obvious (for seeing to a ball, the reflection at ball center is weaker, stronger by proximal edge).
S102. it after receiving specified camera three-dimensional coordinate, is searched and the specified camera three-dimensional coordinate from background data base
Corresponding perspective images data, wherein the specified camera three-dimensional coordinate is camera spherical coordinate or camera cylindrical coordinates.
In the step S102, the specified camera three-dimensional coordinate can be, but not limited to the seat from human-computer interaction interface
Mark state modulator axis or page operation instruction etc., wherein the page operation instruction can be, but not limited to comprising the mobile behaviour in visual angle
Make instruction and/or three-dimensional coordinate input instruction etc..It can not only so see from each visual angle and simulate the true of real world completely
Effect, moreover it is possible to adjust a certain item parameter and be reacted to generate a series of linkage to result.In addition, if also receiving specified time axis
Coordinate, then the perspective images data found must also be corresponding with specified time axial coordinate, to realize the big number of time dimension
According to interaction, wherein the specified time axial coordinate equally can be, but not limited to the control of the coordinate parameters from human-computer interaction interface
Axis or page operation instruction etc..
S103. by the perspective images data transmission found to foreground display end.
It further include following steps before the step S103:The perspective images data found are converted into picture format
File or WebGL formatted file etc..In the step S103, transmitting path based on RJ45 cable or RS232 either passed
The cable network of the transmission such as defeated line, can also be based on WiFi, 3G, 4G and the mobile radio network of even future 5G technology.This
Outside, the foreground display end can be, but not limited to set for existing LED display, LCD display, holographic film, pseudo- line holographic projections
Standby, 3D electric fan, AR equipment, VR equipment, projector touch interactive device or Three-dimensional holographic image documentation equipment etc..
S104. output displaying is carried out to the perspective images data received by foreground display end.
In the step 104, optimization, the output exhibition method of the perspective images data can be specific as follows:According to
The customized visual field sizes range of user selects the image data of corresponding regional area to carry out defeated from the perspective images data
It shows out.It, can be to ZB grades of (i.e. 10,000,000,000,000 hundred million byte levels from there through the lossless information transfer scheme of abovementioned steps S101~S104
Data volume not) is browsed low with display end, without data degradation because the data seen of orientation in office all with
The data seen on background data base are same data, are not only the same formats, whole process is not to original number
According to any information reduction of progress.I.e. in other words, the information content of 100TB or so can be browsed on the picture of 145KB.
To sum up, it is realized based on camera Matrix Technology in the display end progress lossless friendship of big data using provided by the present embodiment
The method mutually applied, has the following technical effect that:
(1) it present embodiments provides one kind and losslessization is carried out to magnanimity byte data from background data base to foreground display end
The 3D hologram image data of target object is placed in background data base, is then based on specified view by the method for interactive application
Image data losslessization at corresponding visual angle is sent to foreground display end and carries out output displaying, due in the whole process by point
It does not need to cut down collected initial data, it (includes concave-convex surface that characterization physical details can be retained in display data
With texture etc.) initial data of information, it is ensured that export the authenticity or accuracy of vision;
(2) during acquiring virtual reality image data, it can not only accomplish the comprehensive data acquisition to the original.
It can also accomplish to correspond to the accurate match of the physical messages such as the room and time of each point in Point Cloud Processing link, into
And it can avoid carrying out the work such as image compensation in the later period;
(3) can make acquisition and treated three-dimensional holographic data (including room and time data etc.) can stable delivery,
Ensure in the following various data processing circumstances, is not in the calculating such as data degradation, disorder, or even collapse error problem, i.e.,
After one-pass molding, the following various holographic applications environment can be coped with, realize the working effect of " once and for all ", be the various of future
Holographic applications save a large amount of model development cost;
(4) it can be achieved visually to carry out material object the purpose that complete and lossless digitlization reduction reproduces, with fidelity
Researching value, the digitlization particularly suitable for historical relic, historic site, spot, important security place and traditional architecture etc. are protected
The application fields such as shield property engineering field and electronics/3D hologram interaction books field;
(5) when the hologram interactive application in transfer complex environment, without setting up processing large capacity near scene
Low bi-dimensional end (the benefit matched can be used in the ancillary hardware server apparatus of data, such as user under current 4G communication environment
In smart phone) hologram of ZB (i.e. 10,000,000,000,000 hundred million bytes) rank is browsed, it thus can look forward to logical in following 5G
Interrogate under environment, can also long-range real-time delivery naked eye grade hologram interactive application data;
(6) output element is being applied, the simple interactivity hologram that both may be output as Entertainment or browse level is answered
With, can be also the corresponding output format of complex interactions hologram application and development such as scientific research or engineering, promote convenient for practical and
Using.
The present invention is not limited to above-mentioned optional embodiment, anyone can show that other are each under the inspiration of the present invention
The product of kind form.Above-mentioned specific embodiment should not be understood the limitation of pairs of protection scope of the present invention, protection of the invention
Range should be subject to be defined in claims, and specification can be used for interpreting the claims.
Claims (10)
1. a kind of realize that, in the method that display end carries out the lossless interactive application of big data, feature exists based on camera Matrix Technology
In including the following steps:
S101. the camera matrix perspective data of target object is stored in advance in background data base, wherein the camera matrix view
It is spherical coordinate system origin progress several camera spherical coordinates of spherical surface matrix arrangement and right that angular data, which includes using target object center,
The first perspective images data of each camera spherical coordinate are answered, and/or comprising using target object center as cylindrical-coordinate system origin
Carry out several camera cylindrical coordinates of cylinder matrix arrangement and the second perspective images data of corresponding each camera cylindrical coordinates, institute
The sight for stating the first perspective images data is directed toward spherical coordinate system origin by camera spherical coordinate, the second perspective images data
Sight is directed toward cylindrical-coordinate system central axis by camera cylindrical coordinates;
S102. it after receiving specified camera three-dimensional coordinate, is searched from background data base corresponding with the specified camera three-dimensional coordinate
Perspective images data, wherein the specified camera three-dimensional coordinate be camera spherical coordinate or camera cylindrical coordinates;
S103. by the perspective images data transmission found to foreground display end;
S104. output displaying is carried out to the perspective images data received by foreground display end.
2. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end
Method, it is characterised in that:
The first perspective images data are the virtual reality image data that S201~204 is obtained in accordance with the following steps:
S201. using target object center as spherical coordinate system origin, if matrix arrangement arranges that cadre's camera lens is sat towards ball on spherical surface
The ultra high-definition digital camera of mark system origin, and using the spherical coordinate of each portion's ultra high-definition digital camera as camera spherical coordinate;
S202. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains corresponding with each portion's ultra high-definition digital camera
The first multi-view image data;
S203. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the first physical message number of target object
According to, wherein the first physical message data include point cloud information, material information, color information and/or shadow parameter information;
S204. the first multi-view image data matching is carried out with the first physical message data to merge, obtain with it is described
The first consistent first perspective images data of multi-view image clear data degree;
Alternatively, the first perspective images data are the virtual reality image data that S205~207 is obtained in accordance with the following steps:
S205. the threedimensional model of target object is generated using 3 d modeling software;
S206. using the threedimensional model center of target object as spherical coordinate system origin, several views are arranged in matrix arrangement on spherical surface
Line is directed toward the camera view of spherical coordinate system origin, and using the spherical coordinate of each camera view as camera spherical coordinate;
S207. corresponding first perspective images are obtained in each camera view using the virtual camera module in 3-D graphic software
Data.
3. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end
Method, it is characterised in that:
The second perspective images data are the virtual reality image data that S301~304 is obtained in accordance with the following steps:
S301. using target object center as cylindrical-coordinate system origin, if matrix arrangement arrangement cadre's camera lens direction circle on cylinder
The ultra high-definition digital camera of cylindrical coordinate central axis, and sat the cylindrical coordinates of each portion's ultra high-definition digital camera as camera cylinder
Mark;
S302. each portion's ultra high-definition digital camera of synchronously control carries out Image Acquisition, obtains corresponding with each portion's ultra high-definition digital camera
The second multi-view image data;
S303. 3-D scanning is carried out to target object using three-dimensional scanning device, obtains the second physical message number of target object
According to, wherein the second physical message data include point cloud information, material information, color information and/or shadow parameter information;
S304. the second multi-view image data matching is carried out with the second physical message data to merge, obtain with it is described
The second consistent second perspective images data of multi-view image clear data degree;
Alternatively, the second perspective images data are the virtual reality image data that S305~307 is obtained in accordance with the following steps:
S305. the threedimensional model of target object is generated using 3 d modeling software;
S306. using the threedimensional model center of target object as cylindrical-coordinate system origin, several are arranged in matrix arrangement on cylinder
Sight is directed toward the camera view of cylindrical-coordinate system central axis, and sits the cylindrical coordinates of each camera view as camera cylinder
Mark;
S307. corresponding second perspective images are obtained in each camera view using the virtual camera module in 3-D graphic software
Data.
4. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3
The method of application, it is characterised in that:
When perspective images data are virtual reality image data and in Image Acquisition, record acquisition multi-view image data when
Between axial coordinate, the time axial coordinate is then included in camera matrix perspective data together with corresponding obtained perspective images data
In.
5. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3
The method of application, it is characterised in that:
When perspective images data are virtual reality image data and after acquisition, join in conjunction with the physical attribute of threedimensional model
Number carries out computer graphical rendering processing, later period outdoor scene incorporates processing and/or adjusts colors countenance to perspective images data,
In, the physical attribute parameter includes material attribute parameter, environment attribute parameter and/or light source attributes parameter.
6. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3
The method of application, it is characterised in that:
The model Phase One IQ3100MP Trichromatic of the ultra high-definition digital camera, matrix arrangement number
No less than 180.
7. a kind of realized based on camera Matrix Technology carries out the lossless interaction of big data in display end as claimed in claim 2 or claim 3
The method of application, it is characterised in that:
The three-dimensional scanning device is the spatial digitizer based on laser, radar and/or infrared ray.
8. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end
Method, it is characterised in that:
The specified camera three-dimensional coordinate is instructed from the coordinate parameters control shaft or page operation of human-computer interaction interface, wherein
The page operation instruction includes the instruction of visual angle moving operation and/or three-dimensional coordinate input instruction.
9. a kind of realized based on camera Matrix Technology as described in claim 1 carries out the lossless interactive application of big data in display end
Method, it is characterised in that:
It further include following steps before the step S103:The perspective images data found are converted into picture format file
Or WebGL formatted file.
10. a kind of realized based on camera Matrix Technology as described in claim 1 is answered in the display end progress lossless interaction of big data
Method, it is characterised in that:
The foreground display end is that LED display, LCD display, holographic film, pseudo- line holographic projections equipment, 3D electric fan, AR are set
Standby, VR equipment, projector touch interactive device or Three-dimensional holographic image documentation equipment.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810550078.9A CN108881886A (en) | 2018-05-31 | 2018-05-31 | A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810550078.9A CN108881886A (en) | 2018-05-31 | 2018-05-31 | A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN108881886A true CN108881886A (en) | 2018-11-23 |
Family
ID=64336275
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201810550078.9A Pending CN108881886A (en) | 2018-05-31 | 2018-05-31 | A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN108881886A (en) |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110992474A (en) * | 2019-12-13 | 2020-04-10 | 四川中绳矩阵技术发展有限公司 | Method for realizing time domain technology |
| CN111063034A (en) * | 2019-12-13 | 2020-04-24 | 四川中绳矩阵技术发展有限公司 | Time domain interaction method |
| CN111815683A (en) * | 2019-04-12 | 2020-10-23 | 北京京东尚科信息技术有限公司 | Target positioning method and device, electronic equipment and computer readable medium |
| CN114648615A (en) * | 2022-05-24 | 2022-06-21 | 四川中绳矩阵技术发展有限公司 | Method, device and equipment for controlling interactive reproduction of target object and storage medium |
| CN115482325A (en) * | 2022-09-29 | 2022-12-16 | 北京百度网讯科技有限公司 | Image rendering method, device, system, equipment and medium |
Citations (11)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101072366A (en) * | 2007-05-24 | 2007-11-14 | 上海大学 | Free stereo display system and method based on light field and binocular vision technology |
| CN103607578A (en) * | 2013-12-03 | 2014-02-26 | 西安电子科技大学 | Full-view angle picture browsing system based on video encoding |
| CN203708345U (en) * | 2013-11-15 | 2014-07-09 | 中国科学院嘉兴光电工程中心 | Instant three dimension shooting system based on camera array |
| WO2015021381A1 (en) * | 2013-08-08 | 2015-02-12 | University Of Florida Research Foundation, Incorporated | Real-time reconstruction of the human body and automated avatar synthesis |
| CN105678701A (en) * | 2016-01-28 | 2016-06-15 | 鲁鹏 | An archaeological excavation unit sequence three-dimensional visualization method based on multi-view images and laser scanning |
| CN105847778A (en) * | 2016-03-29 | 2016-08-10 | 朱光喜 | 360-degree multi-viewpoint 3D holographic video acquisition method, acquisition device, and realization method |
| CN106097433A (en) * | 2016-05-30 | 2016-11-09 | 广州汉阈数据处理技术有限公司 | Object industry and the stacking method of Image model and system |
| CN106358036A (en) * | 2016-08-31 | 2017-01-25 | 杭州当虹科技有限公司 | Method for watching virtual reality video at preset view angle |
| CN106504339A (en) * | 2016-11-09 | 2017-03-15 | 四川长虹电器股份有限公司 | Historical relic 3D methods of exhibiting based on virtual reality |
| CN107093387A (en) * | 2017-06-26 | 2017-08-25 | 广州天翌云信息科技有限公司 | An interactive holographic three-dimensional display system |
| CN107172442A (en) * | 2017-06-11 | 2017-09-15 | 成都吱吖科技有限公司 | A kind of interactive panoramic video storage method and device based on virtual reality |
-
2018
- 2018-05-31 CN CN201810550078.9A patent/CN108881886A/en active Pending
Patent Citations (11)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101072366A (en) * | 2007-05-24 | 2007-11-14 | 上海大学 | Free stereo display system and method based on light field and binocular vision technology |
| WO2015021381A1 (en) * | 2013-08-08 | 2015-02-12 | University Of Florida Research Foundation, Incorporated | Real-time reconstruction of the human body and automated avatar synthesis |
| CN203708345U (en) * | 2013-11-15 | 2014-07-09 | 中国科学院嘉兴光电工程中心 | Instant three dimension shooting system based on camera array |
| CN103607578A (en) * | 2013-12-03 | 2014-02-26 | 西安电子科技大学 | Full-view angle picture browsing system based on video encoding |
| CN105678701A (en) * | 2016-01-28 | 2016-06-15 | 鲁鹏 | An archaeological excavation unit sequence three-dimensional visualization method based on multi-view images and laser scanning |
| CN105847778A (en) * | 2016-03-29 | 2016-08-10 | 朱光喜 | 360-degree multi-viewpoint 3D holographic video acquisition method, acquisition device, and realization method |
| CN106097433A (en) * | 2016-05-30 | 2016-11-09 | 广州汉阈数据处理技术有限公司 | Object industry and the stacking method of Image model and system |
| CN106358036A (en) * | 2016-08-31 | 2017-01-25 | 杭州当虹科技有限公司 | Method for watching virtual reality video at preset view angle |
| CN106504339A (en) * | 2016-11-09 | 2017-03-15 | 四川长虹电器股份有限公司 | Historical relic 3D methods of exhibiting based on virtual reality |
| CN107172442A (en) * | 2017-06-11 | 2017-09-15 | 成都吱吖科技有限公司 | A kind of interactive panoramic video storage method and device based on virtual reality |
| CN107093387A (en) * | 2017-06-26 | 2017-08-25 | 广州天翌云信息科技有限公司 | An interactive holographic three-dimensional display system |
Cited By (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111815683A (en) * | 2019-04-12 | 2020-10-23 | 北京京东尚科信息技术有限公司 | Target positioning method and device, electronic equipment and computer readable medium |
| CN111815683B (en) * | 2019-04-12 | 2024-05-17 | 北京京东乾石科技有限公司 | Target positioning method and device, electronic device and computer readable medium |
| CN110992474A (en) * | 2019-12-13 | 2020-04-10 | 四川中绳矩阵技术发展有限公司 | Method for realizing time domain technology |
| CN111063034A (en) * | 2019-12-13 | 2020-04-24 | 四川中绳矩阵技术发展有限公司 | Time domain interaction method |
| CN111063034B (en) * | 2019-12-13 | 2023-08-04 | 四川中绳矩阵技术发展有限公司 | Time domain interaction method |
| CN110992474B (en) * | 2019-12-13 | 2023-10-31 | 四川中绳矩阵技术发展有限公司 | Method for realizing time domain technology |
| CN114648615A (en) * | 2022-05-24 | 2022-06-21 | 四川中绳矩阵技术发展有限公司 | Method, device and equipment for controlling interactive reproduction of target object and storage medium |
| CN114648615B (en) * | 2022-05-24 | 2022-07-29 | 四川中绳矩阵技术发展有限公司 | Method, device and equipment for controlling interactive reproduction of target object and storage medium |
| CN115482325A (en) * | 2022-09-29 | 2022-12-16 | 北京百度网讯科技有限公司 | Image rendering method, device, system, equipment and medium |
| CN115482325B (en) * | 2022-09-29 | 2023-10-31 | 北京百度网讯科技有限公司 | Picture rendering method, device, system, equipment and medium |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN103337095B (en) | The tridimensional virtual display methods of the three-dimensional geographical entity of a kind of real space | |
| US10560687B2 (en) | LED-based integral imaging display system as well as its control method and device | |
| CN108881886A (en) | A method of it is realized based on camera Matrix Technology and carries out the lossless interactive application of big data in display end | |
| KR100953931B1 (en) | Mixed Reality Implementation System and Method | |
| CN102568026B (en) | A 3D augmented reality method for autostereoscopic display of multi-viewpoints | |
| CN108919944A (en) | A kind of virtual roaming method carrying out data lossless interaction in display end based on digital city model realization | |
| CN107945282A (en) | The synthesis of quick multi-view angle three-dimensional and methods of exhibiting and device based on confrontation network | |
| US20070291035A1 (en) | Horizontal Perspective Representation | |
| US20050264558A1 (en) | Multi-plane horizontal perspective hands-on simulator | |
| CN107038745B (en) | A 3D tourist landscape roaming interaction method and device | |
| CN106101689A (en) | Utilize the method that mobile phone monocular cam carries out augmented reality to virtual reality glasses | |
| CN105869216A (en) | Method and apparatus for presenting object target | |
| US20050219240A1 (en) | Horizontal perspective hands-on simulator | |
| CN103959340A (en) | Graphics rendering technique for autostereoscopic three dimensional display | |
| EP1740998A2 (en) | Horizontal perspective hand-on simulator | |
| CN103077552A (en) | Three-dimensional displaying method based on multi-view video | |
| CN108043027A (en) | Storage medium, electronic device, the display methods of game picture and device | |
| CN118918293A (en) | CAVE rescue scene construction method and system adopting digital twin and holographic technology | |
| US20170104982A1 (en) | Presentation of a virtual reality scene from a series of images | |
| US20050248566A1 (en) | Horizontal perspective hands-on simulator | |
| JP3951362B2 (en) | Image processing apparatus, game apparatus, method thereof, and recording medium | |
| CN103871094A (en) | Swept-volume-based three-dimensional display system data source generating method | |
| CN108932055A (en) | A kind of method and apparatus of augmented reality content | |
| CN117218266A (en) | 3D white-mode texture map generation method, device, equipment and medium | |
| WO2023177773A1 (en) | Stereoscopic features in virtual reality |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20181123 |