CN104969264A - Method and apparatus for adding annotations to plenoptic light fields - Google Patents
Method and apparatus for adding annotations to plenoptic light fields Download PDFInfo
- Publication number
- CN104969264A CN104969264A CN201280077894.3A CN201280077894A CN104969264A CN 104969264 A CN104969264 A CN 104969264A CN 201280077894 A CN201280077894 A CN 201280077894A CN 104969264 A CN104969264 A CN 104969264A
- Authority
- CN
- China
- Prior art keywords
- annotation
- plenoptic
- view
- annotations
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/21—Indexing scheme for image data processing or generation, in general involving computational photography
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Processing Or Creating Images (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
背景技术 Background technique
本发明涉及增强现实(augmented reality)方法和设备,特别地涉及用于将注解添加到对应于场景(scene)的数据的方法和各种设备。 The present invention relates to augmented reality methods and devices, and in particular, to methods and various devices for adding annotations to data corresponding to scenes.
在手持便携式装置诸如智能手机、掌上型计算机、便携式媒体播放器、个人数字助理(PDA)装置等等的发展中的快速进步已导致包含涉及图像处理的新奇特征和应用。比如增强现实应用是已知的,其中用户将便携式装置指向场景例如风景、建筑物、海报、或博物馆中的绘画,并且显示器将图像与涉及该场景的叠加的信息一起示出。这样的信息能够包含:例如针对山和住所的名称;人名;针对建筑物的历史信息;和商业信息,诸如广告,例如餐馆菜单。这样的系统的示例在EP1246080中和在EP2207113中被描述。 Rapid progress in the development of hand-portable devices such as smartphones, palmtop computers, portable media players, personal digital assistant (PDA) devices, etc. has resulted in the inclusion of novel features and applications involving image processing. For example augmented reality applications are known in which a user points a portable device at a scene such as a landscape, a building, a poster, or a painting in a museum, and the display shows the image with superimposed information relating to the scene. Such information can include: names, for example, of mountains and residences; names of people; historical information for buildings; and commercial information, such as advertisements, eg restaurant menus. Examples of such systems are described in EP1246080 and in EP2207113.
已知的是通过在无线通信网络中的服务器将注解信息提供给便携式装置。包含带有服务器和便携式装置的通信网络的注解系统以及注解方法也是已知的。 It is known to provide annotation information to a portable device by a server in a wireless communication network. Annotation systems involving communication networks with servers and portable devices, as well as annotation methods, are also known.
许多注解方法包含将图像诸如由带有标准CCD或CMOS传感器的标准针孔相机产生的2D图像或计算机生成的图像与储存在数据库中的参考图像的集合相比较的步骤。因为实际查看角度和照明条件能够关于储存在数据库中的图像是不同的,所以比较算法的目标是去除这些参数的影响。 Many annotation methods involve the step of comparing an image, such as a 2D image produced by a standard pinhole camera with a standard CCD or CMOS sensor, or a computer-generated image, with a set of reference images stored in a database. Since actual viewing angles and lighting conditions can vary with respect to images stored in the database, the comparison algorithm aims to remove the influence of these parameters.
比如,WO2008134901描述其中使用与通信终端相关联的数字相机拍摄第一图像的方法。与第一图像相关的询问数据经由通信网络被传输到远程识别服务器,在该远程识别服务器处识别匹配参考图像。通过用注解的图像的至少部分代替第一图像的部分,在通信终端处生成和显示增强图像。用相机拍摄的第一图像的增强发生在平面空间中并且仅处理二维图像和对象。 For example, WO2008134901 describes a method in which a first image is captured using a digital camera associated with a communication terminal. Query data related to the first image is transmitted via the communication network to a remote identification server where a matching reference image is identified. An enhanced image is generated and displayed at the communication terminal by replacing at least a portion of the annotated image for a portion of the first image. The augmentation of the first image taken with the camera takes place in planar space and only deals with two-dimensional images and objects.
光线信息,诸如在空间的每个点中的光线的方向,在传统的图像注解系统中被抛弃。没有光线信息的注解使注解的场景的现实视图更困难。比如,捕捉或显示在对象的表面上的纹理要求光线信息。尽管每个对象在它的表面上具有不同的纹理,但是在当前注解系统中添加纹理信息是不可能的。这导致附着的注解实际不被集成在场景中。 Ray information, such as the direction of the ray in each point in space, is discarded in traditional image annotation systems. Annotations without ray information make a realistic view of the annotated scene more difficult. For example, capturing or displaying textures on the surface of an object requires lighting information. Although each object has different textures on its surface, it is impossible to add texture information in current annotation systems. This results in attached annotations not actually being integrated in the scene.
另外,增强现实应用的快速增长可能在未来引起注解的泛滥。比如在城市中的一些场景含有与不同注解相关联的许多元素,从而导致下述注解的图像:该注解的图像带有非常大数目的注解,该注解覆盖背景图像的大部分。在许多情况中,用户仅对有限数目的那些注解感兴趣,并且另一些注解只会使人分心。因此,经常将期望的是限制注解的数目并且提供一种选择应该被显示的注解的方式。 In addition, the rapid growth of augmented reality applications may cause a flood of annotations in the future. Some scenes such as in a city contain many elements associated with different annotations, resulting in an annotated image with a very large number of annotations covering a large part of the background image. In many cases, the user is only interested in a limited number of those annotations, and the others are only distracting. Therefore, it will often be desirable to limit the number of annotations and provide a way of selecting which annotations should be displayed.
此外,对于注解的场景查看而言计算开支是重要问题。将需要减少计算开支。 Furthermore, computational overhead is an important issue for annotated scene viewing. Computational expenses will need to be reduced.
因此本发明的目标是解决或至少减轻现有增强现实系统的以上提及的问题。 It is therefore an object of the present invention to solve or at least alleviate the above mentioned problems of existing augmented reality systems.
发明内容 Contents of the invention
依据发明,这些目标通过包括下述步骤的方法来实现: According to the invention, these objects are achieved by a method comprising the following steps:
用全光捕捉装置检索表示光场的数据; Retrieving data representing the light field with a plenoptic capture device;
执行用于将捕捉的数据与对应的参考数据匹配的程序代码; executing program code for matching captured data to corresponding reference data;
执行用于检索与所述参考数据的元素相关联的在全光格式中的注解的程序代码; executing program code for retrieving annotations in plenoptic format associated with elements of said reference data;
执行用于从所述捕捉的数据和在全光格式中的所述注解生成注解的数据的程序代码。 Program code is executed for generating annotated data from said captured data and said annotations in a plenoptic format.
发明也通过用于捕捉和注解对应于场景的数据的设备来实现,所述设备包括: The invention is also realized by an apparatus for capturing and annotating data corresponding to a scene, said apparatus comprising:
全光捕捉装置,用于捕捉表示光场的数据; a plenoptic capture device for capturing data representing a light field;
处理器; processor;
显示器; monitor;
程序代码,当执行所述程序代码时,用于引起所述处理器检索与用所述全光捕捉装置捕捉的数据的元素相关联的在全光格式中的至少一个注解,并且用于在所述显示器上再现从捕捉的数据生成的且包含所述至少一个注解的视图。 program code for, when executed, causing the processor to retrieve at least one annotation in a plenoptic format associated with an element of data captured with the plenoptic capture device, and for The view generated from the captured data and including the at least one annotation is rendered on the display.
发明也提供用于确定注解的设备,所述设备包括: The invention also provides apparatus for determining annotations, said apparatus comprising:
处理器; processor;
储存器; memory;
程序代码,当执行所述程序代码时,用于引起所述处理器接收表示光场的数据、将所述数据与一个参考数据匹配、从所述储存器确定与所述参考数据相关联的在全光格式中的注解并且将在全光格式中的所述注解或对应于在全光格式中的注解的图像的数据发送到远程装置。 program code for, when executed, causing the processor to receive data representative of a light field, match the data to a reference data, determine from the memory the current the annotation in the plenoptic format and sending the annotation in the plenoptic format or data corresponding to the image of the annotation in the plenoptic format to the remote device.
在全光格式中的注解的要求保护的添加允许在全光格式中在图像中的注解的更现实的集成;注解似乎是捕捉的场景的元素,而不仅是叠加在图像之上的文本。在全光格式中的注解(在本申请中也被称为“全光注解”)含有比传统的注解更完整的光场描述,包含如何修改光线的信息。 The claimed addition of annotations in plenoptic formats allows for a more realistic integration of annotations in images in plenoptic formats; the annotations appear to be elements of the captured scene, not just text superimposed on top of the image. Annotations in the plenoptic format (also called "plenoptic annotations" in this application) contain a more complete description of the light field than conventional annotations, including information on how to modify the light.
取决于在图像的再现期间由用户选择的或比如基于他的兴趣自动选择的焦距和/或视点,在全光格式中的注解提供也允许选择应该被显示的注解。 Annotation provision in plenoptic format also allows selection of annotations that should be displayed, depending on the focal length and/or viewpoint selected by the user during the reproduction of the image or eg automatically based on his interests.
由于注解与捕捉的数据相比较处于相同的空间(即全光空间)中,所以减少针对注解过程的计算开支。 Since the annotations are in the same space (ie plenoptic space) as compared to the captured data, the computational overhead for the annotation process is reduced.
特别地,减少用于再现在人可理解的格式中的全光数据的计算开支。实际上,由于在全光格式中的图像和全光注解处于相同空间中,所以再现过程对两者是等同的。在一个实施例中,单个再现过程能够被用于再现图像和相关联的注解。在这个情形中,针对全光再现过程选择的投影参数(诸如视点的改变、深度、焦点的选择…)也应用在全光注解上。比如,当改变全光图像的焦点或视点时,相同的变换能够被用于在各种距离处显示全光注解。在另一个实施例中,将注解的效果应用到捕捉的全光图像,并且执行修改的全光图像的再现。 In particular, the computational expense for rendering plenoptic data in a human understandable format is reduced. In fact, since the image in the plenoptic format and the plenoptic annotation are in the same space, the rendering process is equivalent for both. In one embodiment, a single rendering process can be used to render images and associated annotations. In this case, the projection parameters selected for the plenoptic reconstruction process (such as change of viewpoint, depth, choice of focus...) are also applied on the plenoptic annotation. For example, the same transformation can be used to display plenoptic annotations at various distances when changing the focus or viewpoint of the plenoptic image. In another embodiment, the effect of the annotation is applied to the captured plenoptic image and a rendering of the modified plenoptic image is performed.
因此,全光注解,即在全光格式中的注解,提供显示注解的现实方式、允许包含带有纹理的注解的更多类型的注解并且提高计算效率。 Thus, plenoptic annotations, ie annotations in a plenoptic format, provide a realistic way of displaying annotations, allow more types of annotations including annotations with textures, and increase computational efficiency.
不像传统的注解,全光注解可以含有与由全光捕捉装置捕捉的图像一样多的关于光线的信息。因而,可能的是在捕捉的光场中直接合成注解,而没有失去由到2D图像上的投影引起的光线信息。比如,注解能够保留在注解的对象的表面上的光反射的特征,这用传统的注解系统是不可能的。在这个意义上,注解的视图似乎更现实。 Unlike traditional annotations, plenoptic annotations can contain as much information about rays as images captured by plenoptic capture devices. Thus, it is possible to synthesize annotations directly in the captured light field without losing the ray information caused by the projection onto the 2D image. For example, annotations can preserve the characteristics of light reflections on the surfaces of the annotated objects, which is not possible with traditional annotation systems. In this sense, the annotated view seems more realistic.
对光线的直接修改能够促进计算,诸如从多个视点同时生成注解的场景。在注解的场景生成的示例中,对场景的注解处理和其它额外处理诸如模糊或锐化以全光格式直接应用一次,而不是对每个视点在生成的2D图像上附着注解并且应用额外处理。因此,在全光格式中直接合成全光图像和全光注解可以导致减少计算开支。 Direct modification of rays can facilitate computations such as the simultaneous generation of annotated scenes from multiple viewpoints. In the example of annotated scene generation, annotation processing and other additional processing such as blurring or sharpening to the scene is applied directly once in plenoptic format, instead of attaching annotations and applying additional processing on the generated 2D image for each viewpoint. Therefore, direct synthesis of plenoptic images and plenoptic annotations in a plenoptic format can lead to reduced computational overhead.
本发明也涉及用于将注解附着到在全光格式中的参考图像的方法,所述方法包括: The invention also relates to a method for attaching annotations to a reference image in a plenoptic format, the method comprising:
用查看器呈现在全光格式中的所述参考图像; presenting said reference image in a plenoptic format with a viewer;
选择注解; select comment;
用所述查看器选择针对所述注解的位置和一个或多个方向,从所述一个或多个方向能够看到所述注解; selecting with the viewer a location for the annotation and one or more directions from which the annotation can be seen;
在存储器中将所述位置和所述方向与在全光格式中的所述参考图像和所述注解相关联。 The position and the orientation are associated in memory with the reference image and the annotation in a plenoptic format.
这个方法可以用合适的创作系统诸如合适的软件应用或网站来执行。 This method can be performed with a suitable authoring system such as a suitable software application or website.
附图说明 Description of drawings
发明在对作为示例给出且由附图图解的实施例的描述的辅助下将被更好地理解,在附图中: The invention will be better understood with the aid of the description of an embodiment given as an example and illustrated by the accompanying drawings, in which:
图1示意性图解用于捕捉表示带有处于第一距离的对象的场景的光场的数据的全光捕捉装置。 Figure 1 schematically illustrates a plenoptic capture device for capturing data representing a light field of a scene with an object at a first distance.
图2示意性图解用于捕捉表示带有处于第二距离的对象的场景的光场的数据的全光捕捉装置。 Figure 2 schematically illustrates a plenoptic capture device for capturing data representing a light field of a scene with an object at a second distance.
图3示意性图解用于捕捉表示带有处于第三距离的对象的场景的光场的数据的全光捕捉装置。 Fig. 3 schematically illustrates a plenoptic capture device for capturing data representing a light field of a scene with an object at a third distance.
图4示意性图解包括一起体现发明的各种设备元件的系统。 Figure 4 schematically illustrates a system comprising various device elements together embodying the invention.
图5A到5B示出从相同的全光数据再现的注解的视图,其中在再现期间由用户选择的视点在两个视图之间已改变,从而导致以不同方式再现的相同注解。 Figures 5A-5B show annotated views rendered from the same plenoptic data, where the viewpoint selected by the user during the rendering has changed between the two views, resulting in the same annotation rendered differently.
图6A到6B示出从相同的全光数据再现的注解的视图,其中在再现期间由用户选择的视点在两个视图之间已改变,从而导致使第一注解在第一视图上成为可见的并且使第二注解在第二视图上成为可见的。 Figures 6A-6B show views of annotations rendered from the same plenoptic data, where during the rendering the viewpoint selected by the user has changed between the two views, resulting in making the first annotation visible on the first view And make the second annotation visible on the second view.
图7A到7B示出从相同的全光数据再现的注解的视图,其中在再现期间由用户选择的焦距在两个视图之间已改变,从而导致使第一注解在第一视图上成为可见的并且使第二注解在第二视图上成为可见的。 Figures 7A-7B show views of annotations rendered from the same plenoptic data, where during the rendering the focal length selected by the user has changed between the two views, resulting in making the first annotation visible on the first view And make the second annotation visible on the second view.
图8是用于生成和再现带有在全光格式中的注解的视图的方法的框图。 8 is a block diagram of a method for generating and rendering views with annotations in a plenoptic format.
图9是用于当查看者在视图上选择不同查看方向和/或不同焦距时修改注解的再现的方法的框图。 9 is a block diagram of a method for modifying the rendering of annotations when a viewer selects a different viewing direction and/or a different focal length on a view.
图10是用于将在全光格式中的注解与参考数据相关联的方法的框图。 10 is a block diagram of a method for associating annotations in a plenoptic format with reference data.
图11是一系列全光图像诸如由用户在移动中捕捉的视频全光图像或全光图像的连续注解的方法的框图。 11 is a block diagram of a method for sequential annotation of a series of plenoptic images, such as video plenoptic images or plenoptic images captured by a user on the move.
具体实施方式 Detailed ways
传统的相机捕捉场景在传感器上的2D投影,并且生成指示在每个具有或没有颜色的像素上的光的强度的数据。另一方面,照此已知的全光捕捉装置捕捉表示光场的数据,即不仅指示光的强度而且指示关于光场的包含光的方向的更完整的信息的矩阵。 Conventional cameras capture a 2D projection of a scene on a sensor and generate data indicating the intensity of light on each pixel with or without color. On the other hand, plenoptic capture devices known as such capture data representing the light field, ie not only the intensity of the light but also a matrix containing more complete information about the light field including the direction of the light.
完整的光场可以包括用于描述每个光线(或用于描述在给定位置处的光线)的多达7个参数:3个针对位置,2个针对方向,1个针对波长以及(在视频的情形中)1个针对时间。一些当前的全光相机递送全光数据,该全光数据包括:针对位置的2个参数、针对方向的2个参数、以及针对波长的1个参数。传感器生成表示所谓的全光光场的全光数据,即至少指示光线的位置和方向的矩阵。它表示由全光捕捉装置生成的全光数据比由传统2D相机生成的传统2D图像数据含有关于光场的更多信息。 A complete light field can include up to 7 parameters to describe each ray (or to describe a ray at a given position): 3 for position, 2 for direction, 1 for wavelength and (in the video case) 1 for time. Some current plenoptic cameras deliver plenoptic data comprising: 2 parameters for position, 2 parameters for direction, and 1 parameter for wavelength. The sensor generates plenoptic data representing a so-called plenoptic light field, ie a matrix indicating at least the position and direction of light rays. It means that plenoptic data generated by a plenoptic capture device contains more information about the light field than conventional 2D image data generated by a conventional 2D camera.
至今,至少两个公司Lytro和Raytrix提出能够记录这样的全光光场的全光传感器。它们的两款相机在设计方面稍微不同,但是主要思路是分解光的不同方向,该光被认为落在标准相机传感器中的单个光位(photosite)(或像素)上。为了那个目标,如在图1上图解的,微透镜20的阵列被放置在主透镜1后面,代替传统相机的传感器。 So far, at least two companies, Lytro and Raytrix, have proposed plenoptic sensors capable of recording such plenoptic light fields. Their two cameras are slightly different in design, but the main idea is to break down the different directions of light that is thought to fall on a single photosite (or pixel) in a standard camera sensor. For that purpose, as illustrated on Fig. 1, an array of microlenses 20 is placed behind the main lens 1, replacing the sensor of a conventional camera.
那样,微透镜21依据光线的入射角度使光线改变方向,并且改变方向的光线到达传感器21的不同像素210。由形成子图像的N×M像素210中的每个测量的光量取决于射中在那个子图像前面的微透镜20的光束的方向。 In that way, the microlens 21 redirects the light according to the incident angle of the light, and the redirected light reaches different pixels 210 of the sensor 21 . The amount of light measured by each of the NxM pixels 210 forming a sub-image depends on the direction of the light beam hitting the microlens 20 in front of that sub-image.
图1到3图解简单的一维传感器,该简单的一维传感器包括n=9个子图像,每个子图像具有N×M像素(或光位)210的一行,在这个示例中N等于3并且M等于1。许多全光传感器具有更高数目的子图像和用于每个子图像的更高数目的像素,比如9×9像素,从而允许在微透镜20上的光的N×M=81个不同定向之间进行区分。假设场景的所有对象处在焦点上,每个子图像因而包含指示来自各种方向到那个子图像上的光量的亮度值的补片。 1 to 3 illustrate a simple one-dimensional sensor comprising n=9 sub-images each having a row of N×M pixels (or light bits) 210, in this example N equals 3 and M is equal to 1. Many plenoptic sensors have a higher number of sub-images and a higher number of pixels for each sub-image, say 9×9 pixels, allowing between N×M=81 different orientations of light on the microlens 20 Make a distinction. Assuming all objects of the scene are in focus, each sub-image thus contains patches of luminance values indicating the amount of light onto that sub-image from various directions.
在这个构建中,微透镜20的阵列位于由全光捕捉装置的主透镜1形成的图像平面上并且传感器21位于离微透镜距离f处,其中f是微透镜的焦距。这个设计允许高的角度分辨率但是遭受相对差的空间分辨率(每再现的图像的像素的有效数目等于微透镜的数目)。这个问题由其它全光捕捉装置来处理,在该其它全光捕捉装置中微透镜聚焦在主透镜的图像平面上,因而在微透镜和图像平面之间创建间隙。在这样的设计中要付出的代价是较差的角度分辨率。 In this construction, the array of microlenses 20 is located on the image plane formed by the main lens 1 of the plenoptic capture device and the sensor 21 is located at a distance f from the microlenses, where f is the focal length of the microlenses. This design allows high angular resolution but suffers from relatively poor spatial resolution (the effective number of pixels per rendered image is equal to the number of microlenses). This problem is addressed by other plenoptic capture devices where the microlenses are focused on the image plane of the main lens, thus creating a gap between the microlenses and the image plane. The price to pay in such a design is poor angular resolution.
如能够在图1到3上观察到,在这个示例中对应于带有单个点3的场景的全光光场取决于从点3到主透镜1的距离。在图1上,来自这个对象的所有光束到达相同的微透镜20,因而导致全光光场,其中对应于这个微透镜的子图像中的所有像素记录第一正光强度,而对应于其它透镜的所有其它像素记录不同的、空光强度。在其中对象3更靠近透镜1的图2上,源于点3的一些光束到达其它子图像(即与邻近之前射中的微透镜的两个微透镜相关联的子图像)的像素。在其中对象3处于离透镜1更大距离的图3上,源于点3的一些光束到达与邻近之前射中的微透镜的两个微透镜相关联的不同像素。因此,由传感器21递送的数字数据22取决于到对象3的距离。 As can be observed on FIGS. 1 to 3 , the plenoptic light field corresponding to a scene with a single point 3 in this example depends on the distance from the point 3 to the main lens 1 . In Fig. 1, all beams from this object reach the same microlens 20, thus resulting in a plenoptic light field in which all pixels in the sub-image corresponding to this microlens record a first positive light intensity, while those corresponding to other lenses All other pixels record a different, null light intensity. On Fig. 2 where object 3 is closer to lens 1, some beams originating from point 3 reach pixels of other sub-images (ie sub-images associated with the two microlenses adjacent to the previously hit microlens). On FIG. 3 , where object 3 is at a greater distance from lens 1 , some beams originating from point 3 reach different pixels associated with the two microlenses adjacent to the previously hit microlens. Thus, the digital data 22 delivered by the sensor 21 depends on the distance to the object 3 .
全光传感器21因而递送全光数据22,该全光数据22针对对应于微透镜20的每个子图像含有指示在这个子图像以上的透镜上的来自各种方向的光量的(N×M)值的集合。针对给定的聚焦的对象点,子图像的每个像素对应于以特定入射角度φ(在页面的平面中)和θ(与页面的平面垂直)射中传感器的光线的强度测量。 The plenoptic sensor 21 thus delivers plenoptic data 22 containing for each sub-image corresponding to the microlens 20 (N×M) values indicating the amount of light from various directions on the lens above this sub-image collection. For a given focused object point, each pixel of the sub-image corresponds to an intensity measurement of a ray hitting the sensor at a particular angle of incidence φ (in the plane of the page) and θ (perpendicular to the plane of the page).
图4示意性图解体现发明的注解系统的框图。系统包括用户装置4,诸如手持装置、智能手机、平板电脑、相机、眼镜、护目镜、隐形眼镜等。装置4包含:全光捕捉装置41,诸如在图1到3中图解的相机,用于捕捉表示在场景3上的光场的数据;处理器,诸如带有合适的程序代码的微处理器400;和通信模块401,诸如WIFI和/或蜂窝式接口,用于通过诸如互联网6的网络将装置4连接到远程服务器5比如云服务器。服务器5包含:储存器50,带有诸如SQL数据库的数据库、XML文档的集合、在全光格式中的图像的集合等,用于储存表示图像和/或一个或多个全局模型的参考全光数据的收集;和处理器51,包含微处理器,该微处理器带有用于引起微处理器执行在注解方法中需要的操作的计算机代码。注解和对应的位置连同参考全光数据一起也能够被储存在储存器50中。 Figure 4 schematically illustrates a block diagram of an annotation system embodying the invention. The system includes user devices 4 such as handheld devices, smartphones, tablets, cameras, glasses, goggles, contact lenses and the like. The device 4 comprises: plenoptic capture means 41, such as the camera illustrated in Figures 1 to 3, for capturing data representing the light field on the scene 3; a processor, such as a microprocessor 400 with suitable program code ; and a communication module 401 , such as WIFI and/or a cellular interface, for connecting the device 4 to a remote server 5 such as a cloud server through a network such as the Internet 6 . The server 5 comprises: a memory 50 with a database such as a SQL database, a collection of XML documents, a collection of images in a plenoptic format, etc., for storing reference plenograms representing images and/or one or more global models collection of data; and a processor 51 comprising a microprocessor with computer code for causing the microprocessor to perform the operations required in the annotation method. Annotations and corresponding positions can also be stored in memory 50 together with reference plenoptic data.
由用户装置4执行的程序代码能够包含比如能够被用户下载和安装在用户装置4中的应用软件或应用程序(app)。程序代码也能够包含用户装置4的操作代码的部分。程序代码也能够包含在网页中嵌入的或在浏览器中执行的代码,该代码包含比如Java、Javascript、HTML5代码等。程序代码可以被储存为在有形设备可读介质(诸如闪存、硬盘、或永久或半永久存储器的任何类型)中的计算机程序产品。 The program code executed by the user device 4 can include, for example, application software or application programs (apps) that can be downloaded and installed in the user device 4 by the user. The program code can also contain part of the operating code of the user device 4 . Program codes can also include codes embedded in webpages or executed in browsers, such codes include Java, Javascript, HTML5 codes, etc. The program code may be stored as a computer program product in a tangible device readable medium such as flash memory, hard disk, or any type of permanent or semi-permanent memory.
由用户装置4中的微处理器400执行程序代码用于引起这个微处理器将对应于光场的捕捉的数据集合中的至少一些或那些数据集合的特征发送到远程服务器5。程序代码被布置为发送在“全光格式”中的数据,即没有失去关于光线的方向的信息。程序代码也能够引起微处理器400从服务器5接收在全光格式中的注解的数据、或注解的图像、或与之前发送的全光数据相关的注解,并且用于再现对应于带有注解的捕捉的数据的视图。 Execution of program code by the microprocessor 400 in the user device 4 is used to cause this microprocessor to send at least some of the captured data sets corresponding to the light field, or characteristics of those data sets, to the remote server 5 . The program code is arranged to send the data in a "plenoptic format", ie without losing information about the direction of the light rays. The program code can also cause the microprocessor 400 to receive from the server 5 annotated data in plenoptic format, or annotated images, or annotations associated with previously transmitted plenoptic data, and to render images corresponding to the annotated images. A view of the captured data.
全光注解方法可以包括两个部分:离线过程和在线过程。通常,离线过程的主要目的是将注解与在全光格式中的参考图像、或与其它2D、立体的、或3D参考图像相关联。 Plenoptic annotation methods can include two parts: an offline process and an online process. Typically, the main purpose of the offline process is to associate annotations with reference images in plenoptic format, or with other 2D, stereoscopic, or 3D reference images.
离线阶段offline stage
在全光格式中的参考图像的情形中,离线过程可以包括比如下面的步骤: In the case of reference images in plenoptic format, the offline process may include steps such as the following:
1.从装置4接收在全光格式中并且表示光场的参考数据; 1. receiving from the device 4 reference data in plenoptic format and representing the light field;
2.比如用全光查看器呈现全光参考图像的再现的视图; 2. Presenting a reconstructed view of the plenoptic reference image, such as with a plenoptic viewer;
3.选择全光注解, 3. Select Plenoptic Annotation,
4.选择用于在再现的视图中的注解的位置和定向, 4. Select the position and orientation for the annotation in the rendered view,
5.选择注解的一个或多个光场参数, 5. Select one or more light field parameters of the annotation,
6.(可选地)将动作归结于注解, 6. (Optionally) attribute actions to annotations,
7.基于它的位置和定向,在存储器中将参考图像光线与注解光线相关联。 7. Based on its position and orientation, associate the reference image ray with the annotation ray in memory.
这个离线过程能够在服务器5上、在用户装置4中,或在又一个设备诸如个人计算机、平板电脑等中执行。典型地,这个离线过程对与参考图像相关联的每个注解仅被执行一次。如果选择的注解在全光格式中不是初始可获得的,则它可以被转换到全光格式中。 This offline process can be performed on the server 5, in the user device 4, or in yet another device such as a personal computer, tablet or the like. Typically, this offline process is performed only once for each annotation associated with the reference image. If the selected annotation is not originally available in the plenoptic format, it can be converted into the plenoptic format.
离线过程的主要目的是将全光注解添加到全光图像。离线过程可以包括两个阶段。第一个阶段可以通过由在服务器5中的微处理器执行的程序代码来执行,该微处理器可以包含用于引起服务器5执行下面的任务中的至少一些的可执行程序或其它代码: The main purpose of the offline process is to add plenoptic annotations to plenoptic images. The offline process can include two phases. The first stage may be performed by program code executed by a microprocessor in the server 5, which may contain an executable or other code for causing the server 5 to perform at least some of the following tasks:
1.从装置4接收在全光格式中并且表示光场的数据; 1. receiving data from the device 4 in a plenoptic format and representing a light field;
2.从数据库50检索之前储存的模型(参考图像)和/或多个参考数据; 2. Retrieving a previously stored model (reference image) and/or a plurality of reference data from the database 50;
3.将从用户装置接收的数据与参考图像的一个部分、分别与多个参考图像之中的一个匹配, 3. matching data received from the user device to a portion of the reference image, respectively to one of the plurality of reference images,
4.确定与匹配参考图像相关联的注解; 4. Determining annotations associated with matching reference images;
5.将在全光格式中的注解或在全光格式中的注解的图像发送到装置4。 5. Send the annotation in plenoptic format or the image of the annotation in plenoptic format to the device 4 .
在各种实施例中,代替将捕捉的数据发送到远程服务器5用于与在服务器中的参考图像匹配,这个匹配能够用本地储存的参考图像的集合或用在装置中本地储存的模型在用户的装置中本地完成。在这个实施例中,服务器5被装载在用户装置4上。在线过程能够依据用户的要求被执行若干次。 In various embodiments, instead of sending the captured data to the remote server 5 for matching with a reference image in the server, this matching can be performed at the user using a set of reference images stored locally or using a model stored locally in the device. done locally on the device. In this embodiment, the server 5 is hosted on the user device 4 . The online process can be performed several times according to the user's requirements.
在线过程的第二阶段可以通过由在装置4中的微处理器执行的程序代码来执行,该微处理器可以包含用于引起装置4执行下面的任务中的至少一些的可执行程序或其它代码: The second stage of the on-line process may be performed by program code executed by a microprocessor in the device 4, which may contain an executable or other code for causing the device 4 to perform at least some of the following tasks :
1.可能地连同相关联的动作一起从服务器5接收在全光格式中的注解数据; 1. Receiving annotation data in plenoptic format from the server 5, possibly together with associated actions;
2.将接收的注解数据应用到捕捉的全光光场; 2. Apply the received annotation data to the captured plenoptic light field;
3.将注解的光场再现到用户可查看的视图; 3. Reproduce the annotated light field into a user-viewable view;
4.解释用户交互并且执行相关联的注解动作。 4. Interpret user interactions and perform associated annotation actions.
在各种实施例中,代替在装置4上将接收的注解应用到捕捉的全光光场,这个步骤能够在服务器5侧上被完成。在这个情形中,最终再现的视图被传输回到装置4或整个注解的光场被传输回到装置4。 In various embodiments, instead of applying the received annotations to the captured plenoptic light field on the device 4, this step can be done on the server 5 side. In this case the final rendered view is transmitted back to device 4 or the entire annotated light field is transmitted back to device 4 .
因此,用户能够将注解与关于全光参考图像的再现的视图的特定位置和定向相关联,并且指示在这个特定视图中注解应该使用的一个或多个光场参数。在再现视图期间取决于由查看者选择的视点,可以不同地再现相同的注解。因为注解的光场参数可以改变,如果查看者选择不同的视点,则第一注解可以被处于相同位置的第二注解代替。 Thus, the user is able to associate the annotation with a particular position and orientation with respect to the rendered view of the plenoptic reference image, and indicate which light field parameter or parameters the annotation should use in this particular view. Depending on the viewpoint selected by the viewer during rendering of the view, the same annotation may be rendered differently. Because the light field parameters of annotations can change, if the viewer chooses a different viewpoint, the first annotation can be replaced by a second annotation at the same position.
用于离线过程的流程图的示例在图10上被图解。这个流程图图解下述方法:该方法允许用户选择必须与参考图像相关联的注解、以及关于这个注解的位置、定向和光场参数,从而这个注解将被应用到匹配这个全光参考图像的捕捉的全光图像。 An example of a flowchart for an offline process is illustrated on FIG. 10 . This flowchart illustrates the method that allows the user to select the annotation that must be associated with the reference image, as well as the position, orientation and light field parameters for this annotation, so that this annotation will be applied to a captured image that matches this plenoptic reference image. plenoptic image.
这个方法可以使用可以在用户的装置4中本地运行的注解创作系统。注解创作系统也可以被主控在服务器5上,在该服务器5处web平台呈现一些工具以管理注解并且使它们与全光参考图像有关。诸如增强现实使用统计的服务也可以从web平台是可获得的。注解创作系统也可以运行在不同的服务器或设备(包含用户的个人计算机、平板电脑等)中。 This method may use an annotation authoring system that may run locally in the user's device 4 . The annotation authoring system can also be hosted on a server 5 where the web platform presents some tools to manage annotations and relate them to plenoptic reference images. Services such as augmented reality usage statistics may also be available from the web platform. The annotation authoring system can also run on different servers or devices (including users' personal computers, tablets, etc.).
在步骤150中,用户选择参考图像,诸如在全光格式中的图像。该图像被上载在全光创作系统上并且用作用于注解的支持图像。 In step 150, the user selects a reference image, such as an image in a plenoptic format. This image was uploaded on the plenoptic authoring system and used as a supporting image for annotation.
作为全光创作系统的部分,查看器以使得用户能够可视化上载的数据的方式将上载的数据再现给用户。如果数据处在全光格式中,该全光格式不能够照此被人容易地理解,则这可能包含使用全光再现模块来在被用户可理解的空间中再现全光模型。查看器构成下述工具:该工具操纵全光数据并且关于给定视图将注解放置在期望的位置和定向处,但是与全光注解的所有处理和组合直接在全光空间中完成。 As part of the plenoptic authoring system, the viewer renders the uploaded data to the user in a manner that enables the user to visualize the uploaded data. If the data is in a plenoptic format that cannot be easily interpreted by a human as such, this may involve using a plenoptic rendering module to render the plenoptic model in a space intelligible to the user. The viewer constitutes the tool that manipulates the plenoptic data and places annotations at desired positions and orientations for a given view, but all processing and combination with plenoptic annotations is done directly in plenoptic space.
在一个实施例中,全光模型能够被再现为2D视图,从而用户能够一次从一个视点并且一次以一个焦距可视化该2D视图,从而允许他理解和编辑全光模型。为了从一个2D视图导航到其它,控制是可获得的,使得在请求时能够显示另一个2D视图。 In one embodiment, the plenoptic model can be rendered as a 2D view so that the user can visualize the 2D view from one viewpoint at a time and at one focal length at a time, allowing him to understand and edit the plenoptic model. For navigating from one 2D view to the other, controls are available such that another 2D view can be displayed on request.
在另一个实施例中,全光模型可以被再现为部分3D场景,其中能够可视化光线的不同方向。与标准的完整3D场景的主要不同是:当从全光模型再现3D场景探测时,限制该3D场景探测。比如,视图方向以及视图位置被限制到由全光捕捉装置已捕捉的方面。 In another embodiment, the plenoptic model can be rendered as a partial 3D scene, where different directions of light rays can be visualized. The main difference from a standard full 3D scene is that the 3D scene detection is limited when it is reproduced from the plenoptic model. For example, the view direction as well as the view position are restricted to those already captured by the plenoptic capture device.
在步骤151中,用户选择他想要将其与全光模型的特定元素或位置相关联的全光注解。如已经提及的,全光注解被限定在全光空间中并且因而以光线来描述。那些光线能够描述比如文本、图像、视频、或直接作用在全光图像光线上的其它元素。全光注解比如可以从数据库中或文件浏览器中的全光注解的库来检索。全光注解也能够比如通过下述方式飞快地创建:用全光捕捉装置捕捉它、用文本编辑器输入文本、绘制图像和/或记录声音或视频。 In step 151 the user selects the plenoptic annotation that he wants to associate with a specific element or position of the plenoptic model. As already mentioned, plenoptic annotations are defined in plenoptic space and are thus described in terms of rays. Those rays can describe elements such as text, images, video, or other elements acting directly on the plenoptic image rays. Plenoptic annotations can be retrieved, for example, from a library of plenoptic annotations in a database or in a file browser. Plenoptic annotations can also be created on the fly, for example by capturing it with a plenoptic capture device, entering text with a text editor, drawing images and/or recording sound or video.
在一个实施例中,全光注解能够在创作系统上以库或列表被呈现为预览。全光注解预览对应于用于默认视图的注解的再现。这个默认视图能够被随机或在优选实施例中取作对应于关于位置和方向的全光注解范围的中间视图。预览允许用户获得关于全光注解对应于什么的快速和清晰的概念。对于不作用在模型波长上的注解的通用类型,即这些注解照此不是可视化的,预览图解被应用到由创作系统再现的当前模型视图的中央的注解。因此,如果注解的这个类型仅具有使所有的模型光线旋转10°的效果,则预览将由当前模型再现的视图的中央部分组成,其中每个光线已被旋转10°。 In one embodiment, plenoptic annotations can be presented as previews on the authoring system in a library or list. The plenoptic annotation preview corresponds to the rendition of the annotation used for the default view. This default view can be randomized or in a preferred embodiment taken as an intermediate view corresponding to the plenoptic annotation extent with respect to position and orientation. The preview allows the user to get a quick and clear idea of what the plenoptic annotations correspond to. For the general type of annotations that do not act on the model wavelength, ie these annotations are not visualized as such, the preview plot is applied to the central annotation of the current model view rendered by the authoring system. Therefore, if this type of annotation has only the effect of rotating all model rays by 10°, the preview will consist of the central part of the view rendered by the current model, where each ray has been rotated by 10°.
在步骤152中,用户用全光注解创作系统选择在选择的参考模型的再现的视图的坐标系统中的位置,在该位置处他想要添加全光注解。这能够比如通过下述方式来完成:在期望的位置处从显示的视图的顶上的注解预览列表拖动注解,并且可能地平移、旋转、调整大小、裁剪和/或另外编辑注解。替选地,用户也可以输入坐标作为在控制平面中的值。 In step 152, the user selects with the plenoptic annotation authoring system a location in the coordinate system of the rendered view of the selected reference model at which he wants to add a plenoptic annotation. This can be done, for example, by dragging the annotation at the desired location from the annotation preview list on top of the displayed view, and possibly translating, rotating, resizing, cropping and/or otherwise editing the annotation. Alternatively, the user can also enter coordinates as values in the control plane.
在步骤152'中,用户能够调节注解光线的参数以生成注解的另一个视图。当用户使用比如用于改变注解的定向的计算机鼠标指针来改变注解的参数时,注解的光线与全光模型的光线组合,并且针对每个新的位置或新的定向在查看器中生成新的2D视图。当用户鼠标指针和它的移动被投影到全光空间时使这成为可能。指针的移动随后在与对应于2D再现的视图的虚拟平面平行的平面中被应用到注解。 In step 152', the user can adjust the parameters of the annotation ray to generate another view of the annotation. When the user changes the parameters of the annotation using, for example, a computer mouse pointer to change the orientation of the annotation, the rays of the annotation are combined with the rays of the plenoptic model, and for each new position or new orientation a new 2D view. This is made possible when the user mouse pointer and its movement are projected into plenoptic space. The movement of the pointer is then applied to the annotation in a plane parallel to the virtual plane corresponding to the 2D rendered view.
一旦组合全光模型和注解的光线,注解的效果被应用到参考图像的光线。叠加全光注解的过程能够被看作修改光线的过程。捕捉的全光数据能够含有关于光线的方向、针对每个光线的波长(即颜色)的信息,因而注解能够被视为那些参数的修改。比如,在对象的表面上附着文本能够被看作对在表面上的特定区域处的光线的波长的修改。 Once the plenoptic model and the annotated rays are combined, the effects of the annotations are applied to the rays of the reference image. The process of overlaying plenoptic annotations can be viewed as the process of modifying the light. The captured plenoptic data can contain information about the direction of the rays, the wavelength (ie color) for each ray, thus annotations can be considered as modifications of those parameters. For example, attaching text on the surface of an object can be seen as a modification of the wavelength of light at specific regions on the surface.
由注解产生的效果的类型由注解本身确定。在一个实施例中,全光注解比如仅由不透明的文本组成。在这个情形中,模型光线波长完全由针对映射的光线的注解光线波长代替。对于其它注解,通过考虑到改变模型的纹理的注解,模型的光线可以使它们的方向被注解改变以反映新的纹理。在又一个示例中,模型光线位置可以被注解改变。 The type of effect produced by an annotation is determined by the annotation itself. In one embodiment, plenoptic annotations consist, for example, of opaque text only. In this case, the model ray wavelengths are completely replaced by the annotated ray wavelengths for the mapped rays. For other annotations, by taking into account the annotation changing the texture of the model, the model's rays can have their directions changed by the annotation to reflect the new texture. In yet another example, model ray positions may be altered by annotations.
全光注解能够被看作修改光线的过滤器。这提供显示注解的场景的更多可能性。这个处理的一个进一步示例是变更光线的方向。作为实施例,通过将随机性添加到光线的方向中能够将发光效果应用到从捕捉的全光图像中的特定对象入射的光线。能够使注解的对象成为反射性的。另一个示例是表面的属性的修改,诸如纹理信息的修改。由于全光注解允许修改光线的变量,诸如方向和波长,所以可能的是通过组合变量的修改来修改对象的表面好像纹理被添加在其上。比如,全光注解通过修改方向和波长能够实现将带有红颜色的平表面改变到带有黄颜色的粗糙表面。 Plenoptic annotations can be viewed as filters that modify light. This provides more possibilities for displaying annotated scenes. A further example of this process is changing the direction of a ray of light. As an example, a glow effect can be applied to rays incident from a specific object in a captured plenoptic image by adding randomness to the direction of the rays. Ability to make annotated objects reflective. Another example is modification of properties of a surface, such as modification of texture information. Since plenoptic annotations allow modification of variables of light rays, such as direction and wavelength, it is possible to modify the surface of an object as if textures were added to it by combining modifications of variables. For example, plenoptic annotations can change a flat surface with a red color to a rough surface with a yellow color by modifying the direction and wavelength.
描述注解对模型光线的效果的信息可以被储存在全光注解阵列中,如将在步骤154中描述的。 Information describing the effects of annotations on model rays may be stored in a plenoptic annotation array, as will be described in step 154 .
在步骤153中,用户选择一个或多个注解光场参数。这比如能够是注解的波长以改变它的颜色。用户也可以限定针对从不同方向查看的相同注解的不同外观,或甚至与从不同方向查看的相同元素相关联的不同注解。 In step 153, the user selects one or more annotated light field parameters. This could for example be the wavelength of an annotation to change its color. The user may also define different appearances for the same annotation viewed from different directions, or even different annotations associated with the same element viewed from different directions.
替选地,一旦在再现的全光模型上成功地被调节,用户能够选择导航到全光查看器的另一个视图。全光注解在全光模型的新的视图上自动地被报告。用户随后能够决定编辑注解、改变它的光场参数或针对这个特定视图的外观。他能够针对全光模型的所有可获得的视图以相同方式继续进行。 Alternatively, the user can choose to navigate to another view of the plenoptic viewer once successfully adjusted on the rendered plenoptic model. Plenoptic annotations are automatically reported on new views of the plenoptic model. The user can then decide to edit the annotation, changing its light field parameters or appearance for this particular view. He can proceed in the same way for all available views of the plenoptic model.
内插过程可以在全光注解的第一和第二视图之间发生以防止用户免于必须经过全光模型的所有视图来导航。全光注解的这两个视图不必须是连续的。用户必须在两个视图中指定注解的外观并且全光创作系统将自动生成全光注解的居间视图。尚未与注解相关联的全光模型的其它视图将不显示它,从而导致针对场景的特定视点或聚焦平面不再现注解的可能性。 An interpolation process may occur between the first and second views of the plenoptic annotation to prevent the user from having to navigate through all views of the plenoptic model. These two views of the plenoptic annotation do not have to be contiguous. The user must specify the appearance of the annotation in both views and the plenoptic authoring system will automatically generate an intermediate view of the plenoptic annotation. Other views of the plenoptic model that have not been associated with the annotation will not display it, leading to the possibility that the annotation is not reproduced for a particular viewpoint or focal plane of the scene.
全光注解可以包括对应于光线并且以参数的集合来描述的数据。当再现针对第一特定视图的全光注解时,查看器设定一些参数并且允许用户修改其它参数。从这个视图导航到第二个视图,用户改变必须由查看器固定的参数同时能够修改其它参数。内插过程自动计算在这两个视图之间的全光注解的光线参数。 Plenoptic annotations may include data corresponding to rays and described in a set of parameters. When rendering plenoptic annotations for a first specific view, the viewer sets some parameters and allows the user to modify other parameters. Navigating from this view to a second view, the user changes parameters that must be fixed by the viewer while being able to modify other parameters. The interpolation process automatically calculates the ray parameters of the plenoptic annotation between these two views.
在一个实施例中,每个全光注解的参数可以如下:针对空间中的光线位置的3个(或可能2个)参数、针对它们的方向的2个参数、针对它们的波长的1个参数和针对时间的可能1个参数。对于由全光查看器再现的特定视图,位置、方向和时间的参数可以比如由查看器设定。用户能够随后改变未被查看器固定的参数,在这个示例中对应于光线的波长。让我们假定用户将它设定到第一值v1。现在对于注解的另一个视图,即对于位置、方向和时间参数的不同值,让我们假定用户改变针对第二视图的波长值并且将它设定到比如v2。内插过程目的在于为在与第一和第二视图相关联的位置、方向和时间参数中间的视图计算在v1和v2之间的注解值。在其它实施例中,内插也可以考虑为全光数据的包含位置、方向、波长和/或时间的其它参数也计算值。 In one embodiment, the parameters of each plenoptic annotation can be as follows: 3 (or possibly 2) parameters for the position of rays in space, 2 parameters for their direction, 1 parameter for their wavelength and possibly 1 parameter for time. For a particular view rendered by a plenoptic viewer, the parameters of position, orientation and time can be set eg by the viewer. The user can then change a parameter that is not fixed by the viewer, which in this example corresponds to the wavelength of the light. Let us assume that the user sets it to the first value v1. Now for another view of the annotation, ie for different values of the position, direction and time parameters, let us assume that the user changes the wavelength value for the second view and sets it to say v2. The interpolation process aims at computing annotation values between v1 and v2 for views intermediate the position, orientation and time parameters associated with the first and second views. In other embodiments, interpolation can also be considered to calculate values for other parameters of the plenoptic data including position, direction, wavelength and/or time.
内插的具体示例比如包含:全光注解的颜色的改变,比如从橙颜色转到更微红的颜色;注解的可见度的改变,其中对于特定视图,注解是可见的,而对于另一个视图,注解是隐藏的。 Specific examples of interpolation include, for example: a change in the color of a plenoptic annotation, such as from an orange color to a more reddish color; a change in the visibility of an annotation, where for a particular view an annotation is visible, but for another view, Annotations are hidden.
内插的不同方法是可能的,该方法包含比如注解的两个视图之间的线性、二次或更高阶的内插。而且,更高级的内插方法能够考虑场景或注解本身的其它特征以生成注解的新光线。 Different methods of interpolation are possible, including eg linear, quadratic or higher order interpolation between two views of annotations. Also, more advanced interpolation methods can take into account other characteristics of the scene or the annotation itself to generate new rays for the annotation.
在步骤153'中,当注解被显示在捕捉的图像上时,动作也能够被关联到注解中的所有或一些。这些动作能够被用户触发或使用比如定时器自动地被执行。动作包含:启动带有特定URL的web浏览器;使注解活跃,诸如使一个注解移动、出现或消失;播放视频;启动呈现进一步可能动作的菜单;启动幻灯片放映或播放音频文件。允许修改呈现给用户的全光数据的视图的动作也是可能的,比如允许将全光数据的视图聚焦在给定焦距处的动作。 Actions can also be associated to all or some of the annotations when they are displayed on the captured image in step 153'. These actions can be triggered by the user or performed automatically using, for example, a timer. Actions include: launching a web browser with a specific URL; making annotations active, such as making one move, appear or disappear; playing a video; launching a menu presenting further possible actions; starting a slideshow or playing an audio file. Actions that allow modification of the view of plenoptic data presented to the user are also possible, such as actions allowing the view of plenoptic data to be focused at a given focal length.
在步骤154中,全光注解被储存并且被关联在带有对应的位置、定向且带有选择的参考全光模型的存储器中,比如在数据库51中或在用户的装置中。已知需要的注解,以全光格式储存附着到每个参考全光模型的注解是可能的。每个注解被储存为分开的全光文件。 In step 154 the plenoptic annotation is stored and associated in a memory with the corresponding position, orientation and with the selected reference plenoptic model, such as in the database 51 or in the user's device. Knowing the required annotations, it is possible to store annotations attached to each reference plenoptic model in plenoptic format. Each annotation is stored as a separate plenoptic file.
全光注解的参考数据从全光参考数据和对应的一个或多个全光注解生成。这个增强现实模型采取文件的形式,该文件含有为再现回到带有其相关联的注解的全光模型所要求的所有信息。因此它描述在全光参考数据和它的注解之间的关系。全光注解的参考数据能够直接地被再现在全光注解创作系统上以预先可视化结果,而且直接地被再现在客户端侧上以再现某一全光增强现实。 The plenoptic annotation reference data is generated from the plenoptic reference data and the corresponding one or more plenoptic annotations. This augmented reality model takes the form of a file containing all the information required to render back to the plenoptic model with its associated annotations. It thus describes the relationship between plenoptic reference data and its annotations. The plenoptic annotation's reference data can be rendered directly on the plenoptic annotation authoring system to pre-visualize the results, and directly on the client side to render a certain plenoptic augmented reality.
描述注解对模型光线的效果的信息被储存在全光注解数据中。由注解限定的修改作用在模型光线参数上。因此,注解能够描述比如模型光线方向、位置、时间或波长的修改。换句话说,这个信息描述模型光线的功能。 Information describing the effects of annotations on model lighting is stored in the plenoptic annotation data. Modifications defined by annotations act on model light parameters. Thus, annotations can describe modifications such as model ray direction, position, time or wavelength. In other words, this information describes the functionality of the model ray.
在注解创建时,注解的每个光线被分配唯一的识别符。当在创作系统上应用注解时,注解光线唯一识别符被匹配到它们的模型的对应光线。所以,模型的每个光线被分配注解光线识别符,当系统必须逐光线地在模型上应用注解时,该注解光线识别符随后被系统使用,如比如在在线阶段中情形大体上如此。 Each ray of the annotation is assigned a unique identifier when the annotation is created. When applying annotations on the authoring system, annotation ray unique identifiers are matched to their model's corresponding rays. So, each ray of the model is assigned an annotation ray identifier, which is then used by the system when it has to apply annotations on the model ray by ray, as is generally the case eg in the online phase.
注解信息能够被储存在2维阵列中,其中每个光线含有关于其对针对每个参数的模型的效果的信息。注解光线的唯一识别符随后被用来在针对每个参数的阵列中限定对应的光线效果。换句话说,阵列的第一维度对应于光线,该光线由它们的识别符指代,并且第二维度对应于它们的参数,即光场参数。任何注解能够使用这个格式完全被表示,因为针对任何参数的模型光线的任何修改能够被表示在阵列中。 Annotation information can be stored in a 2-dimensional array, where each ray contains information about its effect on the model for each parameter. The unique identifier of the annotation ray is then used to define the corresponding ray effect in the array for each parameter. In other words, the first dimension of the array corresponds to the rays, which are referred to by their identifiers, and the second dimension corresponds to their parameters, ie the light field parameters. Any annotation can be fully represented using this format, since any modification of a model ray for any parameter can be represented in an array.
在一个实施例中,注解能够比如针对一个角度以10°来修改所有模型光线方向。如在下文表1中图解的,该2维阵列随后在对应于方向角度的参数的列中含有10°。该列对所有光线读取10°,因为假定它们全部以相同方式起作用。当期望应用注解对它的对应的模型光线的效果时,系统将首先识别注解和模型光线对、提取对应于注解光线的唯一识别符、查询注解表以查看这个注解光线具有什么效果以便最终将这个改变应用到模型光线。在这个示例中,由注解影响的所有模型光线的角度将以10°被旋转。 In one embodiment, annotations can modify all model ray directions by say 10° for an angle. As illustrated in Table 1 below, the 2-dimensional array then contains 10° in the column corresponding to the parameter of the orientation angle. The column reads 10° for all rays, since they are assumed to all act the same way. When it is desired to apply the effect of an annotation on its corresponding model ray, the system will first identify the annotation and model ray pair, extract the unique identifier corresponding to the annotation ray, query the annotation table to see what effect this annotation ray has in order to finally apply the Changes are applied to the model light. In this example, the angles of all model rays affected by the annotation will be rotated by 10°.
表1.注解阵列。 Table 1. Annotation array.
作为离线阶段的示例,用户可能想要将文本注解添加到含有建筑物的场景。另外,文本注解颜色将需要从一个视点到另一个视点而变化。下面的步骤随后将被用户完成: As an example of the offline phase, a user might want to add text annotations to a scene containing buildings. Also, the text annotation color will need to vary from one viewpoint to another. The following steps will then be completed by the user:
1.建筑物的全光捕捉被上载到全光注解创作系统 1. The plenoptic capture of the building is uploaded to the plenoptic annotation creation system
2.从捕捉的全光图像再现2D视图并且将该2D视图呈现给用户 2. Rendering a 2D view from the captured plenoptic image and presenting the 2D view to the user
3.用户从注解类型列表选择文本注解类型、输入他的文本并且将文本注解拖动到再现的2D视图上 3. The user selects a text annotation type from the annotation type list, enters his text and drags the text annotation onto the rendered 2D view
4.用户能够移动再现的2D视图的视点或注解位置和定向,从而注解正好如用户想要的那样出现 4. The user is able to move the viewpoint or annotation position and orientation of the rendered 2D view so that the annotation appears exactly as the user wants
5.用户为当前再现的视点设定文本颜色 5. The user sets the text color for the currently rendered viewpoint
6.用户将再现的全光图像的视点移动到另一个位置 6. The user moves the viewpoint of the reproduced plenoptic image to another location
7.用户为这个其它视点将文本颜色设定到另一个值 7. The user sets the text color to another value for this other viewpoint
8.全光注解模型随后被保存并且准备用于注解过程的在线阶段。 8. The plenoptic annotation model is then saved and ready for the online phase of the annotation process.
全光注解创作系统执行下面的任务以基于之前描述的用于文本注解的用户动作步骤来生成合适的注解模型: The plenoptic annotation authoring system performs the following tasks to generate a suitable annotation model based on the previously described user action steps for text annotation:
1.基于初始地被设定到默认值的视点设定将2D视图再现给用户 1. The 2D view is rendered to the user based on the viewpoint settings initially set to default values
2.文本注解的全光版本通过从文本对象追踪光线到虚拟视点来生成。这创建光线的集合,每一个由唯一识别符描述。这个光线的集合描述文本。这些光线在存储器中由对应于必须被应用到参考全光图像的修改的阵列来表示。在这个情形中,阵列将含有与注解光线匹配的光线必须采取的波长的值 2. Plenoptic versions of text annotations are generated by tracing rays from text objects to virtual viewpoints. This creates a collection of rays, each described by a unique identifier. This collection of rays describes the text. These rays are represented in memory by arrays corresponding to the modifications that must be applied to the reference plenoptic image. In this case, the array will contain values for the wavelengths that rays must take to match the annotation ray
3.注解初始地位于在创作工具中预限定的默认位置。注解光线与参考全光图像光线组合。在注解和参考图像的光线之间的这些关系通过使用光线唯一识别符为将来使用而被储存 3. Annotations are initially located at default positions predefined in the authoring tool. Annotation rays are combined with reference plenoptic image rays. These relationships between annotations and rays of the reference image are stored for future use using ray unique identifiers
4.当用户使用比如计算机鼠标指针来移动/改变注解的定向时,注解的不同光线与捕捉的全光图像的其它光线组合,并且针对每个位置或定向修改生成新的2D视图。当用户鼠标指针被投影在全光空间中时使这成为可能。指针的平移随后在与对应于2D再现的视图的虚拟平面平行的平面中被应用到注解。当注解被移动时,在注解和参考图像之间的光线的关系依据注解位置或定向改变来改变和更新 4. When the user moves/changes the orientation of the annotation using eg a computer mouse pointer, the different rays of the annotation are combined with other rays of the captured plenoptic image and a new 2D view is generated for each position or orientation modification. This is made possible when the user mouse pointer is projected in plenoptic space. The translation of the pointer is then applied to the annotation in a plane parallel to the virtual plane corresponding to the 2D rendered view. When the annotation is moved, the light relationship between the annotation and the reference image is changed and updated according to the annotation position or orientation change
5.当用户为当前视点选择用于文本的颜色时,注解阵列的波长值被改变以匹配选择的颜色 5. When the user selects a color to use for text for the current viewpoint, the wavelength value of the annotation array is changed to match the selected color
6.当选择新的视点并且选择新的文本颜色时,对应于用来生成这个新的再现的视图的光线的注解阵列的波长值被改变。在第一视点和第二视点中间的波长值使用标准或点对点内插方法来内插 6. When a new viewpoint is selected and a new text color is selected, the wavelength values of the annotation array corresponding to the rays used to generate this new rendered view are changed. Wavelength values intermediate between the first and second viewpoints are interpolated using standard or point-to-point interpolation methods
7.当用户保存模型时,全光注解阵列被保存有上载的全光参考模型,从而它能够被用在在线阶段。 7. When the user saves the model, the plenoptic annotation array is saved with the uploaded plenoptic reference model so that it can be used in the online phase.
在线阶段online phase
如之前解释的,当捕捉全光图像的用户想要该图像被注解时,发生整个注解过程的在线阶段。 As explained before, the online phase of the overall annotation process occurs when a user capturing a plenoptic image wants that image to be annotated.
注解过程的在线阶段被应用到输入全光图像以得到最终全光注解的图像。这由下述构成:将输入图像与一些参考模型匹配、检索匹配的参考模型的注解、将注解与输入全光图像组合、以可理解的形式将注解的视图再现给用户、并且可能地处理用户交互以生成在注解上限定的不同动作。 The online stage of the annotation process is applied to the input plenoptic image to get the final plenoptic annotated image. This consists of matching the input image to some reference model, retrieving annotations for the matched reference model, combining the annotations with the input plenoptic image, rendering the annotated view to the user in an intelligible form, and possibly processing the user Interact to generate different actions defined on annotations.
由于由光线组成的注解内容处在全光格式中并且捕捉的图像也处在全光格式中,所以那两个数据集合处在相同的空间中。因而注解能够直接被应用到全光图像而不需要进一步投影。注解已被应用到的修改的全光空间能够随后被投影到比如2D视图中。这也表示为全光再现过程选择的投影参数(诸如视点的改变、深度、焦点的选择,…)也隐含地应用在全光注解上。比如,当改变再现过程的焦点或视点时,注解将具有应用到它们的效果。 Since the annotation content composed of rays is in plenoptic format and the captured image is also in plenoptic format, those two data sets are in the same space. Annotations can thus be applied directly to the plenoptic image without further projections. The modified plenoptic space to which the annotations have been applied can then be projected into eg a 2D view. This also means that the projection parameters chosen for the plenoptic reconstruction process (such as change of viewpoint, depth, choice of focal point, ...) are also implicitly applied on the plenoptic annotation. For example, annotations will have effects applied to them when changing the focus or viewpoint of the rendering process.
如在图8上图解的在线全光注解过程包括第一步骤100,在该第一步骤100期间表示在全光格式中的光场的数据(全光数据)被检索。全光数据可以由用全光捕捉装置捕捉数据的装置4来检索,或由通过通信链接从装置4接收全光数据的设备诸如服务器5来检索。 The online plenoptic annotation process as illustrated on Fig. 8 comprises a first step 100 during which data representing a light field in a plenoptic format (plenoptic data) is retrieved. The plenoptic data may be retrieved by a device 4 capturing the data with a plenoptic capture device, or by a device such as a server 5 receiving the plenoptic data from the device 4 via a communication link.
在步骤101中,检索的数据与参考数据匹配。这个步骤可以在装置4和/或在服务器5中被执行。这个步骤可能涉及:确定在捕捉的数据中的特征的集合、查找表示带有匹配特征的参考图像的匹配参考数据、并且将捕捉的数据与参考数据配准,如比如在US13645762中描述的。参考数据可以表示在全光格式中的图像或其它图像,并且可以被储存在诸如数据库的存储器51中,该存储器51从多个装置可访问。匹配参考数据的识别可以基于用户的位置、时间、小时、从场景的元素接收的信号、由用户的和/或图像相似度给出的指示。配准过程目的在于查找在用户位置和参考数据之间的几何关系,从而能够推断在捕捉的全光图像的光线和来自匹配的全光参考图像的光线之间的变换。 In step 101, the retrieved data is matched with reference data. This step can be performed in the device 4 and/or in the server 5 . This step may involve determining the set of features in the captured data, finding matching reference data representing a reference image with matching features, and registering the captured data with the reference data, as eg described in US13645762. The reference data may represent images in plenoptic format or other images, and may be stored in a memory 51, such as a database, accessible from a plurality of devices. The identification of matching reference data may be based on the user's location, time, hour, signals received from elements of the scene, indications given by the user and/or image similarity. The registration process aims to find the geometric relationship between the user position and the reference data, so that the transformation between the rays of the captured plenoptic image and the rays from the matched plenoptic reference image can be inferred.
在步骤102中,比如从存储器51检索与匹配参考数据相关联的全光注解。这个注解处在全光格式中,即以光线来描述。那些注解光线可以表示比如文本、静态图像、视频图像、标识语、和/或直接作用在全光图像光线上的其它元素。 In step 102, plenoptic annotations associated with matching reference data are retrieved, eg from memory 51 . This annotation is in plenoptic format, i.e. described in terms of rays. Those annotation rays may represent, for example, text, still images, video images, logos, and/or other elements acting directly on the plenoptic image rays.
注解可以包含在全光空间中的声音,例如附着到全光参考图像的光线的特定组的声音,从而声音将仅对其中选择的光线在全光图像中也是可见的和/或在焦点上的一些方向来播放。 Annotations can contain sounds in the plenoptic space, e.g. sounds for a specific group of rays attached to the plenoptic reference image, whereby the sound will only be for which selected rays are also visible in the plenoptic image and/or in focus Some directions to play.
在步骤103中,在全光格式中的检索的注解与捕捉的全光数据组合以生成注解的数据,该注解的数据表示在全光格式中的注解的图像。这个组合可以在服务器5中或在装置4中进行。在后者的情形中,服务器5可以将注解的数据发送到装置4,该装置4随后进行组合。因为将参考图像的光线投影到捕捉的全光图像的变换从匹配步骤(步骤101)是已知的,所以使这个注解组合成为可能。因此注解也能够被应用到捕捉的全光图像。 In step 103, the retrieved annotations in plenoptic format are combined with the captured plenoptic data to generate annotated data representing the annotated image in plenoptic format. This combination can take place in the server 5 or in the device 4 . In the latter case, the server 5 can send the annotated data to the device 4, which then assembles it. This combination of annotations is made possible because the transformation that projects the ray projections of the reference image to the captured plenoptic image is known from the matching step (step 101 ). Annotations can thus also be applied to captured plenoptic images.
全光注解使用下面的方法能够被应用到捕捉的全光图像: Plenoptic annotations can be applied to captured plenoptic images using the following methods:
1.查找变换,该变换用于将参考全光图像光线投影到在图8的步骤100中检索的在线全光图像光线上; 1. Find the transformation used to project the reference plenoptic image ray onto the online plenoptic image ray retrieved in step 100 of FIG. 8 ;
2.针对在离线阶段中限定的参考全光图像的每个检索的注解: 2. Annotations for each retrieval of the reference plenoptic images defined in the offline phase:
1.通过读取在离线阶段中限定的注解阵列,识别和选择参考全光图像的哪些光线必须依据注解来修改 1. Identify and select which rays of the reference plenoptic image must be modified according to the annotations by reading the annotation array defined in the offline phase
2.将在点(1)中识别的光线投影到在线全光图像上。这创建在参考全光图像的选择的光线和来自捕捉的全光图像的光线之间的对应 2. Project the rays identified in point (1) onto the online plenoptic image. This creates a correspondence between selected rays of the reference plenoptic image and rays from the captured plenoptic image
3.针对在点(2)处已被选择的捕捉的全光图像的每个光线,应用变换到如在全光注解阵列中限定的光线。阵列被用作查找表,其中由于步骤(1)和(2)的选择过程能够识别的光线和变换的参数(诸如波长、方向…)被用作查找关键码。 3. For each ray of the captured plenoptic image that has been selected at point (2), apply the transformation to the ray as defined in the plenoptic annotation array. The array is used as a lookup table, where the parameters of the rays and transformations (such as wavelength, direction...) that can be identified due to the selection process of steps (1) and (2) are used as lookup keys.
作为示例,如果注解光线表示文本,则注解阵列将含有单个非空光场参数,该单个非空光场参数是对应于文本颜色的波长。因而捕捉的全光图像光线将通过以储存在注解阵列中的因子来增大/减小光线的波长而修改。这个因子通过使用在配准过程中计算的光线之间的变换而在阵列中被查找。 As an example, if the annotation ray represents text, the annotation array will contain a single non-null light field parameter which is the wavelength corresponding to the color of the text. The captured plenoptic image rays will thus be modified by increasing/decreasing the wavelength of the rays by the factor stored in the annotation array. This factor is looked up in the array using the transformation between rays computed during registration.
在步骤104中,从注解的数据再现视图,比如2D或立体视图,并且将视图比如显示在显示器40上或用另一个设备显示给用户/查看者。这个视图再现过程下面连同图9被更详细描述。 In step 104 a view, such as a 2D or stereoscopic view, is rendered from the annotated data and displayed to the user/viewer, such as on a display 40 or with another device. This view rendering process is described in more detail below in conjunction with FIG. 9 .
在步骤105中,使与注解的交互成为可能。系统能够对不同的事件做出反应以执行之前在注解过程的离线部分中限定的特定动作。这样的事件能够是用户与注解的交互。借助于触摸屏、手追踪传感器或任何其它输入装置,用户能够指向给定注解并且与给定注解交互。这个交互将生成交互事件,该交互事件能够触发在注解过程的离线阶段中限定的特定动作。 In step 105, interaction with annotations is enabled. The system can react to different events to perform specific actions previously defined in the offline part of the annotation process. Such an event could be a user interaction with an annotation. A user can point to and interact with a given annotation by means of a touch screen, hand tracking sensor, or any other input means. This interaction will generate interaction events capable of triggering specific actions defined in the offline phase of the annotation process.
事件的另一个可能类型是当检测到场景中的特定改变时触发的事件。如在这个部分稍后解释的,能够检测到由捕捉的全光图像中的参考模型的对象的闭塞。这个闭塞事件能够触发之前在注解过程的离线阶段中限定的动作。作为触发注解动作的可能事件的另一个示例,能够使用声音识别模块以基于检测到的声音的特定类型来触发特定动作。 Another possible type of event is an event that fires when a particular change in the scene is detected. As explained later in this section, occlusions of objects by reference models in captured plenoptic images can be detected. This occlusion event can trigger actions previously defined in the offline phase of the annotation process. As another example of possible events that trigger annotation actions, a sound recognition module could be used to trigger certain actions based on a certain type of sound detected.
图9图解视图的再现以及对于查看者随后修改再现的各种可能性。如之前指示的,在步骤104中从下述再现增强现实视图:从捕捉的视图生成的注解的数据和在全光格式中的注解数据,如之前以图8描述的。再现的视图可以是如由针孔相机产生的标准2D视图、立体视图、视频、全光数据的全息投影、或优选地以用于重新聚焦和/或改变视点的一些命令来呈现图像的动态图像模块。动态图像模块能够是能够再现全光图像作为命令值的功能或作为Flash对象的HTML5/Javascript网页,或允许若干图像的动态呈现的任何其它技术。可以在步骤104期间被再现的视图的示例在图5A、6A、和7A上示出。在图5A和6A上的视图包含带有注解61的对象60。在图7A的视图上也看到处于不同深度并且因此在焦点外的额外对象62。重新聚焦或改变视点能够被用户手动地(比如通过选择在图像上或围绕图像的对象或位置)或自动地(比如当用户移动时)触发。 Figure 9 illustrates the rendering of a view and the various possibilities for the viewer to subsequently modify the rendering. As previously indicated, the augmented reality view is rendered in step 104 from the annotated data generated from the captured view and the annotation data in plenoptic format as previously described with FIG. 8 . The rendered view can be a standard 2D view as produced by a pinhole camera, a stereoscopic view, a video, a holographic projection of plenoptic data, or preferably a dynamic image that presents the image with some commands for refocusing and/or changing viewpoint module. The dynamic image module can be a function capable of rendering a plenoptic image as a command value or an HTML5/Javascript webpage as a Flash object, or any other technology allowing the dynamic presentation of several images. Examples of views that may be rendered during step 104 are shown on Figures 5A, 6A, and 7A. The view on FIGS. 5A and 6A contains object 60 with annotation 61 . An additional object 62 at a different depth and thus out of focus is also seen on the view of Fig. 7A. Refocusing or changing viewpoint can be triggered manually by the user (such as by selecting an object or location on or around the image) or automatically (such as when the user moves).
在步骤105中,用户输入用于修改视点的命令以在步骤107期间从相同全光数据产生新奇视图,该新奇视图对应于从不同视点观察的相同场景。用于从全光数据生成如从不同视点或查看方向看到的场景的各种2D图像的算法照此是已知的,并且比如在US6222937中被描述。由这个命令产生的并且被视点选择模块403执行的修改的2D图像的示例被图解在图5B上。如能够看到,不仅对象60的透视图而且注解61的透视图已被这个命令修改。实际上,由于注解直接被应用在由输入全光数据表示的全光空间上,当视图从全光空间生成时,注解看似以与全光图像相同方式被变换。这产生更现实的注解。 In step 105, the user inputs a command for modifying the viewpoint to generate a novel view from the same plenoptic data during step 107, the novel view corresponding to the same scene viewed from a different viewpoint. Algorithms for generating from plenoptic data various 2D images of a scene as seen from different viewpoints or viewing directions are known as such and are described eg in US6222937. An example of a modified 2D image resulting from this command and executed by the viewpoint selection module 403 is illustrated on FIG. 5B . As can be seen, not only the perspective of the object 60 but also the perspective of the annotation 61 has been modified by this command. In fact, since the annotations are applied directly on the plenoptic space represented by the input plenoptic data, when the view is generated from the plenoptic space, the annotations appear to be transformed in the same way as the plenoptic images. This produces more realistic annotations.
一些注解可以仅从查看方向的第一集合是可见的,但是从其它方向不是可见的。因此,如以图6B图解的,在步骤105期间视点的改变可以导致新的视图,其中使一个注解61成为不可见但是显露与相同对象相关联的新的注解64。多个注解可以与参考图像的单个位置相关联,但是与不同的查看方向相关联。由于在注解过程的离线阶段中设定的不同注解光场参数,当从与第二不同视图方向相比的第一查看方向再现注解时,注解本身也可以看起来不同。外观的改变能够被注解本身限定,但是它也能够是输入全光图像的功能。 Some annotations may only be visible from the first set of viewing directions, but not from other directions. Thus, as illustrated in Figure 6B, a change of viewpoint during step 105 may result in a new view in which one annotation 61 is made invisible but a new annotation 64 associated with the same object is revealed. Multiple annotations may be associated with a single location of the reference image, but with different viewing directions. Due to different annotation light field parameters set in the offline phase of the annotation process, the annotation itself may also look different when rendered from a first viewing direction compared to a second different viewing direction. The change in appearance can be defined by the annotation itself, but it can also be a function of the input plenoptic image.
在图9的步骤106中,用户输入命令,该命令用于重新聚焦图像并且用于从在全光格式中的数据生成聚焦在不同距离处的新图像。这个命令可以被重新聚焦模块402执行。如能够在图7A和7B上看到,这可以导致在第一焦距处可见的第一注解61消失或在图7B上示出的第二焦距处变为不那么锐利,然而第二注解63仅出现在这个第二焦距处。 In step 106 of Figure 9, the user enters commands for refocusing the image and for generating a new image focused at a different distance from the data in the plenoptic format. This command can be executed by the refocusing module 402 . As can be seen on FIGS. 7A and 7B , this can cause the first annotation 61 visible at the first focal length to disappear or become less sharp at the second focal length shown on FIG. 7B , whereas the second annotation 63 only occurs at this second focal length.
在步骤105和106中用来改变再现的视图的不同命令也能够关于用户移动自动地发布。在一个实施例中,用户移动用嵌入在全光捕捉装置中的惯性测量单元(IMU)追踪。通过使用这个模块,当用户移动时再现的视图自动地被更新。比如,当用户在左边移动时,查看方向稍微地平移到左边。当用户向前移动时,相同的原理适用,其中聚焦范围也向前移动,与之前再现的视图比较产生在背景平面中的更锐利的对象以及在前景平面中的更柔和的对象。本发明不被约束到使用IMU来追踪用户移动。诸如直接使用全光图像内容以追踪用户移动的其它装置也能够被使用。 The different commands to change the rendered view in steps 105 and 106 can also be issued automatically with respect to the user movement. In one embodiment, user movement is tracked with an inertial measurement unit (IMU) embedded in the plenoptic capture device. By using this module, the rendered view is automatically updated as the user moves. For example, when the user moves to the left, the viewing direction pans slightly to the left. The same principle applies when the user moves forward, where the focus range also moves forward, resulting in sharper objects in the background plane and softer objects in the foreground plane compared to the previously rendered view. The invention is not limited to the use of IMUs to track user movement. Other means such as directly using the plenoptic image content to track user movement can also be used.
在另一个实施例中,在线全光注解过程被持续应用到由移动中的用户的全光捕捉装置产生的全光图像流。这个持续处理允许用户持续地移动或移动他的全光捕捉装置并且使全光注解实时地更新。全光图像流以及视图的再现(图8的步骤104)必须实时被处理,从而用户感知注解好像它们是场景的部分。在这个实施例中,在不需要具有另一个全光捕捉的情况下能够然后修改查看方向的事实允许以需要从流来处理的全光图像的低得多的数目来实现相同效果。实际上,如果我们假定单个全光捕捉允许在特定查看方向范围内的视图的再现,并且只要用户不移动出该范围,则来自流的全光图像不需要被处理并且仅图8的步骤104需要再次被执行。这开启下述新的可能性:当用户正接近查看范围的边界时通过异步地处理新的全光图像帧来进行更计算高效的实时追踪,从而当新的帧应该被处理时用户感知不到延迟。 In another embodiment, an online plenoptic annotation process is continuously applied to the plenoptic image stream generated by the user's plenoptic capture device on the move. This continuous process allows the user to continuously move or move his plenoptic capture device and have the plenoptic annotations updated in real time. The plenoptic image stream and rendering of the view (step 104 of Fig. 8) must be processed in real-time so that the user perceives the annotations as if they were part of the scene. In this embodiment, the fact that the viewing direction can then be modified without needing to have another plenoptic capture allows to achieve the same effect with a much lower number of plenoptic images that need to be processed from the stream. In fact, if we assume that a single plenoptic capture allows the rendering of views within a certain range of viewing directions, and as long as the user does not move out of this range, then the plenoptic image from the stream does not need to be processed and only step 104 of FIG. is executed again. This opens new possibilities for more computationally efficient real-time tracking by processing new plenoptic image frames asynchronously as the user is approaching the boundaries of the field of view, so that the user is not aware when a new frame should be processed Delay.
用于注解活跃的全光图像的方法的示例被图解在图11上: An example of a method for annotating an active plenoptic image is illustrated on Figure 11:
图20的步骤200、201、202、203与图8中的步骤100、101、102、103类似或等效。 Steps 200 , 201 , 202 and 203 in FIG. 20 are similar or equivalent to steps 100 , 101 , 102 and 103 in FIG. 8 .
在步骤204中,查看方向参数被计算,作为步骤201的配准过程的结果。 In step 204 , viewing direction parameters are calculated as a result of the registration process of step 201 .
在步骤205中,基于在之前步骤中计算的查看方向再现视图。 In step 205, the view is rendered based on the viewing direction calculated in the previous step.
在步骤206中,惯性测量单元(IMU)被用来关于步骤200已被计算的时间来确定用户移动。随后采取下述决定:要么回到步骤200用于处理新的全光图像,要么直接到步骤204以基于IMU移动评估来更新查看方向参数。移动量被用来确定之前捕捉的全光数据是否能够被用来生成新奇视图。这典型地取决于全光捕捉装置的视场。 In step 206 an inertial measurement unit (IMU) is used to determine user movement with respect to the time that step 200 has been calculated. A decision is then taken to either go back to step 200 for processing a new plenoptic image, or go directly to step 204 to update the view direction parameter based on the IMU movement evaluation. The amount of movement is used to determine whether previously captured plenoptic data can be used to generate novel views. This typically depends on the field of view of the plenoptic capture device.
全光注解的再现可以考虑可能的闭塞。如果要注解的目标元素从捕捉装置视野被位于输入全光图像的另一个对象隐藏,则全光注解可能被闭塞。 Reproduction of plenoptic annotations can account for possible occlusions. The plenoptic annotation may be occluded if the target element to be annotated is hidden from the capture device view by another object located in the input plenoptic image.
在一个实施例中,再现模块利用捕捉的数据的全光格式将注解在视觉上隐藏在不相关对象后面。再现模块从全光参考数据知道捕捉的光线的属性,该捕捉的光线的属性应该来自捕捉的全光图像的每个元素。如果捕捉的光线具有与元素的期望的光线不同的属性,则它能够表示闭塞对象在元素前面,并且因而不必须对于这个元素显示注解。 In one embodiment, the rendering module utilizes the plenoptic format of the captured data to visually hide annotations behind irrelevant objects. The rendering module knows from the plenoptic reference data the properties of the captured rays that should come from each element of the captured plenoptic image. If the captured ray has different properties than the element's expected ray, it can indicate that the occluding object is in front of the element, and thus the annotation does not have to be displayed for this element.
在类似的方式中,如果对应于在捕捉的图像中的元素的光线具有与在参考图像中的对应的方向不同的方向,则这能够表示元素处在不同深度。再现模块能够使用这个信息以检测闭塞。额外地,光线的颜色信息也能够被用来确定捕捉的要素是否被闭塞。然而,颜色信息是不足够的,因为闭塞对象可能具有与目标元素相同的颜色。 In a similar manner, if the light rays corresponding to an element in the captured image have a different direction than the corresponding direction in the reference image, this can indicate that the element is at a different depth. The rendering module can use this information to detect occlusions. Additionally, the color information of the light can also be used to determine whether the captured feature is occluded. However, color information is not sufficient since occluded objects may have the same color as the target element.
应用application
在与注解相同的空间中注解全光图像的过程和在全光格式中的注解的提供为增强现实带来新的应用。 The process of annotating plenoptic images in the same space as annotations and the provision of annotations in a plenoptic format brings new applications for augmented reality.
应用的第一示例是在社会背景中的全光注解系统的使用。实际上,对象/场景的全光图像能够被用户用他们的全光捕捉装置来捕捉。捕捉的全光图像随后能够被用户使用各种各样的注解(包含之前捕捉的并且被用作注解的全光图像)来注解。它们的注解的场景随后能够使用社交网络被分享给用户的朋友,从而当他们用他们自己的全光捕捉装置来捕捉它时那些朋友能够体验注解的场景。在这个情形中使用全光注解过程的优点以下述事实被补充支持(leverage):因为注解是全光图像,所以注解已经处在全光空间中。因此在相同的全光空间中进行注解过程是更计算高效的并且产生更现实的注解场景。 A first example of application is the use of a plenoptic annotation system in a social context. In fact, plenoptic images of objects/scenes can be captured by users with their plenoptic capture devices. The captured plenoptic images can then be annotated by the user using various annotations, including plenoptic images that were previously captured and used as annotations. Their annotated scene can then be shared to the user's friends using social networking, so that those friends can experience the annotated scene when they capture it with their own plenoptic capture device. The advantage of using the plenoptic annotation process in this case is leveraged by the fact that since the annotation is a plenoptic image, the annotation is already in plenoptic space. Doing the annotation process in the same plenoptic space is therefore more computationally efficient and yields more realistically annotated scenes.
利用全光空间的不同信息的应用的第二示例是在体系结构设计的领域中的特殊设计的全光注解的使用。如在发明的之前部分中描述的,全光注解由与在线阶段中的全光图像光线组合的光线组成。在注解过程的离线部分中限定组合该光线的方式。这个组合能够使得来自全光图像的光线不被来自注解的其它光线代替,而是比如仅改变它们的方向。通过限定注解(这不仅修改全光图像的光线的波长而且修改比如它们的方向),使模拟捕捉的场景的材料或纹理的改变成为可能。在体系结构设计的这个情形中,全光注解能够有利地被使用,以模拟比如在不同材料被应用到墙壁的情况下特定房间或特定建筑物将看起来怎样。在另一个实施例中,天气条件的模拟能够被应用到捕捉的全光图像。模拟下雨的注解能够被应用到场景。这将产生注解的图像,该注解的图像带有应用到它的下雨效果,从而用户能够在下雨或其它不同天气条件的情形中在视觉上看到场景将像如何,其中由于全光信息,不同的光反射和折射以现实方式合适地被处置和计算。 A second example of an application utilizing different information of plenoptic spaces is the use of specially designed plenoptic annotations in the field of architectural design. As described in the previous section of the invention, the plenoptic annotation consists of rays combined with the plenoptic image rays in the online phase. The way this ray is combined is defined in the offline part of the annotation process. This combination enables the rays from the plenoptic image not to be replaced by other rays from the annotations, but instead just change their direction, for example. By defining annotations, which modify not only the wavelength of the rays of the plenoptic image but also eg their direction, it is possible to simulate changes in the material or texture of the captured scene. In this case of architectural design, plenoptic annotations can advantageously be used to simulate, for example, how a particular room or a particular building would look if different materials were applied to the walls. In another embodiment, a simulation of weather conditions can be applied to the captured plenoptic images. Annotations that simulate rain can be applied to the scene. This will produce an annotated image with a rain effect applied to it so that the user can visually see how the scene will look like in the event of rain or other different weather conditions, where due to the plenoptic information, Different light reflections and refractions are properly handled and calculated in a realistic manner.
作为另一个示例,寻宝在传统的二维增强现实解决方案中是流行的应用。它由下述构成:将注解附着到物理对象,并且通过将线索交给朋友或其他人,让他们搜寻这些注解(被称为宝藏)。换句话说,当某人接近隐藏的对象时,他能够用他的全光捕捉装置来扫描周围的对象以确定它们是否与注解相关联。通过使用全光注解,由于我们能够将注解可见度限制到一些查看方向或焦距,所以寻宝变为更令人激动。比如,用户能够将注解附着到雕像,并且决定仅当将来搜寻者被放置在雕像前面并且因此他从那个角度看到雕像时使这个注解可见。类似地,我们能够使用全光空间的重新聚焦属性以确保搜寻者被聚焦在雕像本身上并且因此在这个情形中仅显示注解。它使寻宝成为更吸引人的,因为它避免用户在随机扫描周围时发现宝藏而是迫使他真正地解谜。 As another example, treasure hunting is a popular application in traditional 2D augmented reality solutions. It consists of attaching annotations to physical objects and having friends or others search for these annotations (called treasures) by giving clues to them. In other words, when someone approaches a hidden object, he can use his plenoptic capture device to scan surrounding objects to determine if they are associated with annotations. By using plenoptic annotations, treasure hunting becomes more exciting as we are able to limit annotation visibility to a few viewing directions or focal lengths. For example, a user could attach an annotation to a statue and decide to only make this annotation visible if in the future the seeker is placed in front of the statue and therefore he sees the statue from that angle. Similarly, we can use the refocusing properties of plenoptic space to ensure that the seeker is focused on the statue itself and thus only display the annotation in this case. It makes treasure hunting more engaging, because it prevents the user from randomly scanning around to find treasure and instead forces him to actually solve the puzzle.
另一个应用涉及在都市环境中的城市指南。比如,让我们考虑用户正在他访问的城市中并且寻找旅游点诸如历史纪念碑、观光点、雕像、博物馆、本地餐馆…。使用他的增强现实系统,用户当然不想要使所有信息立即出现在他的屏幕上:他只会被视觉上重叠在屏幕上的所有这些内容而变得混乱。相反,能够使全光注解依赖于用户视点和焦点。比如,由用户以特定视角(或以特定视角范围)捕捉的图像的元素能够以比用户面对的元素更低的重要性来显示。在一个实施例中,低重要性注解仅能够在屏幕上被显示为标题或点(当用户在该标题或点上点击时,它们能够被扩展),而更重要的感兴趣的点呈现更多细节或具有图像上的更大尺寸或重点。 Another application involves city guides in urban environments. For example, let us consider that a user is in a city he is visiting and looking for tourist spots such as historical monuments, sightseeing spots, statues, museums, local restaurants... . With his augmented reality system, the user certainly doesn't want to have all the information appearing on his screen at once: he'll just be confused by all this stuff visually superimposed on the screen. Instead, plenoptic annotations can be made dependent on user viewpoint and focus. For example, elements of an image captured by a user at a particular viewing angle (or at a particular range of viewing angles) can be displayed with lower importance than elements facing the user. In one embodiment, low-importance annotations can only be displayed on the screen as titles or points (when the user clicks on the title or point, they can be expanded), while more important points of interest are presented more Details may have greater size or emphasis on the image.
选择查看方向(从该查看方向注解不是可见的)的能力对车辆驾驶者是有吸引力的,该车辆驾驶者比如可能想要在导航仪显示器上得到增强现实图像,但是不想要被附着到元素的注解诸如对交通不相关的广告、商店等分心。在这个情形中,那些让人分心的注解可以与选择的定向的范围相关联,从而它们将不会被显示在从道路捕捉的图像上。 The ability to choose the viewing direction from which annotations are not visible is attractive to vehicle drivers who might want, for example, augmented reality images on a navigator display, but do not want to be attached to elements Annotations such as distractions from traffic-irrelevant advertisements, shops, etc. In this case, those distracting annotations may be associated with a range of orientations selected so that they will not be displayed on images captured from the road.
术语和限定Terms and Definitions
以上描述的方法的各种操作可以被能够执行操作的任何合适的装置诸如各种硬件和/或(一个或多个)软件部件、电路、和/或(一个或多个)模块来执行。通常,在申请中描述的任何操作可以被能够执行操作的对应的功能装置来执行。各种装置、逻辑块、和模块可以包含各种硬件和/或(一个或多个)软件部件和/或(一个或多个)模块,其包含但不被限制到:电路、专用集成电路(ASIC)、或通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现场可编程门阵列信号(FPGA)或其它可编程逻辑器件(PLD)、分立门或晶体管逻辑、分立硬件部件或设计成执行在本文中描述的功能的其任何组合。通用处理器可以是微处理器,但是在替选方式中处理器可以是任何商业可获得的处理器、控制器、微控制器或状态机。处理器也可以被实施为计算装置的组合,例如DSP和微处理器的组合、多个微处理器、连同DSP核的一个或多个微处理器、或任何其它这样的配置。服务器可以被实施为单个机器,机器的集合、虚拟服务器、或云服务器。 Various operations of the methods described above may be performed by any suitable means capable of performing operations, such as various hardware and/or software component(s), circuits, and/or module(s). In general, any operations described in the application may be performed by corresponding functional means capable of performing the operations. The various means, logic blocks, and modules may comprise various hardware and/or software component(s) and/or module(s), including but not limited to: circuits, application specific integrated circuits ( ASIC), or general-purpose processor, digital signal processor (DSP), application-specific integrated circuit (ASIC), field-programmable gate array signal (FPGA) or other programmable logic device (PLD), discrete gate or transistor logic, discrete hardware components or any combination thereof designed to perform the functions described herein. A general-purpose processor may be a microprocessor, but in the alternative the processor may be any commercially available processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, eg, a combination of a DSP and a microprocessor, multiple microprocessors, one or more microprocessors together with a DSP core, or any other such configuration. A server can be implemented as a single machine, a collection of machines, a virtual server, or a cloud server.
如在本文中使用的,表述“全光数据”指定下述任何数据:该数据用全光捕捉装置生成、或从其它类型的数据来计算、并且描述场景的光场图像,即其中储存不仅光的亮度和颜色而且这个光的方向的图像。由于丢失光的这个方向,所以从这样的全光数据再现的2D或立体投影不被视为全光图像。 As used herein, the expression "plenoptic data" designates any data that is generated with a plenoptic capture device, or calculated from other types of data, and that describes a light field image of a scene in which not only light The brightness and color of this light and the direction of the image. Due to the loss of this direction of light, 2D or stereoscopic projections rendered from such plenoptic data are not considered plenoptic images.
如在本文中使用的,表述“全光空间”可以指定多维空间,光场即描述在空间中的每个方向中的光量的函数能够以该多维空间来描述。全光空间可以由针对光线的位置的至少两个参数、针对它的定向的两个参数和针对它的波长的一个参数以及(在视频的情形中)可能地针对时间的一个参数来描述。 As used herein, the expression "plenoptic space" may designate a multidimensional space in which a light field, ie a function describing the amount of light in each direction in space, can be described. The plenoptic space can be described by at least two parameters for the position of a ray, two parameters for its orientation and one parameter for its wavelength and (in the case of video) possibly one parameter for time.
如在本文中使用的,术语“注解”涵盖各种各样的可能的元素,该元素包含比如文本、静态图像、视频图像、标识语、声音和/或能够被叠加或以另外的方式被合并到由全光数据表示的全光空间中的其它元素。更一般地,术语注解涵盖用于变更由全光数据表示的全光空间光线的不同参数的不同方式。注解可以是动态的并且随时间推移改变它们的位置和/或外观。此外,注解可以是用户交互的并且对用户的操作作出反应(例如在用户交互时移动或变换)。 As used herein, the term "annotation" encompasses a wide variety of possible elements including, for example, text, still images, video images, logos, sounds, and/or to other elements in the plenoptic space represented by the plenoptic data. More generally, the term annotation covers different ways of altering different parameters of plenoptic spatial rays represented by plenoptic data. Annotations can be dynamic and change their position and/or appearance over time. Additionally, annotations can be user-interactive and react to user actions (eg, move or transform upon user interaction).
如在本文中使用的,术语“像素”可以指定一个单个单色光位,或用于检测在不同颜色中的光的多个邻近光位。比如,用于检测红色、绿色和蓝色光的三个邻近光位能够形成单个像素。 As used herein, the term "pixel" may designate a single single-color light bit, or multiple adjacent light-bits for detecting light in different colors. For example, three adjacent light bits for detecting red, green and blue light can form a single pixel.
如在本文中使用的,术语“确定”涵盖各种各样的动作。比如,“确定”可以包含推算、计算、处理、推导、调查、查找(例如在表、数据库或另一个数据结构中查找)、查明、评估等等。而且,“确定”可以包含接收(例如接收信息)、访问(例如访问在存储器中的数据)等等。而且,“确定”可以包含分辨、选择、选定、建立等等。 As used herein, the term "determining" encompasses a wide variety of actions. For example, "determining" may include deriving, calculating, processing, deriving, investigating, looking up (eg, in a table, database, or another data structure), ascertaining, evaluating, and the like. Also, "determining" may encompass receiving (eg, receiving information), accessing (eg, accessing data in memory), and the like. Also, "determining" may include resolving, selecting, selecting, establishing, and the like.
捕捉场景的图像涉及使用数字针孔相机来测量到达相机的图像传感器的光的亮度。捕捉全光数据可以涉及使用全光捕捉装置,或可以涉及从虚拟3D模型或对场景和光源的其它描述来生成光场数据。检索图像可以涉及捕捉图像、或通过通信链接从不同的装置检索图像。 Capturing an image of a scene involves using a digital pinhole camera to measure the brightness of light reaching the camera's image sensor. Capturing plenoptic data may involve using a plenoptic capture device, or may involve generating light field data from a virtual 3D model or other description of the scene and light sources. Retrieving an image may involve capturing the image, or retrieving the image from a different device over a communication link.
表达“再现视图”,比如“从全光数据再现2D视图”,涵盖下述动作:计算或生成图像,比如从包含在全光数据中的信息计算2D图像或全息图像。 The expression "reconstructing a view", such as "reproducing a 2D view from plenoptic data", covers the action of computing or generating an image, such as computing a 2D image or a holographic image from information contained in the plenoptic data.
连同本公开内容一起描述的方法或算法的步骤可以在硬件中、在由处理器执行的软件模块中、或在该两者的组合中直接被体现。软件模块可以驻留在本领域中已知的储存介质的任何形式中。可以被使用的储存介质的一些示例包含:随机存取存储器(RAM)、只读存储器(ROM)、闪存、EPROM存储器、EEPROM存储器、寄存器、硬盘、可移动磁盘、CD-ROM等等。软件模块可以包括单个指令、或许多指令,并且可以在若干不同代码段内、在不同程序之中、并且跨过多个储存介质来分布。软件模块可以由下述构成:可执行程序、在完整程序中使用的部分或例程或库、多个互连的程序、由许多智能手机、平台电脑或计算机执行的“应用程序”、窗口小部件、Flash应用、HTML代码的部分等。储存介质可以被耦合到处理器,使得处理器能够从储存介质读取信息并且将信息写到储存介质。在替选方式中,储存介质可以对处理器是不可或缺的。数据库可以被实施为数据的任何结构化的集合,其包含SQL数据库、XML文档的集合、语义的数据库、或通过IP网络可获得的信息的集合、或任何其它合适的结构。 The steps of a method or algorithm described in connection with this disclosure may be embodied directly in hardware, in a software module executed by a processor, or in a combination of both. A software module may reside in any form of storage medium known in the art. Some examples of storage media that may be used include: random access memory (RAM), read only memory (ROM), flash memory, EPROM memory, EEPROM memory, registers, hard disk, removable disk, CD-ROM, and the like. A software module may comprise a single instruction, or many instructions, and may be distributed over several different code segments, among different programs, and across multiple storage media. A software module may consist of an executable program, parts or routines or libraries used in a complete program, multiple interconnected programs, an "application" executed by many smartphones, pc or computers, widgets Widgets, Flash applications, sections of HTML code, etc. A storage medium may be coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. A database may be implemented as any structured collection of data, including a SQL database, a collection of XML documents, a semantic database, or a collection of information available over an IP network, or any other suitable structure.
因而,特定方面可以包括用于执行在本文中呈现的操作的计算机程序产品。比如,这样的计算机程序产品可以包括计算机可读介质,该计算机可读介质具有在其上被储存(和/或被编码)的指令,该指令被一个或多个处理器可执行以执行在本文中描述的操作。对于特定方面,计算机程序产品可以包含封装材料。 Accordingly, certain aspects may include a computer program product for performing the operations presented herein. For example, such a computer program product may include a computer-readable medium having stored thereon (and/or encoded) instructions executable by one or more processors to perform the operations described in . For certain aspects, a computer program product may include packaging materials.
要理解的是权利要求不被限制到以上图解的精确的配置和部件。在以上描述的方法和设备的布置、操作和细节中可以进行各种修改、改变和变动,而没有脱离权利要求的范围。 It is to be understood that the claims are not limited to the precise configuration and components illustrated above. Various modifications, changes and variations may be made in the arrangement, operation and details of the methods and apparatus described above without departing from the scope of the claims.
Claims (28)
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/EP2012/076643 WO2014094874A1 (en) | 2012-12-21 | 2012-12-21 | Method and apparatus for adding annotations to a plenoptic light field |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN104969264A true CN104969264A (en) | 2015-10-07 |
Family
ID=47553021
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201280077894.3A Pending CN104969264A (en) | 2012-12-21 | 2012-12-21 | Method and apparatus for adding annotations to plenoptic light fields |
Country Status (5)
| Country | Link |
|---|---|
| EP (1) | EP2936442A1 (en) |
| JP (1) | JP2016511850A (en) |
| KR (1) | KR20150106879A (en) |
| CN (1) | CN104969264A (en) |
| WO (1) | WO2014094874A1 (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10921896B2 (en) | 2015-03-16 | 2021-02-16 | Facebook Technologies, Llc | Device interaction in augmented reality |
| CN113494698A (en) * | 2020-04-02 | 2021-10-12 | 阿诺德和里克特电影技术公司 | Spotlight system, spotlight, optical component of a spotlight and method for determining the spatial light distribution of an optical component |
Families Citing this family (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP3059949A1 (en) | 2015-02-23 | 2016-08-24 | Thomson Licensing | Method and apparatus for generating lens-related metadata |
| EP3099077B1 (en) * | 2015-05-29 | 2020-07-15 | InterDigital CE Patent Holdings | Method for displaying a content from 4d light field data |
| EP3151534A1 (en) | 2015-09-29 | 2017-04-05 | Thomson Licensing | Method of refocusing images captured by a plenoptic camera and audio based refocusing image system |
| JP7209474B2 (en) * | 2018-03-30 | 2023-01-20 | 株式会社スクウェア・エニックス | Information processing program, information processing method and information processing system |
| US11182872B2 (en) | 2018-11-02 | 2021-11-23 | Electronics And Telecommunications Research Institute | Plenoptic data storage system and operating method thereof |
| KR102577447B1 (en) | 2018-11-02 | 2023-09-13 | 한국전자통신연구원 | Plenoptic data storage system and operating method thereof |
| US10565773B1 (en) | 2019-01-15 | 2020-02-18 | Nokia Technologies Oy | Efficient light field video streaming |
| JP7667656B2 (en) * | 2020-12-25 | 2025-04-23 | 時男 後藤 | 3D annotation system for surgical support |
Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2008134901A1 (en) * | 2007-05-08 | 2008-11-13 | Eidgenössische Technische Zürich | Method and system for image-based information retrieval |
Family Cites Families (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6009188A (en) | 1996-02-16 | 1999-12-28 | Microsoft Corporation | Method and system for digital plenoptic imaging |
| US8432414B2 (en) | 1997-09-05 | 2013-04-30 | Ecole Polytechnique Federale De Lausanne | Automated annotation of a view |
| JP2002098548A (en) * | 2000-09-22 | 2002-04-05 | Casio Comput Co Ltd | Guide information transmitting device and recording medium |
| JP2006255021A (en) * | 2005-03-15 | 2006-09-28 | Toshiba Corp | Image display device and image display method |
| US20120127203A1 (en) * | 2010-11-18 | 2012-05-24 | Canon Kabushiki Kaisha | Mixed reality display |
-
2012
- 2012-12-21 EP EP12813361.8A patent/EP2936442A1/en not_active Withdrawn
- 2012-12-21 WO PCT/EP2012/076643 patent/WO2014094874A1/en active Application Filing
- 2012-12-21 JP JP2015548228A patent/JP2016511850A/en active Pending
- 2012-12-21 CN CN201280077894.3A patent/CN104969264A/en active Pending
- 2012-12-21 KR KR1020157016284A patent/KR20150106879A/en not_active Withdrawn
Patent Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2008134901A1 (en) * | 2007-05-08 | 2008-11-13 | Eidgenössische Technische Zürich | Method and system for image-based information retrieval |
Non-Patent Citations (3)
| Title |
|---|
| HIROSHI KAWASAKI 等: ""Image-based rendering for mixed reality"", 《IMAGE PROCESSING》 * |
| INA FRIED: ""With New Features, Lytro Aims to Show Its Futuristic Camera Is No One-Trick Pony"", 《HTTP://ALLTHINGSD.COM/20121115/WITH-NEW-FEATURES-LYTRO-AIMS-TO-SHOW-ITS-FUTURISTIC-CAMERA-IS-NO-ONE-TRICK-PONY/》 * |
| MARC LEVOY 等: ""Light Field Rendering"", 《PROCEEDINGS OF THE 23RD ANNUAL CONFERENCE ON COMPUTER GRAPHICS AND INTERACTIVE TECHNIQUES》 * |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10921896B2 (en) | 2015-03-16 | 2021-02-16 | Facebook Technologies, Llc | Device interaction in augmented reality |
| CN113494698A (en) * | 2020-04-02 | 2021-10-12 | 阿诺德和里克特电影技术公司 | Spotlight system, spotlight, optical component of a spotlight and method for determining the spatial light distribution of an optical component |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20150106879A (en) | 2015-09-22 |
| WO2014094874A1 (en) | 2014-06-26 |
| EP2936442A1 (en) | 2015-10-28 |
| JP2016511850A (en) | 2016-04-21 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20140181630A1 (en) | Method and apparatus for adding annotations to an image | |
| JP7559033B2 (en) | Augmented and Virtual Reality | |
| US10755485B2 (en) | Augmented reality product preview | |
| CN104969264A (en) | Method and apparatus for adding annotations to plenoptic light fields | |
| US9704295B2 (en) | Construction of synthetic augmented reality environment | |
| CN106464853B (en) | Image processing device and method | |
| US8719687B2 (en) | Method for summarizing video and displaying the summary in three-dimensional scenes | |
| KR20140082610A (en) | Method and apaaratus for augmented exhibition contents in portable terminal | |
| JP2016537901A (en) | Light field processing method | |
| CN111277890B (en) | Virtual gift acquisition method and three-dimensional panoramic living broadcast room generation method | |
| CN101189643A (en) | 3D image generation and display system | |
| JP2016500169A (en) | Annotation method and apparatus | |
| Zoellner et al. | Cultural heritage layers: Integrating historic media in augmented reality | |
| CN111652986B (en) | Stage effect presentation method and device, electronic equipment and storage medium | |
| CN117671119A (en) | Static image dynamic processing method, device, equipment and storage medium | |
| CN117218266A (en) | 3D white-mode texture map generation method, device, equipment and medium | |
| JP4379594B2 (en) | Re-experience space generator | |
| AU2018203909A1 (en) | A User Interface | |
| Morvan et al. | Handling occluders in transitions from panoramic images: A perceptual study | |
| CN109348132B (en) | Panoramic shooting method and device | |
| KR20240178845A (en) | Apparatus and Method for Providing virtual space making solution | |
| CN116684540A (en) | Method, device and medium for presenting augmented reality data | |
| Huang et al. | Generation of Animated Stereo Panoramic Images for Image-Based Virtual Reality Systems |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20151007 |
|
| WD01 | Invention patent application deemed withdrawn after publication |