US20170315365A1 - Head mounted display content capture and sharing - Google Patents
Head mounted display content capture and sharing Download PDFInfo
- Publication number
- US20170315365A1 US20170315365A1 US15/144,112 US201615144112A US2017315365A1 US 20170315365 A1 US20170315365 A1 US 20170315365A1 US 201615144112 A US201615144112 A US 201615144112A US 2017315365 A1 US2017315365 A1 US 2017315365A1
- Authority
- US
- United States
- Prior art keywords
- video
- virtual reality
- goggle
- cameras
- view
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012545 processing Methods 0.000 claims abstract description 45
- 238000004891 communication Methods 0.000 claims abstract description 14
- 239000002131 composite material Substances 0.000 claims abstract description 14
- 238000000034 method Methods 0.000 claims description 23
- 230000001413 cellular effect Effects 0.000 claims description 7
- 230000008878 coupling Effects 0.000 claims description 5
- 238000010168 coupling process Methods 0.000 claims description 5
- 238000005859 coupling reaction Methods 0.000 claims description 5
- 230000006870 function Effects 0.000 description 18
- 238000010586 diagram Methods 0.000 description 12
- 230000015654 memory Effects 0.000 description 12
- 230000007246 mechanism Effects 0.000 description 9
- 210000003128 head Anatomy 0.000 description 7
- 230000000694 effects Effects 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 3
- 230000001934 delay Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000005452 bending Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000009194 climbing Effects 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000013213 extrapolation Methods 0.000 description 1
- 230000033001 locomotion Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0176—Head mounted characterised by mechanical features
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/131—Protocols for games, networked simulations or virtual reality
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04847—Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- H04L67/38—
-
- H04W4/008—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W4/00—Services specially adapted for wireless communication networks; Facilities therefor
- H04W4/80—Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B13/00—Optical objectives specially designed for the purposes specified below
- G02B13/06—Panoramic objectives; So-called "sky lenses" including panoramic objectives having reflecting surfaces
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0132—Head-up displays characterised by optical features comprising binocular systems
- G02B2027/0134—Head-up displays characterised by optical features comprising binocular systems of stereoscopic type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
Definitions
- the present disclosure is related to content capture and sharing of content, and in particular to a head mounted display for capturing and sharing virtual reality content.
- the captured video may be stitched together via a separate computer or cloud based application and made available to a virtual reality head mounted display for viewing the resulting virtual reality video.
- the process of capturing, sensing, and then viewing can be cumbersome and involve using several different devices.
- a virtual reality goggle includes a goggle shaped body having multiple cameras mounted thereon.
- a display is supported by the goggle shaped body.
- a support extends between sides of the goggle shaped body and includes multiple cameras.
- Circuitry is coupled to receive video from the multiple cameras, the video comprising a composite field of view of approximately 360 degrees about the goggle, the circuitry to couple to a wireless communication device to transmit the received video to networked processing services, and to receive stitched stereoscopic three dimensional virtual reality video from the networked processing services, the circuitry further coupled to provide the received stitched stereoscopic three dimensional virtual reality video for display.
- a method includes capturing video from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees about the user, transmitting the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving the stitched stereoscopic three dimensional virtual reality video from the networked processing services, and displaying the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle.
- a computer readable storage device has instruction stored thereon for execution by circuitry to perform operations.
- the operations include capturing video from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees about the user, transmitting the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving the stitched stereoscopic three dimensional virtual reality video from the networked processing services, and displaying the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle.
- FIG. 1 is a block perspective view of a virtual reality head mounted display system according to an example embodiment.
- FIG. 2 is a block schematic diagram of a head mounted virtual reality (VR) capture, viewing, and sharing system according to an example embodiment.
- VR virtual reality
- FIGS. 3A and 3B are respective front and rear perspective block diagrams of a head mounted display system according to an example embodiment.
- FIG. 4 is a partial longitudinal block view of a head mounted display strap showing a camera and CSI ribbon cable supported in or on the strap according to an example embodiment.
- FIG. 5 is a detailed block diagram illustrating a strap containing a CSI ribbon cable coupled to a stem via a snap on connector according to an example embodiment.
- FIG. 6 is a block schematic diagram illustrating devices and circuitry of a head mounted display system according to an example embodiment.
- FIG. 7 is an example graphical user interface for use by a user in operating the VR system according to an example embodiment.
- FIG. 8 is an example set up screen providing checkboxes for selecting a social media sharing mechanism according to an example embodiment.
- FIG. 9 is a block flow diagram illustrating a method of capturing, uploading, viewing, and sharing 3D VR video according to an example embodiment.
- FIG. 10 is a block diagram illustrating circuitry for implementing algorithms and performing methods according to example embodiments.
- the functions or algorithms described herein may be implemented in software in one embodiment.
- the software may consist of computer executable instructions stored on computer readable media or computer readable storage device such as one or more non-transitory memories or other type of hardware based storage devices, either local or networked.
- modules which may be software, hardware, firmware or any combination thereof. Multiple functions may be performed in one or more modules as desired, and the embodiments described are merely examples.
- the software may be executed on a digital signal processor, ASIC, microprocessor, or other type of processor operating on a computer system, such as a smartphone or portable computer, personal computer, server or other computer system, turning such computer system into a specifically programmed machine.
- FIG. 1 is a block perspective view of a virtual reality head mounted display (HMD) system 100 .
- System 100 includes a goggle shaped body 110 containing a display 115 mounted within the body 110 where it may be viewed by a user wearing the system 100 .
- Display 115 is not visible in FIG. 1 , but is shown in block form in FIG. 2 .
- the body 110 may be shaped to fit on a user like a pair of goggles, with suitable contours for being supported by a nose bridge of a user and spacing the display 115 within the body 110 a distance from the eyes of the user such that the display 115 can be viewed by the user.
- Separate displays may be used for each eye to provide the ability to provide a three dimensional virtual reality view. Suitable displays, including see through displays with optional projected images for a control interface, are currently available and may be incorporated in to the system 100 .
- System 100 also includes multiple cameras disposed to capture multiple fields of view about the head of the user wearing system 100 .
- Side stems 120 , 122 are coupled to the body 110 and positioned to extend from a front of a face of the user to the ears of the user to provide further support, in a manner similar to stems of eyeglasses.
- the stems 120 and 122 may optionally be coupled to the body via hinges at 123 , facilitating folding of the system 100 into a smaller carrying footprint.
- the hinges 123 may simply be a thinner portion of the stems, allowing bending of the stems at the hinges 123 .
- Sides of the body 110 may also include hinges to facilitate folding.
- a strap 125 may extend from ends of the stems around the back of the head of the user when system 100 is worn by the user.
- the cameras may include a camera 130 mounted on a left side of the body 110 , having a field of view to a user's left side.
- a camera 132 (not visible in FIG. 1 ) may similarly be mounted on a right side of the body to provide a field of view to a user's right side.
- Two cameras indicated at 134 and 136 may be mounted on the strap 125 to provide fields of view behind a user, referred to as rearward views with respect to a wearer, or the body 110 itself.
- One or more further cameras 137 , 138 may be mounted on a top of the body 110 to provide a field of view above the head of the user.
- Each camera may also include one or more microphones to capture audio.
- a sixth camera or cameras 139 may be supported on a front face of the body 110 to capture video with a field of view in front of the user wearing the system 100 , referred to as a forward view with respect to a wearer, or the body 110 itself.
- the stems 120 and 122 combined with the strap 125 comprise a support that both holds the goggle shaped body 110 on a wearer and provides structure to support the multiple cameras that provide side and rearward fields of view.
- the support may comprise other structures, such as longer stems that may support a rearward facing camera at ends of the stems to provide the rear field of view.
- the support may be a single strap that extends from the goggle shaped body around the head of a user with multiple cameras supported to provide side and rear fields of view. Other structures may also be utilized that help hold the goggle on a wearer and provide the desired fields of view.
- the various fields of view of the cameras overlap to some extent to allow extraction of depth information from the captured video via common triangulation techniques. Areas not overlapping may be estimated by stitching functions using extrapolation and other techniques. Stitching is a known process that combines images with overlapping fields of view to produce a view that has a wider field of view than any single image. Several cloud based services are available to perform stitching services as further described below.
- Additional cameras may be positioned on the body 110 , stems 120 , 122 , and strap 125 to provide additional overlapping fields of view for a significant part if not all of a 360 degree lateral view about the user's head, facilitating creation of a more robust 3D virtual reality (VR) view from a user's perspective.
- the video captured by the multiple cameras comprises a composite field of view of approximately 360 degrees about a wearer of the system 100 . As indicated above, there may be gaps in the approximately 360 degree view that can be estimated. If the gaps are too large for estimation, the composite field of view is no longer considered to be approximately 360 degrees.
- Side stems 120 may contain one or more cameras as indicated on stem 120 at 147 and 148 . Side stem 122 may also contain similar cameras.
- System 100 in one embodiment may include circuitry 140 that receives the video, optionally including audio, from each camera.
- a Wi-Fi, cellular, or other network device 142 may be used by the circuitry 140 to upload the captured video to processing resources 145 coupled to the network, such as cloud based resources.
- the circuitry 140 and network device 142 may be built into and supported by the body 110 in some embodiments.
- the display is designed into the body 110 of system 100 .
- Videos from the cameras are received by circuitry 140 , which may use network device 142 via a cellular or Wi-Fi connection to provide the videos to the processing device 145 .
- Processing device 145 will run services like JUMP Assembler to stitch the videos. Stitched 3D videos are sent back to the system 100 to be displayed via the built in displays as driven by circuitry 140 .
- videos from the cameras may be sent via circuitry 140 to a smart phone 143 via a micro USB interface.
- the smart phone 143 may provide both a display and a connection to the processing device 145 .
- Received stitched video from processing device 145 may be displayed via a split screen on a display of the smart phone 143 that is visible through the goggle shaped body 110 .
- the screen may be split in two to provide a separate display for each eye of a wearer of the goggle body 110 to provide a stereoscopic view of the stitched 3D virtual reality video.
- the display may be designed into the body 110 of the system 100 .
- Videos from the cameras may be sent through network device 142 via a wirelessHD (a high definition video content wireless protocol) connection to a portable computing device to be carried close to but separate from the body 110 , such as in a pocket of a user.
- the computing device may perform stitching and provide the stitched stereoscopic video back to the circuitry 140 for driving an internal display of the body 110 .
- smart phone 143 may be supported by the body 110 of system 110 , either as shown on a front exterior face of the body 110 , or on or in a side or bottom portion of the body 110 that does not obstruct vision of the user through a see through display 115 .
- the smart phone may include the camera 139 providing the forward view. Camera 139 may also be separate from the smart phone and supported directly by the body 110 .
- the smart phone 143 may separate from the body 110 and be coupled via a wired or short wireless mechanism to circuitry 140 .
- the smart phone may provide the communication functions of the wireless communication device 142 .
- the smart phone 143 may have a touch interface to provide an icon based touch interface via one or more apps all integrated into system 100 for controlling the capture and sending of video to the processing resources 145 , and receiving stitched stereoscopic virtual reality 3D video.
- stereoscopic refers to the use of two images of an object or scene that are taken at slightly different angles, and when viewed together, create an impression of depth and solidity.
- the circuitry 140 may then drive the display 115 to provide the user with a virtual reality experience by providing stereoscopic images to each of the separate displays 115 for each eye.
- the stereoscopic images may also be shared with others via user interfaces provided via the touch interface, which may be a touchscreen of a smartphone, or a touch panel of the head mounted display system 100 tied to circuitry 140 .
- the touch interface functions may be implemented via a virtual reality gaze input, voice, or some other input method.
- all the functions of video capture, uploading, viewing, and sharing may be performed by a single integrated unit.
- the circuitry 140 is disposed on a side of body 110 such that it does not interfere with a see through display 115 .
- the circuitry 140 may be used to implement transceiver functions and user interface functions via integrated circuits and mechanical, gaze tracking cameras, microphones, or touch sensors 153 on the body 110 for controlling the capture of video by the cameras, uploading, viewing, and sharing.
- a user interface may be provided by the display 115 and interacted with by user motions as captured by the cameras and provided to the circuitry 140 .
- Typical controls for the capture may include “Record” and “Stop”. Editing controls may also be provided for clipping portions of video.
- Each captured video feed may include synchronized time information and optionally include an identification of the camera that captured the video.
- Typical controls for uploading may include selection of video from a directory of captured video, which may be named by a user or simply time stamped, and an upload icon or other selection mechanism. Sharing may be done by entering user contact information, or selecting users to share the video with from an existing list.
- the uploading, sharing, and viewing of the captured video and stitched stereoscopic 3D VR video may be done as it is captured, providing a near real time experience for the user and for the users the stitched video is shared with, delayed only by network and processing delays, which delays may decrease with improved performance.
- the capture, uploading, and sharing may be set up for a single actuation at the beginning of an activity in some embodiments.
- the sharing may be via any social media mechanism or may simply involve the texting or emailing of a link to the stitched 3D VR video provided by the processing resources 145 .
- a user may wear system 100 while skydiving or any other activity.
- Starting the video capture may be accomplished by simply selecting a record and share feature at any point in the skydive.
- Selection of the record and share features starts the video capture and uploading to the cloud for stitching, and sharing it with others who may watch and listen on their own head mounted 3D VR displays, which may or may not provide similar integrated capture, upload, view, and share functions.
- Many other activities such as ziplining, skiing, hiking, climbing, biking, etc., are all candidates for quickly capturing a 3D experience and sharing a 3D VR version with others.
- FIG. 2 is a block schematic diagram of a head mounted VR capture, viewing, and sharing system 200 , showing components which may be used in implementing system 100 .
- the diagram illustrates multiple cameras 210 coupled to the circuitry 220 , which drives a left display 225 and a right display 230 to provide 3D content to a user.
- the circuitry also provides a wireless connection indicated at 235 for sending captured video and receiving stitched video for viewing and some editing functions.
- the circuitry may also run a browser for interfacing with remote resources via a web based interface, wherein the remote resources may include web sites that provide services, such as video stitching to generate 3D VR content and links to that content that can be viewed by a user and shared with other users, such as via text, email, social media, and other methods.
- the circuitry 220 may also generate a graphical user interface for display by the displays 225 and 230 .
- the displays may be see-through displays with the ability to project the graphical user interface to different portions of a field of view provided by the displays.
- the circuitry 200 may comprise a smart phone or other circuitry with processing resources and transceiver, and may be supported by the body 110 outside the field of view so as not to obstruct a user's view during participation in an activity.
- a typical field of view may be displayed on non-transparent displays that are not see-through, so the user can observe their actual environment.
- FIGS. 3A and 3B are front and back perspective representations of a virtual reality head mounted display system 300 .
- System 300 includes a head mounted display 310 supported by two stems 315 and 320 coupled to sides of the head mounted display 310 such that they support the head mounted display 310 on the head of a user.
- a head mounted display strap 325 may also be coupled to ends of the steps 315 and 320 to extend around the head of the user when worn.
- the head mounted display strap 325 may include multiple cameras 330 , 332 and microphones 335 , 337 mounted to provide images and sound from the back side of a user when wearing the head mounted display system 300 .
- a pair of CSI (camera interface specification by the Mipi® alliance) ribbon cables are shown at 340 , 342 mounted on or in the strip 325 .
- the cables 340 and 342 are coupled to the cameras 330 , 332 to provide signals to the head mounted display via a snap on ribbon cable receptacles 345 , 347 positioned in ends of the stems 315 , 320 .
- Microphones 335 , 337 may utilize an I 2 S (integrated interchip sound—a communication protocol designed to carry digital data) bus to carry signals to the stems.
- the stems contain conductors for providing the camera and microphone signals to circuitry 220 , which may be mounted in the stems 315 , 332 in one embodiment.
- One or more additional cameras 350 , 352 may be mounted on or otherwise supported by a top portion 353 of the head mounted display 310 and cameras 354 , 356 , 358 , 360 supported by stems 310 , 320 in further embodiments.
- One or more forward looking cameras 361 , 362 may also be supported by a front portion 363 of the head mounted display, such as near a periphery or other portion of the head mounted display 310 that do not significantly affect a user's field of view through the head mounted display in embodiments, where the display is a see through display.
- stems 320 and 322 are coupled to the head mounted display 310 via hinges 364 and 365 which permit the stems 320 and 322 to folded inward as indicated at 367 to provide for compact storage of the system 300 .
- the strap 325 is flexible, as are the cables 340 , 342 .
- the top portion 353 of the head mounted display 310 may also be folded downward as indicated by arrow 370 by use of a hinged portion 372 coupling the top portion 353 to the front portion 363 of the head mounted display 310 .
- FIG. 4 is a partial longitudinal block view of a head mounted display strap 400 showing a camera 410 and CSI ribbon cable 415 supported in or on the strap 400 .
- a microphone 420 is also supported in or on the strap 400 and is coupled to an I 2 S interface cable 425 . Both the CSI ribbon cable 415 and I 2 S interface cable 425 are shown extending through the strap 400 towards one of the stems as indicated by arrow 430 .
- FIG. 5 is a detailed block diagram illustrating a strap 500 containing a CSI ribbon cable 505 coupled to a stem 510 via a snap on connector 515 .
- One or more clamps 520 may be used to secure the cable 505 , and thereby the strap 500 , to the stem 510 .
- the I 2 S audio cable may be similarly connected via the snap on connector 515 .
- FIG. 6 is a block diagram illustrating devices and circuitry generally at 600 .
- Multiple cameras 610 , 612 , 614 , 616 , and microphones 620 , 622 , 624 , 626 are shown coupled via CSI and I 2 S cables to processors indicated at 630 , 632 .
- processors 630 , 632 reside in respective stems of the head mounted display system and perform functions related to capturing and processing video and audio.
- the processors 630 and 632 may be couple via USB (universal serial bus) cables 634 , 636 to a USB hub 640 .
- USB universal serial bus
- the USB hub 640 which may be supported by the head mounted display 310 , combines signals from the cables 634 , 636 and provides them on a communications connection 642 to a smart phone 645 for further processing, such as sending wirelessly for stitching.
- the smart phone may be supported by the head mounted display as illustrated in FIG. 1 , or may be supported by one or more of the stems, or may be otherwise supported by a user, such as a wearer of the head mounted display system 100 .
- the communications connection 642 may be a wired or wireless connection in various embodiments that has a sufficient bandwidth for transferring video and audio information.
- FIG. 7 is an example graphical user interface 700 for use by a user in operating system 100 , 200 , and 300 .
- Several icons are included for invoking functions performed by system 100 , 200 , and 300 .
- a record icon 710 may be selected to begin recording and a stop icon 715 may be selected to stop the recording.
- Many different methods may be used for selecting an icon, such as gaze based, external navigation and selection buttons or touchscreens supported on the body 110 , or other methods.
- a stitching icon 720 may be selected to begin stitching the just recorded video in one embodiment.
- the stitching icon 720 may also present a further screen for selecting recordings from different times, which may be shown in list form, for stitching, and may further provide an interface to set up stitching resources.
- the video may be provided via a network connection to stitching resources as it is being captured.
- a view icon 725 may be used to either automatically review 3D VR content that was just stitched, or may also be used to select content from a list of content that is available.
- the 3D VR content may be obtained for viewing via a link provided by the stitching service.
- interface 700 may be provided by the stitching resources and viewed via a browser or browser app running on the system 100 .
- the content may be viewed via a browser running on the circuitry, or by an app designed for viewing the 3D VR content.
- the app, or a remote server or other resources such as the resources providing the stitching may provide some basic editing functions, such as clipping in order to shorten the 3D VR content to be shared, or even deleting the content.
- the user may desire to share a 3D VR content as an activity is occurring.
- An auto record function is available to allow the user to have the system record video, send the video for stitching, and share the stereoscopic 3D VR content as quickly as it is stitched and available for viewing.
- Such a sequence may be controlled by selection of an auto-record setup icon indicated at 735 . Selection of the icon allows selection of a stitching resource and sharing mechanism prior to beginning recording. Once those selections are made, an auto record and share icon 740 may be selected, and video recording will begin, be communicated with the stitching resource, and automatically shared as previously specified.
- FIG. 8 is an example set up screen 800 providing checkboxes 815 , 817 for selecting a social media sharing mechanism, such as Facebook, LinkedIn, or Twitter.
- Checkboxes 819 , 821 may be used to select users for sharing via one or more email or text based addresses, or other sharing mechanisms.
- An icon 825 may be provided to add users.
- the selections in one embodiment results in the creation of a record and auto share script for execution by the circuitry.
- the script is prewritten with variables corresponding to the selections, including script to control the uploading of video to a stitching service, providing credentials for use of such services.
- the variables are provided with values corresponding to the boxes that are checked via the set up screen 800 .
- a continue icon may navigate the user back to screen 700 , where by pressing a single record and auto share icon 740 , the record and auto share script is executed, causing recording, transmission, and sharing to be initiated and performed without further action required by the user.
- the sharing will be coordinating with the stitching resources via the circuitry 200 by obtaining a link from the stitching resources where the stereoscopic 3D VR content may be viewed in near real time as available and providing that link to the sharing mechanisms and addresses provided.
- FIG. 9 is a block flow diagram illustrating a method 900 of capturing, uploading, viewing, and sharing 3D VR video according to an example embodiment.
- a user determines that they want to capture a 360 degree 3D video and puts on a virtual reality head mounted display system 100 with an optional smart phone or circuitry 140 , 200 , 300 .
- the user may use the graphical user interface 700 to instruct the system 100 to start capturing video.
- the user may hold still as indicated at 925 as the cameras and microphones record video and audio.
- the user may move around while recording in some embodiments.
- the video and audio may be uploaded in real-time through a Wi-Fi or cellular network at 830 in some embodiments.
- the user may use the graphical user interface 700 to instruct the system 100 to stop video capture.
- the user may receive the stitched together video comprising stereoscopic 3D VR content at 940 , review it, and optionally edit it, such as by clipping it.
- a user may open a sharing page in VR at 945 , which may be hosted by resources performing the video stitching, and may select recipients for sharing at 950 through such sharing page.
- the user may send a stereoscopic 3D VR link to recipients for their viewing pleasure.
- FIG. 10 is a block diagram illustrating circuitry for implementing algorithms and performing methods according to example embodiments. All components need not be used in various embodiments, and as described above, the circuitry may be divided into different components that operate together, including circuitry for driving the display, providing networking communications, executing browser functions, providing interfaces. FIG. 10 is also representative of a smart phone, such as a mobile phone that may be used for one or more of the functions performed by the circuitry.
- One example computing device in the form of a computer 1000 may include a processing unit 1002 , memory 1003 , removable storage 1010 such as an SD card or similar storage device, and non-removable storage 1012 .
- a processing unit 1002 may include a central processing unit 1003 , memory 1003 , removable storage 1010 such as an SD card or similar storage device, and non-removable storage 1012 .
- removable storage 1010 such as an SD card or similar storage device
- non-removable storage 1012 non-removable storage 1012 .
- the example computing device is illustrated and described as computer 1000 , the computing device may be in different forms in different embodiments.
- the computing device may be a smart phone or circuitry utilizing one or more elements as illustrated and described with regard to FIG. 10 .
- Memory 1003 may include volatile memory 1014 and non-volatile memory 1008 .
- Computer 1000 may include—or have access to a computing environment that includes—a variety of computer-readable media, such as volatile memory 1014 and non-volatile memory 1008 , removable storage 1010 and non-removable storage 1012 .
- Computer storage includes random access memory (RAM), read only memory (ROM), erasable programmable read-only memory (EPROM) and electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technologies or any other medium capable of storing computer-readable instructions.
- Computer 1000 may include or have access to a computing environment that includes input 1006 , output 1004 , and a communication connection 1016 .
- Output 1004 may include a display device, such as a touchscreen, that also may serve as an input device.
- the input 1006 may include one or more of a touchscreen, touchpad, microphone, camera, one or more device-specific buttons, one or more sensors integrated within or coupled via wired or wireless data connections to the computer 1000 , and other input devices.
- the computer 1000 in one embodiment operates in a networked environment using a communication connection to connect to one or more remote computers, such as database servers and video stitching services.
- Computer-readable instructions stored on a computer-readable medium are executable by the processing unit 1002 of the computer 1000 .
- RAM and ROM are some examples of articles including a non-transitory computer-readable medium such as a storage device.
- the terms computer-readable medium and storage device do not include carrier waves to the extent carrier waves are deemed too transitory.
- a virtual reality goggle includes a goggle shaped body 110 having multiple cameras 130 , 132 , 137 , 138 mounted thereon.
- a display 225 , 230 is supported by the goggle shaped body 110 in a position viewable by a wearer of the goggle shaped body 110 .
- a support 120 , 122 , 125 extends between sides of the goggle shaped body 110 and includes multiple cameras 134 , 136 , 147 , 148 .
- Circuitry 140 coupled to receive video from the multiple cameras, the video comprising a composite field of view of approximately 360 degrees about the goggle, the circuitry 140 to couple to a wireless communication device 142 to transmit the received video to networked processing services 145 , and to receive stitched stereoscopic three dimensional virtual reality video from the networked processing services 145 , the circuitry 140 further coupled to provide the received stitched stereoscopic three dimensional virtual reality video on the display 225 , 230 .
- the virtual reality goggle of example 1 wherein the goggle shaped body 110 has a first body camera 130 supported on a left side of the body and a second body camera 132 supported on a right side of the body providing left and right oriented fields of view.
- the virtual reality goggle of example 4 wherein the support comprises two stems 120 , 122 , each stem coupled to different sides of the goggle shaped body 110 at one end of each stem, and a strap 125 coupling the other ends of each stem to each other, wherein the multiple cameras 134 , 136 included in the support are supported by the strap 125 and provide rearward oriented fields of view.
- circuitry 140 is configured to provide a user interface 700 including a record option selector 710 and an option to share 730 the stereoscopic three dimensional virtual reality video.
- a method 900 includes capturing video 920 from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees, transmitting 930 the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving the stitched stereoscopic three dimensional virtual reality video from the networked processing services 940 , and displaying 945 the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle.
- the virtual reality goggle comprises a goggle shaped body 110 with a support for holding the body on the wearer, and wherein the video including a composite field of view of approximately 360 degrees about the goggle shaped body is provided by multiple cameras including a first body camera 130 supported on a left side of the body and a second body camera 132 supported on a right side of the body providing left and right oriented fields of view, a third body camera 139 supported on a front face of the body providing a forward oriented field of view, a fourth body camera 137 , 138 supported on a top face of the body providing an upward oriented field of view, and wherein the support comprises two stems 120 , 122 , each stem coupled to different sides of the body at one end of each stem, and a strap 125 coupling the other ends of each stem to each other, wherein the multiple cameras 134 , 136 included in the support are supported by the strap and provide rearward oriented fields of view.
- a computer readable storage device 1003 has instruction stored thereon for execution by circuitry 1002 to perform operations.
- the operations include capturing 920 video from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees, transmitting 930 the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving 940 the stitched stereoscopic three dimensional virtual reality video from the networked processing services, and displaying 940 the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle.
- the computer readable storage device of example 16 wherein the operations further comprise providing a user interface 700 , 945 on the display, the user interface including a record option selector 710 and an option 730 to share the stereoscopic three dimensional virtual reality video.
- a record and auto share option 740 which when selected, causes the goggle to begin recording video, automatically send the video to the networked processing services, and share the video with selected users.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Hardware Design (AREA)
- Optics & Photonics (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computer Graphics (AREA)
- Software Systems (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- User Interface Of Digital Computer (AREA)
- Studio Devices (AREA)
- Cameras In General (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
- The present disclosure is related to content capture and sharing of content, and in particular to a head mounted display for capturing and sharing virtual reality content.
- There are many stand-alone products currently available to capture video from multiple directions to capture sufficient video for a 360 degree virtual reality view. The captured video may be stitched together via a separate computer or cloud based application and made available to a virtual reality head mounted display for viewing the resulting virtual reality video. The process of capturing, sensing, and then viewing can be cumbersome and involve using several different devices.
- A virtual reality goggle includes a goggle shaped body having multiple cameras mounted thereon. A display is supported by the goggle shaped body. A support extends between sides of the goggle shaped body and includes multiple cameras. Circuitry is coupled to receive video from the multiple cameras, the video comprising a composite field of view of approximately 360 degrees about the goggle, the circuitry to couple to a wireless communication device to transmit the received video to networked processing services, and to receive stitched stereoscopic three dimensional virtual reality video from the networked processing services, the circuitry further coupled to provide the received stitched stereoscopic three dimensional virtual reality video for display.
- A method includes capturing video from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees about the user, transmitting the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving the stitched stereoscopic three dimensional virtual reality video from the networked processing services, and displaying the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle.
- A computer readable storage device has instruction stored thereon for execution by circuitry to perform operations. The operations include capturing video from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees about the user, transmitting the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving the stitched stereoscopic three dimensional virtual reality video from the networked processing services, and displaying the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle.
-
FIG. 1 is a block perspective view of a virtual reality head mounted display system according to an example embodiment. -
FIG. 2 is a block schematic diagram of a head mounted virtual reality (VR) capture, viewing, and sharing system according to an example embodiment. -
FIGS. 3A and 3B are respective front and rear perspective block diagrams of a head mounted display system according to an example embodiment. -
FIG. 4 is a partial longitudinal block view of a head mounted display strap showing a camera and CSI ribbon cable supported in or on the strap according to an example embodiment. -
FIG. 5 is a detailed block diagram illustrating a strap containing a CSI ribbon cable coupled to a stem via a snap on connector according to an example embodiment. -
FIG. 6 is a block schematic diagram illustrating devices and circuitry of a head mounted display system according to an example embodiment. -
FIG. 7 is an example graphical user interface for use by a user in operating the VR system according to an example embodiment. -
FIG. 8 is an example set up screen providing checkboxes for selecting a social media sharing mechanism according to an example embodiment. -
FIG. 9 is a block flow diagram illustrating a method of capturing, uploading, viewing, and sharing 3D VR video according to an example embodiment. -
FIG. 10 is a block diagram illustrating circuitry for implementing algorithms and performing methods according to example embodiments. - In the following description, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration specific embodiments which may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the invention, and it is to be understood that other embodiments may be utilized and that structural, logical and electrical changes may be made without departing from the scope of the present invention. The following description of example embodiments is, therefore, not to be taken in a limited sense, and the scope of the present invention is defined by the appended claims.
- The functions or algorithms described herein may be implemented in software in one embodiment. The software may consist of computer executable instructions stored on computer readable media or computer readable storage device such as one or more non-transitory memories or other type of hardware based storage devices, either local or networked. Further, such functions correspond to modules, which may be software, hardware, firmware or any combination thereof. Multiple functions may be performed in one or more modules as desired, and the embodiments described are merely examples. The software may be executed on a digital signal processor, ASIC, microprocessor, or other type of processor operating on a computer system, such as a smartphone or portable computer, personal computer, server or other computer system, turning such computer system into a specifically programmed machine.
-
FIG. 1 is a block perspective view of a virtual reality head mounted display (HMD)system 100.System 100 includes a goggleshaped body 110 containing adisplay 115 mounted within thebody 110 where it may be viewed by a user wearing thesystem 100.Display 115 is not visible inFIG. 1 , but is shown in block form inFIG. 2 . Thebody 110 may be shaped to fit on a user like a pair of goggles, with suitable contours for being supported by a nose bridge of a user and spacing thedisplay 115 within the body 110 a distance from the eyes of the user such that thedisplay 115 can be viewed by the user. Separate displays may be used for each eye to provide the ability to provide a three dimensional virtual reality view. Suitable displays, including see through displays with optional projected images for a control interface, are currently available and may be incorporated in to thesystem 100. -
System 100 also includes multiple cameras disposed to capture multiple fields of view about the head of theuser wearing system 100.Side stems 120, 122 are coupled to thebody 110 and positioned to extend from a front of a face of the user to the ears of the user to provide further support, in a manner similar to stems of eyeglasses. Thestems 120 and 122 may optionally be coupled to the body via hinges at 123, facilitating folding of thesystem 100 into a smaller carrying footprint. Thehinges 123 may simply be a thinner portion of the stems, allowing bending of the stems at thehinges 123. Sides of thebody 110 may also include hinges to facilitate folding. Astrap 125 may extend from ends of the stems around the back of the head of the user whensystem 100 is worn by the user. - In one embodiment, the cameras may include a
camera 130 mounted on a left side of thebody 110, having a field of view to a user's left side. A camera 132 (not visible inFIG. 1 ) may similarly be mounted on a right side of the body to provide a field of view to a user's right side. Two cameras indicated at 134 and 136 may be mounted on thestrap 125 to provide fields of view behind a user, referred to as rearward views with respect to a wearer, or thebody 110 itself. One or morefurther cameras body 110 to provide a field of view above the head of the user. Each camera may also include one or more microphones to capture audio. A sixth camera orcameras 139 may be supported on a front face of thebody 110 to capture video with a field of view in front of the user wearing thesystem 100, referred to as a forward view with respect to a wearer, or thebody 110 itself. - In one embodiment, the
stems 120 and 122 combined with thestrap 125 comprise a support that both holds the goggleshaped body 110 on a wearer and provides structure to support the multiple cameras that provide side and rearward fields of view. The support may comprise other structures, such as longer stems that may support a rearward facing camera at ends of the stems to provide the rear field of view. In one embodiment, the support may be a single strap that extends from the goggle shaped body around the head of a user with multiple cameras supported to provide side and rear fields of view. Other structures may also be utilized that help hold the goggle on a wearer and provide the desired fields of view. - In some embodiments, the various fields of view of the cameras overlap to some extent to allow extraction of depth information from the captured video via common triangulation techniques. Areas not overlapping may be estimated by stitching functions using extrapolation and other techniques. Stitching is a known process that combines images with overlapping fields of view to produce a view that has a wider field of view than any single image. Several cloud based services are available to perform stitching services as further described below.
- Additional cameras may be positioned on the
body 110, stems 120, 122, andstrap 125 to provide additional overlapping fields of view for a significant part if not all of a 360 degree lateral view about the user's head, facilitating creation of a more robust 3D virtual reality (VR) view from a user's perspective. In one embodiment, the video captured by the multiple cameras comprises a composite field of view of approximately 360 degrees about a wearer of thesystem 100. As indicated above, there may be gaps in the approximately 360 degree view that can be estimated. If the gaps are too large for estimation, the composite field of view is no longer considered to be approximately 360 degrees. Side stems 120 may contain one or more cameras as indicated on stem 120 at 147 and 148. Side stem 122 may also contain similar cameras. -
System 100 in one embodiment may includecircuitry 140 that receives the video, optionally including audio, from each camera. A Wi-Fi, cellular, orother network device 142 may be used by thecircuitry 140 to upload the captured video toprocessing resources 145 coupled to the network, such as cloud based resources. Thecircuitry 140 andnetwork device 142 may be built into and supported by thebody 110 in some embodiments. - In one embodiment, the display is designed into the
body 110 ofsystem 100. Videos from the cameras are received bycircuitry 140, which may usenetwork device 142 via a cellular or Wi-Fi connection to provide the videos to theprocessing device 145.Processing device 145 will run services like JUMP Assembler to stitch the videos. Stitched 3D videos are sent back to thesystem 100 to be displayed via the built in displays as driven bycircuitry 140. - In some embodiments, videos from the cameras may be sent via
circuitry 140 to asmart phone 143 via a micro USB interface. Thesmart phone 143 may provide both a display and a connection to theprocessing device 145. Received stitched video fromprocessing device 145 may be displayed via a split screen on a display of thesmart phone 143 that is visible through the goggle shapedbody 110. The screen may be split in two to provide a separate display for each eye of a wearer of thegoggle body 110 to provide a stereoscopic view of the stitched 3D virtual reality video. - In still further embodiments, the display may be designed into the
body 110 of thesystem 100. Videos from the cameras may be sent throughnetwork device 142 via a wirelessHD (a high definition video content wireless protocol) connection to a portable computing device to be carried close to but separate from thebody 110, such as in a pocket of a user. The computing device may perform stitching and provide the stitched stereoscopic video back to thecircuitry 140 for driving an internal display of thebody 110. - In one embodiment,
smart phone 143 may be supported by thebody 110 ofsystem 110, either as shown on a front exterior face of thebody 110, or on or in a side or bottom portion of thebody 110 that does not obstruct vision of the user through a see throughdisplay 115. When mounted as shown, the smart phone may include thecamera 139 providing the forward view.Camera 139 may also be separate from the smart phone and supported directly by thebody 110. In a further embodiment, thesmart phone 143 may separate from thebody 110 and be coupled via a wired or short wireless mechanism tocircuitry 140. The smart phone may provide the communication functions of thewireless communication device 142. - The
smart phone 143 may have a touch interface to provide an icon based touch interface via one or more apps all integrated intosystem 100 for controlling the capture and sending of video to theprocessing resources 145, and receiving stitched stereoscopicvirtual reality 3D video. The term, “stereoscopic” refers to the use of two images of an object or scene that are taken at slightly different angles, and when viewed together, create an impression of depth and solidity. Thecircuitry 140 may then drive thedisplay 115 to provide the user with a virtual reality experience by providing stereoscopic images to each of theseparate displays 115 for each eye. - The stereoscopic images may also be shared with others via user interfaces provided via the touch interface, which may be a touchscreen of a smartphone, or a touch panel of the head mounted
display system 100 tied tocircuitry 140. In some embodiment, the touch interface functions may be implemented via a virtual reality gaze input, voice, or some other input method. Thus, all the functions of video capture, uploading, viewing, and sharing may be performed by a single integrated unit. - In further embodiments, the
circuitry 140 is disposed on a side ofbody 110 such that it does not interfere with a see throughdisplay 115. Thecircuitry 140 may be used to implement transceiver functions and user interface functions via integrated circuits and mechanical, gaze tracking cameras, microphones, ortouch sensors 153 on thebody 110 for controlling the capture of video by the cameras, uploading, viewing, and sharing. - In still further embodiments, a user interface may be provided by the
display 115 and interacted with by user motions as captured by the cameras and provided to thecircuitry 140. Typical controls for the capture may include “Record” and “Stop”. Editing controls may also be provided for clipping portions of video. Each captured video feed may include synchronized time information and optionally include an identification of the camera that captured the video. Typical controls for uploading may include selection of video from a directory of captured video, which may be named by a user or simply time stamped, and an upload icon or other selection mechanism. Sharing may be done by entering user contact information, or selecting users to share the video with from an existing list. - In one embodiment, the uploading, sharing, and viewing of the captured video and stitched stereoscopic 3D VR video may be done as it is captured, providing a near real time experience for the user and for the users the stitched video is shared with, delayed only by network and processing delays, which delays may decrease with improved performance. The capture, uploading, and sharing may be set up for a single actuation at the beginning of an activity in some embodiments. The sharing may be via any social media mechanism or may simply involve the texting or emailing of a link to the stitched 3D VR video provided by the
processing resources 145. - In one example use of the
system 100, a user may wearsystem 100 while skydiving or any other activity. Starting the video capture may be accomplished by simply selecting a record and share feature at any point in the skydive. Selection of the record and share features starts the video capture and uploading to the cloud for stitching, and sharing it with others who may watch and listen on their own head mounted 3D VR displays, which may or may not provide similar integrated capture, upload, view, and share functions. Many other activities, such as ziplining, skiing, hiking, climbing, biking, etc., are all candidates for quickly capturing a 3D experience and sharing a 3D VR version with others. - Trying to perform the functions necessary to share 3D VR while skydiving using prior video capture mechanisms which may require separate applications and devices to first capture the video, finding a Wi-Fi hotspot, uploading, all while switching through multiple devices and apps, such as social media, would be nearly impossible and if possible, extremely inconvenient.
-
FIG. 2 is a block schematic diagram of a head mounted VR capture, viewing, andsharing system 200, showing components which may be used in implementingsystem 100. The diagram illustratesmultiple cameras 210 coupled to thecircuitry 220, which drives aleft display 225 and aright display 230 to provide 3D content to a user. The circuitry also provides a wireless connection indicated at 235 for sending captured video and receiving stitched video for viewing and some editing functions. The circuitry may also run a browser for interfacing with remote resources via a web based interface, wherein the remote resources may include web sites that provide services, such as video stitching to generate 3D VR content and links to that content that can be viewed by a user and shared with other users, such as via text, email, social media, and other methods. Thecircuitry 220 may also generate a graphical user interface for display by thedisplays - As previously mentioned, in some embodiments, the displays may be see-through displays with the ability to project the graphical user interface to different portions of a field of view provided by the displays. The
circuitry 200 may comprise a smart phone or other circuitry with processing resources and transceiver, and may be supported by thebody 110 outside the field of view so as not to obstruct a user's view during participation in an activity. In further embodiments, a typical field of view may be displayed on non-transparent displays that are not see-through, so the user can observe their actual environment. -
FIGS. 3A and 3B are front and back perspective representations of a virtual reality head mounteddisplay system 300.System 300 includes a head mounteddisplay 310 supported by twostems display 310 such that they support the head mounteddisplay 310 on the head of a user. A head mounteddisplay strap 325 may also be coupled to ends of thesteps - In addition, the head mounted
display strap 325 may includemultiple cameras microphones display system 300. A pair of CSI (camera interface specification by the Mipi® alliance) ribbon cables are shown at 340, 342 mounted on or in thestrip 325. Thecables cameras ribbon cable receptacles 345, 347 positioned in ends of the stems 315, 320.Microphones - The stems contain conductors for providing the camera and microphone signals to
circuitry 220, which may be mounted in thestems additional cameras top portion 353 of the head mounteddisplay 310 andcameras cameras front portion 363 of the head mounted display, such as near a periphery or other portion of the head mounteddisplay 310 that do not significantly affect a user's field of view through the head mounted display in embodiments, where the display is a see through display. - In one embodiment, stems 320 and 322 are coupled to the head mounted
display 310 viahinges system 300. Thestrap 325 is flexible, as are thecables top portion 353 of the head mounteddisplay 310 may also be folded downward as indicated byarrow 370 by use of a hingedportion 372 coupling thetop portion 353 to thefront portion 363 of the head mounteddisplay 310. -
FIG. 4 is a partial longitudinal block view of a head mounteddisplay strap 400 showing acamera 410 andCSI ribbon cable 415 supported in or on thestrap 400. Amicrophone 420 is also supported in or on thestrap 400 and is coupled to an I2S interface cable 425. Both theCSI ribbon cable 415 and I2S interface cable 425 are shown extending through thestrap 400 towards one of the stems as indicated byarrow 430. -
FIG. 5 is a detailed block diagram illustrating astrap 500 containing aCSI ribbon cable 505 coupled to astem 510 via a snap onconnector 515. One or more clamps 520 may be used to secure thecable 505, and thereby thestrap 500, to thestem 510. The I2S audio cable may be similarly connected via the snap onconnector 515. -
FIG. 6 is a block diagram illustrating devices and circuitry generally at 600.Multiple cameras microphones processors processors USB hub 640. TheUSB hub 640, which may be supported by the head mounteddisplay 310, combines signals from the cables 634, 636 and provides them on a communications connection 642 to asmart phone 645 for further processing, such as sending wirelessly for stitching. The smart phone may be supported by the head mounted display as illustrated inFIG. 1 , or may be supported by one or more of the stems, or may be otherwise supported by a user, such as a wearer of the head mounteddisplay system 100. The communications connection 642 may be a wired or wireless connection in various embodiments that has a sufficient bandwidth for transferring video and audio information. -
FIG. 7 is an examplegraphical user interface 700 for use by a user inoperating system system record icon 710 may be selected to begin recording and astop icon 715 may be selected to stop the recording. Many different methods may be used for selecting an icon, such as gaze based, external navigation and selection buttons or touchscreens supported on thebody 110, or other methods. - A
stitching icon 720 may be selected to begin stitching the just recorded video in one embodiment. Thestitching icon 720 may also present a further screen for selecting recordings from different times, which may be shown in list form, for stitching, and may further provide an interface to set up stitching resources. Note that in some embodiments, the video may be provided via a network connection to stitching resources as it is being captured. - A
view icon 725 may be used to either automatically review 3D VR content that was just stitched, or may also be used to select content from a list of content that is available. The 3D VR content may be obtained for viewing via a link provided by the stitching service. In some embodiments,interface 700 may be provided by the stitching resources and viewed via a browser or browser app running on thesystem 100. The content may be viewed via a browser running on the circuitry, or by an app designed for viewing the 3D VR content. The app, or a remote server or other resources such as the resources providing the stitching may provide some basic editing functions, such as clipping in order to shorten the 3D VR content to be shared, or even deleting the content. - In some uses of the
system share icon 740 may be selected, and video recording will begin, be communicated with the stitching resource, and automatically shared as previously specified. -
FIG. 8 is an example set upscreen 800 providingcheckboxes Checkboxes icon 825 may be provided to add users. The selections in one embodiment results in the creation of a record and auto share script for execution by the circuitry. In one embodiment, the script is prewritten with variables corresponding to the selections, including script to control the uploading of video to a stitching service, providing credentials for use of such services. The variables are provided with values corresponding to the boxes that are checked via the set upscreen 800. - A continue icon may navigate the user back to
screen 700, where by pressing a single record andauto share icon 740, the record and auto share script is executed, causing recording, transmission, and sharing to be initiated and performed without further action required by the user. The sharing will be coordinating with the stitching resources via thecircuitry 200 by obtaining a link from the stitching resources where the stereoscopic 3D VR content may be viewed in near real time as available and providing that link to the sharing mechanisms and addresses provided. -
FIG. 9 is a block flow diagram illustrating amethod 900 of capturing, uploading, viewing, and sharing 3D VR video according to an example embodiment. At 910, a user determines that they want to capture a 360degree 3D video and puts on a virtual reality head mounteddisplay system 100 with an optional smart phone orcircuitry graphical user interface 700 to instruct thesystem 100 to start capturing video. In one embodiment, the user may hold still as indicated at 925 as the cameras and microphones record video and audio. The user may move around while recording in some embodiments. The video and audio may be uploaded in real-time through a Wi-Fi or cellular network at 830 in some embodiments. At 535, the user may use thegraphical user interface 700 to instruct thesystem 100 to stop video capture. - The user may receive the stitched together video comprising stereoscopic 3D VR content at 940, review it, and optionally edit it, such as by clipping it. A user may open a sharing page in VR at 945, which may be hosted by resources performing the video stitching, and may select recipients for sharing at 950 through such sharing page. At 955, the user may send a stereoscopic 3D VR link to recipients for their viewing pleasure.
-
FIG. 10 is a block diagram illustrating circuitry for implementing algorithms and performing methods according to example embodiments. All components need not be used in various embodiments, and as described above, the circuitry may be divided into different components that operate together, including circuitry for driving the display, providing networking communications, executing browser functions, providing interfaces.FIG. 10 is also representative of a smart phone, such as a mobile phone that may be used for one or more of the functions performed by the circuitry. - One example computing device in the form of a
computer 1000 may include aprocessing unit 1002,memory 1003,removable storage 1010 such as an SD card or similar storage device, andnon-removable storage 1012. Although the example computing device is illustrated and described ascomputer 1000, the computing device may be in different forms in different embodiments. For example, the computing device may be a smart phone or circuitry utilizing one or more elements as illustrated and described with regard toFIG. 10 . -
Memory 1003 may includevolatile memory 1014 andnon-volatile memory 1008.Computer 1000 may include—or have access to a computing environment that includes—a variety of computer-readable media, such asvolatile memory 1014 andnon-volatile memory 1008,removable storage 1010 andnon-removable storage 1012. Computer storage includes random access memory (RAM), read only memory (ROM), erasable programmable read-only memory (EPROM) and electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technologies or any other medium capable of storing computer-readable instructions. -
Computer 1000 may include or have access to a computing environment that includesinput 1006,output 1004, and a communication connection 1016.Output 1004 may include a display device, such as a touchscreen, that also may serve as an input device. Theinput 1006 may include one or more of a touchscreen, touchpad, microphone, camera, one or more device-specific buttons, one or more sensors integrated within or coupled via wired or wireless data connections to thecomputer 1000, and other input devices. Thecomputer 1000 in one embodiment operates in a networked environment using a communication connection to connect to one or more remote computers, such as database servers and video stitching services. - Computer-readable instructions stored on a computer-readable medium are executable by the
processing unit 1002 of thecomputer 1000. RAM and ROM are some examples of articles including a non-transitory computer-readable medium such as a storage device. The terms computer-readable medium and storage device do not include carrier waves to the extent carrier waves are deemed too transitory. - 1. In example 1, a virtual reality goggle includes a goggle shaped
body 110 havingmultiple cameras display body 110 in a position viewable by a wearer of the goggle shapedbody 110. Asupport body 110 and includesmultiple cameras Circuitry 140 coupled to receive video from the multiple cameras, the video comprising a composite field of view of approximately 360 degrees about the goggle, thecircuitry 140 to couple to awireless communication device 142 to transmit the received video tonetworked processing services 145, and to receive stitched stereoscopic three dimensional virtual reality video from thenetworked processing services 145, thecircuitry 140 further coupled to provide the received stitched stereoscopic three dimensional virtual reality video on thedisplay - 2. The virtual reality goggle of example 1 wherein the goggle shaped
body 110 has afirst body camera 130 supported on a left side of the body and asecond body camera 132 supported on a right side of the body providing left and right oriented fields of view. - 3. The virtual reality goggle of example 2 wherein the goggle shaped
body 110 has athird body camera 139 supported on a front face of the body providing a forward oriented field of view. - 4. The virtual reality goggle of example 3 wherein the goggle shaped
body 110 has afourth body camera - 5. The virtual reality goggle of example 4 wherein the support comprises two stems 120, 122, each stem coupled to different sides of the goggle shaped
body 110 at one end of each stem, and astrap 125 coupling the other ends of each stem to each other, wherein themultiple cameras strap 125 and provide rearward oriented fields of view. - 6. The virtual reality goggle of any of examples 1-5 wherein the
wireless communication device 142 comprises a cellular transceiver. - 7. The virtual reality goggle of example 6 wherein the cellular transceiver comprises a
smart phone 143 supported by the goggle shapedbody 110. - 8. The virtual reality goggle of any of examples 1-7 wherein the
circuitry 140 is configured to provide auser interface 700 including arecord option selector 710 and an option to share 730 the stereoscopic three dimensional virtual reality video. - 9. The virtual reality goggle of example 8 wherein the
user interface option 735 to identify the networked processing services and theusers - 10. The virtual reality goggle of any of examples 8-9 wherein the
user interface 700 includes a record andauto share option 740, which when selected, causes the goggle to begin recording video, automatically send the video to thenetworked processing services 145, and share the video with selectedusers - 11. In example 11, a
method 900 includes capturingvideo 920 from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees, transmitting 930 the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving the stitched stereoscopic three dimensional virtual reality video from thenetworked processing services 940, and displaying 945 the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle. - 12. The method of example 11 and further comprising providing a
user interface 700 on the display, the user interface including arecord option selector 710 and anoption 730 to share the stereoscopic three dimensional virtual reality video. - 13. The method of example 12 wherein the
user interface 700 includes a set upoption 735 to identify the networked processing services and theusers - 14. The method of example 13 wherein the
user interface 700 includes a record andauto share option 740, which when selected, causes the goggle to begin recordingvideo 920, automatically send the video to thenetworked processing services 930, and share the video with selected users 955. - 15. The method of any of examples 11-14 wherein the virtual reality goggle comprises a goggle shaped
body 110 with a support for holding the body on the wearer, and wherein the video including a composite field of view of approximately 360 degrees about the goggle shaped body is provided by multiple cameras including afirst body camera 130 supported on a left side of the body and asecond body camera 132 supported on a right side of the body providing left and right oriented fields of view, athird body camera 139 supported on a front face of the body providing a forward oriented field of view, afourth body camera strap 125 coupling the other ends of each stem to each other, wherein themultiple cameras - 16. In example 16, a computer
readable storage device 1003 has instruction stored thereon for execution bycircuitry 1002 to perform operations. The operations include capturing 920 video from multiple cameras supported by a virtual reality goggle while being worn by a user, the video including a composite field of view of approximately 360 degrees, transmitting 930 the received video to networked processing services for stitching the received video into a stereoscopic three dimensional virtual reality video, receiving 940 the stitched stereoscopic three dimensional virtual reality video from the networked processing services, and displaying 940 the received stitched stereoscopic three dimensional virtual reality video in a display of the virtual reality goggle. - 17. The computer readable storage device of example 16 wherein the operations further comprise providing a
user interface record option selector 710 and anoption 730 to share the stereoscopic three dimensional virtual reality video. - 18. The computer readable storage device of any of examples 16-17 wherein the user interface includes a set up
option 735 to identify the networked processing services and theusers - 19. The computer readable storage device of example 18 wherein the
user interface 700 includes a record andauto share option 740, which when selected, causes the goggle to begin recording video, automatically send the video to the networked processing services, and share the video with selected users. - 20. The computer readable storage device of any of examples 16-19 wherein the circuitry comprises a
smart phone 140. - Although a few embodiments have been described in detail above, other modifications are possible. For example, the logic flows depicted in the figures do not require the particular order shown, or sequential order, to achieve desirable results. Other steps may be provided, or steps may be eliminated, from the described flows, and other components may be added to, or removed from, the described systems. Other embodiments may be within the scope of the following claims.
Claims (20)
Priority Applications (11)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/144,112 US9823477B1 (en) | 2016-05-02 | 2016-05-02 | Head mounted display content capture and sharing |
EP17792461.0A EP3443738B1 (en) | 2016-05-02 | 2017-04-28 | Head mounted display content capture and sharing |
CN201780026779.6A CN109076165B (en) | 2016-05-02 | 2017-04-28 | Head mounted display content capture and sharing |
PCT/CN2017/082541 WO2017190635A1 (en) | 2016-05-02 | 2017-04-28 | Head mounted display content capture and sharing |
EP21189935.6A EP3923124A3 (en) | 2016-05-02 | 2017-04-28 | Head mounted display content capture and sharing |
JP2018557391A JP6724164B2 (en) | 2016-05-02 | 2017-04-28 | Shoot and share head-mounted display content |
CN202110158076.7A CN113281905A (en) | 2016-05-02 | 2017-04-28 | Head mounted display content capture and sharing |
US15/799,364 US10473941B2 (en) | 2016-05-02 | 2017-10-31 | Head mounted display content capture and sharing |
US16/668,985 US20200166761A1 (en) | 2016-05-02 | 2019-10-30 | Head mounted display content capture and sharing |
JP2020107844A JP6966603B2 (en) | 2016-05-02 | 2020-06-23 | Shooting and sharing head-mounted display content |
JP2021172371A JP2022009355A (en) | 2016-05-02 | 2021-10-21 | Shooting and sharing head-mounted display content |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/144,112 US9823477B1 (en) | 2016-05-02 | 2016-05-02 | Head mounted display content capture and sharing |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/799,364 Continuation US10473941B2 (en) | 2016-05-02 | 2017-10-31 | Head mounted display content capture and sharing |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170315365A1 true US20170315365A1 (en) | 2017-11-02 |
US9823477B1 US9823477B1 (en) | 2017-11-21 |
Family
ID=60157458
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/144,112 Active US9823477B1 (en) | 2016-05-02 | 2016-05-02 | Head mounted display content capture and sharing |
US15/799,364 Active US10473941B2 (en) | 2016-05-02 | 2017-10-31 | Head mounted display content capture and sharing |
US16/668,985 Abandoned US20200166761A1 (en) | 2016-05-02 | 2019-10-30 | Head mounted display content capture and sharing |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/799,364 Active US10473941B2 (en) | 2016-05-02 | 2017-10-31 | Head mounted display content capture and sharing |
US16/668,985 Abandoned US20200166761A1 (en) | 2016-05-02 | 2019-10-30 | Head mounted display content capture and sharing |
Country Status (5)
Country | Link |
---|---|
US (3) | US9823477B1 (en) |
EP (2) | EP3443738B1 (en) |
JP (3) | JP6724164B2 (en) |
CN (2) | CN109076165B (en) |
WO (1) | WO2017190635A1 (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180017742A1 (en) * | 2016-05-11 | 2018-01-18 | Inneos LLC | Active optical cable for wearable device display |
US20180268585A1 (en) * | 2017-03-15 | 2018-09-20 | Boe Technology Group Co., Ltd. | Character input method, character input device, and wearable device |
US10454579B1 (en) | 2016-05-11 | 2019-10-22 | Zephyr Photonics Inc. | Active optical cable for helmet mounted displays |
WO2020180859A1 (en) * | 2019-03-05 | 2020-09-10 | Facebook Technologies, Llc | Apparatus, systems, and methods for wearable head-mounted displays |
US11163333B2 (en) * | 2019-03-29 | 2021-11-02 | Htc Corporation | Head-mounted display |
CN114503059A (en) * | 2019-09-30 | 2022-05-13 | 美国斯耐普公司 | Automated eye-mounted device sharing system |
US20240134191A1 (en) * | 2020-05-26 | 2024-04-25 | Magic Leap, Inc. | Monovision display for wearable device |
US20240345401A1 (en) * | 2018-03-29 | 2024-10-17 | Adlens Limited | Improvements in or relating to variable focusing power optical devices |
JP7596333B2 (en) | 2022-06-30 | 2024-12-09 | キヤノン株式会社 | HEAD-MOUNTED DISPLAY DEVICE, IMAGE PROCESSING DEVICE, CONTROL METHOD FOR HEAD-MOUNTED DISPLAY DEVICE, AND PROGRAM |
US20250036344A1 (en) * | 2023-07-24 | 2025-01-30 | Google Llc | Wearable display with coupled mobile interface module |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11494986B2 (en) * | 2017-04-20 | 2022-11-08 | Samsung Electronics Co., Ltd. | System and method for two dimensional application usage in three dimensional virtual reality environment |
KR102145852B1 (en) * | 2018-12-14 | 2020-08-19 | (주)이머시브캐스트 | Camera-based mixed reality glass apparatus and mixed reality display method |
US11800231B2 (en) * | 2019-09-19 | 2023-10-24 | Apple Inc. | Head-mounted display |
CN112416125A (en) * | 2020-11-17 | 2021-02-26 | 青岛小鸟看看科技有限公司 | VR head-mounted all-in-one machine |
CN115474177A (en) * | 2022-03-03 | 2022-12-13 | 北京罗克维尔斯科技有限公司 | Communication method, device, electronic device and storage medium of vehicle-mounted VR equipment |
CN116320366A (en) * | 2023-05-18 | 2023-06-23 | 中数元宇数字科技(上海)有限公司 | Video stream data pushing method, device, equipment and storage medium |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130242262A1 (en) * | 2005-10-07 | 2013-09-19 | Percept Technologies Inc. | Enhanced optical and perceptual digital eyewear |
US20140348484A1 (en) * | 2013-05-23 | 2014-11-27 | Chaotic Moon, LLC | Safety accessory with situational awareness and data retention |
US20150105035A1 (en) * | 2012-04-13 | 2015-04-16 | Eduardo Soares de Oliveira | Web-Connected Safety Helmet |
US20150348327A1 (en) * | 2014-05-30 | 2015-12-03 | Sony Computer Entertainment America Llc | Head Mounted Device (HMD) System Having Interface With Mobile Computing Device for Rendering Virtual Reality Content |
US20160093108A1 (en) * | 2014-09-30 | 2016-03-31 | Sony Computer Entertainment Inc. | Synchronizing Multiple Head-Mounted Displays to a Unified Space and Correlating Movement of Objects in the Unified Space |
US20160210785A1 (en) * | 2013-10-03 | 2016-07-21 | Sulon Technologies Inc. | Augmented reality system and method for positioning and mapping |
US20160309134A1 (en) * | 2015-04-19 | 2016-10-20 | Pelican Imaging Corporation | Multi-baseline camera array system architectures for depth augmentation in vr/ar applications |
US20170052595A1 (en) * | 2015-08-21 | 2017-02-23 | Adam Gabriel Poulos | Holographic Display System with Undo Functionality |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005172851A (en) * | 2003-12-05 | 2005-06-30 | Sony Corp | Image display apparatus |
JP4717728B2 (en) * | 2005-08-29 | 2011-07-06 | キヤノン株式会社 | Stereo display device and control method thereof |
US20100259619A1 (en) | 2009-04-10 | 2010-10-14 | Nicholson Timothy J | Hmd with elevated camera |
JP5560600B2 (en) * | 2009-07-09 | 2014-07-30 | 株式会社島津製作所 | Head motion tracker device |
JP2011197736A (en) * | 2010-03-17 | 2011-10-06 | Mitsubishi Electric Corp | Vision field support device |
US20130050069A1 (en) * | 2011-08-23 | 2013-02-28 | Sony Corporation, A Japanese Corporation | Method and system for use in providing three dimensional user interface |
JP5483761B2 (en) * | 2012-06-29 | 2014-05-07 | 株式会社ソニー・コンピュータエンタテインメント | Video output device, stereoscopic video observation device, video presentation system, and video output method |
CN103020983B (en) | 2012-09-12 | 2017-04-05 | 深圳先进技术研究院 | A kind of human-computer interaction device and method for target following |
CN105359063B (en) * | 2013-06-09 | 2018-08-17 | 索尼电脑娱乐公司 | Utilize the head-mounted display of tracking |
US9787895B2 (en) * | 2014-02-17 | 2017-10-10 | Sony Corporation | Information processing device, information processing method, and program for generating circumferential captured images |
US11205305B2 (en) * | 2014-09-22 | 2021-12-21 | Samsung Electronics Company, Ltd. | Presentation of three-dimensional video |
US10257494B2 (en) * | 2014-09-22 | 2019-04-09 | Samsung Electronics Co., Ltd. | Reconstruction of three-dimensional video |
CN104473717A (en) | 2014-12-04 | 2015-04-01 | 上海交通大学 | Wearable guide apparatus for totally blind people |
US9851564B2 (en) * | 2015-01-20 | 2017-12-26 | Microsoft Technology Licensing, Llc | Head-mounted display device with protective visor |
CN105404005A (en) | 2015-12-10 | 2016-03-16 | 合肥虔视光电科技有限公司 | Head-mounted display for augmented reality |
-
2016
- 2016-05-02 US US15/144,112 patent/US9823477B1/en active Active
-
2017
- 2017-04-28 WO PCT/CN2017/082541 patent/WO2017190635A1/en active Application Filing
- 2017-04-28 JP JP2018557391A patent/JP6724164B2/en active Active
- 2017-04-28 CN CN201780026779.6A patent/CN109076165B/en active Active
- 2017-04-28 CN CN202110158076.7A patent/CN113281905A/en active Pending
- 2017-04-28 EP EP17792461.0A patent/EP3443738B1/en active Active
- 2017-04-28 EP EP21189935.6A patent/EP3923124A3/en not_active Withdrawn
- 2017-10-31 US US15/799,364 patent/US10473941B2/en active Active
-
2019
- 2019-10-30 US US16/668,985 patent/US20200166761A1/en not_active Abandoned
-
2020
- 2020-06-23 JP JP2020107844A patent/JP6966603B2/en active Active
-
2021
- 2021-10-21 JP JP2021172371A patent/JP2022009355A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130242262A1 (en) * | 2005-10-07 | 2013-09-19 | Percept Technologies Inc. | Enhanced optical and perceptual digital eyewear |
US20150105035A1 (en) * | 2012-04-13 | 2015-04-16 | Eduardo Soares de Oliveira | Web-Connected Safety Helmet |
US20140348484A1 (en) * | 2013-05-23 | 2014-11-27 | Chaotic Moon, LLC | Safety accessory with situational awareness and data retention |
US20160210785A1 (en) * | 2013-10-03 | 2016-07-21 | Sulon Technologies Inc. | Augmented reality system and method for positioning and mapping |
US20150348327A1 (en) * | 2014-05-30 | 2015-12-03 | Sony Computer Entertainment America Llc | Head Mounted Device (HMD) System Having Interface With Mobile Computing Device for Rendering Virtual Reality Content |
US20160093108A1 (en) * | 2014-09-30 | 2016-03-31 | Sony Computer Entertainment Inc. | Synchronizing Multiple Head-Mounted Displays to a Unified Space and Correlating Movement of Objects in the Unified Space |
US20160309134A1 (en) * | 2015-04-19 | 2016-10-20 | Pelican Imaging Corporation | Multi-baseline camera array system architectures for depth augmentation in vr/ar applications |
US20170052595A1 (en) * | 2015-08-21 | 2017-02-23 | Adam Gabriel Poulos | Holographic Display System with Undo Functionality |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180017742A1 (en) * | 2016-05-11 | 2018-01-18 | Inneos LLC | Active optical cable for wearable device display |
US10454579B1 (en) | 2016-05-11 | 2019-10-22 | Zephyr Photonics Inc. | Active optical cable for helmet mounted displays |
US20180268585A1 (en) * | 2017-03-15 | 2018-09-20 | Boe Technology Group Co., Ltd. | Character input method, character input device, and wearable device |
US10621766B2 (en) * | 2017-03-15 | 2020-04-14 | Boe Technology Group Co., Ltd. | Character input method and device using a background image portion as a control region |
US20240345401A1 (en) * | 2018-03-29 | 2024-10-17 | Adlens Limited | Improvements in or relating to variable focusing power optical devices |
WO2020180859A1 (en) * | 2019-03-05 | 2020-09-10 | Facebook Technologies, Llc | Apparatus, systems, and methods for wearable head-mounted displays |
JP2022522579A (en) * | 2019-03-05 | 2022-04-20 | フェイスブック・テクノロジーズ・リミテッド・ライアビリティ・カンパニー | Equipment, systems, and methods for wearable head-mounted displays |
US11163333B2 (en) * | 2019-03-29 | 2021-11-02 | Htc Corporation | Head-mounted display |
CN114503059A (en) * | 2019-09-30 | 2022-05-13 | 美国斯耐普公司 | Automated eye-mounted device sharing system |
US20240134191A1 (en) * | 2020-05-26 | 2024-04-25 | Magic Leap, Inc. | Monovision display for wearable device |
JP7596333B2 (en) | 2022-06-30 | 2024-12-09 | キヤノン株式会社 | HEAD-MOUNTED DISPLAY DEVICE, IMAGE PROCESSING DEVICE, CONTROL METHOD FOR HEAD-MOUNTED DISPLAY DEVICE, AND PROGRAM |
US20250036344A1 (en) * | 2023-07-24 | 2025-01-30 | Google Llc | Wearable display with coupled mobile interface module |
Also Published As
Publication number | Publication date |
---|---|
US20200166761A1 (en) | 2020-05-28 |
CN109076165A (en) | 2018-12-21 |
US20180052329A1 (en) | 2018-02-22 |
US10473941B2 (en) | 2019-11-12 |
JP2022009355A (en) | 2022-01-14 |
EP3443738A1 (en) | 2019-02-20 |
EP3923124A2 (en) | 2021-12-15 |
EP3443738B1 (en) | 2021-08-25 |
EP3443738A4 (en) | 2019-07-17 |
JP6724164B2 (en) | 2020-07-15 |
WO2017190635A1 (en) | 2017-11-09 |
JP2020171031A (en) | 2020-10-15 |
CN109076165B (en) | 2021-02-12 |
US9823477B1 (en) | 2017-11-21 |
EP3923124A3 (en) | 2022-03-16 |
JP6966603B2 (en) | 2021-11-17 |
JP2019519968A (en) | 2019-07-11 |
CN113281905A (en) | 2021-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10473941B2 (en) | Head mounted display content capture and sharing | |
US11563886B2 (en) | Automated eyewear device sharing system | |
US9423619B2 (en) | Head mounted display and method of outputting a content using the same in which the same identical content is displayed | |
CN103049175B (en) | Preview screen rendering method, device and terminal | |
EP3460745B1 (en) | Spherical content editing method and electronic device supporting same | |
US10136056B2 (en) | Panoramic imaging apparatus and system, method of generating panoramic image using panoramic imaging system, computer-readable recording medium, and computer program stored in computer-readable recording medium | |
US11288871B2 (en) | Web-based remote assistance system with context and content-aware 3D hand gesture visualization | |
KR20180094340A (en) | Mobile terminal and method for controlling the same | |
US20150244756A1 (en) | Method, Apparatus and System for Determining Terminal That is to Share Real-Time Video | |
CN104869315A (en) | Photographing control method and terminal | |
CN109496293A (en) | Extend content display method, device, system and storage medium | |
EP4493991A1 (en) | Sharing received objects with co-located users | |
US20240397033A1 (en) | Hyper-connected and synchronized ar glasses | |
JP2025510949A (en) | Cloud desktop display method, apparatus, device and storage medium | |
KR102262019B1 (en) | Method and system for extended reality communication soliciting network speed | |
KR20170011934A (en) | Mobile terminal and control method for the mobile terminal | |
CN111091498A (en) | Image processing method, image processing apparatus, electronic device, and medium | |
WO2022181378A1 (en) | Image processing device, image processing method, and program | |
TWI592894B (en) | transmitting device having a function of transmitting programmable figures and programmable messages and related server, system, and method | |
KR20130012655A (en) | Supporter used in photographing 3-dimensional image and method for controlling photographing 3-dimensional image | |
TWM532130U (en) | Transmitting device having a function of transmitting programmable figures and programmable messages and related server and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FUTUREWEI TECHNOLOGIES, INC., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SHEN, TIAN;REEL/FRAME:039423/0231 Effective date: 20160613 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |