US20070120954A1 - Teleconferencing method and system - Google Patents
Teleconferencing method and system Download PDFInfo
- Publication number
- US20070120954A1 US20070120954A1 US11/537,177 US53717706A US2007120954A1 US 20070120954 A1 US20070120954 A1 US 20070120954A1 US 53717706 A US53717706 A US 53717706A US 2007120954 A1 US2007120954 A1 US 2007120954A1
- Authority
- US
- United States
- Prior art keywords
- station
- image
- participants
- situated
- recited
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 239000002131 composite material Substances 0.000 description 64
- 230000001953 sensory effect Effects 0.000 description 53
- 206010015037 epilepsy Diseases 0.000 description 15
- 230000005236 sound signal Effects 0.000 description 15
- 238000003860 storage Methods 0.000 description 15
- 230000005540 biological transmission Effects 0.000 description 14
- KRQUFUKTQHISJB-YYADALCUSA-N 2-[(E)-N-[2-(4-chlorophenoxy)propoxy]-C-propylcarbonimidoyl]-3-hydroxy-5-(thian-3-yl)cyclohex-2-en-1-one Chemical compound CCC\C(=N/OCC(C)OC1=CC=C(Cl)C=C1)C1=C(O)CC(CC1=O)C1CCCSC1 KRQUFUKTQHISJB-YYADALCUSA-N 0.000 description 13
- 230000000295 complement effect Effects 0.000 description 12
- 102100033118 Phosphatidate cytidylyltransferase 1 Human genes 0.000 description 9
- 101710178747 Phosphatidate cytidylyltransferase 1 Proteins 0.000 description 9
- 102100033126 Phosphatidate cytidylyltransferase 2 Human genes 0.000 description 9
- 101710178746 Phosphatidate cytidylyltransferase 2 Proteins 0.000 description 9
- 238000005034 decoration Methods 0.000 description 8
- 230000002708 enhancing effect Effects 0.000 description 8
- 230000006835 compression Effects 0.000 description 5
- 238000007906 compression Methods 0.000 description 5
- 230000006837 decompression Effects 0.000 description 4
- 230000004069 differentiation Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 239000004575 stone Substances 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 230000014759 maintenance of location Effects 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 239000000853 adhesive Substances 0.000 description 2
- 230000001070 adhesive effect Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 235000013305 food Nutrition 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 244000025254 Cannabis sativa Species 0.000 description 1
- 241000196324 Embryophyta Species 0.000 description 1
- 229920006328 Styrofoam Polymers 0.000 description 1
- 239000011111 cardboard Substances 0.000 description 1
- 239000000919 ceramic Substances 0.000 description 1
- 239000004927 clay Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 239000006260 foam Substances 0.000 description 1
- 239000005338 frosted glass Substances 0.000 description 1
- 229910052602 gypsum Inorganic materials 0.000 description 1
- 239000010440 gypsum Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 235000012054 meals Nutrition 0.000 description 1
- 230000003278 mimic effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000000465 moulding Methods 0.000 description 1
- 239000000123 paper Substances 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 229940085606 rembrandt Drugs 0.000 description 1
- 230000010076 replication Effects 0.000 description 1
- 239000008261 styrofoam Substances 0.000 description 1
- 239000002023 wood Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/142—Constructional details of the terminal equipment, e.g. arrangements of the camera and the display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/56—Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/56—Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities
- H04M3/567—Multimedia conference systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/147—Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
Definitions
- the present invention is related to a video conferencing system and method and, more particularly, to a teleconferencing system which is capable of producing a “video mirror” at a station such that any participants at one or more remote stations may be imaged and displayed in the video mirror at the station so that they appear to be present or face-to-face with any participants at the station.
- Visual telephone systems presently provide communication between at least two locations for allowing a video conference among participants situated at each station.
- An objective in some video conferencing arrangements is to provide a plurality of television cameras at one location. The outputs of those cameras are transmitted along with audio signals to a corresponding plurality of television monitors at a second location such that the participants at the first location are perceived to be present or face-to-face with participants at the second location.
- the number of conferees included in the video picture from each camera is normally limited to a few people, typically one to four.
- the apparatuses and methods employed heretofore to achieve proper positioning, focus and alignment have been complex and costly.
- the images captured by the plurality of cameras must be arranged and displayed so that they generate a non-overlapping and/or contiguous field of view, for example, as described in U.S. Pat. No. 4,890,314 which issued to Judd et al. on Dec. 26, 1989 and which is hereby incorporated by reference and made a part hereof.
- the prior art systems have also been deficient because they have failed to provide means for generating an image, such as an image of a plurality of participants, at one station, differentiating the image to provide a differentiated image and subsequently compositing the differentiated image with a predetermined composite image to provide a composited image which complements or becomes visually complementary, contiguous or integrated with the remote station when the image is displayed at the remote station.
- Another problem with prior art video conferencing systems is eye contact among participants at the stations.
- a camera is placed somewhere above the display monitor at which a participant is observing a display of the participant from the remote station. Consequently, the camera captures the participant at an angle above the participants viewing level or head. Thus, when an image of that participant is displayed at the remote station, it appears as if the participant is looking down (e.g., towards the ground).
- Previous solutions to this problem have required complex optical systems and methods using, for example, a plurality of lenses and mirrors. The solutions have usually been designed for use when the camera is capturing an image of a single participant, and they fall short when simultaneously capturing images of multiple participants.
- a primary object of the present invention to provide a face-to-face teleconferencing system which enables a plurality of participants at a plurality of stations to teleconference such that the participants generally appear face-to-face with one or more participants at remote stations in the teleconferencing system.
- Another object of the present invention is to provide a “video mirror” at a station.
- Yet another object of the invention is to provide an imaging system which provides a simplified means capturing substantially eye level images of participants at stations while also providing means for simultaneously displaying images at such stations.
- Still another object of this invention is to provide a system and method for compositing a plurality of signals corresponding to a plurality of images from at least one station to provide a contiguous or seamless composite image.
- Still another object is to provide a method and system for providing a plurality of teleconferencing stations that have complementary predetermined sensory settings which facilitate creating a face-to-face environment when images of such settings and participants are displayed at remote stations.
- Another object of the invention is to provide a method and apparatus for generating a video mirror such that an image having a predetermined sensory setting of participants or subjects captured at one station may be displayed at a remote station having a different predetermined sensory setting, yet the remote participants will appear face-to-face in the same predetermined setting as the participants or subjects at the one station.
- this invention comprises a teleconferencing method consisting of the steps of receiving image data at a teleconferencing station, the image data including data corresponding to at least one person and displaying an image corresponding to the image data at the teleconferencing station so that at least one participant at the teleconferencing station views said image of at least one person, thereby providing a face-to-face presence environment whereat the image of at least one person appears substantially life size.
- this invention comprises a conferencing system comprising a first station comprising a display having a viewing area, a table facing the viewing area, furniture on which participants may sit and face the viewing area, and an image system for capturing a first station image of at least a portion of the first station and of at least one of the participants, the viewing area being directly between the participants and the image system.
- this invention comprises a conferencing system comprising a plurality of walls defining a conferencing station, a display on at least one of the plurality of walls for providing a stationary or moving image for defining a desired video background during a teleconference.
- this invention comprises a conferencing system comprising a first station comprising a display having a viewing area, furniture comprising a table facing the viewing area and seating on which participants may sit and face the viewing area, and an image system for capturing a first station image of at least a portion of the first station and of at least one of the participants, the participants being situated a predetermined position relative to the image system.
- this invention comprises a method of causing participants in a video conference to be situated at a predetermined position relative to a display, the method comprising the steps of situating a table in relation to the display, situating a plurality of seats adjacent to the table, the tables and the plurality of seats being situated a predetermined distance of at least a focal distance of a camera associated with the display.
- FIGS. 1A and 1B taken together, show a teleconferencing system according to one embodiment of this invention
- FIG. 2 is a partly broken away top view of a first station of the teleconferencing system shown in FIG. 1A ;
- FIGS. 3A and 3B taken together, show another embodiment of the present invention wherein the stations have different predetermined sensory settings
- FIGS. 4A and 4B taken together, show still another embodiment of the invention having stations which have predetermined sensory settings which are designed, decorated and defined to be complementary and/or substantially identical;
- FIGS. 5A and 5B taken together, provide a visual illustration of the images corresponding to some of the signals generated by the teleconferencing system.
- FIGS. 6A-6D taken together, show a schematic diagram of a method according to an embodiment of this invention.
- a teleconferencing system 10 having a first station or suite 12 and a second station or suite 14 .
- the first station 12 comprises a first conference or sensory area 16
- the second station 14 comprises a second conference or sensory area 18 - 1 , respectively.
- the first and second stations 12 and 14 also comprise a first video area 20 and a second video area 22 - 1 , respectively, associated with the first and second conference areas 16 and 18 - 1 .
- the first video area 20 is generally integral with a wall 32 h in the first station 12 .
- the second video area 22 - 1 is generally integral with a wall 32 h - 1 in the second station 14 .
- the first and second stations are geographically remote from each other, but they could be situated on the same premises if desired.
- the first station 12 is shown assembled or constructed into a generally elongated octagonal shape.
- the first station 12 comprises a plurality of modular members 32 a - 32 h which include walls 32 a , 32 c - e , 32 g - h , doors in wall members 32 b and 32 f and entry facade 32 f - 321 .
- the first station 12 also comprises a ceiling 34 ( FIG.
- each of the members 32 a - 32 h and the ceiling 34 is molded or formed to provide or define an environment having a unique architectural setting and/or sensory setting.
- the wall member 32 a may be formed to provide a plurality of stones 36 , a plurality of columns 38 , and an arch 40 to facilitate defining a first predetermined setting 12 a having a Roman/Italian motif, theme or aura.
- One or more of the members 32 a - 32 h may be provided with inlays, wall decorations (like picture 58 in FIGS. 1A and 2 ), or even a permanent frosted glass window and frame arrangement 42 mounted therein.
- members 32 b and 32 f may be provided with sliding doors 44 which facilitate entering and exiting the first station 12 and which are designed to complement or further enhance the Roman/Italian motif.
- member 32 h ( FIGS. 1A and 2 ) is formed to provide a stone and pillar appearance and texture complementary to the stone and pillar appearance and texture of the wall members, such as member 32 a .
- the member 32 a may be shaped to frame or mask a rear projection screen 46 , as shown.
- the rear projection screen 46 comprises a high resolution lenticular rear projection screen which is either integral with or mounted directly to member 32 h to provide a first video area 20 having a usable projection area of about 52 inches by 92 inches with an associated aspect ratio of 16:9.
- Each of the members 32 a - 32 h and ceiling 34 are created in separate modular units using a plurality of molds (not shown).
- a suitable material for molding the members 32 a - 32 h and ceiling 34 to provide a granite-like appearance may be Gypsum, but they could be formed from other suitable material such as stone or clay-based materials, ceramic, paper, cardboard, foam, wood, Styrofoam and the like.
- the member 32 d may be provided with a shelf or mantle 33 .
- the various members 32 a - 32 h are assembled together as shown in FIG.
- first station 12 After the first station 12 is assembled and the ceiling 34 is secured thereto, it has a length of about 14 feet, 6 inches (indicated by double arrow L in FIG. 2 ) and a width of about 12 feet, 0 inches (indicated by double arrow W in FIG. 2 ).
- the first station 12 has an approximate height from floor to ceiling 34 of about 8 feet, 6 inches.
- the members 32 a , 32 c , 32 e and 32 g have a width (indicated by double arrow Y in FIG. 2 ) of about 5 feet, 0 inch.
- the back wall member 32 d and front wall member 32 h comprises a width of about 7 feet, 8 inches (indicated by double arrow X in FIG. 2 ).
- the first station 12 may be further decorated, designed or ornamented with a plurality of subjects, decorations or ornaments which facilitate providing the first predetermined sensory setting 12 a which defines a first aura, motif or theme.
- the second station 14 may be further provided or ornamented with a plurality of subjects, decorations or ornaments which facilitate providing a second predetermined sensory setting 14 a which defines a second aura, motif or theme. For example, as illustrated in FIG.
- the predetermined sensory setting 12 a of the first station 12 may be further decorated with a table 50 , table decorations, pillar and wall decorations, carpet (not shown), plants 54 and other wall decorations (not shown) to further enhance the Roman/Italian motif, theme or aura.
- the first and second predetermined sensory settings 12 a and 14 a may also comprise appropriate lighting fixtures 56 and appropriate furnishings, such as chairs 60 and tables 61 , which complement the predetermined setting to further facilitate defining the Roman/Italian theme or motif for the stations 12 and 14 .
- first and second stations 12 and 14 are assembled and ornamented or decorated to provide their respective first and second predetermined sensory settings 12 a and 14 a , they define an aura, theme or motif which facilitates providing or creating a very sensual and impressionable environment. Providing such a station, such as station 12 , with a strong sensory environment facilitates enhancing the virtual presence illusion created by teleconferencing system 10 of the present invention.
- first station 12 and second station 14 are shown in the embodiment in FIGS. 1A and 1B as having complementary or similar first and second predetermined sensory settings 12 a and 14 a , they could be provided with first and second predetermined sensory settings 12 a and 14 a having different themes, motifs or auras.
- first and second predetermined sensory settings 12 a and 14 a having different themes, motifs or auras.
- FIGS. 1A and 1B illustrate a first and second set of stations 12 and 14 having a Roman/Italian motif
- another set of stations such as station 12 ′ and station 14 ′ in the embodiment illustrated in FIGS. 3A and 3B , may have at least one station having a different predetermined setting.
- the second station 14 ′ in FIG. 3B provides a setting 14 a ′ which defines a Chinese aura, theme or motif.
- the members 32 a - 32 h , ceiling 34 and associated predetermined sensory setting are provided to be transportable and capable of being assembled at any suitable location, such as an existing rectangular room, suite or conference area having dimensions of at least 20 feet ⁇ 20 feet ⁇ 9 feet. While it may be desirable to provide the first and second stations 12 and 14 in the teleconferencing system 10 with substantially the same dimensions, it should be appreciated that they could be provided with differing dimensions, depending on, for example, the number of participants at each station. It should also be appreciated that the second station 14 and other stations described herein would preferably be manufactured and assembled in the same or similar manner as the first station 12 .
- the stations in the teleconference system 10 may be decorated with wall, ceiling and floor coverings to provide, for example, the first predetermined sensory setting 12 a without using the pre-formed or molded modular members 32 a - 32 h described above, although the use of such members may be preferable in this embodiment.
- the teleconferencing system 10 also comprises conferencing means or a conferencing system means for teleconnecting the first and second stations 12 and 14 together to facilitate capturing an image or images at one of said stations and displaying at least a portion of the image or a sub-image at another of the stations such that it becomes generally visually integrated with the predetermined sensory setting at that station, thereby facilitating creating a “video mirror” and a “face-to-face” environment for the participant situated at that station.
- the conferencing system associated with the first station 12 comprises image sensor means, imager or image sensors for sensing images at the first station 12 .
- the image sensor means comprises a plurality of cameras which are operably associated with the rear projection screen 46 of first station 12 .
- the plurality of cameras comprise a first camera head 62 and second camera head 64 which are operatively coupled to a first camera control unit 66 and second camera control unit 68 , respectively.
- the first and second camera control units 66 and 68 are remotely situated from the first and second camera heads 62 and 64 . This facilitates permitting the first and second cameras 62 and 64 to be placed directly in the projection path of the rear projection screen 46 , without substantially interfering with the video image being projected.
- the first camera head 62 and second camera head 64 are situated approximately 16 inches above the surface of table 50 which generally corresponds to the eye level of the seated participants situated at table 50 .
- the first and second cameras 62 and 64 are situated behind the rear projection screen 46 in operative relationship with a pair of 1-1 ⁇ 4 inch diameter openings 66 and 68 , respectively.
- the first and second cameras 62 and 64 are mounted on a suitable narrow or non-interfering bracket (not shown) such that they can be positioned behind the rear projection screen 46 in operative relationship with openings 66 and 68 , respectively.
- the first and second cameras 62 and 64 are 1-1 ⁇ 4 inch by 1-1 ⁇ 4 inch 3-CCD camera heads which generate images having an aspect ratio of about 3:4 and a picture resolution of about 494 ⁇ 700 pixels.
- One suitable 3-CCD camera heads 62 and 64 and associated camera control units 66 and 68 may be Model No. GP-US502 manufactured by Panasonic Broadcast and Television Systems Company of Japan. It should be appreciated that while the teleconferencing system 10 shown and described in relation to FIGS. 1A and 1B show image sensor means comprising a plurality of camera heads 62 and 64 and camera control units 66 and 68 situated at a station, a single camera may be used (as shown and described relative to the embodiment shown in FIGS.
- the camera heads 62 and 64 and associated camera control units 66 and 68 are configured and positioned at the first station 12 to facilitate providing maximum vertical eye contact among participates in the teleconference, while minimally interrupting the substantially life-size video projection on the rear projection screen 46 .
- the conferencing means also comprises a first differentiator or differential key generator 70 ( FIG. 1A ) and a second differentiator or differential key generator 72 , respectively.
- the camera control unit 66 generates an RGB analog signal I- 62 which is received by the first differentiator 70
- the camera control unit 68 generates an RGB signal I- 64 which is received by the second differentiator 72 .
- the first and second differentiators 70 and 72 provide means for processing the image signals generated by the camera control units 66 and 68 to remove or differentiate any undesired portion of the images corresponding to the signals I- 62 and I- 64 .
- the image of the participants situated at the first station 12 from at least a portion of the first predetermined sensory setting 12 a , such as the background behind the participants, in order to provide a differential signal VS- 1 that has that portion of the first predetermined sensory setting 12 A removed.
- This facilitates transmitting the video image of the participants at the first station 12 to the remote second station 14 and also facilitates compositing the image with other images, as described below.
- Suitable differentiators 70 and 72 may comprise the differential key generator shown and described in U.S. Pat. No. 4,800,432, issued on Jan. 24, 1989 to Barnett et al. and assigned to The Grass Valley Group, Inc., which is incorporated herein by reference and made a part hereof.
- the differential key generators 70 and 72 convert the I- 62 and I- 64 signals from RGB analog signals to digital image signals having corresponding images 104 and 106 ( FIG. 5A ), respectively.
- the differential key generators 70 and 72 compare the digital image signals to an associated differential reference signals DRS- 62 and DRS- 64 , respectively, which generally corresponds to images 108 and 110 in FIG. 5A .
- these images 108 and 110 comprise at least a portion of the first predetermined sensory setting 12 a such as the background.
- the differential reference signals DRS- 62 and DRS- 64 are stored in appropriate storage 74 and 76 ( FIG. 1A ) associated with the differential key generators 70 , 72 , respectively.
- the differential reference signals DRS- 62 and DRS- 64 comprise a reference frame of a video image grabbed by one or both cameras 62 or 64 situated at the first station 12 from a video sequence of the first predetermined sensory setting 12 a of the first station 12 background where no participants, chairs, or other foreground elements are in place.
- the first and second differentiators 70 and 72 generate differentiated video signals VS- 1 and VS- 2 ( FIG. 1A ), respectively.
- the VS- 1 and VS- 2 signals generally correspond to the individuals situated at the first station 12 when viewed in the direction of arrow A in FIG. 2 .
- the images 112 and 114 ( FIG. 5 ) associated with the VS- 1 and VS- 2 signals respectively, notice that the background area shown in images 104 and 106 has been removed and is tagged as a “zero” image area.
- tagging at least a portion of the image represented by the VS- 1 signal as “zero” background facilitates compressing the VS- 1 and VS- 2 signals and providing corresponding compressed CDS- 1 and CDS- 2 signals, thereby reducing the amount of transmission band width needed.
- This tagging also facilitates compositing or overlaying another predetermined image to provide a seamless composited image as described in detail below.
- the video signals VS- 1 and VS- 2 are received by a first compression/decompression means or CODEC 78 and a second compression/decompression means or CODEC 80 , respectively.
- the CODECs 78 and 80 also receive an audio signal AS-A 1 and AS-A 2 from suitable microphones 82 and 83 , respectively, which may be positioned or concealed at an appropriate location in the first station 12 , such as underneath or on top of table 50 , as illustrated in FIG. 1A .
- the function of the first and second CODEC 78 and 80 is to compress video and audio signals for transmitting to remote stations, such as the second station 14 , and also to decompress compressed video and audio signals received from remote stations.
- the CODECs 78 and 80 are configured with suitable compression and decompression algorithms which are known to those of ordinary skill in the art.
- the CODEC Model No. Rembrandt II VP available from Compression Labs, Inc. of San Jose, Calif. is suitable for use in the embodiment described herein, but it should be noted that other suitable compression/decompression means may be employed.
- the CODEC 78 receives the video signal VS- 1 and audio signal AS-A 1
- CODEC 80 receives the video signal VS- 2 and audio signal AS-A 2
- the CODECs 78 and 80 generate digital signals CDS- 1 and CDS- 2 , respectively, in response thereto which are in turn transmitted to remote station 14 via a transmission network 84 .
- the transmission network 84 may be configured as a private network, public circuit switch service, and it may utilize telecommunication and/or satellite technology.
- the transmission network 84 preferably includes a plurality of T- 1 lines (not shown) which are capable of accommodating bit streams having a suitable band width, such as 1.544 megabytes per second.
- the teleconferencing system 10 and conference means associated with the first station 12 also comprises enhancing means for enhancing the resolution of an image or sub-image received from a remote station, such as the second station 14 .
- enhancing means comprises a first line doubler 86 and a second line doubler 88 which are operatively coupled to the first CODEC 78 and second CODEC 80 , respectively.
- the first and second line doublers 86 and 88 enhance the resolution and picture quality of at least a portion of the image corresponding to video signals VS- 3 and VS- 4 received from the CODECs 78 and 80 , respectively, by about 50-150%.
- the VS- 3 and VS- 4 signals correspond to images or sub-images received from remote station(s), such as station 14 , as described in detail below.
- One suitable line doubler is the Model No. LD 100 available from Faroudja Laboratories, Inc. of Sunnyvale, Calif., but other suitable enhancing means may be provided to provide greater or less enhancement of the images to be displayed.
- suitable enhancing means may be provided to provide greater or less enhancement of the images to be displayed.
- lenses, mirrors, optical pixel interpolation or other electrical means may be employed as desired.
- the present invention may be performed without the use of any enhancing means without departing from the scope of the invention.
- the first and second line doublers 86 and 88 generate enhanced video signals which are input into compositing means, compositor or video compositing multiplexer 92 for compositing the enhanced video signals associated with the images or sub-images received from the remote station(s) with one or more predetermined composite signals, such as predetermined composite signal A, corresponding to a predetermined composite image or sub-image which are stored in a suitable storage device 94 associated with the compositor 92 .
- the predetermined composite signal A corresponds to an image of at least a portion of first predetermined sensory setting 12 a , such as the background of the first station 12 .
- the video compositing multiplexer 92 composites the signals received from the first and second line doublers 86 and 88 with the predetermined composite signal A and generates a RGB analog composite signal in response thereto. It has been found that Model No. E-Space-1 available from Miranda Technologies, Inc. of Montreal and Quebec, Canada, is one suitable video compositing multiplexer 92 .
- the teleconferencing system 10 comprises a projector 96 coupled to the video compositing multiplexer 92 which receives the RGB composite signal and projects a corresponding image 90 ( FIG. 1A ) corresponding to the composite signal on the rear projection screen 46 .
- the Model No. 3300 available from AMPRO Corporation of Titusville, Fla. has been found to be a suitable projector 96 .
- a liquid crystal display (LCD) or other electronic screen may be suitable to display images at a station. This may eliminate the need for the projector 96 .
- LCD liquid crystal display
- the projector 96 could be used with an optical system or a plurality of mirrors (not shown), or prisms (not shown) such that the projector can be positioned, for example, to the side or below the rear projection screen 46 or in a manner that permits the projector 96 to project the image towards a mirror (not shown), which causes the image to be projected on the rear projection screen 46 .
- the composite signal and its corresponding image 90 generally comprise a video image of at least a portion of the first predetermined sensory setting 12 a combined or composited with a differentiated image, such as an image of the participants from the second station 14 which correspond to the VS- 3 and VS- 4 ( FIG. 1B ) signals. Consequently, the resultant image 90 projected on screen 46 at the first station 12 complements or blends with the architectural motif, aura, theme or design defined by the first predetermined sensory setting 12 a at the first station 12 , such that the projected image 90 appears visually integrated with the first predetermined sensory setting 12 a of the first station 12 .
- the sub-images or images received from the remote station(s) typically have a resolution on the order of about 352 ⁇ 288 pixels and the predetermined composite signal A comprises a resolution on the order of about 1280 ⁇ 1024 pixels.
- the resultant composite image 90 may comprise, for example, an image of the participants situated at the second station 14 having a first resolution and a background image of the first station 12 having a second resolution, which is higher than the first resolution. This enables compositor 92 to provide a composite image 90 which, when displayed on screen 46 , gives the illusion or effect of a “video mirror” to the participants situated at the first station 12 .
- the teleconferencing system 10 also includes audio means comprising a plurality of speakers 100 and 102 ( FIGS. 1A and 2 ) which, in turn, receive audio signals AS-B 1 and AS-B 2 from CODECs 78 and 80 , respectively.
- the audio signal AS-B 1 and AS-B 2 generally correspond to the audio associated with the sound (e.g., voices, music and the like) associated with the remote station(s), such as second station 14 .
- the rear projection screen 46 and projector 96 are configured and selected to enable the teleconferencing system 10 to project the composited image 90 ( FIG. 1A ) at a predetermined scale, such as substantially full scale.
- the compositor 92 comprises a scaler 95 which is integral therewith for scaling the composited signal associated with the composited image 90 to a desired or predetermined scale, such as substantially full scale.
- the second station 14 comprises similar components as the first station and such like components are labelled with the same reference numeral as their corresponding component in the first station 12 , except that the components associated with the second station 14 have a “ ⁇ 1” designator added thereto. Such components operate and function in substantially the same manner as described above with regard to the first station 12 with the following being some differences.
- the differential reference signals DRS- 3 and DRS- 4 ( FIG. 5 ) associated with the second station 14 generally correspond to an image or sub-image of at least a portion of the second predetermined sensory setting 14 a , such as the background 98 - 1 , of the second station 14 .
- Such sub-image or image may include at least a portion of the background 98 - 1 without any participants, chairs or other foreground subjects situated in the second station 14 .
- a predetermined composite signal B may be stored in the storage 94 - 1 associated with the compositor 92 - 1 second station 14 .
- the predetermined composite signal B may correspond to an image or sub-image of at least a portion of the second predetermined sensory setting 14 a of the second station 14 .
- Such sub-image or image may include, for example, an image of the walls 32 a - 1 to 32 h - 1 and conference area 18 or background of the second station 14 . Notice that in the embodiment shown in FIGS.
- the second station 14 has a second predetermined sensory setting 14 a which mirrors or is complementary to the first predetermined sensory setting 12 a .
- the first and second predetermined sensory settings 12 a and 14 a may be different.
- FIGS. 6A-6D A method of operating the teleconferencing system 10 will now be described in relation to FIGS. 6A-6D .
- the modular components such as members 32 a to 32 h and ceiling 34 for first station 10 , decorations and the like, are configured, assembled and decorated (block 99 in FIG. 6A ) at a desired location to provide a conference station comprising a predetermined sensory setting defining a predetermined theme, motif or aura.
- the theme, motif or aura may be complementary (as shown in FIGS. 1A and 1B ) or they can be completely different, as shown in FIGS. 3A and 3B (described below).
- FIGS. 3A and 3B As ease of illustration, it will be assumed that the stations are assembled and decorated as shown and described relative to the embodiment in FIGS. 1A and 1B .
- the modular stations 12 and 14 may be desired (decision point 101 in FIG. 6A ) to use differentiator (e.g., differentiator 72 in FIG. 1A ). As discussed herein relative to the embodiments shown in FIGS. 4A and 4B , it may not always be desired to generate a differential reference image, thereby making it unnecessary to generate the differential reference signal. If differentiation is desired, then the camera heads 62 or 64 generate at least one video image (block 103 ) of at least a portion of the first predetermined sensory setting 12 A at the first station 12 .
- the differentiators 72 and 74 grab or capture at least one differential reference image or sub-image from those images and generate (block 107 ) the differential reference signals DRS- 62 and DRS- 64 , respectively. These signals are stored in suitable storage 74 and 76 for use by the differentiators 70 and 72 , respectively. Likewise, cameras 62 - 1 and 64 - 1 at the second station 14 generate video images of at least a portion of the second predetermined setting 14 a at the second station 14 . The differentiators 70 - 1 and 72 - 1 grab or capture at least one differential reference image or sub-image from those images and generate differential reference signals (not shown) corresponding thereto. These signals are then stored (block 109 ) in suitable storage 74 - 1 and 76 - 1 for use by differential key generators 70 - 1 and 72 - 1 , respectively.
- the differential reference signals DRS- 62 and DRS- 64 comprise an image of at least a portion of the first predetermined sensory setting 12 a , such as an image of the first station 12 without any participants, chairs or other subjects which are not stationary during the teleconference.
- the differential reference signals associated with the differentiators 70 - 1 and 72 - 1 comprise at least a portion of the second predetermined sensory setting 14 a at the second station 14 , such as an image of the background 98 - 1 without the participants, chairs and other subjects which are not stationary during the teleconference.
- Such predetermined composite image would preferably include a substantial portion of the first predetermined sensory setting 12 a , including the background and/or conference area 16 of the first station 12 . If compositing is desired, then the predetermined composite signal A is generated (block 111 in FIG. 6B ). The corresponding predetermined composite signal A may then be stored in suitable storage 94 .
- the predetermined composite image at the second station 14 and corresponding predetermined composite signal B may be generated and stored as predetermined composite signal B in suitable storage 94 - 1 .
- the predetermined composite image associated with the second station 14 includes an image of at least a portion of the second predetermined sensory setting 14 a , including the background 98 - 1 .
- the predetermined composite signals A and B are generated by a suitable still camera (not shown) to provide a still image (not shown) of the station 12 or 14 being photographed.
- the still image would subsequently be scanned and digitized for storage by a suitable scanner (not shown).
- the still camera and scanner would preferably be capable of generating images having a resolution on the order of about 1280 ⁇ 1024 pixels.
- the resultant composite image (such as image 90 in FIG. 1A ) may comprise an image having a high resolution background, for example, combined with a comparatively lower resolution image of the remote station participants. This, in turn, facilitates enhancing the “video mirror” effect wherein a mimic or replication of a common architectural technique of mirroring a wall of a given room which makes the overall room appear to be extended beyond its actual wall line.
- the first and second suites 12 and 14 may then be teleconnected (block 113 ) or connected by satellite or other suitable means via the transmission network 84 .
- one or more participants may be situated at the first and second stations 12 and 14 .
- the predetermined distance B generally corresponds to a preferred or optimum focal distance at which optimum imaging by cameras 62 and 64 may be performed. In the embodiment being described, it has been found that the predetermined distance should be about 5 feet, 6 inches.
- the participants are situated at the second station 14 in a similar manner and the face-to-face teleconference may then begin.
- the first and second cameras 62 and 64 capture (block 117 in FIG. 6B ) live images of the participants situated at the first station 12 and generate corresponding RGB analog signals I- 62 and I- 64 which are received by the differential key generators 70 and 72 , respectively. If differentiation was selected (decision point 147 in FIG. 6C ), processing continues at block 119 otherwise it proceeds at block 123 .
- the differential key generators 70 and 72 generate (block 121 in FIG. 6C ) the digital differential signal VS- 1 and VS- 2 , respectively, after comparing (block 119 in FIG. 6C ) the I- 62 and I- 64 signals received from cameras 62 and 64 to their respective differential reference signals DRS 62 and DRS- 64 which are received from storages 74 and 76 .
- the differential signals VS- 1 and VS- 2 are then received by CODECs 78 and 80 which also receive the audio signals AS-A 1 and AS-A 2 which correspond to the audio, including sounds, music and voices, associated with the first station 12 .
- the CODECs 78 and 80 digitize the audio signals AS-A 1 and AS-A 2 , combine the audio signals with their respective video signal VS- 1 or VS- 2 , and generate (block 123 ) the compressed CDS- 1 and CDS- 2 signals in response thereto.
- the CDS- 1 and CDS- 2 signals are then transmitted (block 125 ) to the second station 14 via the transmission network 84 ( FIG. 1B ).
- the CDS- 1 and CDS- 2 signals are received and decompressed (block 127 in FIG. 6C ) by CODECs 78 - 1 and 80 - 1 , respectively, associated with the second station 14 to provide decompressed VS- 1 and VS- 2 signals.
- the CODECs 78 - 1 and 80 - 1 also decompress the audio signals AS-A 1 and AS-A 2 received from the first station 10 which are transmitted to speakers 100 - 1 and 102 - 1 , respectively, at the second station 14 .
- CODECs 78 - 1 and 80 - 1 decompress the CDS- 1 and CDS- 2 signals to provide VS- 1 and VS- 2 signals.
- the decompressed video signals VS- 1 and VS- 2 are then received by line doublers 86 - 1 and 88 - 1 . If it is desired to enhance the signals (decision point 129 ), then the line doublers 86 - 1 and 88 - 1 process or manipulate the signals (block 131 ) in order to enhance the resolution of the image corresponding to those signals.
- the video compositor 92 - 1 composites images (block 135 ) corresponding to those signals with at least one predetermined composite image, such as image 122 ( FIG. 5B ) corresponding to the predetermined composite signal B provided from storage 94 - 1 ( FIG. 1B ) to provide a composite signal.
- the composite signal generally corresponds to the composited image 91 - 1 to be displayed on the rear projection screen 46 - 1 at the second station 14 .
- the compositor 92 - 1 may (decision point 137 , block 139 in FIG. 6D ) scale the composited image to a desired scale, such as full scale, using scaler 95 - 1 . Thereafter, the compositor 95 - 1 transmits a corresponding RGB analog signal to projector 96 - 1 which displays (block 141 ) the scaled, composited image on the rear projection screen 46 - 1 ( FIG. 1B ).
- the teleconference may then be continued or terminated as desired (decision point 143 , block 145 ).
- the composited image is substantially full scale when projected and includes a high resolution image of at least a portion of the second predetermined sensory setting 14 a .
- the image appears to blend or become visually integrated with the second predetermined sensory setting 14 a . This, in turn, gives the participants situated at the second station 14 the perception that the first station participants are present or face-to-face with them in the second station 14 .
- images and signals relative to the second station 14 images are captured, processed and displayed at the first station 12 . So that images of the participants at the second station 14 are displayed at the first station 12 such that they appear to have a face-to-face presence at the first station 12 .
- images of the second station 14 participants may be differentiated and composited such that, when they are displayed at the first station 12 , the image completes or provides “the other half” of the first station 12 and becomes generally visually integrated therewith.
- it may be desirable to enhance the face-to-face presence by providing, for example, first and second predetermined sensory settings 12 a and 14 a which define a dining environment wherein food or meals may be served.
- the face-to-face presence may be further enhanced if the participants at both stations 12 and 14 order food and drinks from identical menus.
- trained maitre-de and/or waiters may be used to actively promote the perception of a face-to-face dinner using a scripted dialog and interaction with remote participants, maitre-de and/or waiters.
- the stations 12 and 14 may be used by the same or different participants without the need to reconstruct or re-assemble the stations.
- FIGS. 5A and 5B provide a visual illustration of the images corresponding to some of the signals described above utilizing the method and embodiment described above.
- images 104 and 106 generally correspond to the actual images captured by the first and second cameras 62 and 64 , respectively.
- associated image signals I- 62 and I- 64 are transmitted to the differential key generators 70 and 72 , respectively.
- the differential key generators 70 and 72 compare the images 104 and 106 to the images 108 and 110 associated with the differential reference signals DRS- 62 and DRS- 64 which are received from storages 74 and 76 , respectively, and which were previously generated by cameras 62 and 64 from an identical fixed camera position.
- the differential key generators 70 and 72 generate differential signals VS- 1 and VS- 2 which have corresponding images 112 and 114 .
- these images 112 and 114 comprise an image of the participants which are situated at the first station 12 with the background area having been removed or tagged as a “zero” area.
- this “zero” area becomes “filled-in” with the desired or predetermined composite image which may include, for example, an image of at least a portion of the predetermined setting or background of the second station 14 .
- the video signals VS- 1 and VS- 2 are fed into CODECs 78 and 80 which compresses the signals along with audio signal AS-A 1 and AS-A 2 and generates signals CDS- 1 and CDS- 2 .
- the CDS- 1 and CDS- 2 signals are then transmitted, via transmission network 84 , to the second station 14 and received by the CODECs 78 - 1 and 80 - 1 associated with the second station 14 .
- the CODEC 78 - 1 and 80 - 1 decompresses the CDS- 1 and CDS- 2 signals, respectively, from the first station 12 and feeds them into associated line doublers 86 - 1 and 88 - 1 .
- the line doublers 86 - 1 and 88 - 1 facilitate enhancing the images associated with the video signals to provide enhanced video signals EVS- 1 and EVS- 2 ( FIG. 5B ), respectively.
- the enhanced video signals EVS- 1 and EVS- 2 are then received by the video compositing multiplexer 92 - 1 associated with the second station 14 wherein the signals are combined to provide an intermediate composite signal ICS having an associated intermediate composite signal image 120 having an aspect ratio of about 8:3.
- the video compositing multiplexer 92 - 1 also receives the predetermined composite signal B having a predetermined composite signal B image 122 from storage 94 - 1 .
- the video compositing multiplexer 92 - 1 composites or combines the images 120 and 122 to generate the composite signal having an associated or corresponding composite image 124 as shown in FIG. 5B .
- the predetermined composite signal B image 122 generally corresponds to at least a portion of the predetermined setting or background of the second station 14 and has an aspect ratio of 16:9.
- the video compositing multiplexer 92 - 1 causes the “zero” area of the intermediate composite signal image 120 to be “filled in” with the predetermined composite signal B image.
- the composite image 124 may then be scaled to a predetermined size or scale, such as full scale, using scaler 94 - 1 , so that the composite image 124 may be scaled to a substantially full scale or real-life size image as desired.
- the composite image signal corresponding to the composite image 124 is transmitted to the projector 96 - 1 and then displayed on the rear projection screen 46 - 1 at the second station 14 .
- the composite image 124 may be appropriately framed or masked (such as with an archway 125 in FIGS. 1B and 5B ) when it is projected at the second station 14 to enhance the face-to-face, real time environment.
- the audio and video signals transmitted between the first and second stations 12 and 14 may be, in this illustration, transmitted over separate T- 1 lines (not shown) in the transmission network 84 in order to effect a substantially simultaneous and/or “real time” video conference.
- the participants may be geographically remotely located, yet the participants situated at the first station 12 will feel as if the second station 14 participants are located face-to-face or present with them at the first station 12 , while the participants situated at the second station 14 will feel as if the first station participants are face-to-face or present with them at the second station.
- predetermined composite signal B and associated predetermined composite signal image 122 when the predetermined composite signal B and associated predetermined composite signal image 122 is composited with the intermediate composite signal and associated intermediate composite signal image 120 , it overlays that signal to provide a seamless composite image 124 , which facilitates reducing or eliminating the need to match up the borders or seams of the camera images with any high degree of accuracy.
- cameras 62 and 64 and 62 - 1 and 64 - 1 preferably be situated such that they capture an entire participant rather than, for example, half of a participant. Thus, it may be desired to position the participants in a location such that any particular participants will not be in the field of view of more than one camera.
- the invention provides an apparatus and method for providing a video mirror at each station 12 and 14 which facilitates creating a face-to-face and non-interrupted image of any participants in the video conference. Because the image of the participants is differentiated, less transmission bandwidth, computer memory and the like is required. Also, the differentiators and compositors of the present invention enable a user to create a composite image 124 ( FIG. 5B ) having at least a portion thereof imaged at a greater resolution than the portion which was transmitted over transmission network 84 . This facilitates reducing the effect of limitations or transmission restrictions of the transmission network 84 which, in turn, facilitates increasing the quality of images displayed at a station.
- the composite image 124 may have an aspect ratio which is different from the aspect ratio of the cameras 62 and 64 .
- This enables the system and method of the present invention to utilize cameras which generate images having smaller or even larger aspect ratios.
- This also enables the system and method to use cameras having standard or common aspect ratios, such as 4:3.
- FIGS. 3A and 3B when taken together, illustrate another embodiment of the invention.
- the operation and components of the embodiment shown in FIGS. 3A and 3B are substantially the same as the operation of components of the embodiment described above relative to FIGS. 1A and 1B with the same reference numerals being used for the same components with the addition of single prime (′) designator. Consequently this embodiment is similar to the embodiment shown in FIGS. 1A and 1B , except that the second predetermined setting 14 a ′ in FIG. 3B and its associated theme, aura or motif is substantially different from the second predetermined setting 14 a shown in FIG. 1B .
- the first predetermined sensory setting 12 a ′ comprises a plurality of decorations 120 defining the Chinese theme, motif or aura.
- the predetermined composite signal A stored in storage 94 - 1 ′ and the differential reference signals stored in storages 74 - 1 ′ and 76 - 1 would generally correspond to an image of at least a portion of that setting 14 a′.
- an image of the participants situated at the first station 12 ′ is composited by compositor 92 - 1 ′ with a predetermined composite image of at least a portion of the second predetermined sensory setting 14 a ′ of the second station 14 ′ and projected onto the rear projection screen 46 - 1 ′ at the second station 14 ′.
- the first station 12 ′ participants appear to be face-to-face with the second station 14 ′ participants because they have a relatively high resolution video image behind them which complements or becomes integrated with the second predetermined sensory setting 14 a ′.
- the image 91 - 1 ′ ( FIG. 3B ) of the ladies at the first station 12 ′ includes a Chinese background which blends or complements the actual predetermined sensory setting 14 a′.
- the image of the participants situated at the second station 14 ′ is projected on the rear projection screen 46 ′ at the first station 12 ′, they appear to be in the same room as the participants situated at the first station 12 ′ because the Roman/Italian video background which is seen behind the second station 14 ′ participants generally complements and becomes visually integrated with the actual Roman/Italian theme, motif or aura defined by the first predetermined sensory setting 12 ′ of the first station 12 ′.
- FIGS. 4A and 4B when taken together, illustrate another embodiment of the invention.
- the components of the embodiment shown in FIGS. 4A and 4B which are substantially identical to the components in the embodiment shown in FIGS. 1A and 1B which have the same reference numerals with the addition of a double prime (“′′”) designators.
- ′′ double prime
- two remote modular stations such as stations 12 ′′ and 14 ′′ may be provided and designed to have first and second predetermined sensory settings 12 a ′′ and 14 a ′′ which are substantially identical.
- images may be captured in the manner described above at station 12 ′′ received by CODECs 78 ′′ and 80 ′′ and then transmitted, via transmission 84 ′′, to associated CODECs 78 - 1 ′′ and 80 - 1 ′′, respectively.
- the CODECs 78 - 1 ′′ and 80 - 1 ′′ then generate a decompressed signal which may be enhanced by line doublers 86 - 1 ′′ and 88 - 1 ′′, respectively; scaled to an appropriate scale by scaler 95 - 1 ′′; and then projected by projector 96 - 1 ′′ onto rear projection screen 46 - 1 ′′.
- this embodiment does not utilize the differentiating and compositing features of the previous embodiment, but may still achieve a face-to-face conference environment because the second predetermined sensory setting 14 a ′′ is configured to be identical to or complementary with the first predetermined sensory setting 12 a ′′.
- entire images or sub-images of the stations 12 and 14 are displayed at remote station(s). Because the stations 12 ′′ and 14 ′′ are assembled, decorated and designed to be complementary or identical, they appear visually integrated to participants situated in the stations 12 and 14 .
- the first and second predetermined sensory settings 12 a ′′ and 14 a ′′ are designed and arranged in a geometric fashion such that as cameras 62 ′′ and 64 ′′ capture images of the participants, they also capture images of the first and second predetermined sensory setting 12 a ′′ and 14 a ′′, respectively, at the most advantageous perspective for display at the remote station(s).
- this causes the first station 12 ′′ participants to perceive that the second station 14 ′′ participants are situated or present with the first station 12 ′′ participants at the first station 14 ′′.
- first station 12 ′′ participants appear to be face-to-face with the second station 14 ′′ participants at the second station 14 ′′ when the images associated with the first station 12 ′′ are displayed on screen 46 - 1 ′′. Consequently, by providing complementary or identical first and second predetermined sensory settings 12 a ′′ and 14 a ′′, a face-to-face conference may be created. As with previous embodiments, it may also be desired to differentiate, enhance, composite or scale the images as described with previous embodiments, but this is not required with the embodiment being described.
- stations can be provided with predetermined settings which are completely different, yet, by utilizing the apparatus and method of the present invention, the images of the participants in these stations may be projected at remote stations so that they appear to be virtually face-to-face with the remote station participants at one or more remote stations.
- the screen 46 for station 12 has been shown as being integral with a portion of a wall 32 h ( FIGS. 1A and 2A ), it could comprise a larger or smaller portion of that wall 32 h , or it could be provided as part of one or more other walls, or even as part of the ceiling 34 .
- one or more of the compositors may comprise a stationary or moving image database (not shown) for providing a plurality of predetermined composite signals which define a particular or desired video background.
- a stationary or moving image database for providing a plurality of predetermined composite signals which define a particular or desired video background.
- participants may elect to use the arched background of their proximity, choose an event-related scene, or decide to meet in a setting completely unrelated to their site or station.
- a station having a Manhattan eatery motif may be provided with a screen configured as a window (not shown).
- Certain moving video backgrounds of a busy New York avenue may be deposited and displayed on the screen to give the illusion that the participants situated at the station are dining in a popular Manhattan eatery.
- the predetermined settings could define any type of aura, theme or motif which is suitable for video conferencing and in which it is desired to provide a “real-life” or face-to-face presence illusion.
- the apparatus and method of this invention could be used in a business setting, education setting, seminar setting, home environment, religious setting, celebration setting (such as a birthday, retirement party, holiday or anniversary), or any other suitable setting as desired.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Telephonic Communication Services (AREA)
- Input Circuits Of Receivers And Coupling Of Receivers And Audio Equipment (AREA)
- Cosmetics (AREA)
- Computer And Data Communications (AREA)
Abstract
Description
- This application is a continuation of application Ser. No. 10/716,972, filed Nov. 19, 2003; U.S. Pat. No. 7,116,350 issued Oct. 3, 2006; which is a continuation of application Ser. No. 10/233,363 filed Aug. 30, 2002; now U.S. Pat. No. 6,654,045 issued Nov. 25, 2003; which is a continuation of application Ser. No. 09/695,460 filed Oct. 24, 2000, now U.S. Pat. No. 6,445,405 issued Set. 3, 2002, which is a continuation of application Ser. No. 09/047,771 filed Mar. 25, 1998, now U.S. Pat. No. 6,160,573, which is a continuation of Ser. No. 08/740,839 filed Nov. 4, 1996, now U.S. Pat. No. 5,751,337, which is a continuation of Ser. No. 08/308,603 filed Sep. 19, 1994, issued as U.S. Pat. No. 5,572,248, all of which are incorporated herein by reference and made a part hereof.
- 1. Field of the Invention
- The present invention is related to a video conferencing system and method and, more particularly, to a teleconferencing system which is capable of producing a “video mirror” at a station such that any participants at one or more remote stations may be imaged and displayed in the video mirror at the station so that they appear to be present or face-to-face with any participants at the station.
- 2. Description of Related Art
- Visual telephone systems presently provide communication between at least two locations for allowing a video conference among participants situated at each station. An objective in some video conferencing arrangements is to provide a plurality of television cameras at one location. The outputs of those cameras are transmitted along with audio signals to a corresponding plurality of television monitors at a second location such that the participants at the first location are perceived to be present or face-to-face with participants at the second location. In achieving good face-to-face presence, the number of conferees included in the video picture from each camera is normally limited to a few people, typically one to four. There are usually a like number of monitors at the receiving station, each strategically focused, aligned and positioned so that their displays appear contiguous, seamless and properly aligned. The apparatuses and methods employed heretofore to achieve proper positioning, focus and alignment have been complex and costly.
- Further, the images captured by the plurality of cameras must be arranged and displayed so that they generate a non-overlapping and/or contiguous field of view, for example, as described in U.S. Pat. No. 4,890,314 which issued to Judd et al. on Dec. 26, 1989 and which is hereby incorporated by reference and made a part hereof.
- The prior art systems have also been deficient because they have failed to provide means for generating an image, such as an image of a plurality of participants, at one station, differentiating the image to provide a differentiated image and subsequently compositing the differentiated image with a predetermined composite image to provide a composited image which complements or becomes visually complementary, contiguous or integrated with the remote station when the image is displayed at the remote station.
- Another problem with prior art video conferencing systems is eye contact among participants at the stations. Typically, a camera is placed somewhere above the display monitor at which a participant is observing a display of the participant from the remote station. Consequently, the camera captures the participant at an angle above the participants viewing level or head. Thus, when an image of that participant is displayed at the remote station, it appears as if the participant is looking down (e.g., towards the ground). Previous solutions to this problem have required complex optical systems and methods using, for example, a plurality of lenses and mirrors. The solutions have usually been designed for use when the camera is capturing an image of a single participant, and they fall short when simultaneously capturing images of multiple participants.
- The prior art stations themselves were not architecturally designed in a modular form so that they could be easily assembled, decorated and combined with a video image or sub-image from the remote station in a manner which would enhance the virtual presence environment.
- It is, therefore, a primary object of the present invention to provide a face-to-face teleconferencing system which enables a plurality of participants at a plurality of stations to teleconference such that the participants generally appear face-to-face with one or more participants at remote stations in the teleconferencing system.
- Another object of the present invention is to provide a “video mirror” at a station.
- Yet another object of the invention is to provide an imaging system which provides a simplified means capturing substantially eye level images of participants at stations while also providing means for simultaneously displaying images at such stations.
- Still another object of this invention is to provide a system and method for compositing a plurality of signals corresponding to a plurality of images from at least one station to provide a contiguous or seamless composite image.
- Still another object is to provide a method and system for providing a plurality of teleconferencing stations that have complementary predetermined sensory settings which facilitate creating a face-to-face environment when images of such settings and participants are displayed at remote stations.
- Another object of the invention is to provide a method and apparatus for generating a video mirror such that an image having a predetermined sensory setting of participants or subjects captured at one station may be displayed at a remote station having a different predetermined sensory setting, yet the remote participants will appear face-to-face in the same predetermined setting as the participants or subjects at the one station.
- In one aspect, this invention comprises a teleconferencing method consisting of the steps of receiving image data at a teleconferencing station, the image data including data corresponding to at least one person and displaying an image corresponding to the image data at the teleconferencing station so that at least one participant at the teleconferencing station views said image of at least one person, thereby providing a face-to-face presence environment whereat the image of at least one person appears substantially life size.
- In another aspect this invention comprises a conferencing system comprising a first station comprising a display having a viewing area, a table facing the viewing area, furniture on which participants may sit and face the viewing area, and an image system for capturing a first station image of at least a portion of the first station and of at least one of the participants, the viewing area being directly between the participants and the image system.
- In yet another aspect this invention comprises a conferencing system comprising a plurality of walls defining a conferencing station, a display on at least one of the plurality of walls for providing a stationary or moving image for defining a desired video background during a teleconference.
- In still another aspect this invention comprises a conferencing system comprising a first station comprising a display having a viewing area, furniture comprising a table facing the viewing area and seating on which participants may sit and face the viewing area, and an image system for capturing a first station image of at least a portion of the first station and of at least one of the participants, the participants being situated a predetermined position relative to the image system.
- In yet another aspect this invention comprises a method of causing participants in a video conference to be situated at a predetermined position relative to a display, the method comprising the steps of situating a table in relation to the display, situating a plurality of seats adjacent to the table, the tables and the plurality of seats being situated a predetermined distance of at least a focal distance of a camera associated with the display.
- These advantages and objects, and others, may be more readily understood in connection with the following specification, claims and drawings.
-
FIGS. 1A and 1B , taken together, show a teleconferencing system according to one embodiment of this invention; -
FIG. 2 is a partly broken away top view of a first station of the teleconferencing system shown inFIG. 1A ; -
FIGS. 3A and 3B , taken together, show another embodiment of the present invention wherein the stations have different predetermined sensory settings; -
FIGS. 4A and 4B , taken together, show still another embodiment of the invention having stations which have predetermined sensory settings which are designed, decorated and defined to be complementary and/or substantially identical; -
FIGS. 5A and 5B , taken together, provide a visual illustration of the images corresponding to some of the signals generated by the teleconferencing system; and -
FIGS. 6A-6D , taken together, show a schematic diagram of a method according to an embodiment of this invention. - Referring now to
FIGS. 1A and 1B , ateleconferencing system 10 is shown having a first station orsuite 12 and a second station orsuite 14. Thefirst station 12 comprises a first conference orsensory area 16, and thesecond station 14 comprises a second conference or sensory area 18-1, respectively. The first andsecond stations first video area 20 and a second video area 22-1, respectively, associated with the first andsecond conference areas 16 and 18-1. Thefirst video area 20 is generally integral with awall 32 h in thefirst station 12. Likewise, the second video area 22-1 is generally integral with awall 32 h-1 in thesecond station 14. In the embodiment being described, the first and second stations are geographically remote from each other, but they could be situated on the same premises if desired. - For ease of illustration, the construction and modular assembly of the stations in
teleconferencing system 10 will be described in relation to thefirst station 12. As shown in the sectional top view ofFIG. 2 , thefirst station 12 is shown assembled or constructed into a generally elongated octagonal shape. Thefirst station 12 comprises a plurality ofmodular members 32 a-32 h which includewalls wall members entry facade 32 f-321. Thefirst station 12 also comprises a ceiling 34 (FIG. 1A ) which is mounted on themembers 32 a-32 h with suitable fasteners, such as nuts, bolts, adhesives, brackets, or any other suitable fastening means. Notice that theceiling 34 has a dropped orsunken portion 34 a which supportsappropriate lighting fixtures 56. - In the embodiment being described, each of the
members 32 a-32 h and theceiling 34 is molded or formed to provide or define an environment having a unique architectural setting and/or sensory setting. For example, as illustrated inFIG. 1A , thewall member 32 a may be formed to provide a plurality ofstones 36, a plurality ofcolumns 38, and an arch 40 to facilitate defining a first predetermined setting 12 a having a Roman/Italian motif, theme or aura. One or more of themembers 32 a-32 h may be provided with inlays, wall decorations (likepicture 58 inFIGS. 1A and 2 ), or even a permanent frosted glass window andframe arrangement 42 mounted therein. Furthermore,members FIG. 2 ) may be provided with slidingdoors 44 which facilitate entering and exiting thefirst station 12 and which are designed to complement or further enhance the Roman/Italian motif. - In the embodiment being described, notice that
member 32 h (FIGS. 1A and 2 ) is formed to provide a stone and pillar appearance and texture complementary to the stone and pillar appearance and texture of the wall members, such asmember 32 a. Also, themember 32 a may be shaped to frame or mask arear projection screen 46, as shown. The function and operation of therear projection screen 46 will be described later herein. In the embodiment being described, therear projection screen 46 comprises a high resolution lenticular rear projection screen which is either integral with or mounted directly tomember 32 h to provide afirst video area 20 having a usable projection area of about 52 inches by 92 inches with an associated aspect ratio of 16:9. - Each of the
members 32 a-32 h andceiling 34 are created in separate modular units using a plurality of molds (not shown). In the embodiment being described, a suitable material for molding themembers 32 a-32 h andceiling 34 to provide a granite-like appearance may be Gypsum, but they could be formed from other suitable material such as stone or clay-based materials, ceramic, paper, cardboard, foam, wood, Styrofoam and the like. As illustrated in 1A and 2, themember 32 d may be provided with a shelf ormantle 33. Thevarious members 32 a-32 h are assembled together as shown inFIG. 2 and secured together with suitable support braces 48 which may be secured to thewalls 32 a-32 h with any suitable fastener such as screws, bolts, an adhesive or the like. After thefirst station 12 is assembled and theceiling 34 is secured thereto, it has a length of about 14 feet, 6 inches (indicated by double arrow L inFIG. 2 ) and a width of about 12 feet, 0 inches (indicated by double arrow W inFIG. 2 ). Thefirst station 12 has an approximate height from floor toceiling 34 of about 8 feet, 6 inches. Further, themembers FIG. 2 ) of about 5 feet, 0 inch. Finally, theback wall member 32 d andfront wall member 32 h comprises a width of about 7 feet, 8 inches (indicated by double arrow X inFIG. 2 ). - After the
members 32 a-32 h andceiling 34 are assembled, thefirst station 12 may be further decorated, designed or ornamented with a plurality of subjects, decorations or ornaments which facilitate providing the first predetermined sensory setting 12 a which defines a first aura, motif or theme. Likewise, thesecond station 14 may be further provided or ornamented with a plurality of subjects, decorations or ornaments which facilitate providing a second predetermined sensory setting 14 a which defines a second aura, motif or theme. For example, as illustrated inFIG. 1A , the predetermined sensory setting 12 a of thefirst station 12 may be further decorated with a table 50, table decorations, pillar and wall decorations, carpet (not shown), plants 54 and other wall decorations (not shown) to further enhance the Roman/Italian motif, theme or aura. The first and second predeterminedsensory settings appropriate lighting fixtures 56 and appropriate furnishings, such aschairs 60 and tables 61, which complement the predetermined setting to further facilitate defining the Roman/Italian theme or motif for thestations - It should be appreciated that once the first and
second stations sensory settings station 12, with a strong sensory environment facilitates enhancing the virtual presence illusion created byteleconferencing system 10 of the present invention. - It should also be appreciated, however, that although the
first station 12 andsecond station 14 are shown in the embodiment inFIGS. 1A and 1B as having complementary or similar first and second predeterminedsensory settings sensory settings FIGS. 1A and 1B illustrate a first and second set ofstations station 12′ andstation 14′ in the embodiment illustrated inFIGS. 3A and 3B , may have at least one station having a different predetermined setting. For example, thesecond station 14′ inFIG. 3B provides a setting 14 a′ which defines a Chinese aura, theme or motif. - It should also be appreciated that the
members 32 a-32 h,ceiling 34 and associated predetermined sensory setting are provided to be transportable and capable of being assembled at any suitable location, such as an existing rectangular room, suite or conference area having dimensions of at least 20 feet×20 feet×9 feet. While it may be desirable to provide the first andsecond stations teleconferencing system 10 with substantially the same dimensions, it should be appreciated that they could be provided with differing dimensions, depending on, for example, the number of participants at each station. It should also be appreciated that thesecond station 14 and other stations described herein would preferably be manufactured and assembled in the same or similar manner as thefirst station 12. Also, the stations in theteleconference system 10 may be decorated with wall, ceiling and floor coverings to provide, for example, the first predetermined sensory setting 12 a without using the pre-formed or moldedmodular members 32 a-32 h described above, although the use of such members may be preferable in this embodiment. - The
teleconferencing system 10 also comprises conferencing means or a conferencing system means for teleconnecting the first andsecond stations FIG. 1A , the conferencing system associated with thefirst station 12 comprises image sensor means, imager or image sensors for sensing images at thefirst station 12. For the embodiment shown inFIGS. 1A and 2 , the image sensor means comprises a plurality of cameras which are operably associated with therear projection screen 46 offirst station 12. In this regard, the plurality of cameras comprise afirst camera head 62 andsecond camera head 64 which are operatively coupled to a firstcamera control unit 66 and secondcamera control unit 68, respectively. Notice that the first and secondcamera control units second cameras rear projection screen 46, without substantially interfering with the video image being projected. - In the embodiment being described, the
first camera head 62 andsecond camera head 64 are situated approximately 16 inches above the surface of table 50 which generally corresponds to the eye level of the seated participants situated at table 50. As illustrated inFIG. 2 , the first andsecond cameras rear projection screen 46 in operative relationship with a pair of 1-¼inch diameter openings second cameras rear projection screen 46 in operative relationship withopenings second cameras camera control units teleconferencing system 10 shown and described in relation toFIGS. 1A and 1B show image sensor means comprising a plurality of camera heads 62 and 64 andcamera control units FIGS. 4A and 4B ) or even multiple cameras could be used depending on such things as the size of the station, the number of participants situated at the station, and/or the aspect ratio of each camera head selected. It should also be appreciated that the camera heads 62 and 64 and associatedcamera control units first station 12 to facilitate providing maximum vertical eye contact among participates in the teleconference, while minimally interrupting the substantially life-size video projection on therear projection screen 46. - The conferencing means also comprises a first differentiator or differential key generator 70 (
FIG. 1A ) and a second differentiator or differentialkey generator 72, respectively. Thecamera control unit 66 generates an RGB analog signal I-62 which is received by thefirst differentiator 70, and thecamera control unit 68 generates an RGB signal I-64 which is received by thesecond differentiator 72. The first andsecond differentiators camera control units first station 12 from at least a portion of the first predetermined sensory setting 12 a, such as the background behind the participants, in order to provide a differential signal VS-1 that has that portion of the first predetermined sensory setting 12A removed. This, in turn, facilitates transmitting the video image of the participants at thefirst station 12 to the remotesecond station 14 and also facilitates compositing the image with other images, as described below. -
Suitable differentiators - The differential
key generators images 104 and 106 (FIG. 5A ), respectively. The differentialkey generators images FIG. 5A . As described in detail later herein, theseimages appropriate storage 74 and 76 (FIG. 1A ) associated with the differentialkey generators cameras first station 12 from a video sequence of the first predetermined sensory setting 12 a of thefirst station 12 background where no participants, chairs, or other foreground elements are in place. - In response to the comparison, the first and
second differentiators FIG. 1A ), respectively. As illustrated inFIG. 5 , the VS-1 and VS-2 signals generally correspond to the individuals situated at thefirst station 12 when viewed in the direction of arrow A inFIG. 2 . As illustrated in theimages 112 and 114 (FIG. 5 ) associated with the VS-1 and VS-2 signals, respectively, notice that the background area shown inimages - Advantageously, tagging at least a portion of the image represented by the VS-1 signal as “zero” background facilitates compressing the VS-1 and VS-2 signals and providing corresponding compressed CDS-1 and CDS-2 signals, thereby reducing the amount of transmission band width needed. This tagging also facilitates compositing or overlaying another predetermined image to provide a seamless composited image as described in detail below.
- The video signals VS-1 and VS-2 are received by a first compression/decompression means or
CODEC 78 and a second compression/decompression means orCODEC 80, respectively. TheCODECs suitable microphones first station 12, such as underneath or on top of table 50, as illustrated inFIG. 1A . The function of the first andsecond CODEC second station 14, and also to decompress compressed video and audio signals received from remote stations. Consequently, theCODECs - The
CODEC 78 receives the video signal VS-1 and audio signal AS-A1, andCODEC 80 receives the video signal VS-2 and audio signal AS-A2. TheCODECs remote station 14 via atransmission network 84. - The
transmission network 84 may be configured as a private network, public circuit switch service, and it may utilize telecommunication and/or satellite technology. In the embodiment being described, thetransmission network 84 preferably includes a plurality of T-1 lines (not shown) which are capable of accommodating bit streams having a suitable band width, such as 1.544 megabytes per second. - The
teleconferencing system 10 and conference means associated with thefirst station 12 also comprises enhancing means for enhancing the resolution of an image or sub-image received from a remote station, such as thesecond station 14. In the embodiment being described, enhancing means comprises afirst line doubler 86 and asecond line doubler 88 which are operatively coupled to thefirst CODEC 78 andsecond CODEC 80, respectively. In this embodiment, the first andsecond line doublers CODECs station 14, as described in detail below. One suitable line doubler is the Model No.LD 100 available from Faroudja Laboratories, Inc. of Sunnyvale, Calif., but other suitable enhancing means may be provided to provide greater or less enhancement of the images to be displayed. For example, lenses, mirrors, optical pixel interpolation or other electrical means may be employed as desired. It should also be noted that the present invention may be performed without the use of any enhancing means without departing from the scope of the invention. - The first and
second line doublers video compositing multiplexer 92 for compositing the enhanced video signals associated with the images or sub-images received from the remote station(s) with one or more predetermined composite signals, such as predetermined composite signal A, corresponding to a predetermined composite image or sub-image which are stored in asuitable storage device 94 associated with thecompositor 92. In the embodiment being described, the predetermined composite signal A corresponds to an image of at least a portion of first predetermined sensory setting 12 a, such as the background of thefirst station 12. Thevideo compositing multiplexer 92 composites the signals received from the first andsecond line doublers video compositing multiplexer 92. - The
teleconferencing system 10 comprises aprojector 96 coupled to thevideo compositing multiplexer 92 which receives the RGB composite signal and projects a corresponding image 90 (FIG. 1A ) corresponding to the composite signal on therear projection screen 46. The Model No. 3300 available from AMPRO Corporation of Titusville, Fla. has been found to be asuitable projector 96. Although the embodiment has been described usingprojector 96 andrear projection screen 46, other suitable means may be employed for projecting or displaying the composited image. For example, a liquid crystal display (LCD) or other electronic screen may be suitable to display images at a station. This may eliminate the need for theprojector 96. - The
projector 96 could be used with an optical system or a plurality of mirrors (not shown), or prisms (not shown) such that the projector can be positioned, for example, to the side or below therear projection screen 46 or in a manner that permits theprojector 96 to project the image towards a mirror (not shown), which causes the image to be projected on therear projection screen 46. - As described in detail below, the composite signal and its
corresponding image 90 generally comprise a video image of at least a portion of the first predetermined sensory setting 12 a combined or composited with a differentiated image, such as an image of the participants from thesecond station 14 which correspond to the VS-3 and VS-4 (FIG. 1B ) signals. Consequently, theresultant image 90 projected onscreen 46 at thefirst station 12 complements or blends with the architectural motif, aura, theme or design defined by the first predetermined sensory setting 12 a at thefirst station 12, such that the projectedimage 90 appears visually integrated with the first predetermined sensory setting 12 a of thefirst station 12. This, in turn, causes any image of the participants situated at thesecond station 14 and included in theimage 90 to appear to be face-to-face with participants at thefirst station 12 during the teleconference. The operation of thecompositor 92 is described in more detail later herein. - It should be appreciated that the sub-images or images received from the remote station(s) typically have a resolution on the order of about 352×288 pixels and the predetermined composite signal A comprises a resolution on the order of about 1280×1024 pixels. Thus, the resultant
composite image 90 may comprise, for example, an image of the participants situated at thesecond station 14 having a first resolution and a background image of thefirst station 12 having a second resolution, which is higher than the first resolution. This enablescompositor 92 to provide acomposite image 90 which, when displayed onscreen 46, gives the illusion or effect of a “video mirror” to the participants situated at thefirst station 12. - The
teleconferencing system 10 also includes audio means comprising a plurality ofspeakers 100 and 102 (FIGS. 1A and 2 ) which, in turn, receive audio signals AS-B1 and AS-B2 fromCODECs second station 14. - It should also be appreciated that the
rear projection screen 46 andprojector 96 are configured and selected to enable theteleconferencing system 10 to project the composited image 90 (FIG. 1A ) at a predetermined scale, such as substantially full scale. In this regard, thecompositor 92 comprises ascaler 95 which is integral therewith for scaling the composited signal associated with the compositedimage 90 to a desired or predetermined scale, such as substantially full scale. - Referring now to
FIG. 1B , thesecond station 14 comprises similar components as the first station and such like components are labelled with the same reference numeral as their corresponding component in thefirst station 12, except that the components associated with thesecond station 14 have a “−1” designator added thereto. Such components operate and function in substantially the same manner as described above with regard to thefirst station 12 with the following being some differences. The differential reference signals DRS-3 and DRS-4 (FIG. 5 ) associated with thesecond station 14 generally correspond to an image or sub-image of at least a portion of the second predetermined sensory setting 14 a, such as the background 98-1, of thesecond station 14. Such sub-image or image may include at least a portion of the background 98-1 without any participants, chairs or other foreground subjects situated in thesecond station 14. Also, like the predetermined composite signal A stored in thestorage 94 associated with thefirst station 10, a predetermined composite signal B may be stored in the storage 94-1 associated with the compositor 92-1second station 14. The predetermined composite signal B may correspond to an image or sub-image of at least a portion of the second predetermined sensory setting 14 a of thesecond station 14. Such sub-image or image may include, for example, an image of thewalls 32 a-1 to 32 h-1 and conference area 18 or background of thesecond station 14. Notice that in the embodiment shown inFIGS. 1A and 1B , thesecond station 14 has a second predetermined sensory setting 14 a which mirrors or is complementary to the first predetermined sensory setting 12 a. As described above, however, the first and second predeterminedsensory settings - A method of operating the
teleconferencing system 10 will now be described in relation toFIGS. 6A-6D . The modular components, such asmembers 32 a to 32 h andceiling 34 forfirst station 10, decorations and the like, are configured, assembled and decorated (block 99 inFIG. 6A ) at a desired location to provide a conference station comprising a predetermined sensory setting defining a predetermined theme, motif or aura. As mentioned earlier herein, the theme, motif or aura may be complementary (as shown inFIGS. 1A and 1B ) or they can be completely different, as shown inFIGS. 3A and 3B (described below). For ease of illustration, it will be assumed that the stations are assembled and decorated as shown and described relative to the embodiment inFIGS. 1A and 1B . - Once the
modular stations decision point 101 inFIG. 6A ) to use differentiator (e.g.,differentiator 72 inFIG. 1A ). As discussed herein relative to the embodiments shown inFIGS. 4A and 4B , it may not always be desired to generate a differential reference image, thereby making it unnecessary to generate the differential reference signal. If differentiation is desired, then the camera heads 62 or 64 generate at least one video image (block 103) of at least a portion of the first predetermined sensory setting 12A at thefirst station 12. Thedifferentiators suitable storage differentiators second station 14 generate video images of at least a portion of the second predetermined setting 14 a at thesecond station 14. The differentiators 70-1 and 72-1 grab or capture at least one differential reference image or sub-image from those images and generate differential reference signals (not shown) corresponding thereto. These signals are then stored (block 109) in suitable storage 74-1 and 76-1 for use by differential key generators 70-1 and 72-1, respectively. - As mentioned above, it is preferred that the differential reference signals DRS-62 and DRS-64 comprise an image of at least a portion of the first predetermined sensory setting 12 a, such as an image of the
first station 12 without any participants, chairs or other subjects which are not stationary during the teleconference. Likewise, it is preferred that the differential reference signals associated with the differentiators 70-1 and 72-1 comprise at least a portion of the second predetermined sensory setting 14 a at thesecond station 14, such as an image of the background 98-1 without the participants, chairs and other subjects which are not stationary during the teleconference. - If differentiation of signals is not selected or at the end of the differentiation process, it may be desired to generate a composite image (decision point 97) for one or more of the stations. As discussed below, however, this may not always be required to achieve certain advantages of the invention. Such predetermined composite image would preferably include a substantial portion of the first predetermined sensory setting 12 a, including the background and/or
conference area 16 of thefirst station 12. If compositing is desired, then the predetermined composite signal A is generated (block 111 inFIG. 6B ). The corresponding predetermined composite signal A may then be stored insuitable storage 94. In the same manner, the predetermined composite image at thesecond station 14 and corresponding predetermined composite signal B may be generated and stored as predetermined composite signal B in suitable storage 94-1. In the embodiment being described, the predetermined composite image associated with thesecond station 14 includes an image of at least a portion of the second predetermined sensory setting 14 a, including the background 98-1. - In the embodiment being described, the predetermined composite signals A and B are generated by a suitable still camera (not shown) to provide a still image (not shown) of the
station image 90 inFIG. 1A ) may comprise an image having a high resolution background, for example, combined with a comparatively lower resolution image of the remote station participants. This, in turn, facilitates enhancing the “video mirror” effect wherein a mimic or replication of a common architectural technique of mirroring a wall of a given room which makes the overall room appear to be extended beyond its actual wall line. - Once the
stations second suites transmission network 84. - Next, one or more participants may be situated at the first and
second stations FIG. 2 , notice that the participants seated at thefirst station 12 are situated a predetermined distance B from a participant'sside 46 a of therear projection screen 46. The predetermined distance B generally corresponds to a preferred or optimum focal distance at which optimum imaging bycameras second station 14 in a similar manner and the face-to-face teleconference may then begin. - For ease of illustration, the imaging and display of
first station 12 participants at thesecond station 14 will be described. The first andsecond cameras FIG. 6B ) live images of the participants situated at thefirst station 12 and generate corresponding RGB analog signals I-62 and I-64 which are received by the differentialkey generators decision point 147 inFIG. 6C ), processing continues atblock 119 otherwise it proceeds atblock 123. The differentialkey generators FIG. 6C ) the digital differential signal VS-1 and VS-2, respectively, after comparing (block 119 inFIG. 6C ) the I-62 and I-64 signals received fromcameras reference signals DRS 62 and DRS-64 which are received fromstorages - The differential signals VS-1 and VS-2 are then received by
CODECs first station 12. TheCODECs second station 14 via the transmission network 84 (FIG. 1B ). - The CDS-1 and CDS-2 signals are received and decompressed (block 127 in
FIG. 6C ) by CODECs 78-1 and 80-1, respectively, associated with thesecond station 14 to provide decompressed VS-1 and VS-2 signals. The CODECs 78-1 and 80-1 also decompress the audio signals AS-A1 and AS-A2 received from thefirst station 10 which are transmitted to speakers 100-1 and 102-1, respectively, at thesecond station 14. - Substantially simultaneously with the broadcasting of the audio signals at the
second station 14, CODECs 78-1 and 80-1 decompress the CDS-1 and CDS-2 signals to provide VS-1 and VS-2 signals. The decompressed video signals VS-1 and VS-2 are then received by line doublers 86-1 and 88-1. If it is desired to enhance the signals (decision point 129), then the line doublers 86-1 and 88-1 process or manipulate the signals (block 131) in order to enhance the resolution of the image corresponding to those signals. After the signals VS-1 and VS-2 are processed, it may be desired to composite (decision point 133 inFIG. 6D ) those signals with one or more other signals. In this illustration, for example, the video compositor 92-1 composites images (block 135) corresponding to those signals with at least one predetermined composite image, such as image 122 (FIG. 5B ) corresponding to the predetermined composite signal B provided from storage 94-1 (FIG. 1B ) to provide a composite signal. As mentioned above, the composite signal generally corresponds to the composited image 91-1 to be displayed on the rear projection screen 46-1 at thesecond station 14. - The compositor 92-1 may (
decision point 137, block 139 inFIG. 6D ) scale the composited image to a desired scale, such as full scale, using scaler 95-1. Thereafter, the compositor 95-1 transmits a corresponding RGB analog signal to projector 96-1 which displays (block 141) the scaled, composited image on the rear projection screen 46-1 (FIG. 1B ). - The teleconference may then be continued or terminated as desired (
decision point 143, block 145). - Because the composited image is substantially full scale when projected and includes a high resolution image of at least a portion of the second predetermined sensory setting 14 a, the image appears to blend or become visually integrated with the second predetermined sensory setting 14 a. This, in turn, gives the participants situated at the
second station 14 the perception that the first station participants are present or face-to-face with them in thesecond station 14. - In the same or similar manner, images and signals relative to the
second station 14 images are captured, processed and displayed at thefirst station 12. So that images of the participants at thesecond station 14 are displayed at thefirst station 12 such that they appear to have a face-to-face presence at thefirst station 12. Thus, images of thesecond station 14 participants may be differentiated and composited such that, when they are displayed at thefirst station 12, the image completes or provides “the other half” of thefirst station 12 and becomes generally visually integrated therewith. Although not required, it may be desirable to enhance the face-to-face presence by providing, for example, first and second predeterminedsensory settings stations - Once the teleconferencing is terminated, the
stations -
FIGS. 5A and 5B provide a visual illustration of the images corresponding to some of the signals described above utilizing the method and embodiment described above. In this regard,images second cameras key generators key generators images images storages cameras - As illustrated in
FIG. 5A , the differentialkey generators corresponding images images first station 12 with the background area having been removed or tagged as a “zero” area. As described herein, this “zero” area becomes “filled-in” with the desired or predetermined composite image which may include, for example, an image of at least a portion of the predetermined setting or background of thesecond station 14. It has been found that removing a portion of the image, such as the background, by tagging it as zero, in the manner described herein, facilitates compressing the signals VS-1 and VS-2 and reducing the amount of bandwidth needed to transmit the images overtransmission network 84 and between the first andsecond stations - As mentioned above, the video signals VS-1 and VS-2 are fed into
CODECs transmission network 84, to thesecond station 14 and received by the CODECs 78-1 and 80-1 associated with thesecond station 14. As illustrated inFIG. 5B , the CODEC 78-1 and 80-1 decompresses the CDS-1 and CDS-2 signals, respectively, from thefirst station 12 and feeds them into associated line doublers 86-1 and 88-1. As mentioned earlier herein, the line doublers 86-1 and 88-1 facilitate enhancing the images associated with the video signals to provide enhanced video signals EVS-1 and EVS-2 (FIG. 5B ), respectively. - As stated earlier, the enhanced video signals EVS-1 and EVS-2 are then received by the video compositing multiplexer 92-1 associated with the
second station 14 wherein the signals are combined to provide an intermediate composite signal ICS having an associated intermediatecomposite signal image 120 having an aspect ratio of about 8:3. - The video compositing multiplexer 92-1 also receives the predetermined composite signal B having a predetermined composite
signal B image 122 from storage 94-1. The video compositing multiplexer 92-1 composites or combines theimages composite image 124 as shown inFIG. 5B . As stated earlier, the predetermined compositesignal B image 122 generally corresponds to at least a portion of the predetermined setting or background of thesecond station 14 and has an aspect ratio of 16:9. - Notice that when the predetermined composite
signal B image 122 is combined with the intermediatecomposite signal image 120, the video compositing multiplexer 92-1 causes the “zero” area of the intermediatecomposite signal image 120 to be “filled in” with the predetermined composite signal B image. - The
composite image 124 may then be scaled to a predetermined size or scale, such as full scale, using scaler 94-1, so that thecomposite image 124 may be scaled to a substantially full scale or real-life size image as desired. The composite image signal corresponding to thecomposite image 124 is transmitted to the projector 96-1 and then displayed on the rear projection screen 46-1 at thesecond station 14. As illustrated inFIGS. 1B and 5B , thecomposite image 124 may be appropriately framed or masked (such as with anarchway 125 inFIGS. 1B and 5B ) when it is projected at thesecond station 14 to enhance the face-to-face, real time environment. - The audio and video signals transmitted between the first and
second stations transmission network 84 in order to effect a substantially simultaneous and/or “real time” video conference. Thus, in the illustration shown inFIGS. 1A and 1B , the participants may be geographically remotely located, yet the participants situated at thefirst station 12 will feel as if thesecond station 14 participants are located face-to-face or present with them at thefirst station 12, while the participants situated at thesecond station 14 will feel as if the first station participants are face-to-face or present with them at the second station. - It should be appreciated that when the predetermined composite signal B and associated predetermined
composite signal image 122 is composited with the intermediate composite signal and associated intermediatecomposite signal image 120, it overlays that signal to provide a seamlesscomposite image 124, which facilitates reducing or eliminating the need to match up the borders or seams of the camera images with any high degree of accuracy. In this regard, it is preferable thatcameras - Advantageously, the invention provides an apparatus and method for providing a video mirror at each
station FIG. 5B ) having at least a portion thereof imaged at a greater resolution than the portion which was transmitted overtransmission network 84. This facilitates reducing the effect of limitations or transmission restrictions of thetransmission network 84 which, in turn, facilitates increasing the quality of images displayed at a station. - In addition, notice that the composite image 124 (
FIG. 5B ) may have an aspect ratio which is different from the aspect ratio of thecameras -
FIGS. 3A and 3B , when taken together, illustrate another embodiment of the invention. The operation and components of the embodiment shown inFIGS. 3A and 3B are substantially the same as the operation of components of the embodiment described above relative toFIGS. 1A and 1B with the same reference numerals being used for the same components with the addition of single prime (′) designator. Consequently this embodiment is similar to the embodiment shown inFIGS. 1A and 1B , except that the second predetermined setting 14 a′ inFIG. 3B and its associated theme, aura or motif is substantially different from the second predetermined setting 14 a shown inFIG. 1B . InFIG. 3B , the first predetermined sensory setting 12 a′ comprises a plurality ofdecorations 120 defining the Chinese theme, motif or aura. Also, the predetermined composite signal A stored in storage 94-1′ and the differential reference signals stored in storages 74-1′ and 76-1 would generally correspond to an image of at least a portion of that setting 14 a′. - As with the illustration described above relative to
FIGS. 1A and 1B , the video and audio signals would be processed in substantially the same manner. In general, an image of the participants situated at thefirst station 12′ is composited by compositor 92-1′ with a predetermined composite image of at least a portion of the second predetermined sensory setting 14 a′ of thesecond station 14′ and projected onto the rear projection screen 46-1′ at thesecond station 14′. Thefirst station 12′ participants appear to be face-to-face with thesecond station 14′ participants because they have a relatively high resolution video image behind them which complements or becomes integrated with the second predetermined sensory setting 14 a′. Thus, as shown inFIG. 3B , the image 91-1′ (FIG. 3B ) of the ladies at thefirst station 12′ includes a Chinese background which blends or complements the actual predetermined sensory setting 14 a′. - Likewise, when the image of the participants situated at the
second station 14′ is projected on therear projection screen 46′ at thefirst station 12′, they appear to be in the same room as the participants situated at thefirst station 12′ because the Roman/Italian video background which is seen behind thesecond station 14′ participants generally complements and becomes visually integrated with the actual Roman/Italian theme, motif or aura defined by the first predetermined sensory setting 12′ of thefirst station 12′. -
FIGS. 4A and 4B , when taken together, illustrate another embodiment of the invention. The components of the embodiment shown inFIGS. 4A and 4B which are substantially identical to the components in the embodiment shown inFIGS. 1A and 1B which have the same reference numerals with the addition of a double prime (“″”) designators. As illustrated inFIGS. 4A and 4B , two remote modular stations such asstations 12″ and 14″ may be provided and designed to have first and second predeterminedsensory settings 12 a″ and 14 a″ which are substantially identical. Thus, as shown inFIGS. 4A and 4B , images may be captured in the manner described above atstation 12″ received byCODECs 78″ and 80″ and then transmitted, viatransmission 84″, to associated CODECs 78-1″ and 80-1″, respectively. The CODECs 78-1″ and 80-1″ then generate a decompressed signal which may be enhanced by line doublers 86-1″ and 88-1″, respectively; scaled to an appropriate scale by scaler 95-1″; and then projected by projector 96-1″ onto rear projection screen 46-1″. - Notice that the image comprising the
second station 14″ participants and second predetermined sensory setting 14 a″ is displayed onscreen 46″ at thefirst station 12″. Thus, this embodiment does not utilize the differentiating and compositing features of the previous embodiment, but may still achieve a face-to-face conference environment because the second predetermined sensory setting 14 a″ is configured to be identical to or complementary with the first predetermined sensory setting 12 a″. In this embodiment, entire images or sub-images of thestations 12 and 14 (including images of both participants and background) are displayed at remote station(s). Because thestations 12″ and 14″ are assembled, decorated and designed to be complementary or identical, they appear visually integrated to participants situated in thestations sensory settings 12 a″ and 14 a″, including the background, are designed and arranged in a geometric fashion such that ascameras 62″ and 64″ capture images of the participants, they also capture images of the first and second predetermined sensory setting 12 a″ and 14 a″, respectively, at the most advantageous perspective for display at the remote station(s). As with prior embodiments, this causes thefirst station 12″ participants to perceive that thesecond station 14″ participants are situated or present with thefirst station 12″ participants at thefirst station 14″. Likewise, thefirst station 12″ participants appear to be face-to-face with thesecond station 14″ participants at thesecond station 14″ when the images associated with thefirst station 12″ are displayed on screen 46-1″. Consequently, by providing complementary or identical first and second predeterminedsensory settings 12 a″ and 14 a″, a face-to-face conference may be created. As with previous embodiments, it may also be desired to differentiate, enhance, composite or scale the images as described with previous embodiments, but this is not required with the embodiment being described. - Thus, it should be apparent that stations can be provided with predetermined settings which are completely different, yet, by utilizing the apparatus and method of the present invention, the images of the participants in these stations may be projected at remote stations so that they appear to be virtually face-to-face with the remote station participants at one or more remote stations.
- Various changes or modifications in the invention described may occur to those skilled in the art without departing from the spirit or scope of the invention. For example, the
screen 46 forstation 12 has been shown as being integral with a portion of awall 32 h (FIGS. 1A and 2A ), it could comprise a larger or smaller portion of thatwall 32 h, or it could be provided as part of one or more other walls, or even as part of theceiling 34. - It should also be appreciated that while the embodiments have been shown and described comprising two stations, images from more than two remote stations may be displayed at a station, thereby permitting a teleconference convention among more than two stations.
- Although not shown, one or more of the compositors, such as
compositors 12 or 12-1 (FIG. 1A ) may comprise a stationary or moving image database (not shown) for providing a plurality of predetermined composite signals which define a particular or desired video background. For example, participants may elect to use the arched background of their proximity, choose an event-related scene, or decide to meet in a setting completely unrelated to their site or station. For example, a station having a Manhattan eatery motif may be provided with a screen configured as a window (not shown). Certain moving video backgrounds of a busy New York avenue may be deposited and displayed on the screen to give the illusion that the participants situated at the station are dining in a popular Manhattan eatery. - It should also be appreciated that while the embodiments being shown and described herein refer to teleconferencing environments that have predetermined settings and motifs or auras relating to dining, the predetermined settings could define any type of aura, theme or motif which is suitable for video conferencing and in which it is desired to provide a “real-life” or face-to-face presence illusion. For example, the apparatus and method of this invention could be used in a business setting, education setting, seminar setting, home environment, religious setting, celebration setting (such as a birthday, retirement party, holiday or anniversary), or any other suitable setting as desired.
- The above description of the invention is intended to be illustrative and not limiting, and is not intended that the invention be restricted thereto but that it be limited only by the spirit and scope of the appended claims.
Claims (23)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/537,177 US20070120954A1 (en) | 1994-09-19 | 2006-09-29 | Teleconferencing method and system |
Applications Claiming Priority (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/308,603 US5572248A (en) | 1994-09-19 | 1994-09-19 | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US08/740,839 US5751337A (en) | 1994-09-19 | 1996-11-04 | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US09/047,771 US6160573A (en) | 1994-09-19 | 1998-03-25 | Teleconference method and system for providing face-to-face teleconference environment |
US09/695,460 US6445405B1 (en) | 1994-09-19 | 2000-10-24 | Teleconferencing method and system |
US10/233,363 US6654045B2 (en) | 1994-09-19 | 2002-08-30 | Teleconferencing method and system |
US10/716,972 US7116350B2 (en) | 1994-09-19 | 2003-11-19 | Teleconferencing method and system |
US11/537,177 US20070120954A1 (en) | 1994-09-19 | 2006-09-29 | Teleconferencing method and system |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/716,972 Continuation US7116350B2 (en) | 1994-09-19 | 2003-11-19 | Teleconferencing method and system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070120954A1 true US20070120954A1 (en) | 2007-05-31 |
Family
ID=23194627
Family Applications (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US08/308,603 Expired - Lifetime US5572248A (en) | 1994-09-19 | 1994-09-19 | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US08/740,839 Expired - Lifetime US5751337A (en) | 1994-09-19 | 1996-11-04 | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US09/047,771 Expired - Lifetime US6160573A (en) | 1994-09-19 | 1998-03-25 | Teleconference method and system for providing face-to-face teleconference environment |
US09/695,460 Expired - Lifetime US6445405B1 (en) | 1994-09-19 | 2000-10-24 | Teleconferencing method and system |
US10/233,363 Expired - Fee Related US6654045B2 (en) | 1994-09-19 | 2002-08-30 | Teleconferencing method and system |
US10/716,972 Expired - Fee Related US7116350B2 (en) | 1994-09-19 | 2003-11-19 | Teleconferencing method and system |
US11/537,177 Abandoned US20070120954A1 (en) | 1994-09-19 | 2006-09-29 | Teleconferencing method and system |
Family Applications Before (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US08/308,603 Expired - Lifetime US5572248A (en) | 1994-09-19 | 1994-09-19 | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US08/740,839 Expired - Lifetime US5751337A (en) | 1994-09-19 | 1996-11-04 | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US09/047,771 Expired - Lifetime US6160573A (en) | 1994-09-19 | 1998-03-25 | Teleconference method and system for providing face-to-face teleconference environment |
US09/695,460 Expired - Lifetime US6445405B1 (en) | 1994-09-19 | 2000-10-24 | Teleconferencing method and system |
US10/233,363 Expired - Fee Related US6654045B2 (en) | 1994-09-19 | 2002-08-30 | Teleconferencing method and system |
US10/716,972 Expired - Fee Related US7116350B2 (en) | 1994-09-19 | 2003-11-19 | Teleconferencing method and system |
Country Status (9)
Country | Link |
---|---|
US (7) | US5572248A (en) |
EP (2) | EP1187480B1 (en) |
JP (1) | JPH10508990A (en) |
KR (1) | KR100403885B1 (en) |
AT (2) | ATE311724T1 (en) |
AU (1) | AU697578B2 (en) |
CA (1) | CA2200477C (en) |
DE (2) | DE69534660T2 (en) |
WO (1) | WO1996009722A1 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060268360A1 (en) * | 2005-05-12 | 2006-11-30 | Jones Peter W J | Methods of creating a virtual window |
US20080268899A1 (en) * | 2007-04-24 | 2008-10-30 | Lg Electronics Inc. | Video communication terminal and method of displaying images |
US20100302344A1 (en) * | 2009-05-28 | 2010-12-02 | Microsoft Corporation | Establishing eye contact in video conferencing |
WO2013043289A1 (en) * | 2011-09-23 | 2013-03-28 | Tangome, Inc. | Augmenting a video conference |
US8508576B2 (en) | 2009-04-14 | 2013-08-13 | Huawei Device Co., Ltd. | Remote presenting system, device, and method |
US8625769B1 (en) | 2009-10-13 | 2014-01-07 | Signal Perfection, Ltd. | System for managing, scheduling, monitoring and controlling audio and video communication and data collaboration |
WO2014100250A3 (en) * | 2012-12-18 | 2014-08-14 | Nissi Vilcovsky | Devices, systems and methods of capturing and displaying appearances |
WO2014130977A1 (en) * | 2013-02-25 | 2014-08-28 | Herold Williams | Nonlinear scaling in video conferencing |
US20140267575A1 (en) * | 2013-03-14 | 2014-09-18 | Polycom, Inc. | Method and system for providing a virtual cafeteria |
US8860775B2 (en) | 2009-04-14 | 2014-10-14 | Huawei Device Co., Ltd. | Remote presenting system, device, and method |
US8976160B2 (en) | 2005-03-01 | 2015-03-10 | Eyesmatch Ltd | User interface and authentication for a virtual mirror |
US8982109B2 (en) | 2005-03-01 | 2015-03-17 | Eyesmatch Ltd | Devices, systems and methods of capturing and displaying appearances |
US8982110B2 (en) | 2005-03-01 | 2015-03-17 | Eyesmatch Ltd | Method for image transformation, augmented reality, and teleperence |
US9269157B2 (en) | 2005-03-01 | 2016-02-23 | Eyesmatch Ltd | Methods for extracting objects from digital images and for performing color change on the object |
US9334244B2 (en) | 2010-05-12 | 2016-05-10 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase |
US9365557B2 (en) | 2008-12-19 | 2016-06-14 | Vertex Pharmaceuticals Incorporated | Substituted pyrazin-2-amines as inhibitors of ATR kinase |
US9630956B2 (en) | 2010-05-12 | 2017-04-25 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase |
US9791456B2 (en) | 2012-10-04 | 2017-10-17 | Vertex Pharmaceuticals Incorporated | Method for measuring ATR inhibition mediated increases in DNA damage |
US9862709B2 (en) | 2011-09-30 | 2018-01-09 | Vertex Pharmaceuticals Incorporated | Processes for making compounds useful as inhibitors of ATR kinase |
US10250888B2 (en) | 2015-10-08 | 2019-04-02 | Samsung Electronics Co., Ltd. | Electronic device configured to non-uniformly encode/decode image data according to display shape |
US10478430B2 (en) | 2012-04-05 | 2019-11-19 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase and combination therapies thereof |
US10813929B2 (en) | 2011-09-30 | 2020-10-27 | Vertex Pharmaceuticals Incorporated | Treating cancer with ATR inhibitors |
US11464774B2 (en) | 2015-09-30 | 2022-10-11 | Vertex Pharmaceuticals Incorporated | Method for treating cancer using a combination of DNA damaging agents and ATR inhibitors |
Families Citing this family (283)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5818616A (en) * | 1993-08-23 | 1998-10-06 | Canon Kabushiki Kaisha | Optical communication apparatus and conference system |
US7301536B2 (en) * | 1993-09-10 | 2007-11-27 | Geovector Corporation | Electro-optic vision systems |
US5802281A (en) | 1994-09-07 | 1998-09-01 | Rsi Systems, Inc. | Peripheral audio/video communication system that interfaces with a host computer and determines format of coded audio/video signals |
US5572248A (en) * | 1994-09-19 | 1996-11-05 | Teleport Corporation | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US5815197A (en) * | 1995-02-16 | 1998-09-29 | Sumitomo Electric Industries, Ltd. | Two-way interactive system, terminal equipment and image pickup apparatus having mechanism for matching lines of sight between interlocutors through transmission means |
US5754190A (en) * | 1995-06-07 | 1998-05-19 | Advanced Micro Devices | System for reproducing images utilizing image libraries |
JPH09134319A (en) * | 1995-10-03 | 1997-05-20 | Sony Electron Inc | User interface for personal communication routing system and rule processing |
US6219045B1 (en) * | 1995-11-13 | 2001-04-17 | Worlds, Inc. | Scalable virtual world chat client-server system |
US5880731A (en) * | 1995-12-14 | 1999-03-09 | Microsoft Corporation | Use of avatars with automatic gesturing and bounded interaction in on-line chat session |
US5757424A (en) * | 1995-12-19 | 1998-05-26 | Xerox Corporation | High-resolution video conferencing system |
WO1997022918A1 (en) * | 1995-12-20 | 1997-06-26 | Mediamaxx Incorporated | Computer-controlled system for producing three-dimensional navigable photographs of areas and method thereof |
US6011587A (en) * | 1996-03-07 | 2000-01-04 | Kokusai Denshin Denwa Kabushiki Kaisha | Packet video bitrate conversion system |
US5956491A (en) | 1996-04-01 | 1999-09-21 | Marks; Daniel L. | Group communications multiplexing system |
WO1997039581A1 (en) * | 1996-04-18 | 1997-10-23 | Campbell Jack J | Method and apparatus for presenting in a computer display images representing remote moving objects |
US5959667A (en) * | 1996-05-09 | 1999-09-28 | Vtel Corporation | Voice activated camera preset selection system and method of operation |
US5808613A (en) * | 1996-05-28 | 1998-09-15 | Silicon Graphics, Inc. | Network navigator with enhanced navigational abilities |
US5856842A (en) * | 1996-08-26 | 1999-01-05 | Kaiser Optical Systems Corporation | Apparatus facilitating eye-contact video communications |
US6545700B1 (en) | 1997-06-25 | 2003-04-08 | David A. Monroe | Virtual video teleconferencing system |
US6388654B1 (en) * | 1997-10-03 | 2002-05-14 | Tegrity, Inc. | Method and apparatus for processing, displaying and communicating images |
US6259470B1 (en) * | 1997-12-18 | 2001-07-10 | Intel Corporation | Image capture system having virtual camera |
US6844893B1 (en) * | 1998-03-09 | 2005-01-18 | Looking Glass, Inc. | Restaurant video conferencing system and method |
US8446449B1 (en) | 1998-03-09 | 2013-05-21 | Looking Glass, Inc. | Restaurant videoconferencing system and method |
US6593955B1 (en) * | 1998-05-26 | 2003-07-15 | Microsoft Corporation | Video telephony system |
EP0961523B1 (en) * | 1998-05-27 | 2010-08-25 | Sony France S.A. | Music spatialisation system and method |
US6339842B1 (en) | 1998-06-10 | 2002-01-15 | Dennis Sunga Fernandez | Digital television with subscriber conference overlay |
US6414707B1 (en) * | 1998-10-16 | 2002-07-02 | At&T Corp. | Apparatus and method for incorporating virtual video conferencing environments |
US6346950B1 (en) * | 1999-05-20 | 2002-02-12 | Compaq Computer Corporation | System and method for display images using anamorphic video |
US6292713B1 (en) * | 1999-05-20 | 2001-09-18 | Compaq Computer Corporation | Robotic telepresence system |
GB9918704D0 (en) * | 1999-08-10 | 1999-10-13 | White Peter M | Device and method for eye to eye communication overa network |
AU6171200A (en) * | 1999-08-10 | 2001-03-05 | Peter Mcduffie White | Communications system |
US20040193722A1 (en) * | 1999-08-30 | 2004-09-30 | Donovan Kevin Remington Joseph Bartholomew | Universal instant messaging system for the internet |
US7057635B1 (en) * | 2000-01-27 | 2006-06-06 | Atheros Communications, Inc. | High-speed RF link for a multi-user meeting |
KR100366384B1 (en) * | 2000-02-26 | 2002-12-31 | (주) 고미드 | Information search system based on communication of users |
AU2001234576A1 (en) * | 2000-10-04 | 2002-04-15 | Ezonics Corporation | A method and system for internet-based video chat |
RU2225025C2 (en) * | 2000-12-06 | 2004-02-27 | ОПТИВА, Инк. | Liquid-crystal device for information display |
US6361173B1 (en) | 2001-02-16 | 2002-03-26 | Imatte, Inc. | Method and apparatus for inhibiting projection of selected areas of a projected image |
US6930702B1 (en) * | 2001-04-11 | 2005-08-16 | Applied Minds, Inc. | Device for positioning and operating audio and video equipment |
US7502770B2 (en) * | 2001-04-11 | 2009-03-10 | Metaweb Technologies, Inc. | Knowledge web |
JP4803906B2 (en) * | 2001-06-07 | 2011-10-26 | 株式会社キーエンス | Method for generating barcode for setting optical information reader, method for changing setting of optical information reader, program for generating barcode for setting, and recording medium |
US7333726B2 (en) * | 2001-07-05 | 2008-02-19 | Wave7 Optics, Inc. | Method and system for supporting multiple service providers within a single optical network |
DE10132568A1 (en) * | 2001-07-10 | 2003-01-30 | Holger Berthold | communication system |
US6798457B2 (en) | 2001-09-26 | 2004-09-28 | Digeo, Inc. | Camera positioning system and method for eye-to-eye communication |
US6943843B2 (en) * | 2001-09-27 | 2005-09-13 | Digeo, Inc. | Camera positioning system and method for eye-to eye communication |
US7892088B2 (en) * | 2001-10-18 | 2011-02-22 | Steve Brandstetter | Gaming device having a second separate bonusing event |
FR2832282B1 (en) * | 2001-11-12 | 2004-07-30 | France Telecom | MODULAR AUDIOVISUAL SYSTEM WITH CONCATENABLE MODULES FOR PRESENTING A LOCAL SCENE AND A REMOTE SCENE |
US20040207718A1 (en) * | 2001-11-14 | 2004-10-21 | Boyden James H. | Camera positioning system and method for eye -to-eye communication |
US20030112325A1 (en) * | 2001-12-13 | 2003-06-19 | Digeo, Inc. | Camera positioning system and method for eye-to-eye communication |
US7126627B1 (en) | 2002-03-06 | 2006-10-24 | Lewis Thomas B | Video conferencing device and method |
US6783247B2 (en) | 2002-03-14 | 2004-08-31 | White Peter Mcduffie | Life-size communications systems with front projection |
US20030195834A1 (en) * | 2002-04-10 | 2003-10-16 | Hillis W. Daniel | Automated online purchasing system |
US7844610B2 (en) * | 2003-12-12 | 2010-11-30 | Google Inc. | Delegated authority evaluation system |
US8069175B2 (en) * | 2002-04-10 | 2011-11-29 | Google Inc. | Delegating authority to evaluate content |
US7386799B1 (en) | 2002-11-21 | 2008-06-10 | Forterra Systems, Inc. | Cinematic techniques in avatar-centric communication during a multi-user online simulation |
AU2003297193A1 (en) | 2002-12-13 | 2004-07-09 | Applied Minds, Inc. | Meta-web |
US8012025B2 (en) * | 2002-12-13 | 2011-09-06 | Applied Minds, Llc | Video game controller hub with control input reduction and combination schemes |
US20040135880A1 (en) * | 2003-01-13 | 2004-07-15 | Royer George R. | Integrated televised meeting procedure |
US7525576B2 (en) * | 2003-02-17 | 2009-04-28 | Axis, Ab | Method and apparatus for panning and tilting a camera |
US7352809B2 (en) * | 2003-02-21 | 2008-04-01 | Polycom, Inc. | System and method for optimal transmission of a multitude of video pictures to one or more destinations |
US7593546B2 (en) * | 2003-03-11 | 2009-09-22 | Hewlett-Packard Development Company, L.P. | Telepresence system with simultaneous automatic preservation of user height, perspective, and vertical gaze |
US6888562B2 (en) | 2003-03-26 | 2005-05-03 | Broadcom Corporation | Integral eye-path alignment on telephony and computer video devices using a pinhole image sensing device |
EP1621010A1 (en) * | 2003-05-02 | 2006-02-01 | Allan Robert Staker | Interactive system and method for video compositing |
US8299979B2 (en) * | 2003-05-14 | 2012-10-30 | Broadcom Corporation | Integral eye-path alignment on telephony and computer video devices using two or more image sensing devices |
US20050021620A1 (en) * | 2003-05-30 | 2005-01-27 | Todd Simon | Web data conferencing system and method with full motion interactive video |
JP4292891B2 (en) * | 2003-06-26 | 2009-07-08 | ソニー株式会社 | Imaging apparatus, image recording apparatus, and image recording method |
US20050010874A1 (en) * | 2003-07-07 | 2005-01-13 | Steven Moder | Virtual collaborative editing room |
US7119829B2 (en) * | 2003-07-31 | 2006-10-10 | Dreamworks Animation Llc | Virtual conference room |
US20050131918A1 (en) * | 2003-12-12 | 2005-06-16 | W. Daniel Hillis | Personalized profile for evaluating content |
US7305078B2 (en) * | 2003-12-18 | 2007-12-04 | Electronic Data Systems Corporation | Speaker identification during telephone conferencing |
US8824730B2 (en) * | 2004-01-09 | 2014-09-02 | Hewlett-Packard Development Company, L.P. | System and method for control of video bandwidth based on pose of a person |
US8599239B2 (en) | 2004-04-21 | 2013-12-03 | Telepresence Technologies, Llc | Telepresence systems and methods therefore |
US8208007B2 (en) | 2004-04-21 | 2012-06-26 | Telepresence Technologies, Llc | 3-D displays and telepresence systems and methods therefore |
US7057637B2 (en) | 2004-04-21 | 2006-06-06 | White Peter Mcduffie | Reflected backdrop for communications systems |
US20050271194A1 (en) * | 2004-06-07 | 2005-12-08 | Woods Paul R | Conference phone and network client |
US7865834B1 (en) * | 2004-06-25 | 2011-01-04 | Apple Inc. | Multi-way video conferencing user interface |
US7500795B2 (en) * | 2004-09-09 | 2009-03-10 | Paul Sandhu | Apparatuses, systems and methods for enhancing telemedicine, video-conferencing, and video-based sales |
US7692683B2 (en) * | 2004-10-15 | 2010-04-06 | Lifesize Communications, Inc. | Video conferencing system transcoder |
EP1943841A2 (en) * | 2004-11-04 | 2008-07-16 | Megamedia, LLC | Apparatus and methods for encoding data for video compositing |
US7619658B2 (en) * | 2004-11-15 | 2009-11-17 | Hewlett-Packard Development Company, L.P. | Methods and systems for producing seamless composite images without requiring overlap of source images |
US7515174B1 (en) | 2004-12-06 | 2009-04-07 | Dreamworks Animation L.L.C. | Multi-user video conferencing with perspective correct eye-to-eye contact |
US7524119B2 (en) * | 2005-02-03 | 2009-04-28 | Paul Sandhu | Apparatus and method for viewing radiographs |
US7986335B2 (en) * | 2005-05-02 | 2011-07-26 | Lifesize Communications, Inc. | Set top box videoconferencing system |
US20070165106A1 (en) * | 2005-05-02 | 2007-07-19 | Groves Randall D | Distributed Videoconferencing Processing |
US20060248210A1 (en) * | 2005-05-02 | 2006-11-02 | Lifesize Communications, Inc. | Controlling video display mode in a video conferencing system |
US20060291637A1 (en) * | 2005-06-13 | 2006-12-28 | David Erickson | Systems and methods for a reliable teleconferencing system |
US20070030984A1 (en) * | 2005-08-02 | 2007-02-08 | Gotfried Bradley L | Conference system |
CN101496387B (en) | 2006-03-06 | 2012-09-05 | 思科技术公司 | System and method for access authentication in a mobile wireless network |
GB0718706D0 (en) | 2007-09-25 | 2007-11-07 | Creative Physics Ltd | Method and apparatus for reducing laser speckle |
US7679639B2 (en) * | 2006-04-20 | 2010-03-16 | Cisco Technology, Inc. | System and method for enhancing eye gaze in a telepresence system |
US7692680B2 (en) * | 2006-04-20 | 2010-04-06 | Cisco Technology, Inc. | System and method for providing location specific sound in a telepresence system |
US7532232B2 (en) * | 2006-04-20 | 2009-05-12 | Cisco Technology, Inc. | System and method for single action initiation of a video conference |
US20070250567A1 (en) * | 2006-04-20 | 2007-10-25 | Graham Philip R | System and method for controlling a telepresence system |
US7707247B2 (en) * | 2006-04-20 | 2010-04-27 | Cisco Technology, Inc. | System and method for displaying users in a visual conference between locations |
US7710448B2 (en) | 2006-04-20 | 2010-05-04 | Cisco Technology, Inc. | System and method for preventing movement in a telepresence system |
US20080001951A1 (en) * | 2006-05-07 | 2008-01-03 | Sony Computer Entertainment Inc. | System and method for providing affective characteristics to computer generated avatar during gameplay |
US20070279483A1 (en) * | 2006-05-31 | 2007-12-06 | Beers Ted W | Blended Space For Aligning Video Streams |
US7889225B2 (en) * | 2006-07-12 | 2011-02-15 | Hewlett-Packard Development Company, L.P. | Videoconferencing environment |
US20080055554A1 (en) * | 2006-08-30 | 2008-03-06 | Keith Tubin | Full Scale Plan Projection |
US7924306B2 (en) * | 2006-09-15 | 2011-04-12 | Hewlett-Packard Development Company, L.P. | Videoconferencing with enhanced illusion of physical presence in a common space |
US7924305B2 (en) * | 2006-09-15 | 2011-04-12 | Hewlett-Packard Development Company, L.P. | Consistent quality for multipoint videoconferencing systems |
US8330791B2 (en) * | 2007-04-30 | 2012-12-11 | Hewlett-Packard Development Company, L.P. | Video conference system with symmetric reference |
CN101743567A (en) | 2007-05-18 | 2010-06-16 | Uab研究基金会 | Virtual interactive presence system and method |
US8253770B2 (en) * | 2007-05-31 | 2012-08-28 | Eastman Kodak Company | Residential video communication system |
US8154583B2 (en) * | 2007-05-31 | 2012-04-10 | Eastman Kodak Company | Eye gazing imaging for video communications |
US8063929B2 (en) * | 2007-05-31 | 2011-11-22 | Eastman Kodak Company | Managing scene transitions for video communication |
US8159519B2 (en) * | 2007-05-31 | 2012-04-17 | Eastman Kodak Company | Personal controls for personal video communications |
US8154578B2 (en) * | 2007-05-31 | 2012-04-10 | Eastman Kodak Company | Multi-camera residential communication system |
US20080316295A1 (en) * | 2007-06-22 | 2008-12-25 | King Keith C | Virtual decoders |
US8139100B2 (en) * | 2007-07-13 | 2012-03-20 | Lifesize Communications, Inc. | Virtual multiway scaler compensation |
US8773495B2 (en) * | 2007-10-12 | 2014-07-08 | Polycom, Inc. | Integrated system for telepresence videoconferencing |
US8074581B2 (en) | 2007-10-12 | 2011-12-13 | Steelcase Inc. | Conference table assembly |
US8379076B2 (en) | 2008-01-07 | 2013-02-19 | Cisco Technology, Inc. | System and method for displaying a multipoint videoconference |
US8797377B2 (en) | 2008-02-14 | 2014-08-05 | Cisco Technology, Inc. | Method and system for videoconference configuration |
US8355041B2 (en) | 2008-02-14 | 2013-01-15 | Cisco Technology, Inc. | Telepresence system for 360 degree video conferencing |
US8319819B2 (en) | 2008-03-26 | 2012-11-27 | Cisco Technology, Inc. | Virtual round-table videoconference |
US8390667B2 (en) | 2008-04-15 | 2013-03-05 | Cisco Technology, Inc. | Pop-up PIP for people not in picture |
US8270075B2 (en) | 2008-05-08 | 2012-09-18 | Musion Ip Limited | Projection apparatuses and associated methods |
US8643691B2 (en) * | 2008-05-12 | 2014-02-04 | Microsoft Corporation | Gaze accurate video conferencing |
US20100035682A1 (en) * | 2008-07-01 | 2010-02-11 | Yoostar Entertainment Group, Inc. | User interface systems and methods for interactive video systems |
GB0910117D0 (en) | 2008-07-14 | 2009-07-29 | Holicom Film Ltd | Method and system for filming |
GB0821996D0 (en) | 2008-12-02 | 2009-01-07 | Musion Ip Ltd | Mobile studio |
US20100031152A1 (en) * | 2008-07-31 | 2010-02-04 | Microsoft Corporation | Creation and Navigation of Infinite Canvas Presentation |
US8108777B2 (en) | 2008-08-11 | 2012-01-31 | Microsoft Corporation | Sections of a presentation having user-definable properties |
US8694658B2 (en) | 2008-09-19 | 2014-04-08 | Cisco Technology, Inc. | System and method for enabling communication sessions in a network environment |
US8514265B2 (en) * | 2008-10-02 | 2013-08-20 | Lifesize Communications, Inc. | Systems and methods for selecting videoconferencing endpoints for display in a composite video image |
US20140361954A1 (en) | 2013-06-07 | 2014-12-11 | Lewis Epstein | Personal control apparatus and method for sharing information in a collaboration workspace |
US10631632B2 (en) | 2008-10-13 | 2020-04-28 | Steelcase Inc. | Egalitarian control apparatus and method for sharing information in a collaborative workspace |
US20100110160A1 (en) * | 2008-10-30 | 2010-05-06 | Brandt Matthew K | Videoconferencing Community with Live Images |
GB0918115D0 (en) | 2009-10-16 | 2009-12-02 | Musion Ip Ltd | A method of manufacturing foil for producing a pepper's ghost illusion |
US9563115B2 (en) | 2008-12-24 | 2017-02-07 | Musion Ip Limited | Method of manufacturing foil for producing a pepper's ghost illusion |
US20100188476A1 (en) * | 2009-01-29 | 2010-07-29 | Optical Fusion Inc. | Image Quality of Video Conferences |
US8643695B2 (en) * | 2009-03-04 | 2014-02-04 | Lifesize Communications, Inc. | Videoconferencing endpoint extension |
US8456510B2 (en) * | 2009-03-04 | 2013-06-04 | Lifesize Communications, Inc. | Virtual distributed multipoint control unit |
US8659637B2 (en) | 2009-03-09 | 2014-02-25 | Cisco Technology, Inc. | System and method for providing three dimensional video conferencing in a network environment |
US8477175B2 (en) | 2009-03-09 | 2013-07-02 | Cisco Technology, Inc. | System and method for providing three dimensional imaging in a network environment |
US8274544B2 (en) * | 2009-03-23 | 2012-09-25 | Eastman Kodak Company | Automated videography systems |
US8237771B2 (en) * | 2009-03-26 | 2012-08-07 | Eastman Kodak Company | Automated videography based communications |
US9335604B2 (en) | 2013-12-11 | 2016-05-10 | Milan Momcilo Popovich | Holographic waveguide display |
US11726332B2 (en) | 2009-04-27 | 2023-08-15 | Digilens Inc. | Diffractive projection apparatus |
US10127524B2 (en) | 2009-05-26 | 2018-11-13 | Microsoft Technology Licensing, Llc | Shared collaboration canvas |
US8830293B2 (en) * | 2009-05-26 | 2014-09-09 | Cisco Technology, Inc. | Video superposition for continuous presence |
US20100306018A1 (en) * | 2009-05-27 | 2010-12-02 | Microsoft Corporation | Meeting State Recall |
US10884607B1 (en) | 2009-05-29 | 2021-01-05 | Steelcase Inc. | Personal control apparatus and method for sharing information in a collaborative workspace |
US8659639B2 (en) | 2009-05-29 | 2014-02-25 | Cisco Technology, Inc. | System and method for extending communications between participants in a conferencing environment |
US9082297B2 (en) | 2009-08-11 | 2015-07-14 | Cisco Technology, Inc. | System and method for verifying parameters in an audiovisual environment |
US8233204B1 (en) | 2009-09-30 | 2012-07-31 | Rockwell Collins, Inc. | Optical displays |
US9341846B2 (en) | 2012-04-25 | 2016-05-17 | Rockwell Collins Inc. | Holographic wide angle display |
US11300795B1 (en) | 2009-09-30 | 2022-04-12 | Digilens Inc. | Systems for and methods of using fold gratings coordinated with output couplers for dual axis expansion |
US11320571B2 (en) | 2012-11-16 | 2022-05-03 | Rockwell Collins, Inc. | Transparent waveguide display providing upper and lower fields of view with uniform light extraction |
US10795160B1 (en) | 2014-09-25 | 2020-10-06 | Rockwell Collins, Inc. | Systems for and methods of using fold gratings for dual axis expansion |
US8350891B2 (en) * | 2009-11-16 | 2013-01-08 | Lifesize Communications, Inc. | Determining a videoconference layout based on numbers of participants |
US8659826B1 (en) | 2010-02-04 | 2014-02-25 | Rockwell Collins, Inc. | Worn display system and method without requiring real time tracking for boresight precision |
US8471889B1 (en) | 2010-03-11 | 2013-06-25 | Sprint Communications Company L.P. | Adjusting an image for video conference display |
US9225916B2 (en) | 2010-03-18 | 2015-12-29 | Cisco Technology, Inc. | System and method for enhancing video images in a conferencing environment |
USD626102S1 (en) | 2010-03-21 | 2010-10-26 | Cisco Tech Inc | Video unit with integrated features |
USD626103S1 (en) | 2010-03-21 | 2010-10-26 | Cisco Technology, Inc. | Video unit with integrated features |
GB201007580D0 (en) | 2010-05-06 | 2010-06-23 | Im Sense Ltd | Making robust images |
US9313452B2 (en) | 2010-05-17 | 2016-04-12 | Cisco Technology, Inc. | System and method for providing retracting optics in a video conferencing environment |
CN102300043B (en) * | 2010-06-23 | 2014-06-11 | 中兴通讯股份有限公司 | Method for adjusting meeting place camera of remote presentation meeting system |
US8896655B2 (en) * | 2010-08-31 | 2014-11-25 | Cisco Technology, Inc. | System and method for providing depth adaptive video conferencing |
US8599934B2 (en) | 2010-09-08 | 2013-12-03 | Cisco Technology, Inc. | System and method for skip coding during video conferencing in a network environment |
US8599865B2 (en) | 2010-10-26 | 2013-12-03 | Cisco Technology, Inc. | System and method for provisioning flows in a mobile network environment |
US8699457B2 (en) | 2010-11-03 | 2014-04-15 | Cisco Technology, Inc. | System and method for managing flows in a mobile network environment |
US9338394B2 (en) | 2010-11-15 | 2016-05-10 | Cisco Technology, Inc. | System and method for providing enhanced audio in a video environment |
US9143725B2 (en) | 2010-11-15 | 2015-09-22 | Cisco Technology, Inc. | System and method for providing enhanced graphics in a video environment |
US8730297B2 (en) | 2010-11-15 | 2014-05-20 | Cisco Technology, Inc. | System and method for providing camera functions in a video environment |
US8902244B2 (en) | 2010-11-15 | 2014-12-02 | Cisco Technology, Inc. | System and method for providing enhanced graphics in a video environment |
US8542264B2 (en) | 2010-11-18 | 2013-09-24 | Cisco Technology, Inc. | System and method for managing optics in a video environment |
US8723914B2 (en) | 2010-11-19 | 2014-05-13 | Cisco Technology, Inc. | System and method for providing enhanced video processing in a network environment |
US9111138B2 (en) | 2010-11-30 | 2015-08-18 | Cisco Technology, Inc. | System and method for gesture interface control |
US8446455B2 (en) | 2010-12-08 | 2013-05-21 | Cisco Technology, Inc. | System and method for exchanging information in a video conference environment |
US8553064B2 (en) | 2010-12-08 | 2013-10-08 | Cisco Technology, Inc. | System and method for controlling video data to be rendered in a video conference environment |
US9383888B2 (en) | 2010-12-15 | 2016-07-05 | Microsoft Technology Licensing, Llc | Optimized joint document review |
US9118612B2 (en) | 2010-12-15 | 2015-08-25 | Microsoft Technology Licensing, Llc | Meeting-specific state indicators |
USD678894S1 (en) | 2010-12-16 | 2013-03-26 | Cisco Technology, Inc. | Display screen with graphical user interface |
USD682854S1 (en) | 2010-12-16 | 2013-05-21 | Cisco Technology, Inc. | Display screen for graphical user interface |
USD678320S1 (en) | 2010-12-16 | 2013-03-19 | Cisco Technology, Inc. | Display screen with graphical user interface |
USD682864S1 (en) | 2010-12-16 | 2013-05-21 | Cisco Technology, Inc. | Display screen with graphical user interface |
USD678308S1 (en) | 2010-12-16 | 2013-03-19 | Cisco Technology, Inc. | Display screen with graphical user interface |
USD682294S1 (en) | 2010-12-16 | 2013-05-14 | Cisco Technology, Inc. | Display screen with graphical user interface |
USD682293S1 (en) | 2010-12-16 | 2013-05-14 | Cisco Technology, Inc. | Display screen with graphical user interface |
USD678307S1 (en) | 2010-12-16 | 2013-03-19 | Cisco Technology, Inc. | Display screen with graphical user interface |
US9864612B2 (en) | 2010-12-23 | 2018-01-09 | Microsoft Technology Licensing, Llc | Techniques to customize a user interface for different displays |
US8384757B2 (en) | 2011-01-11 | 2013-02-26 | Baker Hughes Incorporated | System and method for providing videoconferencing among a plurality of locations |
US8692862B2 (en) | 2011-02-28 | 2014-04-08 | Cisco Technology, Inc. | System and method for selection of video data in a video conference environment |
US8698872B2 (en) | 2011-03-02 | 2014-04-15 | At&T Intellectual Property I, Lp | System and method for notification of events of interest during a video conference |
EP2506568A1 (en) * | 2011-03-30 | 2012-10-03 | Alcatel Lucent | A method, a system, a device, a computer program and a computer program product for conducting a video conference |
US9274349B2 (en) | 2011-04-07 | 2016-03-01 | Digilens Inc. | Laser despeckler based on angular diversity |
US8670019B2 (en) | 2011-04-28 | 2014-03-11 | Cisco Technology, Inc. | System and method for providing enhanced eye gaze in a video conferencing environment |
US8786631B1 (en) | 2011-04-30 | 2014-07-22 | Cisco Technology, Inc. | System and method for transferring transparency information in a video environment |
US8934026B2 (en) | 2011-05-12 | 2015-01-13 | Cisco Technology, Inc. | System and method for video coding in a dynamic environment |
US8767586B2 (en) | 2011-06-20 | 2014-07-01 | At&T Intellectual Property I, L.P. | Methods, systems, and products for network topology |
US9886552B2 (en) | 2011-08-12 | 2018-02-06 | Help Lighting, Inc. | System and method for image registration of multiple video streams |
EP2748670B1 (en) | 2011-08-24 | 2015-11-18 | Rockwell Collins, Inc. | Wearable data display |
WO2016020630A2 (en) | 2014-08-08 | 2016-02-11 | Milan Momcilo Popovich | Waveguide laser illuminator incorporating a despeckler |
US10670876B2 (en) | 2011-08-24 | 2020-06-02 | Digilens Inc. | Waveguide laser illuminator incorporating a despeckler |
US8756348B2 (en) | 2011-09-14 | 2014-06-17 | Barco N.V. | Electronic tool and methods for meetings |
US11258676B2 (en) | 2011-09-14 | 2022-02-22 | Barco N.V. | Electronic tool and methods for meetings |
US10965480B2 (en) | 2011-09-14 | 2021-03-30 | Barco N.V. | Electronic tool and methods for recording a meeting |
CN103931175B (en) * | 2011-09-14 | 2018-04-20 | 巴科股份有限公司 | electronic tool and method for conference audio |
SG11201507376QA (en) * | 2011-09-14 | 2015-10-29 | Barco Nv | Electronic tool and methods for meetings |
US9083769B2 (en) | 2011-09-14 | 2015-07-14 | Barco N.V. | Electronic tool and methods for meetings |
US10050800B2 (en) | 2011-09-14 | 2018-08-14 | Barco N.V. | Electronic tool and methods for meetings for providing connection to a communications network |
CN103141087A (en) * | 2011-09-14 | 2013-06-05 | 巴可有限公司 | Electronic tool and methods for meetings |
US9715067B1 (en) | 2011-09-30 | 2017-07-25 | Rockwell Collins, Inc. | Ultra-compact HUD utilizing waveguide pupil expander with surface relief gratings in high refractive index materials |
US8634139B1 (en) | 2011-09-30 | 2014-01-21 | Rockwell Collins, Inc. | System for and method of catadioptric collimation in a compact head up display (HUD) |
US9599813B1 (en) | 2011-09-30 | 2017-03-21 | Rockwell Collins, Inc. | Waveguide combiner system and method with less susceptibility to glare |
US9366864B1 (en) | 2011-09-30 | 2016-06-14 | Rockwell Collins, Inc. | System for and method of displaying information without need for a combiner alignment detector |
US9544158B2 (en) | 2011-10-05 | 2017-01-10 | Microsoft Technology Licensing, Llc | Workspace collaboration via a wall-type computing device |
US8682973B2 (en) | 2011-10-05 | 2014-03-25 | Microsoft Corporation | Multi-user and multi-device collaboration |
US9996241B2 (en) | 2011-10-11 | 2018-06-12 | Microsoft Technology Licensing, Llc | Interactive visualization of multiple software functionality content items |
US10198485B2 (en) | 2011-10-13 | 2019-02-05 | Microsoft Technology Licensing, Llc | Authoring of data visualizations and maps |
US8947493B2 (en) | 2011-11-16 | 2015-02-03 | Cisco Technology, Inc. | System and method for alerting a participant in a video conference |
CN102368816A (en) * | 2011-12-01 | 2012-03-07 | 中科芯集成电路股份有限公司 | Intelligent front end system of video conference |
US8682087B2 (en) | 2011-12-19 | 2014-03-25 | Cisco Technology, Inc. | System and method for depth-guided image filtering in a video conference environment |
US20150010265A1 (en) | 2012-01-06 | 2015-01-08 | Milan, Momcilo POPOVICH | Contact image sensor using switchable bragg gratings |
US9523852B1 (en) | 2012-03-28 | 2016-12-20 | Rockwell Collins, Inc. | Micro collimator system and method for a head up display (HUD) |
GB2501471A (en) | 2012-04-18 | 2013-10-30 | Barco Nv | Electronic conference arrangement |
US9020203B2 (en) | 2012-05-21 | 2015-04-28 | Vipaar, Llc | System and method for managing spatiotemporal uncertainty |
US8804321B2 (en) | 2012-05-25 | 2014-08-12 | Steelcase, Inc. | Work and videoconference assembly |
US9151150B2 (en) | 2012-10-23 | 2015-10-06 | Baker Hughes Incorporated | Apparatus and methods for well-bore proximity measurement while drilling |
US9933684B2 (en) | 2012-11-16 | 2018-04-03 | Rockwell Collins, Inc. | Transparent waveguide display providing upper and lower fields of view having a specific light output aperture configuration |
US9681154B2 (en) | 2012-12-06 | 2017-06-13 | Patent Capital Group | System and method for depth-guided filtering in a video conference environment |
US9710968B2 (en) | 2012-12-26 | 2017-07-18 | Help Lightning, Inc. | System and method for role-switching in multi-reality environments |
US10062106B2 (en) | 2013-03-15 | 2018-08-28 | Mirza M. AHMAD | Menu sharing systems and methods for teledining |
US9674413B1 (en) | 2013-04-17 | 2017-06-06 | Rockwell Collins, Inc. | Vision system and method having improved performance and solar mitigation |
WO2014182508A1 (en) | 2013-05-06 | 2014-11-13 | Yoostar Entertainment Group, Inc. | Audio-video compositing and effects |
US9843621B2 (en) | 2013-05-17 | 2017-12-12 | Cisco Technology, Inc. | Calendaring activities based on communication processing |
US9603457B2 (en) | 2013-05-31 | 2017-03-28 | Steelcase Inc. | Lounge assemblies for supporting portable electronics devices |
US9940750B2 (en) | 2013-06-27 | 2018-04-10 | Help Lighting, Inc. | System and method for role negotiation in multi-reality environments |
WO2015015138A1 (en) | 2013-07-31 | 2015-02-05 | Milan Momcilo Popovich | Method and apparatus for contact image sensing |
US9244281B1 (en) | 2013-09-26 | 2016-01-26 | Rockwell Collins, Inc. | Display system and method using a detached combiner |
US9210377B2 (en) | 2013-10-30 | 2015-12-08 | At&T Intellectual Property I, L.P. | Methods, systems, and products for telepresence visualizations |
US10075656B2 (en) | 2013-10-30 | 2018-09-11 | At&T Intellectual Property I, L.P. | Methods, systems, and products for telepresence visualizations |
US10732407B1 (en) | 2014-01-10 | 2020-08-04 | Rockwell Collins, Inc. | Near eye head up display system and method with fixed combiner |
US9524588B2 (en) | 2014-01-24 | 2016-12-20 | Avaya Inc. | Enhanced communication between remote participants using augmented and virtual reality |
US9519089B1 (en) | 2014-01-30 | 2016-12-13 | Rockwell Collins, Inc. | High performance volume phase gratings |
US9479730B1 (en) | 2014-02-13 | 2016-10-25 | Steelcase, Inc. | Inferred activity based conference enhancement method and system |
US9244280B1 (en) | 2014-03-25 | 2016-01-26 | Rockwell Collins, Inc. | Near eye display system and method for display enhancement or redundancy |
US10348724B2 (en) | 2014-04-07 | 2019-07-09 | Barco N.V. | Ad hoc one-time pairing of remote devices using online audio fingerprinting |
US10074374B2 (en) | 2014-04-07 | 2018-09-11 | Barco N.V. | Ad hoc one-time pairing of remote devices using online audio fingerprinting |
US10359736B2 (en) | 2014-08-08 | 2019-07-23 | Digilens Inc. | Method for holographic mastering and replication |
US10241330B2 (en) | 2014-09-19 | 2019-03-26 | Digilens, Inc. | Method and apparatus for generating input images for holographic waveguide displays |
US10088675B1 (en) | 2015-05-18 | 2018-10-02 | Rockwell Collins, Inc. | Turning light pipe for a pupil expansion system and method |
US9715110B1 (en) | 2014-09-25 | 2017-07-25 | Rockwell Collins, Inc. | Automotive head up display (HUD) |
CN111323867A (en) | 2015-01-12 | 2020-06-23 | 迪吉伦斯公司 | Environmentally isolated waveguide display |
US9541998B2 (en) | 2015-01-29 | 2017-01-10 | Samsung Electronics Co., Ltd. | Electronic system with gaze alignment mechanism and method of operation thereof |
US9632226B2 (en) | 2015-02-12 | 2017-04-25 | Digilens Inc. | Waveguide grating device |
US9628754B2 (en) | 2015-04-02 | 2017-04-18 | Telepresence Technologies, Llc | TelePresence architectural systems and methods therefore |
US10247943B1 (en) | 2015-05-18 | 2019-04-02 | Rockwell Collins, Inc. | Head up display (HUD) using a light pipe |
US10126552B2 (en) | 2015-05-18 | 2018-11-13 | Rockwell Collins, Inc. | Micro collimator system and method for a head up display (HUD) |
US11366316B2 (en) | 2015-05-18 | 2022-06-21 | Rockwell Collins, Inc. | Head up display (HUD) using a light pipe |
US10108010B2 (en) | 2015-06-29 | 2018-10-23 | Rockwell Collins, Inc. | System for and method of integrating head up displays and head down displays |
JP6598269B2 (en) | 2015-10-05 | 2019-10-30 | ディジレンズ インコーポレイテッド | Waveguide display |
US10598932B1 (en) | 2016-01-06 | 2020-03-24 | Rockwell Collins, Inc. | Head up display for integrating views of conformally mapped symbols and a fixed image source |
US10859768B2 (en) | 2016-03-24 | 2020-12-08 | Digilens Inc. | Method and apparatus for providing a polarization selective holographic waveguide device |
EP3433658B1 (en) | 2016-04-11 | 2023-08-09 | DigiLens, Inc. | Holographic waveguide apparatus for structured light projection |
USD838129S1 (en) | 2016-04-15 | 2019-01-15 | Steelcase Inc. | Worksurface for a conference table |
USD808197S1 (en) | 2016-04-15 | 2018-01-23 | Steelcase Inc. | Support for a table |
US10219614B2 (en) | 2016-04-15 | 2019-03-05 | Steelcase Inc. | Reconfigurable conference table |
USD862127S1 (en) | 2016-04-15 | 2019-10-08 | Steelcase Inc. | Conference table |
WO2017195514A1 (en) * | 2016-05-13 | 2017-11-16 | ソニー株式会社 | Image processing device, image processing system, and image processing method, and program |
US9729820B1 (en) | 2016-09-02 | 2017-08-08 | Russell Holmes | Systems and methods for providing real-time composite video from multiple source devices |
US10089793B2 (en) | 2016-09-02 | 2018-10-02 | Russell Holmes | Systems and methods for providing real-time composite video from multiple source devices featuring augmented reality elements |
US11513350B2 (en) | 2016-12-02 | 2022-11-29 | Digilens Inc. | Waveguide device with uniform output illumination |
US10264213B1 (en) | 2016-12-15 | 2019-04-16 | Steelcase Inc. | Content amplification system and method |
WO2018129398A1 (en) | 2017-01-05 | 2018-07-12 | Digilens, Inc. | Wearable heads up displays |
US10295824B2 (en) | 2017-01-26 | 2019-05-21 | Rockwell Collins, Inc. | Head up display with an angled light pipe |
MX2019009971A (en) * | 2017-02-27 | 2020-02-12 | Espinosa Baruque Pablo | Display of a three dimensional recodring in a system for rehabilitation. |
US10942430B2 (en) | 2017-10-16 | 2021-03-09 | Digilens Inc. | Systems and methods for multiplying the image resolution of a pixelated display |
US20190212588A1 (en) | 2018-01-08 | 2019-07-11 | Digilens, Inc. | Systems and Methods for Manufacturing Waveguide Cells |
US10914950B2 (en) | 2018-01-08 | 2021-02-09 | Digilens Inc. | Waveguide architectures and related methods of manufacturing |
US10732569B2 (en) | 2018-01-08 | 2020-08-04 | Digilens Inc. | Systems and methods for high-throughput recording of holographic gratings in waveguide cells |
JP2019180027A (en) * | 2018-03-30 | 2019-10-17 | 株式会社リコー | Communication terminal, image communication system, display method, and program |
US11402801B2 (en) | 2018-07-25 | 2022-08-02 | Digilens Inc. | Systems and methods for fabricating a multilayer optical structure |
WO2020149956A1 (en) | 2019-01-14 | 2020-07-23 | Digilens Inc. | Holographic waveguide display with light control layer |
EP3924759A4 (en) | 2019-02-15 | 2022-12-28 | Digilens Inc. | METHODS AND APPARATUS FOR MAKING A HOLOGRAPHIC WAVEGUIDE DISPLAY WITH INTEGRATED GRIDINGS |
US20220283377A1 (en) | 2019-02-15 | 2022-09-08 | Digilens Inc. | Wide Angle Waveguide Display |
WO2020186113A1 (en) | 2019-03-12 | 2020-09-17 | Digilens Inc. | Holographic waveguide backlight and related methods of manufacturing |
JP2022535460A (en) | 2019-06-07 | 2022-08-08 | ディジレンズ インコーポレイテッド | Waveguides incorporating transmission and reflection gratings, and associated fabrication methods |
KR20220038452A (en) | 2019-07-29 | 2022-03-28 | 디지렌즈 인코포레이티드. | Method and apparatus for multiplying the image resolution and field of view of a pixelated display |
US11442222B2 (en) | 2019-08-29 | 2022-09-13 | Digilens Inc. | Evacuated gratings and methods of manufacturing |
EP4288831A4 (en) | 2021-03-05 | 2025-01-15 | Digilens Inc | EVACUATED PERIODIC STRUCTURES AND MANUFACTURING METHODS |
KR102426374B1 (en) * | 2022-01-11 | 2022-07-27 | 최해용 | Virtual face-to-face table device |
Citations (64)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3233346A (en) * | 1962-09-13 | 1966-02-08 | Cornberg Sol | Carrel |
US3869992A (en) * | 1972-09-11 | 1975-03-11 | Edward J Kramer | Clusterable learning module |
US4004084A (en) * | 1976-03-03 | 1977-01-18 | Bell Telephone Laboratories, Incorporated | Video conferencing system using spatial reduction and temporal resolution |
USD252635S (en) * | 1977-04-27 | 1979-08-14 | Control Data Corporation | Portable learning unit |
US4400724A (en) * | 1981-06-08 | 1983-08-23 | The United States Of America As Represented By The Secretary Of The Army | Virtual space teleconference system |
US4546383A (en) * | 1982-06-18 | 1985-10-08 | Inria Institute National De Recherche En Informatique Et En Automatique | Method and apparatus for visual telecommunications, in particular for use by the deaf |
US4650929A (en) * | 1984-02-29 | 1987-03-17 | Heinrich-Hertz-Institut Fur Nachrichtentechnik Berlin Gmbh | Communication system for videoconferencing |
US4703348A (en) * | 1984-04-25 | 1987-10-27 | Matsushita Electric Works, Ltd. | Picture transmission system using secondary differential variable sampling rate coding |
US4707077A (en) * | 1986-01-30 | 1987-11-17 | Hughes Aircraft Company | Real time image subtraction with a single liquid crystal light valve |
US4758887A (en) * | 1985-10-09 | 1988-07-19 | Weyel Kg | Conference table |
US4800432A (en) * | 1986-10-24 | 1989-01-24 | The Grass Valley Group, Inc. | Video Difference key generator |
US4823194A (en) * | 1986-08-01 | 1989-04-18 | Hitachi, Ltd. | Method for processing gray scale images and an apparatus thereof |
US4845636A (en) * | 1986-10-17 | 1989-07-04 | Walker Mark E | Remote transaction system |
US4890314A (en) * | 1988-08-26 | 1989-12-26 | Bell Communications Research, Inc. | Teleconference facility with high resolution video display |
US4928301A (en) * | 1988-12-30 | 1990-05-22 | Bell Communications Research, Inc. | Teleconferencing terminal with camera behind display screen |
US4931872A (en) * | 1986-11-24 | 1990-06-05 | Gte Laboratories Incorporated | Methods of and apparatus for the generation of split-screen video displays |
US4942470A (en) * | 1984-07-20 | 1990-07-17 | Nec Corporation | Real time processor for video signals |
US4965666A (en) * | 1987-11-27 | 1990-10-23 | U.S. Philips Corporation | Method of and arrangement for estimating and compensatiing motion in a sequence of pictures and a picture transmission system provided with such an arrangement |
US4965819A (en) * | 1988-09-22 | 1990-10-23 | Docu-Vision, Inc. | Video conferencing system for courtroom and other applications |
US4986194A (en) * | 1986-11-26 | 1991-01-22 | Bollman Clifford J | Multi-adjustable, multi-functional, space saving workplace system |
US4991009A (en) * | 1988-07-08 | 1991-02-05 | Ricoh Company, Ltd. | Dynamic image transmission system |
US4996592A (en) * | 1988-08-12 | 1991-02-26 | Oki Electric Industry Co., Ltd. | Video communication apparatus capable of electronically adjusting the field of view of a video frame |
US5014267A (en) * | 1989-04-06 | 1991-05-07 | Datapoint Corporation | Video conferencing network |
US5016405A (en) * | 1989-04-18 | 1991-05-21 | Rosemount Office Systems, Inc. | Modular workstations |
US5024398A (en) * | 1989-06-28 | 1991-06-18 | The Boeing Company | Office module for passenger aircraft |
US5032901A (en) * | 1988-06-28 | 1991-07-16 | Petro Vlahos | Backing color and luminance nonuniformity compensation for linear image compositing |
US5031620A (en) * | 1986-11-14 | 1991-07-16 | Kabushiki Kaisha Toshiba | Image processing apparatus |
US5038224A (en) * | 1989-12-22 | 1991-08-06 | Bell Communications Research, Inc. | Video imaging device in a teleconferencing system |
US5055927A (en) * | 1988-09-13 | 1991-10-08 | Deutsche Thomson-Brandt Gmbh | Dual channel video signal transmission system |
US5079627A (en) * | 1988-06-15 | 1992-01-07 | Optum Corporation | Videophone |
US5117285A (en) * | 1991-01-15 | 1992-05-26 | Bell Communications Research | Eye contact apparatus for video conferencing |
US5122865A (en) * | 1988-09-15 | 1992-06-16 | Samsung Electronics Co., Ltd. | Chroma key signal generator for a video editing apparatus |
US5127078A (en) * | 1991-02-13 | 1992-06-30 | Terry Weed | Integrated display for image presentation and receiving using fiber optics |
US5135720A (en) * | 1989-05-23 | 1992-08-04 | Seiko Instruments Inc. | Reaction chamber for amino-terminal sequence analysis of proteins or peptides |
US5157491A (en) * | 1988-10-17 | 1992-10-20 | Kassatly L Samuel A | Method and apparatus for video broadcasting and teleconferencing |
US5159451A (en) * | 1991-03-19 | 1992-10-27 | Faroudja Y C | Field memory expansible line doubler for television receiver |
US5168528A (en) * | 1990-08-20 | 1992-12-01 | Itt Corporation | Differential electronic imaging system |
US5173772A (en) * | 1991-12-14 | 1992-12-22 | Samsung Electronics Co., Ltd. | Method for detecting motion vector |
US5187571A (en) * | 1991-02-01 | 1993-02-16 | Bell Communications Research, Inc. | Television system for displaying multiple views of a remote location |
US5214512A (en) * | 1991-02-11 | 1993-05-25 | Ampex Systems Corporation | Keyed, true-transparency image information combine |
US5249039A (en) * | 1991-11-18 | 1993-09-28 | The Grass Valley Group, Inc. | Chroma key method and apparatus |
US5257306A (en) * | 1991-05-29 | 1993-10-26 | Fujitsu Limited | Video teleconferencing system |
US5261735A (en) * | 1992-02-03 | 1993-11-16 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Deployable video conference table |
US5272526A (en) * | 1990-05-30 | 1993-12-21 | Sony Corporation | Television conference system |
US5280540A (en) * | 1991-10-09 | 1994-01-18 | Bell Communications Research, Inc. | Video teleconferencing system employing aspect ratio transformation |
US5282027A (en) * | 1990-04-27 | 1994-01-25 | U.S. Philips Corporation | Image projection display and pick-up apparatus with optical shutter |
US5282341A (en) * | 1992-01-10 | 1994-02-01 | Steelcase Inc. | Dynamic workspace module |
US5315633A (en) * | 1991-12-20 | 1994-05-24 | Unisys Corporation | Digital video switch for video teleconferencing |
US5347306A (en) * | 1993-12-17 | 1994-09-13 | Mitsubishi Electric Research Laboratories, Inc. | Animated electronic meeting place |
US5352033A (en) * | 1992-03-27 | 1994-10-04 | Fisher-Rosemount Systems, Inc. | Operator work station having a monitor support assembly |
US5393964A (en) * | 1990-10-12 | 1995-02-28 | Tpi, Inc. | Telecommunications booth and method of use |
US5394198A (en) * | 1992-12-22 | 1995-02-28 | At&T Corp. | Large-screen display system |
US5397133A (en) * | 1993-09-30 | 1995-03-14 | At&T Corp. | System for playing card games remotely |
US5400069A (en) * | 1993-06-16 | 1995-03-21 | Bell Communications Research, Inc. | Eye contact video-conferencing system and screen |
US5406893A (en) * | 1992-09-25 | 1995-04-18 | Burns; J. Robert | Work-study carrel |
US5438357A (en) * | 1993-11-23 | 1995-08-01 | Mcnelley; Steve H. | Image manipulating teleconferencing system |
US5508743A (en) * | 1991-12-06 | 1996-04-16 | Canon Kabushiki Kaisha | Moving image signal coding apparatus and moving image signal coding control method |
US5541640A (en) * | 1992-06-23 | 1996-07-30 | Larson; Craig R. | Videophone for simultaneous audio and video communication via a standard telephone line |
US5572248A (en) * | 1994-09-19 | 1996-11-05 | Teleport Corporation | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US5689641A (en) * | 1993-10-01 | 1997-11-18 | Vicor, Inc. | Multimedia collaboration system arrangement for routing compressed AV signal through a participant site without decompressing the AV signal |
US5691777A (en) * | 1988-10-17 | 1997-11-25 | Kassatly; Lord Samuel Anthony | Method and apparatus for simultaneous compression of video, audio and data signals |
US5784843A (en) * | 1994-12-30 | 1998-07-28 | Steelcase Inc. | Integrated prefabricated furniture system for fitting-out open plan building space |
US5850250A (en) * | 1994-07-18 | 1998-12-15 | Bell Atlantic Maryland, Inc. | Video distance learning system |
US20070263081A1 (en) * | 2006-04-20 | 2007-11-15 | De Beer Marthinus F | System and method for preventing movement in a telepresence system |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US525306A (en) * | 1894-08-28 | Charles a | ||
JPS5250611A (en) * | 1975-10-22 | 1977-04-22 | Nippon Telegr & Teleph Corp <Ntt> | Imagery communication method |
JPS63142986A (en) * | 1986-12-05 | 1988-06-15 | Nec Corp | Television telephone set |
JPH0468986A (en) * | 1990-07-09 | 1992-03-04 | Sony Corp | Moving picture difference detector |
JP3289730B2 (en) * | 1991-07-12 | 2002-06-10 | 日本電信電話株式会社 | I / O device for image communication |
JPH0568246A (en) * | 1991-09-06 | 1993-03-19 | Nippon Telegr & Teleph Corp <Ntt> | Display and image pickup system for video communication |
US5275526A (en) * | 1992-04-20 | 1994-01-04 | Superior Handling Equipment, Inc. | Truck freight loading and dock device |
-
1994
- 1994-09-19 US US08/308,603 patent/US5572248A/en not_active Expired - Lifetime
-
1995
- 1995-09-18 DE DE69534660T patent/DE69534660T2/en not_active Expired - Lifetime
- 1995-09-18 AU AU36344/95A patent/AU697578B2/en not_active Ceased
- 1995-09-18 KR KR1019970701789A patent/KR100403885B1/en not_active IP Right Cessation
- 1995-09-18 CA CA002200477A patent/CA2200477C/en not_active Expired - Fee Related
- 1995-09-18 EP EP01128723A patent/EP1187480B1/en not_active Expired - Lifetime
- 1995-09-18 AT AT01128723T patent/ATE311724T1/en not_active IP Right Cessation
- 1995-09-18 WO PCT/US1995/011802 patent/WO1996009722A1/en active IP Right Grant
- 1995-09-18 AT AT95933837T patent/ATE220838T1/en not_active IP Right Cessation
- 1995-09-18 EP EP95933837A patent/EP0806115B1/en not_active Expired - Lifetime
- 1995-09-18 JP JP8511000A patent/JPH10508990A/en not_active Ceased
- 1995-09-18 DE DE69527440T patent/DE69527440T2/en not_active Expired - Lifetime
-
1996
- 1996-11-04 US US08/740,839 patent/US5751337A/en not_active Expired - Lifetime
-
1998
- 1998-03-25 US US09/047,771 patent/US6160573A/en not_active Expired - Lifetime
-
2000
- 2000-10-24 US US09/695,460 patent/US6445405B1/en not_active Expired - Lifetime
-
2002
- 2002-08-30 US US10/233,363 patent/US6654045B2/en not_active Expired - Fee Related
-
2003
- 2003-11-19 US US10/716,972 patent/US7116350B2/en not_active Expired - Fee Related
-
2006
- 2006-09-29 US US11/537,177 patent/US20070120954A1/en not_active Abandoned
Patent Citations (69)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3233346A (en) * | 1962-09-13 | 1966-02-08 | Cornberg Sol | Carrel |
US3869992A (en) * | 1972-09-11 | 1975-03-11 | Edward J Kramer | Clusterable learning module |
US4004084A (en) * | 1976-03-03 | 1977-01-18 | Bell Telephone Laboratories, Incorporated | Video conferencing system using spatial reduction and temporal resolution |
USD252635S (en) * | 1977-04-27 | 1979-08-14 | Control Data Corporation | Portable learning unit |
US4400724A (en) * | 1981-06-08 | 1983-08-23 | The United States Of America As Represented By The Secretary Of The Army | Virtual space teleconference system |
US4546383A (en) * | 1982-06-18 | 1985-10-08 | Inria Institute National De Recherche En Informatique Et En Automatique | Method and apparatus for visual telecommunications, in particular for use by the deaf |
US4650929A (en) * | 1984-02-29 | 1987-03-17 | Heinrich-Hertz-Institut Fur Nachrichtentechnik Berlin Gmbh | Communication system for videoconferencing |
US4703348A (en) * | 1984-04-25 | 1987-10-27 | Matsushita Electric Works, Ltd. | Picture transmission system using secondary differential variable sampling rate coding |
US4942470A (en) * | 1984-07-20 | 1990-07-17 | Nec Corporation | Real time processor for video signals |
US4758887A (en) * | 1985-10-09 | 1988-07-19 | Weyel Kg | Conference table |
US4707077A (en) * | 1986-01-30 | 1987-11-17 | Hughes Aircraft Company | Real time image subtraction with a single liquid crystal light valve |
US4823194A (en) * | 1986-08-01 | 1989-04-18 | Hitachi, Ltd. | Method for processing gray scale images and an apparatus thereof |
US4845636A (en) * | 1986-10-17 | 1989-07-04 | Walker Mark E | Remote transaction system |
US4800432A (en) * | 1986-10-24 | 1989-01-24 | The Grass Valley Group, Inc. | Video Difference key generator |
US5031620A (en) * | 1986-11-14 | 1991-07-16 | Kabushiki Kaisha Toshiba | Image processing apparatus |
US4931872A (en) * | 1986-11-24 | 1990-06-05 | Gte Laboratories Incorporated | Methods of and apparatus for the generation of split-screen video displays |
US4986194A (en) * | 1986-11-26 | 1991-01-22 | Bollman Clifford J | Multi-adjustable, multi-functional, space saving workplace system |
US4965666A (en) * | 1987-11-27 | 1990-10-23 | U.S. Philips Corporation | Method of and arrangement for estimating and compensatiing motion in a sequence of pictures and a picture transmission system provided with such an arrangement |
US5079627A (en) * | 1988-06-15 | 1992-01-07 | Optum Corporation | Videophone |
US5032901A (en) * | 1988-06-28 | 1991-07-16 | Petro Vlahos | Backing color and luminance nonuniformity compensation for linear image compositing |
US4991009A (en) * | 1988-07-08 | 1991-02-05 | Ricoh Company, Ltd. | Dynamic image transmission system |
US4996592A (en) * | 1988-08-12 | 1991-02-26 | Oki Electric Industry Co., Ltd. | Video communication apparatus capable of electronically adjusting the field of view of a video frame |
US4890314A (en) * | 1988-08-26 | 1989-12-26 | Bell Communications Research, Inc. | Teleconference facility with high resolution video display |
US5055927A (en) * | 1988-09-13 | 1991-10-08 | Deutsche Thomson-Brandt Gmbh | Dual channel video signal transmission system |
US5122865A (en) * | 1988-09-15 | 1992-06-16 | Samsung Electronics Co., Ltd. | Chroma key signal generator for a video editing apparatus |
US4965819A (en) * | 1988-09-22 | 1990-10-23 | Docu-Vision, Inc. | Video conferencing system for courtroom and other applications |
US5691777A (en) * | 1988-10-17 | 1997-11-25 | Kassatly; Lord Samuel Anthony | Method and apparatus for simultaneous compression of video, audio and data signals |
US5157491A (en) * | 1988-10-17 | 1992-10-20 | Kassatly L Samuel A | Method and apparatus for video broadcasting and teleconferencing |
US4928301A (en) * | 1988-12-30 | 1990-05-22 | Bell Communications Research, Inc. | Teleconferencing terminal with camera behind display screen |
US5014267A (en) * | 1989-04-06 | 1991-05-07 | Datapoint Corporation | Video conferencing network |
US5016405A (en) * | 1989-04-18 | 1991-05-21 | Rosemount Office Systems, Inc. | Modular workstations |
US5135720A (en) * | 1989-05-23 | 1992-08-04 | Seiko Instruments Inc. | Reaction chamber for amino-terminal sequence analysis of proteins or peptides |
US5024398A (en) * | 1989-06-28 | 1991-06-18 | The Boeing Company | Office module for passenger aircraft |
US5038224A (en) * | 1989-12-22 | 1991-08-06 | Bell Communications Research, Inc. | Video imaging device in a teleconferencing system |
US5282027A (en) * | 1990-04-27 | 1994-01-25 | U.S. Philips Corporation | Image projection display and pick-up apparatus with optical shutter |
US5272526A (en) * | 1990-05-30 | 1993-12-21 | Sony Corporation | Television conference system |
US5168528A (en) * | 1990-08-20 | 1992-12-01 | Itt Corporation | Differential electronic imaging system |
US5393964A (en) * | 1990-10-12 | 1995-02-28 | Tpi, Inc. | Telecommunications booth and method of use |
US5117285A (en) * | 1991-01-15 | 1992-05-26 | Bell Communications Research | Eye contact apparatus for video conferencing |
US5187571A (en) * | 1991-02-01 | 1993-02-16 | Bell Communications Research, Inc. | Television system for displaying multiple views of a remote location |
US5214512A (en) * | 1991-02-11 | 1993-05-25 | Ampex Systems Corporation | Keyed, true-transparency image information combine |
US5127078A (en) * | 1991-02-13 | 1992-06-30 | Terry Weed | Integrated display for image presentation and receiving using fiber optics |
US5159451A (en) * | 1991-03-19 | 1992-10-27 | Faroudja Y C | Field memory expansible line doubler for television receiver |
US5257306A (en) * | 1991-05-29 | 1993-10-26 | Fujitsu Limited | Video teleconferencing system |
US5280540A (en) * | 1991-10-09 | 1994-01-18 | Bell Communications Research, Inc. | Video teleconferencing system employing aspect ratio transformation |
US5249039A (en) * | 1991-11-18 | 1993-09-28 | The Grass Valley Group, Inc. | Chroma key method and apparatus |
US5508743A (en) * | 1991-12-06 | 1996-04-16 | Canon Kabushiki Kaisha | Moving image signal coding apparatus and moving image signal coding control method |
US5173772A (en) * | 1991-12-14 | 1992-12-22 | Samsung Electronics Co., Ltd. | Method for detecting motion vector |
US5315633A (en) * | 1991-12-20 | 1994-05-24 | Unisys Corporation | Digital video switch for video teleconferencing |
US5282341A (en) * | 1992-01-10 | 1994-02-01 | Steelcase Inc. | Dynamic workspace module |
US5261735A (en) * | 1992-02-03 | 1993-11-16 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Deployable video conference table |
US5352033A (en) * | 1992-03-27 | 1994-10-04 | Fisher-Rosemount Systems, Inc. | Operator work station having a monitor support assembly |
US5541640A (en) * | 1992-06-23 | 1996-07-30 | Larson; Craig R. | Videophone for simultaneous audio and video communication via a standard telephone line |
US5406893A (en) * | 1992-09-25 | 1995-04-18 | Burns; J. Robert | Work-study carrel |
US5394198A (en) * | 1992-12-22 | 1995-02-28 | At&T Corp. | Large-screen display system |
US5400069A (en) * | 1993-06-16 | 1995-03-21 | Bell Communications Research, Inc. | Eye contact video-conferencing system and screen |
US5397133A (en) * | 1993-09-30 | 1995-03-14 | At&T Corp. | System for playing card games remotely |
US5689641A (en) * | 1993-10-01 | 1997-11-18 | Vicor, Inc. | Multimedia collaboration system arrangement for routing compressed AV signal through a participant site without decompressing the AV signal |
US5438357A (en) * | 1993-11-23 | 1995-08-01 | Mcnelley; Steve H. | Image manipulating teleconferencing system |
US5347306A (en) * | 1993-12-17 | 1994-09-13 | Mitsubishi Electric Research Laboratories, Inc. | Animated electronic meeting place |
US5850250A (en) * | 1994-07-18 | 1998-12-15 | Bell Atlantic Maryland, Inc. | Video distance learning system |
US5572248A (en) * | 1994-09-19 | 1996-11-05 | Teleport Corporation | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US5751337A (en) * | 1994-09-19 | 1998-05-12 | Telesuite Corporation | Teleconferencing method and system for providing face-to-face, non-animated teleconference environment |
US6160573A (en) * | 1994-09-19 | 2000-12-12 | Telesuite Corporation | Teleconference method and system for providing face-to-face teleconference environment |
US6445405B1 (en) * | 1994-09-19 | 2002-09-03 | Telesuite Corporation | Teleconferencing method and system |
US6654045B2 (en) * | 1994-09-19 | 2003-11-25 | Telesuite Corporation | Teleconferencing method and system |
US7116350B2 (en) * | 1994-09-19 | 2006-10-03 | Destiny Conferencing Llc | Teleconferencing method and system |
US5784843A (en) * | 1994-12-30 | 1998-07-28 | Steelcase Inc. | Integrated prefabricated furniture system for fitting-out open plan building space |
US20070263081A1 (en) * | 2006-04-20 | 2007-11-15 | De Beer Marthinus F | System and method for preventing movement in a telepresence system |
Cited By (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9269157B2 (en) | 2005-03-01 | 2016-02-23 | Eyesmatch Ltd | Methods for extracting objects from digital images and for performing color change on the object |
US8982110B2 (en) | 2005-03-01 | 2015-03-17 | Eyesmatch Ltd | Method for image transformation, augmented reality, and teleperence |
US8982109B2 (en) | 2005-03-01 | 2015-03-17 | Eyesmatch Ltd | Devices, systems and methods of capturing and displaying appearances |
US8976160B2 (en) | 2005-03-01 | 2015-03-10 | Eyesmatch Ltd | User interface and authentication for a virtual mirror |
US20060268360A1 (en) * | 2005-05-12 | 2006-11-30 | Jones Peter W J | Methods of creating a virtual window |
US20080268899A1 (en) * | 2007-04-24 | 2008-10-30 | Lg Electronics Inc. | Video communication terminal and method of displaying images |
US9258520B2 (en) | 2007-04-24 | 2016-02-09 | Lg Electronics Inc. | Video communication terminal and method of displaying images |
US8489149B2 (en) * | 2007-04-24 | 2013-07-16 | Lg Electronics Inc. | Video communication terminal and method of displaying images |
US10479784B2 (en) | 2008-12-19 | 2019-11-19 | Vertex Pharmaceuticals Incorporated | Substituted pyrazin-2-amines as inhibitors of ATR kinase |
US10961232B2 (en) | 2008-12-19 | 2021-03-30 | Vertex Pharmaceuticals Incorporated | Substituted pyrazines as ATR kinase inhibitors |
US9365557B2 (en) | 2008-12-19 | 2016-06-14 | Vertex Pharmaceuticals Incorporated | Substituted pyrazin-2-amines as inhibitors of ATR kinase |
US9701674B2 (en) | 2008-12-19 | 2017-07-11 | Vertex Pharmaceuticals Incorporated | Substituted pyrazines as ATR kinase inhibitors |
US8860775B2 (en) | 2009-04-14 | 2014-10-14 | Huawei Device Co., Ltd. | Remote presenting system, device, and method |
US8508576B2 (en) | 2009-04-14 | 2013-08-13 | Huawei Device Co., Ltd. | Remote presenting system, device, and method |
US8514260B2 (en) | 2009-05-28 | 2013-08-20 | Microsoft Corporation | Establishing eye contact in video conferencing |
US20100302344A1 (en) * | 2009-05-28 | 2010-12-02 | Microsoft Corporation | Establishing eye contact in video conferencing |
US8767593B1 (en) | 2009-10-13 | 2014-07-01 | Signal Perfection, Ltd. | Method for managing, scheduling, monitoring and controlling audio and video communication and data collaboration |
US8625769B1 (en) | 2009-10-13 | 2014-01-07 | Signal Perfection, Ltd. | System for managing, scheduling, monitoring and controlling audio and video communication and data collaboration |
US9630956B2 (en) | 2010-05-12 | 2017-04-25 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase |
US9334244B2 (en) | 2010-05-12 | 2016-05-10 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase |
WO2013043289A1 (en) * | 2011-09-23 | 2013-03-28 | Tangome, Inc. | Augmenting a video conference |
US10208027B2 (en) | 2011-09-30 | 2019-02-19 | Vertex Pharmaceuticals Incorporated | Processes for preparing ATR inhibitors |
US10813929B2 (en) | 2011-09-30 | 2020-10-27 | Vertex Pharmaceuticals Incorporated | Treating cancer with ATR inhibitors |
US10822331B2 (en) | 2011-09-30 | 2020-11-03 | Vertex Pharmaceuticals Incorporated | Processes for preparing ATR inhibitors |
US9862709B2 (en) | 2011-09-30 | 2018-01-09 | Vertex Pharmaceuticals Incorporated | Processes for making compounds useful as inhibitors of ATR kinase |
US10478430B2 (en) | 2012-04-05 | 2019-11-19 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase and combination therapies thereof |
US11110086B2 (en) | 2012-04-05 | 2021-09-07 | Vertex Pharmaceuticals Incorporated | Compounds useful as inhibitors of ATR kinase and combination therapies thereof |
US9791456B2 (en) | 2012-10-04 | 2017-10-17 | Vertex Pharmaceuticals Incorporated | Method for measuring ATR inhibition mediated increases in DNA damage |
RU2656817C2 (en) * | 2012-12-18 | 2018-06-06 | Айсмэтч Лтд | Devices, systems and methods of capturing and displaying appearances |
WO2014100250A3 (en) * | 2012-12-18 | 2014-08-14 | Nissi Vilcovsky | Devices, systems and methods of capturing and displaying appearances |
WO2014130977A1 (en) * | 2013-02-25 | 2014-08-28 | Herold Williams | Nonlinear scaling in video conferencing |
US9667914B2 (en) | 2013-02-25 | 2017-05-30 | Herold Williams | Nonlinear scaling in video conferencing |
US20140267575A1 (en) * | 2013-03-14 | 2014-09-18 | Polycom, Inc. | Method and system for providing a virtual cafeteria |
US9392225B2 (en) * | 2013-03-14 | 2016-07-12 | Polycom, Inc. | Method and system for providing a virtual cafeteria |
US11464774B2 (en) | 2015-09-30 | 2022-10-11 | Vertex Pharmaceuticals Incorporated | Method for treating cancer using a combination of DNA damaging agents and ATR inhibitors |
US10250888B2 (en) | 2015-10-08 | 2019-04-02 | Samsung Electronics Co., Ltd. | Electronic device configured to non-uniformly encode/decode image data according to display shape |
Also Published As
Publication number | Publication date |
---|---|
AU3634495A (en) | 1996-04-09 |
WO1996009722A1 (en) | 1996-03-28 |
US6160573A (en) | 2000-12-12 |
US20030020805A1 (en) | 2003-01-30 |
EP1187480A1 (en) | 2002-03-13 |
EP0806115B1 (en) | 2002-07-17 |
EP1187480B1 (en) | 2005-11-30 |
DE69527440D1 (en) | 2002-08-22 |
DE69534660T2 (en) | 2006-07-27 |
US6445405B1 (en) | 2002-09-03 |
US20040100553A1 (en) | 2004-05-27 |
US5751337A (en) | 1998-05-12 |
EP0806115A4 (en) | 1998-03-11 |
ATE311724T1 (en) | 2005-12-15 |
CA2200477A1 (en) | 1996-03-28 |
ATE220838T1 (en) | 2002-08-15 |
AU697578B2 (en) | 1998-10-08 |
US7116350B2 (en) | 2006-10-03 |
US5572248A (en) | 1996-11-05 |
US6654045B2 (en) | 2003-11-25 |
DE69527440T2 (en) | 2003-02-20 |
JPH10508990A (en) | 1998-09-02 |
KR100403885B1 (en) | 2004-04-03 |
DE69534660D1 (en) | 2006-01-05 |
CA2200477C (en) | 2005-02-08 |
EP0806115A1 (en) | 1997-11-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6654045B2 (en) | Teleconferencing method and system | |
US9819907B2 (en) | Communication stage and related systems | |
US7119829B2 (en) | Virtual conference room | |
US9848169B2 (en) | Transparent projection communication terminals | |
US8199185B2 (en) | Reflected camera image eye contact terminal | |
US20100302446A1 (en) | Video Superposition for Continuous Presence | |
US20100079576A1 (en) | Display system and method | |
EP2338277A1 (en) | A control system for a local telepresence videoconferencing system and a method for establishing a video conference call | |
WO2010130084A1 (en) | Telepresence system, method and video capture device | |
KR20100126812A (en) | How to Display Video Image Streams and Client Management Systems | |
JP2001503221A (en) | Video conferencing equipment | |
US11831454B2 (en) | Full dome conference | |
US20040068736A1 (en) | Communication terminal and system | |
US20210367985A1 (en) | Immersive telepresence video conference system | |
Jouppi et al. | Bireality: mutually-immersive telepresence | |
JPH02202275A (en) | Video conference system | |
JP2000217091A (en) | Video conference system | |
EP4203464A1 (en) | Full dome conference | |
EP2198609A1 (en) | Audiovisual terminal | |
WO2015131520A1 (en) | Method and device for displaying layout in telepresence conferencing system | |
Sun et al. | Video bridging for multipoint video conferencing | |
JPH05268688A (en) | Microphone for television conference | |
CA2066163A1 (en) | Video conferencing system for courtroom and other applications | |
JPH01174080A (en) | Picture display device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: DESTINY CAPITAL LLC, OHIO Free format text: ASSET PURCHASE AGREEMENT;ASSIGNOR:TELESUITE CORPORATION;REEL/FRAME:018595/0530 Effective date: 20041104 Owner name: DESTINY CAPITAL LLC, OHIO Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TELESUITE CORPORATION;REEL/FRAME:018596/0867 Effective date: 20050909 Owner name: TELESUITE CORPORATION, OHIO Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ALLEN, DAVID L.;WILLIAMS, HEROLD;REEL/FRAME:018594/0724 Effective date: 20020904 Owner name: DESTINY CONFERENCING LLC, OHIO Free format text: CHANGE OF NAME;ASSIGNOR:DESTINY CAPITAL LLC;REEL/FRAME:018598/0084 Effective date: 20050920 Owner name: DESTINY CONFERENCING CORPORATION, OHIO Free format text: MERGER;ASSIGNOR:DESTINY CONFERENCING LLC;REEL/FRAME:018598/0090 Effective date: 20060720 |
|
AS | Assignment |
Owner name: POLYCOM, INC., CALIFORNIA Free format text: NUNC PRO TUNC ASSIGNMENT;ASSIGNOR:DESTINY CONFERENCING CORPORATION;REEL/FRAME:019331/0591 Effective date: 20070521 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |