US20160184724A1 - Dynamic App Programming Environment with Physical Object Interaction - Google Patents
Dynamic App Programming Environment with Physical Object Interaction Download PDFInfo
- Publication number
- US20160184724A1 US20160184724A1 US14/838,307 US201514838307A US2016184724A1 US 20160184724 A1 US20160184724 A1 US 20160184724A1 US 201514838307 A US201514838307 A US 201514838307A US 2016184724 A1 US2016184724 A1 US 2016184724A1
- Authority
- US
- United States
- Prior art keywords
- physical manipulative
- physical
- computing device
- sound
- manipulative
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000003993 interaction Effects 0.000 title claims description 4
- 241001465754 Metazoa Species 0.000 claims description 18
- 238000000034 method Methods 0.000 claims description 13
- 238000001514 detection method Methods 0.000 claims description 4
- 230000037213 diet Effects 0.000 claims description 4
- 235000005911 diet Nutrition 0.000 claims description 4
- 230000002452 interceptive effect Effects 0.000 claims description 2
- 241000282819 Giraffa Species 0.000 description 9
- 241000283070 Equus zebra Species 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 241000282320 Panthera leo Species 0.000 description 4
- 241000282818 Giraffidae Species 0.000 description 2
- 241001441705 Mochokidae Species 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 1
- 235000021191 food habits Nutrition 0.000 description 1
- 230000005021 gait Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 244000062645 predators Species 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63H—TOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
- A63H29/00—Drive mechanisms for toys in general
- A63H29/22—Electric drives
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B17/00—Teaching reading
- G09B17/003—Teaching reading electrically operated apparatus or devices
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63H—TOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
- A63H17/00—Toy vehicles, e.g. with self-drive; ; Cranes, winches or the like; Accessories therefor
- A63H17/26—Details; Accessories
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63H—TOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
- A63H3/00—Dolls
- A63H3/36—Details; Accessories
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
- G09B5/065—Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63H—TOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
- A63H2200/00—Computerized interactive toys, e.g. dolls
Definitions
- the system of the present invention comprises at least one physical manipulative, a computing device that comprises a display, a user interface, an identification module for identifying the physical manipulative, and a communication module for communicating with a server; a server that comprises a database listing all the types of physical manipulatives used in the game and at least one descriptive attribute for each physical manipulative; where the computing device is configured to identify the physical manipulative, look up at least one descriptive attribute associated with it, and display an animation related to the at least one descriptive attribute.
- the physical manipulative can be an animal figure, a cartoon character figure, a doll, an action figure, a vehicle, or any other toy or shape.
- the descriptive attribute of the physical manipulative can be any or all of the following: personality, habits, sounds, phrases, geographic origin, size, diet, spelling of the name, pronunciation of the name, at least one Internet link.
- the identification module can be a camera, a NFC reader, a QR reader, a bar code reader, a RF receiver, a sound detection device, or any combination of the above.
- the computing device can also detect a motion pattern of the physical manipulative (i.e. a horse being moved in a “galloping” motion) and display an animation relating to the motion pattern.
- a motion pattern of the physical manipulative i.e. a horse being moved in a “galloping” motion
- the computing device can also detect an orientation of the physical manipulative (i.e. upside-down or right-side-up) and display an animation relating to the orientation.
- an orientation of the physical manipulative i.e. upside-down or right-side-up
- the system comprises at least two physical manipulatives
- the computing device can also detect the relative or absolute position or orientation of each physical manipulative, as well as their relative position with respect to each other, and display an animation relating to the interaction between the physical manipulatives.
- system further comprises a sound input device
- the computing device can detect a sound received by the sound input device, look up the sound in the database, determine the physical manipulative associated with the sound, and display an animation relating to the physical manipulative and the sound.
- the system can detect and respond to voice commands.
- the physical manipulative emits sounds and the computing device can detect sounds emitted by the physical manipulative and display an animation relating to the physical manipulative.
- the physical manipulative only emits sounds when it is squeezed, pressed, or moved.
- the energy used to make the physical manipulative emit the sound may be generated by the action of squeezing, pressing, or moving it.
- the method of the present invention comprises detecting the presence of at least one physical manipulative near a computing device, identifying the at least one physical manipulative, using the computing device to look up at least one descriptive attribute associated with the at least one physical manipulative, and using the computing device to display an animation relating to the at least one descriptive attribute.
- the method may also comprise detecting a motion pattern of the physical manipulative near the computing device, and using the computing device to display an animation relating to the at least one descriptive attribute and the motion pattern.
- the method may also comprise detecting the presence of a second physical manipulative near the computing device, using the computing device to look up at least one second descriptive attribute associated with the second physical manipulative, and using the computing device to display an animation relating to both descriptive attributes.
- the method may also comprise detecting the presence of a sound near the computing device, using the computing device to look up a physical manipulative associated with the sound and at least one descriptive attribute associated with the physical manipulative, and using the computing device to display an animation relating to the at least one descriptive attribute.
- the sound is emitted by the physical manipulative itself.
- the descriptive attribute can be at least one of the following: personality, habits, sounds, phrases, geographic origin, size, diet, spelling of a name, pronunciation of a name, link to the Internet.
- the at least one physical manipulative can be an animal figure, a cartoon character figure, a doll, an action figure, a vehicle.
- the step of identifying the at least one physical manipulative may be performed by a QR reader, a bar code reader, a camera, a NFC reader, a RF receiver, or a sound detection device.
- FIG. 1 shows a diagram of an embodiment of the system of the present invention.
- FIG. 2 shows a diagram of an alternate embodiment of the system of the present invention.
- FIG. 3 shows a diagram of the operation of the software side of the present invention.
- FIG. 4A shows a diagram of the operation of the present invention when a new toy is introduced.
- FIG. 4B shows a diagram of the operation of the present invention when a toy is removed.
- the system of the present invention comprises a tablet or other mobile device such as a smartphone (hereafter referred to as a “tablet”), and at least one physical toy, said physical toy incorporating a NFC chip or a similar chip for storing information.
- the tablet comprises a means of reading the NFC chip (or similar chip; throughout this disclosure, it will be understood that any chip that can be used for storing information and can be read by a tablet can be used as an alternative to NFC).
- the tablet preferably also comprises a gaming or educational app that provides an interactive experience for a child.
- FIG. 1 shows an example of the system of the present invention.
- Physical toy 300 comprises a NFC chip 330 that encodes various information about the toy.
- the NFC chip communicates with a tablet 310 , which reads the information on the chip and uses that information to detect what toy is in front of the tablet.
- the tablet then communicates with a server 200 to access various information relating to the toy, the animal represented by the toy, the animal's habits and the noises it makes, links to various websites related to the animal, and so on.
- the server 200 preferably comprises a database that stores any such descriptive attributes. The descriptive attributes are then used to display animations within the app, to perform various actions within the app, to add characters to a game app, to display information in an educational app, and so on.
- FIG. 2 shows an alternate embodiment of the system of the present invention, wherein the physical toy does not communicate directly with the computing device but rather with a short range to long range router/converter 420 .
- the server is not shown in this Figure, but nevertheless is present.
- the router/converter 420 communicates with the toy 400 via NFC link 430 , and then communicates with the tablet or other computing device via a longer-range connection 440 (such as Bluetooth, wi-fi, or any other wireless link).
- the toy does not need to be as close to the tablet in order to be recognized.
- the toy itself may communicate with the computing device via Bluetooth or wi-fi. However, it reduces the cost and complexity of the system to use NFC communication.
- the toy may also comprise a sound generator 340 .
- the sound generator requires no external power or batteries—i.e. either a mechanical sound generator such as a squeaker or a rattle, or an electromechanical sound generator where the power is supplied by a user's handling of the toy (i.e. squeezing the toy or moving it around).
- the sound generator preferably serves two functions—providing an additional entertainment experience for the child, and serving as an additional way for the tablet 310 to identify the toy 330 (assuming that each sound is unique to each toy).
- the sound may also trigger the tablet to display a special animation, play a special sound, or perform some other action within a game.
- the command center comprises a relational database that comprises each physical object that is embedded with a NFC chip, and descriptive attributes of each such physical object.
- the descriptive attributes can be the name of the toy, the dimensions of the physical toy, the spelling and pronunciation of its name, the type of toy it is (animal, cartoon character, vehicle, etc.), the geographic origin of the animal or character, various characteristics of the animal or character, and so on.
- This information is preferably stored in the cloud. The information may be customized to allow for different programming environments and different apps.
- An app developer wishing to include physical devices into their app can thus include the attributes of the physical device into the app experience.
- the tablet senses a NFC chip belonging to a particular toy, it can trigger the app to perform some action or to make some change.
- the giraffe and lion 3D graphic files are pulled from the cloud and dynamically introduced into the app's activities.
- the main characters of the app can then change into giraffes and lions that are identical to the representations of the dolls in color, size, and realism levels.
- the app can also pull characteristics of both animals or characters from the cloud and introduce them into the app's activities.
- the action of the app can then change to incorporate the interaction of a carnivore vs. a large vegetarian, in the case of the giraffe and the lion, for example.
- the characteristics can include behavior, food habits, reactions to external factors, sounds made by the animal, and so on.
- the tablet senses not just the presence of a toy, but also its movement patterns. For example, a child moving the toy in a “running” motion will make the corresponding animal “run” on the screen. A child moving the toy from right to left will make the corresponding animal move from right to left on the screen.
- the movement patterns of the toy are detected and identified by the tablet's built-in camera, or by another camera connected to the tablet.
- the present invention may be used with a wide range of apps.
- game apps may use physical manipulatives to control game play
- educational apps may use physical manipulatives and their movement patterns to deliver educational material to the child (i.e. a child can put a giraffe toy in front of the tablet and learn about giraffes from the app); and so on.
- the present invention is not meant to be limited to any particular app or type of app.
- FIGS. 4A and 4B illustrate the operation of the system of the present invention when a new toy is added or removed from the vicinity of the tablet.
- FIG. 4A shows what happens when a new toy is added.
- the tablet detects the change in physical manipulatives (in this case, the addition of the zebra), and looks up the characteristics associated with the zebra toy.
- the app running on the tablet will then incorporate an animation based on the characteristics of the zebra toy into the app.
- FIG. 4B shows what happens when a toy is removed.
- the tablet detects the change in physical manipulatives (in this case, the removal of the zebra), and removes the zebra animation from the app.
- the app is written in such a way as to change the behavior of each individual character based on what other characters are there.
- the present invention is not limited to any particular app or type of app, and may encompass any type of change in what is displayed based on the addition, removal, or movement of a toy.
- the relational database can be located in the cloud, or can be downloaded onto the tablet for ease of access. Different animals and characters can be added to the database by developers as new toys are added to the system.
- the system of the present invention uses NFC chips to communicate with the toys
- the app may also use the tablet's camera to analyze the toys visually.
- the app may use the camera to identify the toy (by shape, color, or a special marking such as a QR code on the toy), or may use the camera to identify the way the toy is moving or the way the toy is oriented, while using the NFC chip to identify the toy itself.
- the app can identify a toy as a giraffe by its NFC chip. Then, the tablet camera can determine that the child is moving the giraffe in a running motion. The app can then show a running giraffe on the screen, and use the descriptive information regarding the giraffe to give the representation the proper gait as it runs.
- the system can also use the tablet's microphone to identify sounds or speech. This can be used to allow the child to make the appropriate animal noise for the animal toy and to have the animal make the same noise on-screen, to allow the child to give instructions for what he or she wants to see happen on the tablet screen (i.e. “I want to see Spongebob riding a giraffe!”), and so on.
- the toys can generate sound on their own.
- a toy may be equipped with a squeaker or some other mechanical sound generating device, or an electrical sound generating device and some way to either store or generate power for the device.
- the toy may generate sound when squeezed, moved, or touched.
- the tablet may identify the sound emitted by the toy and change the animation or make a sound when the toy emits that sound.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Business, Economics & Management (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
- Electrically Operated Instructional Devices (AREA)
- Toys (AREA)
Abstract
An app programming environment that enables a smartphone or tablet app to incorporate real toys or objects into the app, by use of a NFC chip located in the toy or object and a database comprising information regarding the toy or object used.
Description
- The present application takes priority from Provisional App. No. 62/044,310, filed Aug. 31, 2014, which is incorporated herein by reference.
- Many children's games and toys, including educational games and toys, incorporate tablets and smartphones. However, a smartphone or tablet can only offer a virtual environment rather than any real objects. Since children are concrete, tactile thinkers, a virtual environment is often insufficient to hold their interest or to allow them to learn effectively.
- While some games and toys exist that incorporate real objects into virtual environments, such games and toys typically are limited to particular types of physical objects (i.e. alphabet blocks), and do not provide for unbounded imaginative play.
- A need therefore exists for a way that allows a child to incorporate real toys in a tablet or smartphone based game, and a programming environment that allows app developers to incorporate real toys in their apps in an easy and standardized way.
- A need also exists for a tablet or smartphone based virtual environment that responds to the manipulation of real toys and that is open-ended and responsive to the child's actions, rather than scripted.
- The system of the present invention comprises at least one physical manipulative, a computing device that comprises a display, a user interface, an identification module for identifying the physical manipulative, and a communication module for communicating with a server; a server that comprises a database listing all the types of physical manipulatives used in the game and at least one descriptive attribute for each physical manipulative; where the computing device is configured to identify the physical manipulative, look up at least one descriptive attribute associated with it, and display an animation related to the at least one descriptive attribute.
- The physical manipulative can be an animal figure, a cartoon character figure, a doll, an action figure, a vehicle, or any other toy or shape.
- The descriptive attribute of the physical manipulative can be any or all of the following: personality, habits, sounds, phrases, geographic origin, size, diet, spelling of the name, pronunciation of the name, at least one Internet link.
- The identification module can be a camera, a NFC reader, a QR reader, a bar code reader, a RF receiver, a sound detection device, or any combination of the above.
- In an embodiment, the computing device can also detect a motion pattern of the physical manipulative (i.e. a horse being moved in a “galloping” motion) and display an animation relating to the motion pattern.
- In an embodiment, the computing device can also detect an orientation of the physical manipulative (i.e. upside-down or right-side-up) and display an animation relating to the orientation.
- In an embodiment, the system comprises at least two physical manipulatives, and the computing device can also detect the relative or absolute position or orientation of each physical manipulative, as well as their relative position with respect to each other, and display an animation relating to the interaction between the physical manipulatives.
- In an embodiment, the system further comprises a sound input device, and the computing device can detect a sound received by the sound input device, look up the sound in the database, determine the physical manipulative associated with the sound, and display an animation relating to the physical manipulative and the sound.
- In an embodiment, the system can detect and respond to voice commands.
- In an embodiment, the physical manipulative emits sounds and the computing device can detect sounds emitted by the physical manipulative and display an animation relating to the physical manipulative. In an embodiment, the physical manipulative only emits sounds when it is squeezed, pressed, or moved. The energy used to make the physical manipulative emit the sound may be generated by the action of squeezing, pressing, or moving it.
- The method of the present invention comprises detecting the presence of at least one physical manipulative near a computing device, identifying the at least one physical manipulative, using the computing device to look up at least one descriptive attribute associated with the at least one physical manipulative, and using the computing device to display an animation relating to the at least one descriptive attribute.
- In an embodiment, the method may also comprise detecting a motion pattern of the physical manipulative near the computing device, and using the computing device to display an animation relating to the at least one descriptive attribute and the motion pattern.
- In an embodiment, the method may also comprise detecting the presence of a second physical manipulative near the computing device, using the computing device to look up at least one second descriptive attribute associated with the second physical manipulative, and using the computing device to display an animation relating to both descriptive attributes.
- In an embodiment, the method may also comprise detecting the presence of a sound near the computing device, using the computing device to look up a physical manipulative associated with the sound and at least one descriptive attribute associated with the physical manipulative, and using the computing device to display an animation relating to the at least one descriptive attribute.
- In an embodiment, the sound is emitted by the physical manipulative itself.
- The descriptive attribute can be at least one of the following: personality, habits, sounds, phrases, geographic origin, size, diet, spelling of a name, pronunciation of a name, link to the Internet.
- The at least one physical manipulative can be an animal figure, a cartoon character figure, a doll, an action figure, a vehicle.
- The step of identifying the at least one physical manipulative may be performed by a QR reader, a bar code reader, a camera, a NFC reader, a RF receiver, or a sound detection device.
-
FIG. 1 shows a diagram of an embodiment of the system of the present invention. -
FIG. 2 shows a diagram of an alternate embodiment of the system of the present invention. -
FIG. 3 shows a diagram of the operation of the software side of the present invention. -
FIG. 4A shows a diagram of the operation of the present invention when a new toy is introduced. -
FIG. 4B shows a diagram of the operation of the present invention when a toy is removed. - In the preferred embodiment, the system of the present invention comprises a tablet or other mobile device such as a smartphone (hereafter referred to as a “tablet”), and at least one physical toy, said physical toy incorporating a NFC chip or a similar chip for storing information. The tablet comprises a means of reading the NFC chip (or similar chip; throughout this disclosure, it will be understood that any chip that can be used for storing information and can be read by a tablet can be used as an alternative to NFC). The tablet preferably also comprises a gaming or educational app that provides an interactive experience for a child.
-
FIG. 1 shows an example of the system of the present invention.Physical toy 300 comprises aNFC chip 330 that encodes various information about the toy. The NFC chip communicates with atablet 310, which reads the information on the chip and uses that information to detect what toy is in front of the tablet. The tablet then communicates with aserver 200 to access various information relating to the toy, the animal represented by the toy, the animal's habits and the noises it makes, links to various websites related to the animal, and so on. Theserver 200 preferably comprises a database that stores any such descriptive attributes. The descriptive attributes are then used to display animations within the app, to perform various actions within the app, to add characters to a game app, to display information in an educational app, and so on. -
FIG. 2 shows an alternate embodiment of the system of the present invention, wherein the physical toy does not communicate directly with the computing device but rather with a short range to long range router/converter 420. Note that the server is not shown in this Figure, but nevertheless is present. In that embodiment, the router/converter 420 communicates with thetoy 400 via NFClink 430, and then communicates with the tablet or other computing device via a longer-range connection 440 (such as Bluetooth, wi-fi, or any other wireless link). In that embodiment, the toy does not need to be as close to the tablet in order to be recognized. - It will be understood that while the preferred embodiment uses NFC chips, any other communication method is also acceptable. The toy itself may communicate with the computing device via Bluetooth or wi-fi. However, it reduces the cost and complexity of the system to use NFC communication.
- As shown in
FIGS. 1 and 2 , the toy may also comprise asound generator 340. In the preferred embodiment, the sound generator requires no external power or batteries—i.e. either a mechanical sound generator such as a squeaker or a rattle, or an electromechanical sound generator where the power is supplied by a user's handling of the toy (i.e. squeezing the toy or moving it around). The sound generator preferably serves two functions—providing an additional entertainment experience for the child, and serving as an additional way for thetablet 310 to identify the toy 330 (assuming that each sound is unique to each toy). In an embodiment, the sound may also trigger the tablet to display a special animation, play a special sound, or perform some other action within a game. - The app programming environment of the present invention is shown in
FIG. 3 . The command center comprises a relational database that comprises each physical object that is embedded with a NFC chip, and descriptive attributes of each such physical object. For example, the descriptive attributes can be the name of the toy, the dimensions of the physical toy, the spelling and pronunciation of its name, the type of toy it is (animal, cartoon character, vehicle, etc.), the geographic origin of the animal or character, various characteristics of the animal or character, and so on. This information is preferably stored in the cloud. The information may be customized to allow for different programming environments and different apps. - An app developer wishing to include physical devices into their app can thus include the attributes of the physical device into the app experience. When the tablet senses a NFC chip belonging to a particular toy, it can trigger the app to perform some action or to make some change.
- For example, if an app is running and the tablet senses a plastic doll of a lion and a plastic doll of a giraffe, the giraffe and lion 3D graphic files are pulled from the cloud and dynamically introduced into the app's activities. The main characters of the app can then change into giraffes and lions that are identical to the representations of the dolls in color, size, and realism levels.
- In another embodiment, the app can also pull characteristics of both animals or characters from the cloud and introduce them into the app's activities. The action of the app can then change to incorporate the interaction of a carnivore vs. a large vegetarian, in the case of the giraffe and the lion, for example. The characteristics can include behavior, food habits, reactions to external factors, sounds made by the animal, and so on.
- In an embodiment, the tablet senses not just the presence of a toy, but also its movement patterns. For example, a child moving the toy in a “running” motion will make the corresponding animal “run” on the screen. A child moving the toy from right to left will make the corresponding animal move from right to left on the screen. In the preferred embodiment, the movement patterns of the toy are detected and identified by the tablet's built-in camera, or by another camera connected to the tablet.
- The present invention may be used with a wide range of apps. For example, game apps may use physical manipulatives to control game play; educational apps may use physical manipulatives and their movement patterns to deliver educational material to the child (i.e. a child can put a giraffe toy in front of the tablet and learn about giraffes from the app); and so on. The present invention is not meant to be limited to any particular app or type of app.
-
FIGS. 4A and 4B illustrate the operation of the system of the present invention when a new toy is added or removed from the vicinity of the tablet.FIG. 4A shows what happens when a new toy is added. The tablet detects the change in physical manipulatives (in this case, the addition of the zebra), and looks up the characteristics associated with the zebra toy. The app running on the tablet will then incorporate an animation based on the characteristics of the zebra toy into the app. - Similarly,
FIG. 4B shows what happens when a toy is removed. Once again, the tablet detects the change in physical manipulatives (in this case, the removal of the zebra), and removes the zebra animation from the app. - It must be noted that in the preferred embodiment, the app is written in such a way as to change the behavior of each individual character based on what other characters are there. However, the present invention is not limited to any particular app or type of app, and may encompass any type of change in what is displayed based on the addition, removal, or movement of a toy.
- The relational database can be located in the cloud, or can be downloaded onto the tablet for ease of access. Different animals and characters can be added to the database by developers as new toys are added to the system.
- While in the preferred embodiment, the system of the present invention uses NFC chips to communicate with the toys, the app may also use the tablet's camera to analyze the toys visually. The app may use the camera to identify the toy (by shape, color, or a special marking such as a QR code on the toy), or may use the camera to identify the way the toy is moving or the way the toy is oriented, while using the NFC chip to identify the toy itself.
- For example, in this embodiment, the app can identify a toy as a giraffe by its NFC chip. Then, the tablet camera can determine that the child is moving the giraffe in a running motion. The app can then show a running giraffe on the screen, and use the descriptive information regarding the giraffe to give the representation the proper gait as it runs.
- In an embodiment, the system can also use the tablet's microphone to identify sounds or speech. This can be used to allow the child to make the appropriate animal noise for the animal toy and to have the animal make the same noise on-screen, to allow the child to give instructions for what he or she wants to see happen on the tablet screen (i.e. “I want to see Spongebob riding a giraffe!”), and so on.
- In a related embodiment, the toys can generate sound on their own. For example, a toy may be equipped with a squeaker or some other mechanical sound generating device, or an electrical sound generating device and some way to either store or generate power for the device. The toy may generate sound when squeezed, moved, or touched. In that embodiment, the tablet may identify the sound emitted by the toy and change the animation or make a sound when the toy emits that sound.
- Exemplary embodiments have been described above. It will be understood by a person of reasonable skill in the art that the invention encompasses other embodiments that are equivalent to the ones described above.
Claims (21)
1. A system for interactive play with physical and virtual elements, comprising:
at least one physical manipulative;
a computing device comprising:
a display;
a user interface;
an identification module for identifying the at least one physical manipulative;
a communication module for communicating with a server;
a server comprising a database, said database comprising:
a listing of physical manipulatives;
for each physical manipulative, at least one descriptive attribute;
wherein the computing device is configured to:
identify the physical manipulative;
look up at least one descriptive attribute associated with the physical manipulative in the database;
display or modify an animation related to the at least one descriptive attribute on the display.
2. The system of claim 1 , wherein the at least one physical manipulative is one of the following group: an animal figure, a cartoon character figure, a doll, an action figure, a vehicle.
3. The system of claim 1 , wherein the at least one descriptive attribute is at least one of the following group: personality, habits, sounds, phrases, geographic origin, size, diet, spelling of a name, pronunciation of a name, at least one Internet link.
4. The system of claim 1 , wherein the identification module comprises at least one of the following: a camera, a NFC module, a QR reader, a bar code reader, a RF receiver, a sound detection device.
5. The system of claim 1 , wherein the computing device is further configured to:
detect a motion pattern of the physical manipulative;
display or modify an animation relating to the motion pattern on the display.
6. The system of claim 1 wherein the physical manipulative emits a sound, wherein the identification module is configured to detect the sound and to identify the physical manipulative based on the sound.
7. The system of claim 6 wherein the physical manipulative emits the sound when at least one of the following actions is performed: the physical manipulative is squeezed, the physical manipulative is pressed, the physical manipulative is moved.
8. The system of claim 7 , wherein the action generates energy that is used to make the physical manipulative emit the sound.
9. The system of claim 1 , wherein the computing device is further configured to:
detect an orientation of the physical manipulative;
display an animation relating to the orientation on the display.
10. The system of claim 1 , comprising a first physical manipulative and a second physical manipulative, wherein the computing device is further configured to:
detect one of the following: the relative position of the two physical manipulatives, the absolute position of each physical manipulative, the orientation of each physical manipulative;
display an animation relating to the interaction between the first physical manipulative and the second physical manipulative.
11. The system of claim 1 , further comprising:
a sound input device;
wherein the computing device is further configured to:
detect a sound received by the sound input device;
look up the sound in the database;
determine the physical manipulative associated with the sound;
display an animation relating to the physical manipulative and to the sound.
12. The system of claim 11 , wherein the computing device is further configured to:
detect a voice command received by the sound input device;
display an animation relating to the physical manipulative and the voice command.
13. The system of claim 11 , wherein the physical manipulative comprises a sound generating device, wherein the computing device is further configured to:
detect sounds emitted by the physical manipulative;
display an animation relating to the physical manipulative.
14. A method of entertaining a child, comprising:
detecting the presence of at least one physical manipulative near a computing device;
identifying the at least one physical manipulative;
using the computing device to look up at least one descriptive attribute associated with the at least one physical manipulative;
using the computing device to display an animation relating to the at least one descriptive attribute.
15. The method of claim 14 , further comprising:
detecting a motion pattern of the physical manipulative near the computing device;
using the computing device to display an animation relating to the at least one descriptive attribute and the motion pattern.
16. The method of claim 14 , further comprising:
detecting the presence of a second physical manipulative near the computing device;
using the computing device to look up at least one second descriptive attribute associated with the second physical manipulative;
using the computing device to display an animation relating to the at least one descriptive attribute and the at least one second descriptive attribute.
17. The method of claim 14 , further comprising:
detecting a sound near the computing device;
using the computing device to look up a physical manipulative associated with the sound;
using the computing device to look up at least one descriptive attribute associated with the physical manipulative;
using the computing device to display an animation relating to the at least one descriptive attribute.
18. The method of claim 17 , wherein the sound is emitted by the physical manipulative.
19. The method of claim 14 , where the at least one descriptive attribute is at least one of the following group: personality, habits, sounds, phrases, geographic origin, size, diet, spelling of a name, pronunciation of a name, link to the Internet.
20. The method of claim 14 , where the at least one physical manipulative is at least one of the following group: an animal figure, a cartoon character figure, a doll, an action figure, a vehicle.
21. The method of claim 14 , wherein the step of identifying the at least one physical manipulative is performed by at least one of the following: a QR reader, a bar code reader, a camera, a NFC reader, a RF receiver, a sound detection device.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/838,307 US20160184724A1 (en) | 2014-08-31 | 2015-08-27 | Dynamic App Programming Environment with Physical Object Interaction |
US15/428,154 US10380909B2 (en) | 2014-08-31 | 2017-02-09 | Interactive phonics game system and method |
US16/451,756 US10607501B2 (en) | 2014-08-31 | 2019-06-25 | Interactive phonics game system and method |
US16/794,591 US10922994B2 (en) | 2014-08-31 | 2020-02-19 | Interactive phonics game system and method |
US17/152,032 US11776418B2 (en) | 2014-08-31 | 2021-01-19 | Interactive phonics game system and method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462044310P | 2014-08-31 | 2014-08-31 | |
US14/838,307 US20160184724A1 (en) | 2014-08-31 | 2015-08-27 | Dynamic App Programming Environment with Physical Object Interaction |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/428,154 Continuation-In-Part US10380909B2 (en) | 2014-08-31 | 2017-02-09 | Interactive phonics game system and method |
US15/428,154 Continuation US10380909B2 (en) | 2014-08-31 | 2017-02-09 | Interactive phonics game system and method |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160184724A1 true US20160184724A1 (en) | 2016-06-30 |
Family
ID=56163104
Family Applications (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/838,307 Abandoned US20160184724A1 (en) | 2014-08-31 | 2015-08-27 | Dynamic App Programming Environment with Physical Object Interaction |
US15/428,154 Active US10380909B2 (en) | 2014-08-31 | 2017-02-09 | Interactive phonics game system and method |
US16/451,756 Active US10607501B2 (en) | 2014-08-31 | 2019-06-25 | Interactive phonics game system and method |
US16/794,591 Active US10922994B2 (en) | 2014-08-31 | 2020-02-19 | Interactive phonics game system and method |
US17/152,032 Active 2036-07-20 US11776418B2 (en) | 2014-08-31 | 2021-01-19 | Interactive phonics game system and method |
Family Applications After (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/428,154 Active US10380909B2 (en) | 2014-08-31 | 2017-02-09 | Interactive phonics game system and method |
US16/451,756 Active US10607501B2 (en) | 2014-08-31 | 2019-06-25 | Interactive phonics game system and method |
US16/794,591 Active US10922994B2 (en) | 2014-08-31 | 2020-02-19 | Interactive phonics game system and method |
US17/152,032 Active 2036-07-20 US11776418B2 (en) | 2014-08-31 | 2021-01-19 | Interactive phonics game system and method |
Country Status (1)
Country | Link |
---|---|
US (5) | US20160184724A1 (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9914062B1 (en) * | 2016-09-12 | 2018-03-13 | Laura Jiencke | Wirelessly communicative cuddly toy |
US10074205B2 (en) | 2016-08-30 | 2018-09-11 | Intel Corporation | Machine creation of program with frame analysis method and apparatus |
US10275222B2 (en) * | 2016-03-15 | 2019-04-30 | Intel Corporation | Technologies for physical programming |
US10380909B2 (en) | 2014-08-31 | 2019-08-13 | Square Panda Inc. | Interactive phonics game system and method |
WO2019217987A1 (en) * | 2018-05-15 | 2019-11-21 | Talkiwear Pty Ltd | Systems and methods for facilitating learning through interaction with objects in an environment |
US10652719B2 (en) | 2017-10-26 | 2020-05-12 | Mattel, Inc. | Toy vehicle accessory and related system |
US10748450B1 (en) * | 2016-11-29 | 2020-08-18 | Sproutel, Inc. | System, apparatus, and method for creating an interactive augmented reality experience to simulate medical procedures for pediatric disease education |
US20210187408A1 (en) * | 2018-04-05 | 2021-06-24 | Ftcubic Co., Ltd. | Play kitchen system using smart device based on virtual augmented reality and short range wireless communication |
US11056022B1 (en) * | 2016-11-29 | 2021-07-06 | Sproutel, Inc. | System, apparatus, and method for creating an interactive augmented reality experience to simulate medical procedures for pediatric disease education |
US20220111300A1 (en) * | 2011-05-17 | 2022-04-14 | Learning Squared, Inc. | Educational device |
US11471783B2 (en) | 2019-04-16 | 2022-10-18 | Mattel, Inc. | Toy vehicle track system |
US20230136468A1 (en) * | 2020-03-31 | 2023-05-04 | Tesseract Ventures, Llc | Collectible with themed design and content |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10901560B2 (en) * | 2018-01-08 | 2021-01-26 | Kids2, Inc. | Children's toys with capacitive touch interactivity |
CN109350973A (en) * | 2018-09-04 | 2019-02-19 | 佛山科学技术学院 | A method of controlling a model car with a mobile phone APP |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5853327A (en) * | 1994-07-28 | 1998-12-29 | Super Dimension, Inc. | Computerized game board |
US6290565B1 (en) * | 1999-07-21 | 2001-09-18 | Nearlife, Inc. | Interactive game apparatus with game play controlled by user-modifiable toy |
US6394872B1 (en) * | 1999-06-30 | 2002-05-28 | Inter Robot Inc. | Embodied voice responsive toy |
US20090137185A1 (en) * | 2007-11-28 | 2009-05-28 | Yu Brian Zheng | System, Method, and Apparatus for Interactive Play |
US8287372B2 (en) * | 2006-09-28 | 2012-10-16 | Mattel, Inc. | Interactive toy and display system |
US20120295704A1 (en) * | 2011-05-17 | 2012-11-22 | Paul Reiche | Interactive video game using game-related physical objects for conducting gameplay |
US20130165223A1 (en) * | 2011-12-22 | 2013-06-27 | Robert Leyland | Interactive video game with visual lighting effects |
US20130296058A1 (en) * | 2012-05-01 | 2013-11-07 | Activision Publishing, Inc. | Server based interactive video game with toys |
US20130303047A1 (en) * | 2012-05-08 | 2013-11-14 | Funfare, Llc | Sensor configuration for toy |
US20140030955A1 (en) * | 2010-11-17 | 2014-01-30 | Evgeny Nikolayevich Smetanin | Method and system for determination of one or more limbs of one or more toy devices |
US20140162785A1 (en) * | 2012-12-11 | 2014-06-12 | Paul Reiche | Interactive video game system comprising toys with rewritable memories |
US20140274313A1 (en) * | 2013-03-15 | 2014-09-18 | Karthik Bala | System and method for purchasing physical toys and corresponding virtual toys |
US20140273717A1 (en) * | 2013-03-13 | 2014-09-18 | Hasbro, Inc. | Three way multidirectional interactive toy |
US20150065258A1 (en) * | 2013-09-04 | 2015-03-05 | Christopher John Meade | Microchipped toy for interactive entertainment |
US20150087427A1 (en) * | 2013-09-26 | 2015-03-26 | At&T Mobility Ii Llc | Methods and apparatus to emulate a toy |
Family Cites Families (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1994015272A1 (en) * | 1992-12-22 | 1994-07-07 | Morgan Michael W | Pen-based electronic teaching system |
US5823782A (en) * | 1995-12-29 | 1998-10-20 | Tinkers & Chance | Character recognition educational system |
US20020160340A1 (en) * | 1998-07-31 | 2002-10-31 | Marcus Brian I. | Character recognition educational system |
JP2001307134A (en) * | 2000-04-19 | 2001-11-02 | Sony Corp | Three-dimensional model processor, its method and program providing medium |
GB2424510A (en) * | 2005-03-24 | 2006-09-27 | Nesta | Interactive blocks. |
US20130288563A1 (en) * | 2005-10-21 | 2013-10-31 | Patent Category Corporation | Interactive toy system |
US20090286210A1 (en) * | 2008-05-14 | 2009-11-19 | Fuzzy Logic | Methods and Systems for Providing Interactive Content |
US20110015920A1 (en) * | 2009-07-17 | 2011-01-20 | Locus Publishing Company | Apparatus for chinese language education and method thereof |
GB201019285D0 (en) * | 2010-11-15 | 2010-12-29 | Hepworth Browne Ltd | Interactive system and method of modifying user interaction therein |
US20120258436A1 (en) * | 2011-04-08 | 2012-10-11 | Case Western Reserve University | Automated assessment of cognitive, fine-motor, and memory skills |
US8755743B2 (en) * | 2011-07-29 | 2014-06-17 | Bose Corporation | Earpiece messaging |
WO2013128435A1 (en) * | 2012-02-29 | 2013-09-06 | Reshimo Ltd | Tracking system for objects |
JP2014029747A (en) * | 2012-07-31 | 2014-02-13 | Fujitsu Ltd | Mounting unit attaching device and electronic apparatus including the same |
US9039532B2 (en) * | 2012-10-31 | 2015-05-26 | Activision Publishing, Inc. | Interactive video game with toys having functionality that is unlocked through game play |
US10096003B2 (en) * | 2013-05-31 | 2018-10-09 | Javid Vahid | Apparatus, methods and systems for knowledge based maintenance |
US9433867B2 (en) * | 2013-06-10 | 2016-09-06 | Activision Publishing, Inc. | Video game with backwards-compatible toys |
US20160184724A1 (en) | 2014-08-31 | 2016-06-30 | Andrew Butler | Dynamic App Programming Environment with Physical Object Interaction |
US9259651B1 (en) * | 2015-02-13 | 2016-02-16 | Jumo, Inc. | System and method for providing relevant notifications via an action figure |
-
2015
- 2015-08-27 US US14/838,307 patent/US20160184724A1/en not_active Abandoned
-
2017
- 2017-02-09 US US15/428,154 patent/US10380909B2/en active Active
-
2019
- 2019-06-25 US US16/451,756 patent/US10607501B2/en active Active
-
2020
- 2020-02-19 US US16/794,591 patent/US10922994B2/en active Active
-
2021
- 2021-01-19 US US17/152,032 patent/US11776418B2/en active Active
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5853327A (en) * | 1994-07-28 | 1998-12-29 | Super Dimension, Inc. | Computerized game board |
US6394872B1 (en) * | 1999-06-30 | 2002-05-28 | Inter Robot Inc. | Embodied voice responsive toy |
US6290565B1 (en) * | 1999-07-21 | 2001-09-18 | Nearlife, Inc. | Interactive game apparatus with game play controlled by user-modifiable toy |
US8287372B2 (en) * | 2006-09-28 | 2012-10-16 | Mattel, Inc. | Interactive toy and display system |
US8926395B2 (en) * | 2007-11-28 | 2015-01-06 | Patent Category Corp. | System, method, and apparatus for interactive play |
US20090137185A1 (en) * | 2007-11-28 | 2009-05-28 | Yu Brian Zheng | System, Method, and Apparatus for Interactive Play |
US20140030955A1 (en) * | 2010-11-17 | 2014-01-30 | Evgeny Nikolayevich Smetanin | Method and system for determination of one or more limbs of one or more toy devices |
US20120295704A1 (en) * | 2011-05-17 | 2012-11-22 | Paul Reiche | Interactive video game using game-related physical objects for conducting gameplay |
US20130165223A1 (en) * | 2011-12-22 | 2013-06-27 | Robert Leyland | Interactive video game with visual lighting effects |
US20130296058A1 (en) * | 2012-05-01 | 2013-11-07 | Activision Publishing, Inc. | Server based interactive video game with toys |
US20130303047A1 (en) * | 2012-05-08 | 2013-11-14 | Funfare, Llc | Sensor configuration for toy |
US20140162785A1 (en) * | 2012-12-11 | 2014-06-12 | Paul Reiche | Interactive video game system comprising toys with rewritable memories |
US20140273717A1 (en) * | 2013-03-13 | 2014-09-18 | Hasbro, Inc. | Three way multidirectional interactive toy |
US20140274313A1 (en) * | 2013-03-15 | 2014-09-18 | Karthik Bala | System and method for purchasing physical toys and corresponding virtual toys |
US20150065258A1 (en) * | 2013-09-04 | 2015-03-05 | Christopher John Meade | Microchipped toy for interactive entertainment |
US20150087427A1 (en) * | 2013-09-26 | 2015-03-26 | At&T Mobility Ii Llc | Methods and apparatus to emulate a toy |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220111300A1 (en) * | 2011-05-17 | 2022-04-14 | Learning Squared, Inc. | Educational device |
US10380909B2 (en) | 2014-08-31 | 2019-08-13 | Square Panda Inc. | Interactive phonics game system and method |
US10607501B2 (en) | 2014-08-31 | 2020-03-31 | Square Panda Inc. | Interactive phonics game system and method |
US11776418B2 (en) | 2014-08-31 | 2023-10-03 | Learning Squared, Inc. | Interactive phonics game system and method |
US10922994B2 (en) | 2014-08-31 | 2021-02-16 | Square Panda, Inc. | Interactive phonics game system and method |
US10275222B2 (en) * | 2016-03-15 | 2019-04-30 | Intel Corporation | Technologies for physical programming |
US10074205B2 (en) | 2016-08-30 | 2018-09-11 | Intel Corporation | Machine creation of program with frame analysis method and apparatus |
US9914062B1 (en) * | 2016-09-12 | 2018-03-13 | Laura Jiencke | Wirelessly communicative cuddly toy |
US11056022B1 (en) * | 2016-11-29 | 2021-07-06 | Sproutel, Inc. | System, apparatus, and method for creating an interactive augmented reality experience to simulate medical procedures for pediatric disease education |
US10748450B1 (en) * | 2016-11-29 | 2020-08-18 | Sproutel, Inc. | System, apparatus, and method for creating an interactive augmented reality experience to simulate medical procedures for pediatric disease education |
US10652719B2 (en) | 2017-10-26 | 2020-05-12 | Mattel, Inc. | Toy vehicle accessory and related system |
US20210187408A1 (en) * | 2018-04-05 | 2021-06-24 | Ftcubic Co., Ltd. | Play kitchen system using smart device based on virtual augmented reality and short range wireless communication |
US11907413B2 (en) * | 2018-04-05 | 2024-02-20 | Ftcubic Co., Ltd. | Play kitchen system using smart device based on virtual augmented reality and short range wireless communication |
WO2019217987A1 (en) * | 2018-05-15 | 2019-11-21 | Talkiwear Pty Ltd | Systems and methods for facilitating learning through interaction with objects in an environment |
US11471783B2 (en) | 2019-04-16 | 2022-10-18 | Mattel, Inc. | Toy vehicle track system |
US11964215B2 (en) | 2019-04-16 | 2024-04-23 | Mattel, Inc. | Toy vehicle track system |
US20230136468A1 (en) * | 2020-03-31 | 2023-05-04 | Tesseract Ventures, Llc | Collectible with themed design and content |
Also Published As
Publication number | Publication date |
---|---|
US10380909B2 (en) | 2019-08-13 |
US20200184842A1 (en) | 2020-06-11 |
US11776418B2 (en) | 2023-10-03 |
US20210248919A1 (en) | 2021-08-12 |
US10922994B2 (en) | 2021-02-16 |
US20190318647A1 (en) | 2019-10-17 |
US20170213475A1 (en) | 2017-07-27 |
US10607501B2 (en) | 2020-03-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20160184724A1 (en) | Dynamic App Programming Environment with Physical Object Interaction | |
US11452941B2 (en) | Emoji-based communications derived from facial features during game play | |
US20130059284A1 (en) | Interactive electronic toy and learning device system | |
CN105396295B (en) | A kind of space programming robot toy towards children | |
ES2364956T3 (en) | HANDLING INTERACTIVE DEVICES. | |
EP2744579B1 (en) | Connected multi functional system and method of use | |
Tollervey | Programming with MicroPython: embedded programming with microcontrollers and Python | |
US20110199194A1 (en) | Programmable device and programming method | |
US20170136380A1 (en) | Smart Toys | |
JP2003205483A (en) | Robot system and control method for robot device | |
JP2008279165A (en) | Toy system and computer program | |
JP2017537669A (en) | Game system | |
US20150065258A1 (en) | Microchipped toy for interactive entertainment | |
US20150279371A1 (en) | System and Method for Providing an Audio Interface for a Tablet Computer | |
CN107251570A (en) | Sensing data availability from remote equipment | |
KR101452359B1 (en) | Method for providing of toy assembly video | |
US11550528B2 (en) | Electronic device and method for controlling operation of accessory-mountable robot | |
US20150273321A1 (en) | Interactive Module | |
CN107427724B (en) | Method and apparatus for enhancing gaming experience | |
US9886865B2 (en) | Providing enhanced experience based on device location | |
US20200368630A1 (en) | Apparatus and System for Providing Content to Paired Objects | |
KR102344518B1 (en) | Method for providing mixed rendering content using mixed reality in forest experience environment and apparatus using the same | |
Walker | Ghosts of the Horseshoe: A Mobilization of a Critical Interactive | |
Hanning | Tinkering with ScratchBit: Explorations in blended making | |
KR20190069035A (en) | Apparatus for producing algorithm |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |