WO1997004840A1 - Electronic exercise enhancer - Google Patents

Electronic exercise enhancer Download PDF

Info

Publication number
WO1997004840A1
WO1997004840A1 PCT/US1996/011885 US9611885W WO9704840A1 WO 1997004840 A1 WO1997004840 A1 WO 1997004840A1 US 9611885 W US9611885 W US 9611885W WO 9704840 A1 WO9704840 A1 WO 9704840A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
controller
data
uεer
imaging
Prior art date
Application number
PCT/US1996/011885
Other languages
French (fr)
Inventor
Craig K. Poulton
Original Assignee
Poulton Craig K
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Poulton Craig K filed Critical Poulton Craig K
Priority to EP96925358A priority Critical patent/EP0840638B1/en
Priority to AU65482/96A priority patent/AU6548296A/en
Priority to DE69634915T priority patent/DE69634915D1/en
Publication of WO1997004840A1 publication Critical patent/WO1997004840A1/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B24/00Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B71/00Games or sports accessories not covered in groups A63B1/00 - A63B69/00
    • A63B71/06Indicating or scoring devices for games or players, or for other sports activities
    • A63B71/0619Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
    • A63B71/0622Visual, audio or audio-visual systems for entertaining, instructing or motivating the user
    • A63B2071/0638Displaying moving images of recorded environment, e.g. virtual environment
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B2213/00Exercising combined with therapy
    • A63B2213/004Exercising combined with therapy with electrotherapy
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B2220/00Measuring of physical parameters relating to sporting activity
    • A63B2220/30Speed
    • A63B2220/34Angular speed
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B2220/00Measuring of physical parameters relating to sporting activity
    • A63B2220/50Force related parameters
    • A63B2220/51Force
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B2220/00Measuring of physical parameters relating to sporting activity
    • A63B2220/70Measuring or simulating ambient conditions, e.g. weather, terrain or surface conditions
    • A63B2220/76Wind conditions
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B2225/00Miscellaneous features of sport apparatus, devices or equipment
    • A63B2225/64Heated
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B2225/00Miscellaneous features of sport apparatus, devices or equipment
    • A63B2225/66Cooled
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B69/00Training appliances or apparatus for special sports
    • A63B69/16Training appliances or apparatus for special sports for cycling, i.e. arrangements on or for real bicycles
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S482/00Exercise devices
    • Y10S482/90Ergometer with feedback to load or with feedback comparison

Definitions

  • This invention relates to exercise equipment and, more particularly, to novel systems and methods for enhancing exercises by providing to a user multiple stimuli and by tracking multiple responses of a user, all with programmable electronic control.
  • Exercise continues to be problematic for persons having limited time and limited access to outdoor recreational facilities or large indoor recreational facilities. Meanwhile, more, and more realistic, simulated, training environments are needed for lower cost instruction and practice. For example, flight training requires a very expensive aircraft. Nuclear plant control requires a complex system of hardware and software. combat vehicle training, especially large force maneuvers, requires numerous combat vehicles and supporting equipment. Personal fitness may require numerous machines of substantial size and sophistication placed in a large gym to train athletes in skill or strength, especially if all muscle groups are to be involved. In short, training with real equipment may require substantial real estate and equipment, with commensurate cost.
  • simulated environments often lack many or even most of the realistic stimuli received by a user in the real world including motions over distance, forces, pressures, sensations, temperatures, images, multiple views in the three-dimensions surrounding a user, and so forth.
  • many simulations do not provide the proper activities for a user, including a full range of motions, forces, timing, reflexes, speeds, and the like.
  • What is needed is a system for providing to a user more of the benefits of a real environment in a virtual environment. Also needed is a system for providing coordinated, synchronized, sensory stimulation by multiple devices to more nearly simulate a real three-dimensional spatial environment. Similarly needed is an apparatus and method for tracking a plurality of sensors monitoring a user's performance, integrating the inputs provided by such tracking, and providing a virtual environment simulating time, space, motion, images, forces and the like for the training, conditioning, and experience of a user.
  • a controller capable of changing the stimuli and requirements (such as images, electromuscular and audio stimulation, loads and other resistance to movement, for example) imposed on a user is needed to make training and exercise approach the theoretical limits of comfort, endurance, or optimized improvement, as desired.
  • a system is needed for providing either a choice or a combination of user control, selectable but pre-programmed (template-like or open loop) control, and adaptive (according to a user's condition, comfort, or the like) control of muscle and sensory stimulation, resistances, forces, and other actuation imposed on a user by the system, according to a user's needs or preferences.
  • An electronically controlled exercise enhancer is disclosed in one embodiment of the present invention as including an apparatus having a controller with an associated processor for controlling stimuli delivered to a user and for receiving feedback corresponding to responses of a user.
  • a tracking device may be associated with the controller to communicate with the controller for tracking responses of a user and for providing to the controller certain data corresponding to the condition, exertion, position, and other characteristics of a user.
  • the tracking device may also include a processor for processing signals provided by a plurality of sensors and sending corresponding data to the controller.
  • the plurality of sensors deployed to detect the performance of a user may include, for example, a radar device for detecting position, velocity, motion, or speed; a pressure transducer for detecting stress; strain gauges for detecting forces, motion, or strain in a member of the apparatus associated with performance of a user. Such performance may include strength, force applied to the member, deflection, and the like.
  • Other sensors may include humidity sensors; temperature sensors; calorimeters for detecting energy dissipation, either by rate or integrated over time; a heart rate sensor for detecting pulse; and an imaging device.
  • the imaging device may provide for detecting the position, velocity, or condition of a member. Imaging may also assess a condition of a plane, volume, or an internal or external surface of a bodily member of a user.
  • One or more sensors may be connected to provide analog or digital signals to the tracking device for processing.
  • the tracking device may then transfer corresponding digital data to the controller.
  • the controller may do all signal processing, whereas in other embodiments, distributed processing may be relied upon in the tracker, or even in individual sensors to minimize the bandwidth required for the exchange of data between devices in the apparatus.
  • a stimulus interface device may be associated with the controller for delivering selected stimuli to a user.
  • the stimulus interface device may include a processor for controlling one or more actuators (alternatively called output devices) for providing stimulus to a user.
  • actuators alternatively called output devices
  • certain actuators may also contain processors for certain functions, thus reducing the bandwidth required for communications between the controller and the output devices.
  • the controller may provide processing for data associated with certain actuators.
  • Actuators for the sensory interface device may include aural actuators for presenting sounds to a user, such as speakers, sound synthesizers with speakers, compact disks and players associated with speakers for presenting aural stimuli, or electrodes for providing electrical impulses associated with sound directly to a user.
  • aural actuators for presenting sounds to a user such as speakers, sound synthesizers with speakers, compact disks and players associated with speakers for presenting aural stimuli, or electrodes for providing electrical impulses associated with sound directly to a user.
  • Optical actuators may include cathode ray tubes displaying images in black and white or color, flat panel displays, imaging goggles, or electrodes for direct electrical stimulus delivered to nerves or tissues of a user. Views presented to a user may be identical for both eyes of a user, or may be stereoscopic to show the two views resulting from the parallax of the eyes, thus providing true three-dimensional images to a user.
  • the actuators may include temperature actuators for providing temperature or heat transfer.
  • working fluids warmed or cooled to provide heat transfer thermionic devices for heating and cooling an junction of a bimetallic probe, and the like may be used to provide thermal stimulus to a user.
  • Kinematic actuators may provide movement in one or more degrees of freedom, including translation and rotation with respect to each of the three spatial axes. Moreover, the kinematic actuators may provide a stimulus corresponding to motion, speed, force, pressure or the like.
  • the kinematic actuators may be part of a suite of tactile actuators for replicating or synthesizing stimuli corresponding to each tactile sensation associated with humans' sense or touch of feel. In general a suite of tactile, optical, and aural, and even olfactory and taste actuators may replicate virtually any sensible output for creating a corresponding sensation by a user.
  • the tracking device may be equipped with sensors for sensing position, displacement, motion, deflection, velocity, speed, temperature, pH, humidity, heart rate, images, and the like for accumulating data.
  • Data may correspond to the biological condition and spatial kinematics (position, velocity, forces) of a bodily member of a user. For example, skin tension, pressure, forces in any spatial degree of freedom and the like may be monitored and fed back to the controller.
  • the sensory interface device may produce outputs presented as stimuli to a user.
  • the sensory interface device may include one or more actuators for providing aural, optical, tactile, and electromuscular stimulation to a user.
  • the controller, tracking device, and sensory interface device may all be microprocessor controlled for providing coordinated sensory perceptions of complex events.
  • actuators may represent a coordinated suite of stimuli corresponding to the sensations experienced by a user.
  • a user may experience a panoply of sensory perceptions besides sight.
  • sensations may replicate, from synthesized or sampled data, a cycling tour through varied terrain and vegetation, a rocket launch, a tail spin in an aircraft, a flight by aircraft including takeoff and landing.
  • Sensations may be presented for maneuvers such as aerobatics.
  • a combat engagement may be experienced from within a combat vehicle or simulator.
  • Sensory inputs may include those typical of a turret with slewing control and mounting weaponry with full fire control. Besides motion, sensory inputs may include hits received or made. Sensations may imitate or replicate target acquisition, tracking, and sensing or the like.
  • hand-to-hand combat with a remote user operating a similar apparatus may be simulated by the actuators. Sensors may feed back data to the controller for forwarding to the system of the remote user, corresponding to all the necessary actions, condition, and responses of the user.
  • a mountain hike, a street patrol by police, a police fire fight, an old west gunfight, a mad scramble over rooftops, through tunnels, down cliffs, and the like may all be simulated with properly configured and powered actuators and sensors.
  • Stimuli provided to a user may be provided in a variety of forms, including electromuscular stimulation. Stimuli may by timed by a predetermined timing frequency set according to a pre-programmed regimen set by a user or a trainer as an input to an executable code of a controller.
  • stimuli may be provided with interactively determined timing.
  • Interactively determined timing for electromuscular stimulation means that impulses may be timed and scaled in voltage, frequency, and other parameters according to a user's performance.
  • detection is possible for the motion, speed, position, muscular or joint extension, muscle tension or loading, surface pressure, or the like. Such detection may occur for many body members. Members may include a user's foot, arm, or other bodily member.
  • Sensed inputs may be sensed and used in connection with other factors to control the timing and effect of electromuscular stimulation.
  • the electromuscular stimulation may be employed to enhance the contraction or extension of muscles beyond the degree of physiological stimulation inherent in the user.
  • sensory impact may be provided by actuators electrically stimulating muscles or muscle groups to simulate forces imposed on bodily members by outside influences.
  • a virtual baseball may effectively strike a user.
  • a martial arts player may strike another from a remote location by electromuscular stimulation.
  • two contestants may interact although physically separated by some distance.
  • two contestants may engage in a boxing or martial arts game or contest in which a hit by one contestant faced with a virtual opponent is felt by the opponent.
  • sensory inputs may be provided based on each remote opponents actual movements.
  • impacts may be literally felt by each opponent at the remote location.
  • responses of each opponent may be presented as stimuli to each opponent (user) .
  • Figure l is a schematic block diagram of an apparatus made in accordance with the invention
  • Figures 2-3 are schematic block diagrams of software modules for programmable operation of the apparatus of Figure 1.
  • Figure 4 is a schematic block diagram of one embodiment of the data structures associated with the apparatus of Figure 1 and the software modules of Figures 2-3.
  • Figure 5 is a schematic block diagram of one embodiment of the apparatus of Figure 1 adapted to tracking and actuation, including electromuscular stimulation, of a user of a stationary bicycle exerciser. Best Modes for Carrying Out the Invention
  • Figure 1 illustrates one presently preferred embodiment of a controller for programmably directing the operation of an apparatus made in accordance with the present invention, a tracking device for sensing and feeding back to the controller the condition and responses of a user, and a sensory interface device for providing stimuli to a user through one or more actuators.
  • Figure 2 illustrates in more detail a schematic diagram of one preferred embodiment of software programming modules for the tracking device with its associated sensors, and for the sensory interface device with its associated actuators for providing stimuli to a user.
  • Figure 3 illustrates in more detail a schematic diagram of one preferred embodiment of software modules for programming the controller of Figure 1.
  • Figure 4 illustrates a schematic block diagram of one embodiment of data structures for storing, retrieving and managing data used and produced by the apparatus of Figure 1.
  • the present invention provides an apparatus for presenting one or more selected stimuli to a user, feeding back to a controller the responses of a user, and processing the feedback to provide a new set of stimuli.
  • the apparatus 10 made in accordance with the invention may include a controller 12 for exercising overall control over the apparatus 10 or system 10 of the invention.
  • the controller 12 may be connected to communicate with a tracking device 14 for feeding back data corresponding to performance of a user.
  • the controller 12 may also connect to exchange data with a sensory interface device 16.
  • the sensory interface device 16 may include one or more mechanisms for presenting sensory stimuli to a user.
  • the controller 12, tracking device 14 and interface device 16 may be connected by a link 18, which may include a hardware connection and software protocols such as the general purpose interface bus (GPIB) as described in the IEEE 488 standard, and commonly used as a computer bus.
  • GPIB general purpose interface bus
  • the link 18 may be selected from a universal ace synchronous receiver-transmitter. Since such a system may include a module composed of a single integrated circuit for both receiving and transmitting, asynchronously through a serial communications port, this type of link 18 may be simple, reliable, and inexpensive. Alternatively, a universal synchronous receiver-transmitter (USRT) module may be used for communication over a pair of serial channels. Although slightly more complex, such a link 18 may be used to pass more data.
  • USB universal synchronous receiver-transmitter
  • a link 18 is a network 20, such as a local area network. If the controller 12, tracking device 14 and sensory interface device 16 are each provided with some processor, then each may be a node on the network 20. Thus, a server 22 may be connected to the network 20 for providing data storage, and general file access for any processor in the system 10. A router 24 may also be connected to the network 20 for providing access to a larger internetwork, such as the worldwide web or internet. The operation of servers 22 and routers 24 reduce the duty required of the controller 12, and may also permit interaction between multiple controllers 12 separated across internetworks. For use of an apparatus 10 in an interactive mode, wherein interactive means interaction between users remotely spaced from one another, an individual user might have a substantially easier task trying to find a similarly situated partner for interactive games. Moreover, real-time interaction, training, and teaming between users located at great distances may be accomplished using the system 10.
  • the network interface cards 26A, 26B, 26C, 26D, 26E may be installed in the controller 12, tracking device 14, sensory interface device 16, server 22, and router 24, respectively, for meeting the hardware and software conventions and protocols of the network 20.
  • the controller 12 may include a processor 30 connected to operate with a memory device 32.
  • a memory device 32 may be a random access memory or other volatile memory used during operation of the processor 30.
  • Long term memory of software, data, and the like, may be accommodated by a storage device 34 connected to communicate with the processor 30.
  • the storage device 34 may be a floppy disk drive, a random access memory, but may in one preferred embodiment of the system 10 include one or more hard drives.
  • the - ii - storage device 34 may store applications, data bases, and various files needed by the processor 30 during operation of the system 10.
  • the storage device 34 may download from the server 22 according to the needs of the controller 12 in any particular specific task, game, training session, or the like.
  • An input device 36 may be connected to communicate with a processor 30.
  • a user may program a processor 30 by creating an application to be stored in the storage device 34 and run on the processor 30.
  • An input device 36 may be a keyboard.
  • the input device 36 may be selected from a capacitor membrane keypad, a graphical user interface such as a monitor having menus and screens, or icons presented to a user for selection.
  • An input device may include a graphical pad and stylus for use by a user inputting a figure rather than text or ASCII characters.
  • an output device 38 may be connected to the processor 30 for feeding back to a user certain information needed to control the controller 12 or processor 30.
  • a monitor may be a required output device 38 to operate with the menu and icons of an input device 36 hosted on the same monitor.
  • an output device may include a speaker for producing a sound to indicate that an improper selection, or programming error has been committed by a user operating the input device 36 to program the processor 30.
  • Numerous input device 36 and output devices 38 for interacting with the processor 30 of the controller 12 are available, and within contemplation of the invention.
  • the processor 30, memory device 32, storage device 34, input device 36, and output device 38 may all be connected by a bus 40.
  • the bus may be of any suitable type such as those used in personal computers or other general purpose digital computers.
  • the bus may also be connected to a serial port 42 and a parallel port 44 for communicating with other peripheral devices selected by a user.
  • a parallel port 44 may connect to an additional storage device, a slaved computer, a master computer, or a host of other peripheral devices.
  • a removable media device 46 may be connected to the bus 40.
  • a removable media device such as a floppy disk drive, a BernoulliTM drive, an optical drive, a compact disk laser readable drive, or the like could be connected to the bus 40 or to one of the ports 42, 44.
  • a user could import directly a software program to be loaded into the storage device 34, for later operation on the processor 30.
  • the tracking device 14 and the sensory interface device 16 may be "dumb" apparatus. That is, the tracking device 14 and sensory interface device 16 might have no processors contained within their hardware suites. Thus, the processor 30 of the controller 12 may do all processing of data exchanged by the tracking device, sensory interface device, and controller 12. However, to minimize the required bandwidths of communication lines such as the link 18, the network 20, the bus 40, and so forth, processors may be located in virtually any hardware apparatus.
  • the tracking device 14 may include a processor 50 for performing necessary data manipulation within the tracking device 14.
  • the processor 50 may be connected to a memory device 52 by a bus 54.
  • the tracking device may also include a storage device 56, although a storage device 56 may typically increase the size of the tracking device 14 to an undesirable degree for certain utilities.
  • the tracking device 14 may include a signal converter 58 for interfacing with a suite including one or more sensors 60.
  • the signal converter 58 may be an analog to digital converter, required by certain types of sensors 60. Signal processing may be provided by the processor 50. Nevertheless, certain types of sensors 60 may include a signal processor and signal converter organically included within the packaging of the sensor 60.
  • the sensors 60 may gather information in the form of signals sensed from the activities of the user.
  • the sensors 60 may include a displacement sensor 62 for detecting a change of position in 1, 2, or 3 spacial dimensions.
  • the displacement sensor 62 may be thought of as a sensor of relative position between a fir ⁇ t location and a second location.
  • a position sen ⁇ or 64 may be provided to detect an absolute position in space.
  • a displacement sensor 62 might detect the position or movement of a member of a user's body with respect to a constant frame of reference, whereas a displacement sensor 62 might simply detect motion between a first stop location and a second stop location, the starting location being reset every time the movement stops.
  • Each type of sensor 62, 64 may have certain advantages.
  • a calibrator 66 may be provided for each sensor, or for all the sensors, depending on which types of sensors 60 are used. The calibrator may be used to null the signals from sensors 60 at the beginning of use to as ⁇ ure that biases and drifting do not thwart the function of the system 10.
  • Other sensors 60 may include a velocity sensor 68 for detecting either relative speed, a directionless scalar quantity, or a velocity vector including both speed and direction.
  • a velocity sensor 68 may be configured as a combination of a displacement sensor 62 or position sensor 64 and a clock for corresponding a position to a time.
  • a temperature sensor 70 may be provided, and relative temperatures may also be measured.
  • a temperature-sensing thermocouple may be placed against the skin of a user, or in the air surrounding a user's hand. Thus, temperature may be sensed electronically by temperature sensors 70.
  • relative humidity surrounding a user may be of importance, and may be detected by a humidity sensor 72.
  • a heart rate sensor 74 may be included in the suite of sensor ⁇ 60.
  • Force sensor ⁇ 76 may be of a force variety or of a pressure variety. That is, transducers exist to sense a total integrated force. Alternatively, transducer ⁇ al ⁇ o exist to detect a force per unit of area to which the force is applied, the classical definition of pressure. Thus, the force sensors 76 may include force and pres ⁇ ure monitoring.
  • an imaging sensor 78 may be included as a sensor 60.
  • Imaging sen ⁇ or ⁇ may have a processor or multiple processors organic or integrated within themselves to manage the massive amounts of data received.
  • An imaging sensor may provide certain position data through image processing.
  • the position sensor 64 or displacement sensor 62 may be a radar, such as a Doppler radar mechanism for detecting movement of a foot, leg, the rise and fall of a user's chest during breathing, or the like.
  • a radar system may use a target patch for reflecting its own signal from a surface, such as the ⁇ kin of a user, or the surface of a shoe, the pedal of a bicycle, or the like.
  • a radar may require much lower bandwidths for communicating with the proces ⁇ or 50 or the controller 12 than may be required by an imaging sensor 78. Nevertheless, the application to which the apparatu ⁇ 10 is put may require either an imaging sensor 78 or a simple displacement sensor 62. In another example a linear variable displacement transducer is a common and simple device that has traditionally been used for relative displacement. Thus, one or more of the ⁇ ensors 60 de ⁇ cribed above may be included in the tracking device 14 to monitor the activity and condition of a u ⁇ er of the system 10.
  • a sensory interface device 16 may include a proces ⁇ or 80 and a memory device 82 connected to a bus 84.
  • a storage device 86 may be connected to the bu ⁇ 84 in some configurations, but may be considered too large for highly portable ⁇ en ⁇ ory interface devices 16.
  • the sen ⁇ ory interface device 80 may include a power ⁇ upply 88, and may include more than one power ⁇ upply 88 either centrally located in the sensory interface device or distributed among the various actuators 90.
  • a power supply 88 may be one of several types.
  • a power supply may be an electrical power ⁇ upply.
  • a power ⁇ upply may be a hydraulic power supply, a pneumatic power supply, a magnetic power supply, or a radio frequency power supply.
  • a sensor 60 may use a very small amount of power to detect a motion
  • an actuator 90 may provide a sub ⁇ tantial amount of energy.
  • the actuators 90 may particularly benefit from a calibrator 92.
  • an actuator which provides a ⁇ pecific displacement or motion ⁇ hould be calibrated to be ⁇ ure that it does not move beyond a desired position, since the result could be injury to a user.
  • the actuator ⁇ may be calibrated by a calibrator 92 connected to null out any actuation of the actuator in an inactive, uncommanded mode.
  • an aural actuator 94 may be an aural actuator 94.
  • a simple aural actuator may be a ⁇ ound ⁇ peaker.
  • an aural actuator 94 may include a ⁇ ynthe ⁇ ized ⁇ ound generator a ⁇ well a ⁇ some speaker for projecting the sound.
  • an aural actuator 94 may have within itself the ability to create sound on demand, and thus have its own internal processor, or it may simply duplicate an analog sound signal received from another source.
  • an aural actuator may be a compact disk player, power supply, and all peripheral devices required, with a simple control signal sent by the processor 80 to determine what sound ⁇ are presented to a user by the aural actuator 94.
  • An optical actuator 96 may include a computer monitor that di ⁇ plays images much as a television screen does.
  • an optical actuator may include a pair of goggles comprising a flat panel image display, a radar display, such as an oscillo ⁇ copic catha-ray tube di ⁇ playing a trace of signal, a fibre optic display of an actual image transmitted only by light, or a fibre optic display transmitting a ⁇ ynthetically generated image from a computer or from a compact disk reader.
  • the optical actuator may provide an optical stimulu ⁇ .
  • the optical actuator may actually include electrodes for providing stimulus to optical nerves, or directed to the brain.
  • the optical actuator may be embodied in a sophi ⁇ ticated computer-controlled series of electrodes producing voltages to be received by nerves in the human body.
  • a user may be surrounded by a mosaic of cathode ray tube type monitors or flat panel displays creating a scene to be viewed as if through a cockpit window or other position.
  • a user may wear a pair of stereo goggles, having two images corresponding to the parallax views presented to each eye by a three dimensional image.
  • a manner and mechanism may be ⁇ imilar to those by which stereo aerial photographs are used.
  • a user may be shown multi-dimen ⁇ ional geographical feature ⁇ , stereo views of recorded images.
  • Images may be generated or stored by either analog recording devices such as films.
  • Likewi ⁇ e images may be handled by digital devices such as compact disk ⁇ and computer magnetic memorie ⁇ .
  • Images may be used to provide to a user in a very close environment, stereo views appearing to be three dimensional image ⁇ . For example, ⁇ tereo views may be displayed digitally in the two "lens" displays of goggles adapted for such use.
  • Such devices a ⁇ infrared imaging goggles, or digitized image ⁇ originally produced by infrared imaging goggle ⁇ , may be provided. Any of these optical actuators 96 may be adapted for use with the sensory interface device 16.
  • a tactile actuator 98 may be included for providing to a user a sense of touch.
  • an electromuscular actuator 100 may be a part of, or connected to, the sen ⁇ ory interface device 16 for permitting a u ⁇ er to feel touched.
  • a temperature actuator 102 may pre ⁇ ent different temperature ⁇ of contacting ⁇ urface ⁇ or fluid ⁇ again ⁇ t the skin of a user.
  • the tactile actuator 98, electromuscular actuator 100, and temperature actuator 102 may interact with one another to produce a total tactile experience.
  • the electromuscular actuator 100 may be used to augment exerci ⁇ e, to give a ⁇ en ⁇ ation of impact, or to give feedback to a pro ⁇ thetic device worn by a u ⁇ er in medical rehabilitation.
  • Example ⁇ of tactile actuator ⁇ may include a pressure actuator.
  • a panel, an arm, a probe, or a bladder may have a surface that may be moved with respect to the skin of a u ⁇ er.
  • a user may be moved, or pres ⁇ ured.
  • a u ⁇ er may wear a glove or a boot on a hand or foot, re ⁇ pectively, for ⁇ imulating certain activities.
  • a bladder actuated by a pump may be filled with air, water, or other working fluid to create a pressure. With a surface of the bladder against a retainer on one side, and the skin of a user on the other side, a user may be made to feel pre ⁇ ure over a surface at a uniform level.
  • a glove may have a series of articulated structural members, joints and connectors, actuated by hydraulic or pneumatic cylinders.
  • a user may be made to feel a force exerted against the inside of a user's palm or fingers in response to a grip.
  • a user could be made to feel the grip of a machine by either a force, or a displacement of the articulated members.
  • a user could arm wrestle a machine.
  • a user could arm wrestle a remote user, the pressure actuator 104, force actuator 106, or position actuator 108 inherent in a tactile actuator providing di ⁇ placements and forces in response to the motion of a user.
  • Each user, remote from each other, could nevertheless transfer motion ⁇ and forces digitally across the worldwide web between distant systems 10.
  • the temperature actuator may include a pump or fan for blowing air of a selected temperature over the skin of a u ⁇ er in a ⁇ uit adapted for such use.
  • the temperature actuator may include a bladder touching the skin, the bladder being alternately filled with heated or cooled fluid, either air, water, or other working fluid ⁇ .
  • the temperature actuator 102 may be constructed using thermionic devices.
  • the principle of a thermocouple may be used. A voltage and power are applied to create heat or cooling at a bi ⁇ metallic junction.
  • a temperature actuator 102 may include a thermionic device contacting the ⁇ kin of a u ⁇ er, or providing a source of heat or cold for a working fluid to warm or cool the skin of a user in response to the processor 80.
  • a control module 110 may be operable in the processor 30 of the controller 12.
  • a tracking module 112 may run on a proces ⁇ or 50 of the tracking device 14.
  • An actuation module 114 may include programmed instructions for running on a processor 80 of the sensory interface device 16.
  • the control module 110 may include an input interface module 116 including codes for prompting a user, receiving data, providing data prompts, and otherwise managing the data flow from the input device 36 to the proces ⁇ or 30 of the controller 12.
  • the output interface module 118 of the control module 110 may manage the interaction of the output device 38 with the proce ⁇ sor 30 of the controller 12.
  • the input interface module 116 and output interface module 118 in one presently preferred embodiment, may exchange data with an application module 120 in the control module 110.
  • the application module 120 may operate on the proces ⁇ or 30 of the controller 12 to load and run application ⁇ 122.
  • Each application 122 may corre ⁇ pond to an individual ses ⁇ ion by a u ⁇ er, a particular programmed set of instructions designed for a game, an exercise workout, a rehabilitative regimen, a training session, a training les ⁇ on, or the like. Thu ⁇ , the application module 120 may coordinate the receipt of information from the input interface module 116, output interface module 118, and the application 122 actually running on the proce ⁇ or 30.
  • the application module 120 may be thought of as the highest level programming running on the processor
  • the application module 120 may exchange data with a programming interface module 124 for providing access and control by a user to the application module 120.
  • the programming interface module 124 may be used to control and transfer information provided through a keyboard connected to the controller 12.
  • the programming interface module may include software for downloading applications 122 to be run by the application module 120 on the proces ⁇ or 30 or to be stored in the storage device 34 for later running by the processor 30.
  • the input interface module 116 may include programmed instruction ⁇ for controlling the transfer of information, for example, digital data, between the application module 120 of the control module 110 running on the processor 30, and the tracking device 14.
  • the output interface module 118 may include programmed instructions for transferring information between the application module 120 and the sen ⁇ ory interface device 16.
  • the input interface module 116 and output interface module 118 may deal exclusively with digital data files or data stream ⁇ pa ⁇ ed between the tracking device 14 and the sensory interface device 16 in an embodiment where each of the tracking device 14 and sensory interface device 16 are themselves microprocessor controlled with microprocessors organic (integral) to the respective structures.
  • the control module 10 may include an interaction module 128 for transferring data between control modules 110 of multiple, at least two, systems 10.
  • an interaction module 128 may contain programmed instructions for controlling data flow between an application module 120 in one location and an application module 120 of an entirely different system 10 at another location, thus facilitating a high level of coordination between applications 122 on different system ⁇ 10.
  • a network module 126 may contain programmed instruction ⁇ regarding logging on and off of the network, communication protocol ⁇ over the network, and the like. Thu ⁇ , the application module 120 may be regarded a ⁇ the heart of the software running on the controller 12, or more precisely, on the processor 30 of the controller 12. Meanwhile, the function ⁇ associated with network acces ⁇ may be included in a network module 126, while certain interaction between cooperating systems 10 may be handled by an interaction module 128. Different task ⁇ may be rea ⁇ igned to different software module ⁇ , depending on hardware configuration ⁇ of a ⁇ pecific problem or ⁇ y ⁇ tem 10. Therefore, equivalent ⁇ y ⁇ tem ⁇ 10 may be configured according to the invention. For example, a single application 122 may include all of the functions of the modules 120-128.
  • more than one proces ⁇ or 30 may be u ⁇ ed.
  • a multi-tasking proce ⁇ or may be u ⁇ ed a ⁇ the proce ⁇ or 30.
  • Thu ⁇ , multiple proce ⁇ e ⁇ , thread ⁇ , program ⁇ , or the like, may be made to operate on a variety of proce ⁇ or ⁇ , a plurality of proce ⁇ or ⁇ , or in a ulti- ta ⁇ king arrangement on a multi-tasking proces ⁇ or 30.
  • data may be transferred between a controller 12 and a tracking device 14, the sen ⁇ ory interface device 16, a keyboard, and monitor, a remote controller, and other node ⁇ on a network 20.
  • the tracking module 112 may include a ⁇ ignal generator 130.
  • a ⁇ ignal generator may be any of a variety of mechanisms operating within a sensor, to create a signal.
  • the signal generator 130 may then pass a signal to a signal converter 132.
  • an analog to digital converter may be common in certain transducer ⁇ .
  • a signal generator 130 may itself by microproces ⁇ or-controlled, and may produce a data ⁇ tream needing no conver ⁇ ion by a signal converter 132.
  • a signal converter 132 may convert a signal from a signal generator 130 to a digital data signal that may be proces ⁇ ed by a signal proces ⁇ or 134.
  • a signal processor 134 may operate on the processor 30 of the controller 12, but may benefit from distributive proce ⁇ sing by running on a processor 50 in the tracking device 14. The signal processor 134 may then interact with the control module 110, for example, by passing its data to the input interface module 116 for use by the application module 120 or application 122.
  • the signal generator 130 generates a signal corresponding to a response 136 by a user. For example, if a user moves a finger in a data glove, a displacement sensor 62 or position sensor 64 may detect the respon ⁇ e 136 of a u ⁇ er and generate a ⁇ ignal.
  • a velocity ⁇ en ⁇ or 68 or force sensor 76 may do likewise for a similar motion.
  • the temperature sensor 70 or humidity sensor 72 may detect a response 136 associated with increase body temperature or ⁇ weating.
  • the heart rate sen ⁇ or 74 and imaging ⁇ en ⁇ or 78 may return some signal corresponding to a response 136 by a user.
  • the tracking device 14 with its tracking module 112 may provide data to the controller 110 by which to determine inputs by the control module 110 to the sen ⁇ ory interface device 114.
  • An actuation module 114 run on the processor 80 of the sen ⁇ ory interface device 16 may include a driver 140, also referred to as a software driver, for providing suitable signal ⁇ to the actuator ⁇ 90.
  • the driver 140 may control one or more power supplies 142 for providing energy to the actuators 90.
  • the driver 140 may also provide actuation signals 144 directly to an actuator 90.
  • the driver 140 may provide a controlling instruction to a power supply 142 dedicated to an actuator 90, the power supply, thereby, providing an actuation signal 144.
  • the actuation signal 144 provided to the actuator 90 results in a stimulus signal 146 as an output of the actuator 90.
  • a stimulu ⁇ ⁇ ignal for an aural actuator 94 may be a ⁇ ound produced by a speaker.
  • a stimulu ⁇ ⁇ ignal from an optical actuator 96 may be a visual image on a screen for which an actuation signal is the digital data displaying a CRT image.
  • a stimulus signal for a force actuator 106 or a pressure actuator 104 may be a pressure exerted on the ⁇ kin of a user by the respective actuator 90.
  • a stimulus signal 146 may be a heat flow or temperature driven by a temperature actuator 100.
  • a stimulu ⁇ signal 146 of an electromuscular actuator 100 may actually be an electric voltage, or a specific current.
  • an electromuscular actuator 100 may use application of a voltage directly to each end of a muscle to cause a natural contraction, as if a nerve had commanded that muscle to move.
  • an electromuscular actuator 100 may include a power supply adapted to provide voltages to muscles of a user.
  • a plurality of stimulus signal ⁇ 146 may be available from one or more actuators 90 in response to the actuation signal ⁇ 144 provided by a driver 140 of the actuation module 114.
  • a set up database 150 may be created for containing data associated with each application 122. Multiple set up data bases 150.
  • An operational data base 152 may be set up to contain data that may be necessary and accessible to the controller 12, tracking device 14, sensory interface device 16 or another remote system 10.
  • the set up data base 150 and operational data base 152 may reside on the server 22.
  • certain data may be set up in a sensor table 156.
  • the sensor table 156 may contain data specific to one or more sensors 60 of the tracking device.
  • an actuator table 158 may contain the information for one or more actuators 90.
  • the sensor table 156 and the actuator table 158 may contain information for more than one sensor 60 or actuator 90, respectively, or may be produced in plural, each table 156, 158 corre ⁇ ponding to each ⁇ ensor 60 or actuator 90, respectively.
  • the tables 156, 158 may be used for interpolating and projecting expected inputs and output ⁇ related to sensors 60 and actuators 90 ⁇ o that a device communicating to or from such sensor 60 or actuator 90 may project an expected data value rather than waiting until the value is generated.
  • a predicted response may be programmed to be later corrected by actual data if the direction of movement of a signal changes.
  • the speed of re ⁇ pon ⁇ e of a ⁇ y ⁇ tem 10 may be increased.
  • a linking index 154 may exchange data with a plurality of operational data base ⁇ 152 or with an operational data base and a sensor table 156 or actuator table 158.
  • a high speed indexing linkage may be provided by a linking index 154 or a plurality of linking indices 154 rather than slow-speed searching of an operational data base 152 for specific information needed by a device within the sy ⁇ tem 10.
  • a remote apparatus 11 may be connected through the network 20 or through an internetwork 25 connected to the router 24.
  • the remote system 11 may include one or more corresponding data structures.
  • the remote system 11 may have a corresponding remote set up data base 160, remote operational data bases 162, remote linking data bases 164, remote sensor tables 166, and remote actuator tables 168.
  • interfacing indices may be set up to operate similar to the linking indices 154, 164.
  • a controller 12 may have an interface index 170 for providing high speed indexing of data that may be made rapidly accessible, to eliminate the need to continually update data, or search data in the systems 10, 11.
  • interpolation, projection, and ⁇ imilar technique ⁇ may be u ⁇ ed a ⁇ well as high speed indexing for accessing the needed information in the remote system 11, by a controller 12 having access to an interfacing index 170.
  • An interfacing index 170 may be hosted on both the ⁇ erver 22 and a server a ⁇ sociated with the remote sy ⁇ tem 11.
  • Figure 5 illu ⁇ trate ⁇ one embodiment of an apparatus made in accordance with the invention to include a controller 12 operably connected to a tracking device 14 and a sen ⁇ ory interface device 16 to augment the experience and exerci ⁇ e of a u ⁇ er riding a bicycle.
  • the apparatu ⁇ may include a loading mechani ⁇ m 202 for acting on a wheel 204 of a bicycle 205
  • a ⁇ en ⁇ ing member 208 may be in ⁇ trumented by a wheel and associated dynamometer, or the like, as part of an instrumentation suite 210 for tracking ⁇ peed, energy u ⁇ age, acceleration, and other dynamic ⁇ a ⁇ sociated with the motion of the wheel 204.
  • loads exerted by a user on pedals of the bicycle 205 may be ⁇ ensed by a load transducer 206 connected to the instrumentation suite 210 for transmitting signal ⁇ from the sensors 60 to the tracking device 14.
  • an instrumentation suite 210 may include or connect to any of the sensor ⁇ 60.
  • the in ⁇ trumentation suite 210 may transmit to the tracking device 14 tracking data corresponding to the motion of the sen ⁇ ing member 208.
  • a pickup 212 ⁇ uch a ⁇ may emit or radiate a signal in a frequency range selected, for example, from radio, light, sound, or ultrasound spectra.
  • the signal may be reflected to the pickup 212 by a target 214 attached to a bodily member of a user for detecting position, speed, acceleration, direction, and the like.
  • Other sen ⁇ ors 60 may be similarly positioned to detect desired feedback parameters.
  • a resistance member 216 may be positioned to load the wheel 204 according to a driver 218 connected to the sensory interface device 16.
  • actuators 90 may be configured as resi ⁇ tance members to resist motion by other bodily members of a user, either directly or by resisting motion of mechanical members movable by a user.
  • the resistance member 216 as many actuators 90, devices for providing stimuli, may be controlled by a combination of one or more inputs. Such inputs may be provided by pre-inputs, programmed instruction ⁇ or controlling data pre-programmed into ⁇ etup databa ⁇ es 150, 160, actuator tables 158, 168 or operational databases 152, 162. Inputs may also be provided by user- determined data stored in the actuator table ⁇ 158, 168 or operational databa ⁇ e ⁇ 152, 162.
  • Input ⁇ may al ⁇ o be provided by data corresponding to signal ⁇ collected from the sensors 60 and stored by the tracking device 14 or controller 12 in the sensor tables 156, 166, actuator table ⁇ 158, 168 or operational databa ⁇ es 152, 162.
  • the display 230 may be selected from a goggle apparatus for fitting over the eyes of a user to display an image in one, two, or three dimensions.
  • the display 230 may be a flat panel display, a cathode ray tube (CRT) , or other device for displaying an image.
  • the display 230 may include a "fly's eye" type of mosaic.
  • a wall, several walls, all walls, or the like may be set up to create a room or other chamber.
  • the chamber may be equipped with any number of display devices, such as, for example, television monitors, placed side-by- ⁇ ide and one above another to create a mo ⁇ aic.
  • a u ⁇ er may have the impre ⁇ ion of ⁇ itting in an environment looking out a paned window on the world in all dimensions.
  • images may be displayed on a single monitor of the display 230, or may be displayed on several monitors.
  • a tree, a landscape ⁇ cene at a di ⁇ tance, or the like may u ⁇ e multiple monitor ⁇ to be ⁇ hown in full ⁇ ize a ⁇ envisioned by a user in an environment.
  • a display 230 may be ⁇ elected to include goggle ⁇ like apparatu ⁇ surrounding the eyes and showing up to three dimensions of vision.
  • any number of image presentation monitors may be placed away from the user within a chamber.
  • the di ⁇ play 230 may be controlled by hard wire connection ⁇ or wirele ⁇ connections from a transceiver 219.
  • the transceiver 219 may provide for wireless communication with sen ⁇ ory interface device ⁇ 16, tracking device ⁇ 14, ⁇ ensors 60, or actuator ⁇ 90.
  • the tran ⁇ ceiver 219 may communicate with an activation center 220 to modify or control voltage ⁇ , current ⁇ , or both delivered by electrodes 222, 224 attached to stimulate action by a muscle of the user.
  • Each pair of electrodes 222, 224 may be controlled by a combination of open loop control (e.g. inputs from a pre-programmed code or data), man-in-the-loop control, (e.g. inputs from a user input into the controller 12 by way of the programming interface module 124), feedback control (e.g. inputs from the tracking system 14 to the controller 12) , or any combination selected to optimize the experience, exercise, or training desired.
  • This combination of inputs for control of actuators 90 also may be used to protect a user.
  • the controller 12 may override pre-programmed inputs from a user or other source stored in database ⁇ 150, 152 and table ⁇ 156, 158 or inherent in software modules 110, 112, 114 and the like. That is, the feedback corresponding to the condition of a user as detected by the sensors 60, may be u ⁇ ed to adju ⁇ t exertion and protect a u ⁇ er.
  • the activation center 220 may control other similarly placed pairs of electrodes 226, 228. If wires are used, certain bandwidth limitations may be relaxed, but each sensor 60, actuator 90, or other device may have a proces ⁇ or and memory organic or inherent to itself. Thus, all data that is not likely to change rapidly may be downloaded, including applications, and session data to a lowest level of use. In many cases data may be stored in the controller 12. Session data may be information corresponding to positions, motion, condition, and so forth of an opponent.
  • the session data in the databa ⁇ e ⁇ 160, 162 and tables 166, 168 may be provided to the user and controller 12 as ⁇ ociated with the databases 150, 152 and tables 156, 158 for use during a contest, competition, or the like.
  • the necessary data traffic passed through the transceiver 219 of each of two or more remotely interacting participants may be minimized to improve real time performance of the system 10, and the wireless communications of the transceiver.
  • An environmental suit 232 may provide heating or cooling to create an environment, or to protect a user from the effects of exertion. Actuation of the ⁇ uit 232 may be provided by the ⁇ en ⁇ ory interface device 16 through hard connections or wireles ⁇ ly through the tran ⁇ ceiver 219. Thus, for example, a user cycling indoors may obtain needed additional body cooling to facilitate personal performance ⁇ imilar to that available on an open road at 30 mile-per- hour ⁇ peeds.
  • the environment suit may also be provided with other sensors 60 and actuators 90.
  • An apparatus in accordance with the invention may be used to create a duplicated reality, rather than a virtual reality. That is, two remote users may experience interaction based upon tracking of the activities of each. Thus, the apparatus 10 may track the movements of a first user and transmit to a second user sufficient data to provide an interactive environment for the second user. Meanwhile, another apparatu ⁇ 10 may do the equivalent ⁇ ervice for certain activitie ⁇ of the second user. Feedback on each user may be provided to the other user. Thus, rather than a synthesized environment, a real environment may be properly duplicated.
  • two users may engage in mutual combat in the martial arts.
  • Each user may be faced with an opponent represented by an image moving through the motions of the opponent.
  • the opponent meanwhile, may be tracked by an apparatus 10 in order to provide the information for creating the image to be viewed by the user.
  • an apparatu ⁇ 10 made in accordance with the invention for example, two competitors may run a bicycle course that is a camera-digitized, actual course. Each competitor may experience re ⁇ i ⁇ tance to motion, apparent wind ⁇ peed, and orientation of a bicycle determined by actual condition ⁇ on an actual cour ⁇ e. Thu ⁇ , a duplicated reality may be presented to each user, based on the actual reality experienced by the other user. Effectively, a hybrid actual/duplicate reality exists for each user.
  • Two users may compete on a course not experienced by either. Each may experience the sensations of speed, grade, resistance, and external environment. Each sensation may be exactly as though the user were positioned on the course moving at the user's developed rate of speed. Each user may see the ⁇ urrounding country ⁇ ide pass by at the appropriate speed.
  • the two racers could be removed great distances from one another, and yet compete on the course, each seeing the image of the competitor.
  • the opposing competitor's location, relative to the speed of each user, may be reflected by each re ⁇ pective image of the cour ⁇ e di ⁇ played to the users.
  • Electromuscular stimulation apparatus 100 may be worn to assist a user to exercise at a speed, or at an exertion level above that normally experienced.
  • the EMS may be worn to ensure that muscles do experience total exertion in a limited time.
  • a user may obtain a one hour workout from 30 minutes of activity.
  • one competitor may be handicapped. That is one user may receive greater exertion, a more difficult workout, against a les ⁇ er opponent, without being credited with the exertion by the system.
  • a cyclist may have to exert, for example, ten percent more energy that would actually be required by an actual cour ⁇ e. The motivation of having a competitor close by could then remain, while the better competitor would receive a more appropriate workout.
  • Speed, energy, and so forth may also be similarly handicapped for martial arts contestants in the above example.
  • a skilled mechanic may direct another mechanic at a remote location.
  • a ⁇ killed mechanic may better recognize the nature of an environment or a machine, or may simply not be available to travel to numerous location ⁇ in real time.
  • a principal mechanic on a site may be equipped with cameras.
  • a subject machine may be instrumented.
  • a consulting mechanic located a di ⁇ tance away from the principal mechanic may be readily provided in real time.
  • Data may be tran ⁇ mitted dynamically as the machine or equipment operates.
  • a location or velocity in space may be represented by an image, based upon tracking information provided from the actual device at a remote location.
  • one physical object may be positioned in space relative to another physical object, although one of the objects may be a re-creation or duplication of its real object at a remote location.
  • synthe ⁇ i ⁇ a creation of an imaginary environment by u ⁇ e of computed image ⁇
  • an environment is duplicated (represented by the best available data to duplicate an actual but remote environment) .
  • One advantage of a duplicated environment rather than a synthesized environment is that certain information may be provided in advance to an apparatus 10 controlled by a u ⁇ er. Some le ⁇ er, required amount of necessary operational data may be passed from a remote site.
  • a machine for example, may be represented by images and operational data downloaded into a file stored on a user's computer.
  • the user's computer may provide most of the information needed to re-create an image of the di ⁇ tant machinery.
  • the actual speeds, positioning, and the like, corresponding to the machine may be provided with a limited amount of required data. Such operation may require less data and a far lower bandwidth for transmission.
  • the invention may include a presentation of multiple stimuli to a user, the stimuli including an image pre ⁇ ented vi ⁇ ually.
  • the apparatu ⁇ 10 may then include control of actuator ⁇ 90 by a combination of pre-input ⁇ provided a ⁇ an open loop control contribution by an application, data file, hardware module, or the like. Thu ⁇ , pre-input ⁇ may include open-loop control ⁇ and commands.
  • user-selected inputs may be provided.
  • a user may ⁇ elect option ⁇ or set up a se ⁇ ion through a programming interface module 124.
  • a u ⁇ er may interact with another input device connected to provide input ⁇ through the input module 116.
  • the apparatu ⁇ 10 may obtain a performance of the ⁇ ystem 10 in accordance with the user-selected inputs. Thu ⁇ , a "man-in-the-loop" may exert a certain amount of control.
  • the ⁇ ensors 60 of the tracker device 14 may provide feedback from a user.
  • the feedback in combination with the user-selected data and the pre-inputs, may control actuator ⁇ 90 of the sensory interface device 16.
  • the apparatus 10 may provide stimuli to a user at an appropriate level based on all three different types of inputs.
  • the condition of a user as indicated by feedback from a sensor 60 may be programmed to override a pre-input from the controller 12, or an input from a user through the programming interface module 124.

Abstract

An electronic exercise enhancer apparatus (10) and method for providing stimuli to a user while sensing the performance and condition of the user may rely on a controller (12) for programmably coordinating a tracking device (14) and a sensory interface device (16). The tracking device (14) may be equipped with sensors (60) for sensing position, displacement, motion, deflection, velocity, speed, temperature, humidity, heart rate, internal or external images, and the like. The sensory interface device (16) may produce outputs presented as stimuli to a user. The sensory interface device (16) may include one or more actuators (90) for providing aural, optical, tactile, and electromuscular stimulation to a user. The controller (12), tracking device (14), and sensory interface device (16) may all be microprocessor controlled for providing coordinated sensory perceptions of complex events.

Description

ELECTRONIC EXERCISE ENHANCER
DESCRIPTION
Technical Field
This invention relates to exercise equipment and, more particularly, to novel systems and methods for enhancing exercises by providing to a user multiple stimuli and by tracking multiple responses of a user, all with programmable electronic control.
Background Art
Exercise continues to be problematic for persons having limited time and limited access to outdoor recreational facilities or large indoor recreational facilities. Meanwhile, more, and more realistic, simulated, training environments are needed for lower cost instruction and practice. For example, flight training requires a very expensive aircraft. Nuclear plant control requires a complex system of hardware and software. Combat vehicle training, especially large force maneuvers, requires numerous combat vehicles and supporting equipment. Personal fitness may require numerous machines of substantial size and sophistication placed in a large gym to train athletes in skill or strength, especially if all muscle groups are to be involved. In short, training with real equipment may require substantial real estate and equipment, with commensurate cost.
Many activities may by taught, practiced and tested in a simulated environment. However, simulated environments often lack many or even most of the realistic stimuli received by a user in the real world including motions over distance, forces, pressures, sensations, temperatures, images, multiple views in the three-dimensions surrounding a user, and so forth. Moreover, many simulations do not provide the proper activities for a user, including a full range of motions, forces, timing, reflexes, speeds, and the like.
What is needed is a system for providing to a user more of the benefits of a real environment in a virtual environment. Also needed is a system for providing coordinated, synchronized, sensory stimulation by multiple devices to more nearly simulate a real three-dimensional spatial environment. Similarly needed is an apparatus and method for tracking a plurality of sensors monitoring a user's performance, integrating the inputs provided by such tracking, and providing a virtual environment simulating time, space, motion, images, forces and the like for the training, conditioning, and experience of a user.
Likewise needed is more complete feedback of a user's condition and responses. Such feedback to a controller capable of changing the stimuli and requirements (such as images, electromuscular and audio stimulation, loads and other resistance to movement, for example) imposed on a user is needed to make training and exercise approach the theoretical limits of comfort, endurance, or optimized improvement, as desired. Moreover, a system is needed for providing either a choice or a combination of user control, selectable but pre-programmed (template-like or open loop) control, and adaptive (according to a user's condition, comfort, or the like) control of muscle and sensory stimulation, resistances, forces, and other actuation imposed on a user by the system, according to a user's needs or preferences.
Disclosure of Invention
An electronically controlled exercise enhancer is disclosed in one embodiment of the present invention as including an apparatus having a controller with an associated processor for controlling stimuli delivered to a user and for receiving feedback corresponding to responses of a user. A tracking device may be associated with the controller to communicate with the controller for tracking responses of a user and for providing to the controller certain data corresponding to the condition, exertion, position, and other characteristics of a user.
The tracking device may also include a processor for processing signals provided by a plurality of sensors and sending corresponding data to the controller. The plurality of sensors deployed to detect the performance of a user may include, for example, a radar device for detecting position, velocity, motion, or speed; a pressure transducer for detecting stress; strain gauges for detecting forces, motion, or strain in a member of the apparatus associated with performance of a user. Such performance may include strength, force applied to the member, deflection, and the like. Other sensors may include humidity sensors; temperature sensors; calorimeters for detecting energy dissipation, either by rate or integrated over time; a heart rate sensor for detecting pulse; and an imaging device. The imaging device may provide for detecting the position, velocity, or condition of a member. Imaging may also assess a condition of a plane, volume, or an internal or external surface of a bodily member of a user.
One or more sensors may be connected to provide analog or digital signals to the tracking device for processing. The tracking device may then transfer corresponding digital data to the controller. In one embodiment, the controller may do all signal processing, whereas in other embodiments, distributed processing may be relied upon in the tracker, or even in individual sensors to minimize the bandwidth required for the exchange of data between devices in the apparatus.
A stimulus interface device may be associated with the controller for delivering selected stimuli to a user. The stimulus interface device may include a processor for controlling one or more actuators (alternatively called output devices) for providing stimulus to a user. Alternatively, certain actuators may also contain processors for certain functions, thus reducing the bandwidth required for communications between the controller and the output devices. Alternatively, for certain embodiments where processing capacity in and communications capacity from the controller are adequate, the controller may provide processing for data associated with certain actuators.
Actuators for the sensory interface device may include aural actuators for presenting sounds to a user, such as speakers, sound synthesizers with speakers, compact disks and players associated with speakers for presenting aural stimuli, or electrodes for providing electrical impulses associated with sound directly to a user.
Optical actuators may include cathode ray tubes displaying images in black and white or color, flat panel displays, imaging goggles, or electrodes for direct electrical stimulus delivered to nerves or tissues of a user. Views presented to a user may be identical for both eyes of a user, or may be stereoscopic to show the two views resulting from the parallax of the eyes, thus providing true three-dimensional images to a user.
In certain embodiments, the actuators may include temperature actuators for providing temperature or heat transfer. For example working fluids warmed or cooled to provide heat transfer, thermionic devices for heating and cooling an junction of a bimetallic probe, and the like may be used to provide thermal stimulus to a user.
Kinematic actuators may provide movement in one or more degrees of freedom, including translation and rotation with respect to each of the three spatial axes. Moreover, the kinematic actuators may provide a stimulus corresponding to motion, speed, force, pressure or the like. The kinematic actuators may be part of a suite of tactile actuators for replicating or synthesizing stimuli corresponding to each tactile sensation associated with humans' sense or touch of feel. In general a suite of tactile, optical, and aural, and even olfactory and taste actuators may replicate virtually any sensible output for creating a corresponding sensation by a user. Thus, the tracking device may be equipped with sensors for sensing position, displacement, motion, deflection, velocity, speed, temperature, pH, humidity, heart rate, images, and the like for accumulating data. Data may correspond to the biological condition and spatial kinematics (position, velocity, forces) of a bodily member of a user. For example, skin tension, pressure, forces in any spatial degree of freedom and the like may be monitored and fed back to the controller.
The sensory interface device may produce outputs presented as stimuli to a user. The sensory interface device may include one or more actuators for providing aural, optical, tactile, and electromuscular stimulation to a user. The controller, tracking device, and sensory interface device may all be microprocessor controlled for providing coordinated sensory perceptions of complex events. For example, actuators may represent a coordinated suite of stimuli corresponding to the sensations experienced by a user. For example, a user may experience a panoply of sensory perceptions besides sight.
For example, sensations may replicate, from synthesized or sampled data, a cycling tour through varied terrain and vegetation, a rocket launch, a tail spin in an aircraft, a flight by aircraft including takeoff and landing. Sensations may be presented for maneuvers such as aerobatics. A combat engagement may be experienced from within a combat vehicle or simulator. Sensory inputs may include those typical of a turret with slewing control and mounting weaponry with full fire control. Besides motion, sensory inputs may include hits received or made. Sensations may imitate or replicate target acquisition, tracking, and sensing or the like. Moreover, hand-to-hand combat with a remote user operating a similar apparatus may be simulated by the actuators. Sensors may feed back data to the controller for forwarding to the system of the remote user, corresponding to all the necessary actions, condition, and responses of the user.
Similarly, a mountain hike, a street patrol by police, a police fire fight, an old west gunfight, a mad scramble over rooftops, through tunnels, down cliffs, and the like may all be simulated with properly configured and powered actuators and sensors.
Stimuli provided to a user may be provided in a variety of forms, including electromuscular stimulation. Stimuli may by timed by a predetermined timing frequency set according to a pre-programmed regimen set by a user or a trainer as an input to an executable code of a controller.
Alternatively, stimuli may be provided with interactively determined timing. Interactively determined timing for electromuscular stimulation means that impulses may be timed and scaled in voltage, frequency, and other parameters according to a user's performance.
For example, detection is possible for the motion, speed, position, muscular or joint extension, muscle tension or loading, surface pressure, or the like. Such detection may occur for many body members. Members may include a user's foot, arm, or other bodily member.
Sensed inputs may be sensed and used in connection with other factors to control the timing and effect of electromuscular stimulation. The electromuscular stimulation may be employed to enhance the contraction or extension of muscles beyond the degree of physiological stimulation inherent in the user. Moreover, sensory impact may be provided by actuators electrically stimulating muscles or muscle groups to simulate forces imposed on bodily members by outside influences. Thus, a virtual baseball may effectively strike a user. A martial arts player may strike another from a remote location by electromuscular stimulation.
That iε, in general, two contestants may interact although physically separated by some distance. Thus two contestants may engage in a boxing or martial arts game or contest in which a hit by one contestant faced with a virtual opponent is felt by the opponent. For example, sensory inputs may be provided based on each remote opponents actual movements. Thus impacts may be literally felt by each opponent at the remote location. Likewise, responses of each opponent may be presented as stimuli to each opponent (user) .
Brief Description of Drawings The objects and features of the present invention will become more fully apparent from the following description and appended claims, taken in conjunction with the accompanying drawings. Understanding that these drawings depict only typical embodiments of the invention and are, therefore, not to be considered limiting of its scope, the invention will be described with additional specificity and detail through use of the accompanying drawings in which:
Figure l is a schematic block diagram of an apparatus made in accordance with the invention; Figures 2-3 are schematic block diagrams of software modules for programmable operation of the apparatus of Figure 1.
Figure 4 is a schematic block diagram of one embodiment of the data structures associated with the apparatus of Figure 1 and the software modules of Figures 2-3.
Figure 5 is a schematic block diagram of one embodiment of the apparatus of Figure 1 adapted to tracking and actuation, including electromuscular stimulation, of a user of a stationary bicycle exerciser. Best Modes for Carrying Out the Invention
It will be readily understood that the components of the present invention, as generally described and illustrated in the Figures herein, could be arranged and designed in a wide variety of different configurations. Thus, the following more detailed description of the embodiments of the system and method of the present invention, as represented in Figures 1 through 5, is not intended to limit the scope of the invention, as claimed, but it is merely representative of the presently preferred embodiments of the invention.
The presently preferred embodiments of the invention will be best understood by reference to the drawings, wherein like parts are designated by like numerals throughout.
Figure 1 illustrates one presently preferred embodiment of a controller for programmably directing the operation of an apparatus made in accordance with the present invention, a tracking device for sensing and feeding back to the controller the condition and responses of a user, and a sensory interface device for providing stimuli to a user through one or more actuators.
Reference is next made to Figure 2, which illustrates in more detail a schematic diagram of one preferred embodiment of software programming modules for the tracking device with its associated sensors, and for the sensory interface device with its associated actuators for providing stimuli to a user. Figure 3 illustrates in more detail a schematic diagram of one preferred embodiment of software modules for programming the controller of Figure 1. Figure 4 illustrates a schematic block diagram of one embodiment of data structures for storing, retrieving and managing data used and produced by the apparatus of Figure 1. Those of ordinary skill in the art will, of course, appreciate that various modifications to the detailed schematic diagrams of Figures 1-4 may easily be made without departing from the essential characteristics of the invention, as described in connection with the block diagram of Figure 1 above. Thus, the following description of the detailed schematic diagrams of Figures 2-5 is intended only as an example, and it simply illustrates one presently preferred embodiment of an apparatus and method consistent with the foregoing description of Figure 1 and the invention as claimed herein.
From the above discussion, it will be appreciated that the present invention provides an apparatus for presenting one or more selected stimuli to a user, feeding back to a controller the responses of a user, and processing the feedback to provide a new set of stimuli.
Referring now to Figure 1, the apparatus 10 made in accordance with the invention may include a controller 12 for exercising overall control over the apparatus 10 or system 10 of the invention. The controller 12 may be connected to communicate with a tracking device 14 for feeding back data corresponding to performance of a user. The controller 12 may also connect to exchange data with a sensory interface device 16.
The sensory interface device 16, may include one or more mechanisms for presenting sensory stimuli to a user. The controller 12, tracking device 14 and interface device 16 may be connected by a link 18, which may include a hardware connection and software protocols such as the general purpose interface bus (GPIB) as described in the IEEE 488 standard, and commonly used as a computer bus.
Alternatively, the link 18 may be selected from a universal ace synchronous receiver-transmitter. Since such a system may include a module composed of a single integrated circuit for both receiving and transmitting, asynchronously through a serial communications port, this type of link 18 may be simple, reliable, and inexpensive. Alternatively, a universal synchronous receiver-transmitter (USRT) module may be used for communication over a pair of serial channels. Although slightly more complex, such a link 18 may be used to pass more data.
Another alternative, for a link 18 is a network 20, such as a local area network. If the controller 12, tracking device 14 and sensory interface device 16 are each provided with some processor, then each may be a node on the network 20. Thus, a server 22 may be connected to the network 20 for providing data storage, and general file access for any processor in the system 10. A router 24 may also be connected to the network 20 for providing access to a larger internetwork, such as the worldwide web or internet. The operation of servers 22 and routers 24 reduce the duty required of the controller 12, and may also permit interaction between multiple controllers 12 separated across internetworks. For use of an apparatus 10 in an interactive mode, wherein interactive means interaction between users remotely spaced from one another, an individual user might have a substantially easier task trying to find a similarly situated partner for interactive games. Moreover, real-time interaction, training, and teaming between users located at great distances may be accomplished using the system 10.
The network interface cards 26A, 26B, 26C, 26D, 26E, may be installed in the controller 12, tracking device 14, sensory interface device 16, server 22, and router 24, respectively, for meeting the hardware and software conventions and protocols of the network 20.
The controller 12 may include a processor 30 connected to operate with a memory device 32. Typically, a memory device 32 may be a random access memory or other volatile memory used during operation of the processor 30. Long term memory of software, data, and the like, may be accommodated by a storage device 34 connected to communicate with the processor 30. The storage device 34 may be a floppy disk drive, a random access memory, but may in one preferred embodiment of the system 10 include one or more hard drives. The - ii - storage device 34 may store applications, data bases, and various files needed by the processor 30 during operation of the system 10. The storage device 34 may download from the server 22 according to the needs of the controller 12 in any particular specific task, game, training session, or the like.
An input device 36 may be connected to communicate with a processor 30. For example, a user may program a processor 30 by creating an application to be stored in the storage device 34 and run on the processor 30. An input device 36, therefore, may be a keyboard. Alternatively, the input device 36 may be selected from a capacitor membrane keypad, a graphical user interface such as a monitor having menus and screens, or icons presented to a user for selection. An input device, may include a graphical pad and stylus for use by a user inputting a figure rather than text or ASCII characters.
Similarly, an output device 38 may be connected to the processor 30 for feeding back to a user certain information needed to control the controller 12 or processor 30. For example, a monitor may be a required output device 38 to operate with the menu and icons of an input device 36 hosted on the same monitor.
Also, an output device may include a speaker for producing a sound to indicate that an improper selection, or programming error has been committed by a user operating the input device 36 to program the processor 30. Numerous input device 36 and output devices 38 for interacting with the processor 30 of the controller 12 are available, and within contemplation of the invention.
The processor 30, memory device 32, storage device 34, input device 36, and output device 38 may all be connected by a bus 40. The bus may be of any suitable type such as those used in personal computers or other general purpose digital computers. The bus may also be connected to a serial port 42 and a parallel port 44 for communicating with other peripheral devices selected by a user. For example, a parallel port 44 may connect to an additional storage device, a slaved computer, a master computer, or a host of other peripheral devices.
In addition, a removable media device 46 may be connected to the bus 40. Alternatively, a removable media device such as a floppy disk drive, a Bernoulli™ drive, an optical drive, a compact disk laser readable drive, or the like could be connected to the bus 40 or to one of the ports 42, 44. Thus, a user could import directly a software program to be loaded into the storage device 34, for later operation on the processor 30.
In one embodiment, the tracking device 14 and the sensory interface device 16 may be "dumb" apparatus. That is, the tracking device 14 and sensory interface device 16 might have no processors contained within their hardware suites. Thus, the processor 30 of the controller 12 may do all processing of data exchanged by the tracking device, sensory interface device, and controller 12. However, to minimize the required bandwidths of communication lines such as the link 18, the network 20, the bus 40, and so forth, processors may be located in virtually any hardware apparatus.
The tracking device 14, in one embodiment, for example, may include a processor 50 for performing necessary data manipulation within the tracking device 14. The processor 50 may be connected to a memory device 52 by a bus 54. As in the controller 12, the tracking device may also include a storage device 56, although a storage device 56 may typically increase the size of the tracking device 14 to an undesirable degree for certain utilities.
The tracking device 14 may include a signal converter 58 for interfacing with a suite including one or more sensors 60. For example, the signal converter 58 may be an analog to digital converter, required by certain types of sensors 60. Signal processing may be provided by the processor 50. Nevertheless, certain types of sensors 60 may include a signal processor and signal converter organically included within the packaging of the sensor 60. The sensors 60 may gather information in the form of signals sensed from the activities of the user. The sensors 60 may include a displacement sensor 62 for detecting a change of position in 1, 2, or 3 spacial dimensions. The displacement sensor 62 may be thought of as a sensor of relative position between a firεt location and a second location.
Alternatively, or in addition, a position senεor 64 may be provided to detect an absolute position in space. For example, a displacement sensor 62 might detect the position or movement of a member of a user's body with respect to a constant frame of reference, whereas a displacement sensor 62 might simply detect motion between a first stop location and a second stop location, the starting location being reset every time the movement stops. Each type of sensor 62, 64 may have certain advantages.
A calibrator 66 may be provided for each sensor, or for all the sensors, depending on which types of sensors 60 are used. The calibrator may be used to null the signals from sensors 60 at the beginning of use to asεure that biases and drifting do not thwart the function of the system 10. Other sensors 60 may include a velocity sensor 68 for detecting either relative speed, a directionless scalar quantity, or a velocity vector including both speed and direction. In reality, a velocity sensor 68 may be configured as a combination of a displacement sensor 62 or position sensor 64 and a clock for corresponding a position to a time.
A temperature sensor 70 may be provided, and relative temperatures may also be measured. For example, a temperature-sensing thermocouple may be placed against the skin of a user, or in the air surrounding a user's hand. Thus, temperature may be sensed electronically by temperature sensors 70. In certain circumstances, relative humidity surrounding a user may be of importance, and may be detected by a humidity sensor 72. During exercise, and also various training, rehabilitation, and conceivably in certain high-stress virtual reality games, a heart rate sensor 74 may be included in the suite of sensorε 60.
Force sensorε 76 may be of a force variety or of a pressure variety. That is, transducers exist to sense a total integrated force. Alternatively, transducerε alεo exist to detect a force per unit of area to which the force is applied, the classical definition of pressure. Thus, the force sensors 76 may include force and presεure monitoring.
With the advent of microwave imaging radar, ultraεound, magnetic reεonance imaging, and other non¬ invasive imaging technologies, an imaging sensor 78 may be included as a sensor 60. Imaging senεorε may have a processor or multiple processors organic or integrated within themselves to manage the massive amounts of data received. An imaging sensor may provide certain position data through image processing. However, the position sensor 64 or displacement sensor 62 may be a radar, such as a Doppler radar mechanism for detecting movement of a foot, leg, the rise and fall of a user's chest during breathing, or the like.
A radar system may use a target patch for reflecting its own signal from a surface, such as the εkin of a user, or the surface of a shoe, the pedal of a bicycle, or the like. A radar may require much lower bandwidths for communicating with the procesεor 50 or the controller 12 than may be required by an imaging sensor 78. Nevertheless, the application to which the apparatuε 10 is put may require either an imaging sensor 78 or a simple displacement sensor 62. In another example a linear variable displacement transducer is a common and simple device that has traditionally been used for relative displacement. Thus, one or more of the εensors 60 deεcribed above may be included in the tracking device 14 to monitor the activity and condition of a uεer of the system 10.
A sensory interface device 16 may include a procesεor 80 and a memory device 82 connected to a bus 84. A storage device 86 may be connected to the buε 84 in some configurations, but may be considered too large for highly portable εenεory interface devices 16. The senεory interface device 80 may include a power εupply 88, and may include more than one power εupply 88 either centrally located in the sensory interface device or distributed among the various actuators 90.
A power supply 88 may be one of several types. For example, a power supply may be an electrical power εupply. Alternatively, a power εupply may be a hydraulic power supply, a pneumatic power supply, a magnetic power supply, or a radio frequency power supply. Whereas, a sensor 60 may use a very small amount of power to detect a motion, an actuator 90 may provide a subεtantial amount of energy. The actuators 90 may particularly benefit from a calibrator 92. For example, an actuator which provides a εpecific displacement or motion εhould be calibrated to be εure that it does not move beyond a desired position, since the result could be injury to a user. As with sensorε 60, the actuatorε may be calibrated by a calibrator 92 connected to null out any actuation of the actuator in an inactive, uncommanded mode.
In the one or more actuators 90 included in the sensory interface device 16, or connected as appendages thereto, may be an aural actuator 94. A simple aural actuator may be a εound εpeaker. Alternatively, an aural actuator 94 may include a εyntheεized εound generator aε well aε some speaker for projecting the sound. Thus, an aural actuator 94 may have within itself the ability to create sound on demand, and thus have its own internal processor, or it may simply duplicate an analog sound signal received from another source. One example of an aural actuator may be a compact disk player, power supply, and all peripheral devices required, with a simple control signal sent by the processor 80 to determine what soundε are presented to a user by the aural actuator 94. An optical actuator 96 may include a computer monitor that diεplays images much as a television screen does. Alternatively, an optical actuator may include a pair of goggles comprising a flat panel image display, a radar display, such as an oscilloεcopic catha-ray tube diεplaying a trace of signal, a fibre optic display of an actual image transmitted only by light, or a fibre optic display transmitting a εynthetically generated image from a computer or from a compact disk reader.
Thus, in general, the optical actuator may provide an optical stimuluε. In a medical application, aε compared to a training, or game environment, the optical actuator may actually include electrodes for providing stimulus to optical nerves, or directed to the brain. For example, in a virtual sight device, for use by a person having no natural sight, the optical actuator may be embodied in a sophiεticated computer-controlled series of electrodes producing voltages to be received by nerves in the human body.
By contrast, in a video game providing a virtual reality environment, a user may be surrounded by a mosaic of cathode ray tube type monitors or flat panel displays creating a scene to be viewed as if through a cockpit window or other position. Similarly, a user may wear a pair of stereo goggles, having two images corresponding to the parallax views presented to each eye by a three dimensional image.
Thus, a manner and mechanism may be εimilar to those by which stereo aerial photographs are used. Thus a user may be shown multi-dimenεional geographical featureε, stereo views of recorded images. Images may be generated or stored by either analog recording devices such as films. Likewiεe, images may be handled by digital devices such as compact diskε and computer magnetic memorieε. Images may be used to provide to a user in a very close environment, stereo views appearing to be three dimensional imageε. For example, εtereo views may be displayed digitally in the two "lens" displays of goggles adapted for such use.
In addition, such devices aε infrared imaging goggles, or digitized imageε originally produced by infrared imaging goggleε, may be provided. Any of these optical actuators 96 may be adapted for use with the sensory interface device 16.
A tactile actuator 98 may be included for providing to a user a sense of touch. Moreover, an electromuscular actuator 100 may be a part of, or connected to, the senεory interface device 16 for permitting a uεer to feel touched. In thiε regard, a temperature actuator 102 may preεent different temperatureε of contacting εurfaceε or fluidε againεt the skin of a user. The tactile actuator 98, electromuscular actuator 100, and temperature actuator 102 may interact with one another to produce a total tactile experience. Moreover, the electromuscular actuator 100 may be used to augment exerciεe, to give a εenεation of impact, or to give feedback to a proεthetic device worn by a uεer in medical rehabilitation.
Exampleε of tactile actuatorε may include a pressure actuator. For example, a panel, an arm, a probe, or a bladder, may have a surface that may be moved with respect to the skin of a uεer. Thus, a user may be moved, or presεured. For example, a uεer may wear a glove or a boot on a hand or foot, reεpectively, for εimulating certain activities. A bladder actuated by a pump, may be filled with air, water, or other working fluid to create a pressure. With a surface of the bladder against a retainer on one side, and the skin of a user on the other side, a user may be made to feel preεεure over a surface at a uniform level. Alternatively, a glove may have a series of articulated structural members, joints and connectors, actuated by hydraulic or pneumatic cylinders.
Thus, a user may be made to feel a force exerted against the inside of a user's palm or fingers in response to a grip. Thus, a user could be made to feel the grip of a machine by either a force, or a displacement of the articulated members. Conceivably, a user could arm wrestle a machine. Similarly, a user could arm wrestle a remote user, the pressure actuator 104, force actuator 106, or position actuator 108 inherent in a tactile actuator providing diεplacements and forces in response to the motion of a user. Each user, remote from each other, could nevertheless transfer motionε and forces digitally across the worldwide web between distant systems 10.
The temperature actuator may include a pump or fan for blowing air of a selected temperature over the skin of a uεer in a εuit adapted for such use. Alternatively, the temperature actuator may include a bladder touching the skin, the bladder being alternately filled with heated or cooled fluid, either air, water, or other working fluidε.
Alternatively, the temperature actuator 102 may be constructed using thermionic devices. For example, the principle of a thermocouple may be used. A voltage and power are applied to create heat or cooling at a bi¬ metallic junction.
These thermionic devices, by changing the polarity of the voltage applied, may be made to heat or cool electrically. Thus, a temperature actuator 102 may include a thermionic device contacting the εkin of a uεer, or providing a source of heat or cold for a working fluid to warm or cool the skin of a user in response to the processor 80.
Referring to Figures 2-4, similar to the distributed nature of hardware within the apparatus 10, software for programming, operation, and control, as well aε feedback may be diεtributed among componentε of the εyεtem 10. In general, in one embodiment of an apparatus in accordance with the invention, a control module 110 may be operable in the processor 30 of the controller 12.
Similarly, a tracking module 112 may run on a procesεor 50 of the tracking device 14. An actuation module 114 may include programmed instructions for running on a processor 80 of the sensory interface device 16.
The control module 110 may include an input interface module 116 including codes for prompting a user, receiving data, providing data prompts, and otherwise managing the data flow from the input device 36 to the procesεor 30 of the controller 12. Similarly, the output interface module 118 of the control module 110 may manage the interaction of the output device 38 with the proceεsor 30 of the controller 12. The input interface module 116 and output interface module 118, in one presently preferred embodiment, may exchange data with an application module 120 in the control module 110. The application module 120 may operate on the procesεor 30 of the controller 12 to load and run applicationε 122.
Each application 122 may correεpond to an individual sesεion by a uεer, a particular programmed set of instructions designed for a game, an exercise workout, a rehabilitative regimen, a training session, a training lesεon, or the like. Thuε, the application module 120 may coordinate the receipt of information from the input interface module 116, output interface module 118, and the application 122 actually running on the proceεεor 30.
Likewise, the application module 120 may be thought of as the highest level programming running on the processor
30. Thus, the application module 120 may exchange data with a programming interface module 124 for providing access and control by a user to the application module 120.
For example the programming interface module 124 may be used to control and transfer information provided through a keyboard connected to the controller 12. Similarly, the programming interface module may include software for downloading applications 122 to be run by the application module 120 on the procesεor 30 or to be stored in the storage device 34 for later running by the processor 30. The input interface module 116 may include programmed instructionε for controlling the transfer of information, for example, digital data, between the application module 120 of the control module 110 running on the processor 30, and the tracking device 14. Correspondingly, the output interface module 118 may include programmed instructions for transferring information between the application module 120 and the senεory interface device 16.
The input interface module 116 and output interface module 118 may deal exclusively with digital data files or data streamε paεεed between the tracking device 14 and the sensory interface device 16 in an embodiment where each of the tracking device 14 and sensory interface device 16 are themselves microprocessor controlled with microprocessors organic (integral) to the respective structures. The control module 10 may include an interaction module 128 for transferring data between control modules 110 of multiple, at least two, systems 10. Thus, within the controller 12, an interaction module 128 may contain programmed instructions for controlling data flow between an application module 120 in one location and an application module 120 of an entirely different system 10 at another location, thus facilitating a high level of coordination between applications 122 on different systemε 10. If a controller 12 operates on a network 20, or an internetwork beyond a router 24 connected to a local area network 20 of the controller 12, a network module 126 may contain programmed instructionε regarding logging on and off of the network, communication protocolε over the network, and the like. Thuε, the application module 120 may be regarded aε the heart of the software running on the controller 12, or more precisely, on the processor 30 of the controller 12. Meanwhile, the functionε associated with network accesε may be included in a network module 126, while certain interaction between cooperating systems 10 may be handled by an interaction module 128. Different taskε may be reaεεigned to different software moduleε, depending on hardware configurationε of a εpecific problem or εyεtem 10. Therefore, equivalent εyεtemε 10 may be configured according to the invention. For example, a single application 122 may include all of the functions of the modules 120-128.
In a controller 12, more than one procesεor 30 may be uεed. Likewise, a multi-tasking proceεεor may be uεed aε the proceεεor 30. Thuε, multiple proceεεeε, threadε, programε, or the like, may be made to operate on a variety of proceεεorε, a plurality of proceεεorε, or in a ulti- taεking arrangement on a multi-tasking procesεor 30. Nevertheleεε, at a high level, data may be transferred between a controller 12 and a tracking device 14, the senεory interface device 16, a keyboard, and monitor, a remote controller, and other nodeε on a network 20.
The tracking module 112 may include a εignal generator 130. In general, a εignal generator may be any of a variety of mechanisms operating within a sensor, to create a signal. The signal generator 130 may then pass a signal to a signal converter 132. For example, an analog to digital converter may be common in certain transducerε. In other sophisticated transducerε, a signal generator 130 may itself by microprocesεor-controlled, and may produce a data εtream needing no converεion by a signal converter 132. In general, a signal converter 132 may convert a signal from a signal generator 130 to a digital data signal that may be procesεed by a signal procesεor 134. A signal processor 134 may operate on the processor 30 of the controller 12, but may benefit from distributive proceεsing by running on a processor 50 in the tracking device 14. The signal processor 134 may then interact with the control module 110, for example, by passing its data to the input interface module 116 for use by the application module 120 or application 122.
The signal generator 130 generates a signal corresponding to a response 136 by a user. For example, if a user moves a finger in a data glove, a displacement sensor 62 or position sensor 64 may detect the responεe 136 of a uεer and generate a εignal.
Similarly, a velocity εenεor 68 or force sensor 76 may do likewise for a similar motion. The temperature sensor 70 or humidity sensor 72 may detect a response 136 associated with increase body temperature or εweating. Likewise, the heart rate senεor 74 and imaging εenεor 78 may return some signal corresponding to a response 136 by a user. Thus, the tracking device 14 with its tracking module 112 may provide data to the controller 110 by which to determine inputs by the control module 110 to the senεory interface device 114.
An actuation module 114 run on the processor 80 of the senεory interface device 16 may include a driver 140, also referred to as a software driver, for providing suitable signalε to the actuatorε 90. The driver 140 may control one or more power supplies 142 for providing energy to the actuators 90. The driver 140 may also provide actuation signals 144 directly to an actuator 90. Alternatively, the driver 140 may provide a controlling instruction to a power supply 142 dedicated to an actuator 90, the power supply, thereby, providing an actuation signal 144. The actuation signal 144 provided to the actuator 90 results in a stimulus signal 146 as an output of the actuator 90.
For example, a stimuluε εignal for an aural actuator 94 may be a εound produced by a speaker. A stimuluε εignal from an optical actuator 96 may be a visual image on a screen for which an actuation signal is the digital data displaying a CRT image.
Similarly, a stimulus signal for a force actuator 106 or a pressure actuator 104 may be a pressure exerted on the εkin of a user by the respective actuator 90. A stimulus signal 146 may be a heat flow or temperature driven by a temperature actuator 100. A stimuluε signal 146 of an electromuscular actuator 100 may actually be an electric voltage, or a specific current.
That is, an electromuscular actuator 100 may use application of a voltage directly to each end of a muscle to cause a natural contraction, as if a nerve had commanded that muscle to move. Thus, an electromuscular actuator 100 may include a power supply adapted to provide voltages to muscles of a user.
Thuε, a plurality of stimulus signalε 146 may be available from one or more actuators 90 in response to the actuation signalε 144 provided by a driver 140 of the actuation module 114.
Referring now to Figure 4, the data εtructureε for εtorage, retrieval, transfer, and processing of data associated with the system 10 may be configured in various ways. In one embodiment of an apparatus 10 made in accordance with the invention, a set up database 150 may be created for containing data associated with each application 122. Multiple set up data bases 150.
An operational data base 152 may be set up to contain data that may be necessary and accessible to the controller 12, tracking device 14, sensory interface device 16 or another remote system 10. The set up data base 150 and operational data base 152 may reside on the server 22.
To expedite the tranεfer of data and the rapid interaction between εyεtems 10 remote from one another, as well as between the tracking device 14, sensory interface device 16, and controller 12, certain data may be set up in a sensor table 156. The sensor table 156 may contain data specific to one or more sensors 60 of the tracking device.
Thus, the complete characterization of a sensor 60 may be placed in a sensor table 156 for rapid access and interpolation, during operation of the application 122. Similarly, an actuator table 158 may contain the information for one or more actuators 90. Thus, the sensor table 156 and the actuator table 158 may contain information for more than one sensor 60 or actuator 90, respectively, or may be produced in plural, each table 156, 158 correεponding to each εensor 60 or actuator 90, respectively.
In operation, the tables 156, 158 may be used for interpolating and projecting expected inputs and outputε related to sensors 60 and actuators 90 εo that a device communicating to or from such sensor 60 or actuator 90 may project an expected data value rather than waiting until the value is generated. Thus, a predicted response may be programmed to be later corrected by actual data if the direction of movement of a signal changes. Thus, the speed of reεponεe of a εyεtem 10 may be increased.
To asεist in speeding the transfer of information, the various methods of linking operational data baεes 152 may be provided. For example, a linking index 154 may exchange data with a plurality of operational data baseε 152 or with an operational data base and a sensor table 156 or actuator table 158. Thus, a high speed indexing linkage may be provided by a linking index 154 or a plurality of linking indices 154 rather than slow-speed searching of an operational data base 152 for specific information needed by a device within the syεtem 10.
A remote apparatus 11 may be connected through the network 20 or through an internetwork 25 connected to the router 24. The remote system 11 may include one or more corresponding data structures. For example, the remote system 11 may have a corresponding remote set up data base 160, remote operational data bases 162, remote linking data bases 164, remote sensor tables 166, and remote actuator tables 168. Moreover, interfacing indices may be set up to operate similar to the linking indices 154, 164. Thus, on the server 22, a controller 12 may have an interface index 170 for providing high speed indexing of data that may be made rapidly accessible, to eliminate the need to continually update data, or search data in the systems 10, 11. Thus, interpolation, projection, and εimilar techniqueε may be uεed aε well as high speed indexing for accessing the needed information in the remote system 11, by a controller 12 having access to an interfacing index 170. An interfacing index 170 may be hosted on both the εerver 22 and a server aεsociated with the remote syεtem 11.
Figure 5 illuεtrateε one embodiment of an apparatus made in accordance with the invention to include a controller 12 operably connected to a tracking device 14 and a senεory interface device 16 to augment the experience and exerciεe of a uεer riding a bicycle. The apparatuε may include a loading mechaniεm 202 for acting on a wheel 204 of a bicycle 205
For example a εenεing member 208 may be inεtrumented by a wheel and associated dynamometer, or the like, as part of an instrumentation suite 210 for tracking εpeed, energy uεage, acceleration, and other dynamicε aεsociated with the motion of the wheel 204. Similarly loads exerted by a user on pedals of the bicycle 205 may be εensed by a load transducer 206 connected to the instrumentation suite 210 for transmitting signalε from the sensors 60 to the tracking device 14. In general, an instrumentation suite 210 may include or connect to any of the sensorε 60. The inεtrumentation suite 210 may transmit to the tracking device 14 tracking data corresponding to the motion of the senεing member 208.
A pickup 212 εuch aε, for example, a radar tranεmitting and receiving unit, may emit or radiate a signal in a frequency range selected, for example, from radio, light, sound, or ultrasound spectra. The signal may be reflected to the pickup 212 by a target 214 attached to a bodily member of a user for detecting position, speed, acceleration, direction, and the like. Other senεors 60 may be similarly positioned to detect desired feedback parameters. A resistance member 216 may be positioned to load the wheel 204 according to a driver 218 connected to the sensory interface device 16. Other actuators 90 may be configured as resiεtance members to resist motion by other bodily members of a user, either directly or by resisting motion of mechanical members movable by a user. The resistance member 216, as many actuators 90, devices for providing stimuli, may be controlled by a combination of one or more inputs. Such inputs may be provided by pre-inputs, programmed instructionε or controlling data pre-programmed into εetup databaεes 150, 160, actuator tables 158, 168 or operational databases 152, 162. Inputs may also be provided by user- determined data stored in the actuator tableε 158, 168 or operational databaεeε 152, 162. Inputε may alεo be provided by data corresponding to signalε collected from the sensors 60 and stored by the tracking device 14 or controller 12 in the sensor tables 156, 166, actuator tableε 158, 168 or operational databaεes 152, 162. The display 230 may be selected from a goggle apparatus for fitting over the eyes of a user to display an image in one, two, or three dimensions. Alternatively, the display 230 may be a flat panel display, a cathode ray tube (CRT) , or other device for displaying an image. In other alternative embodiment of the invention, the display 230 may include a "fly's eye" type of mosaic. That is, a wall, several walls, all walls, or the like, may be set up to create a room or other chamber. The chamber may be equipped with any number of display devices, such as, for example, television monitors, placed side-by-εide and one above another to create a moεaic.
Thuε, a uεer may have the impreεεion of εitting in an environment looking out a paned window on the world in all dimensions. Thus, images may be displayed on a single monitor of the display 230, or may be displayed on several monitors. For example, a tree, a landscape εcene at a diεtance, or the like may uεe multiple monitorε to be εhown in full εize aε envisioned by a user in an environment.
Thus a display 230 may be εelected to include goggle¬ like apparatuε surrounding the eyes and showing up to three dimensions of vision. Alternatively, any number of image presentation monitors may be placed away from the user within a chamber.
The diεplay 230 may be controlled by hard wire connectionε or wireleεε connections from a transceiver 219. The transceiver 219 may provide for wireless communication with senεory interface deviceε 16, tracking deviceε 14, εensors 60, or actuatorε 90.
For example, the tranεceiver 219 may communicate with an activation center 220 to modify or control voltageε, currentε, or both delivered by electrodes 222, 224 attached to stimulate action by a muscle of the user. Each pair of electrodes 222, 224 may be controlled by a combination of open loop control (e.g. inputs from a pre-programmed code or data), man-in-the-loop control, (e.g. inputs from a user input into the controller 12 by way of the programming interface module 124), feedback control (e.g. inputs from the tracking system 14 to the controller 12) , or any combination selected to optimize the experience, exercise, or training desired. This combination of inputs for control of actuators 90 also may be used to protect a user. For example, the controller 12 may override pre-programmed inputs from a user or other source stored in databaseε 150, 152 and tableε 156, 158 or inherent in software modules 110, 112, 114 and the like. That is, the feedback corresponding to the condition of a user as detected by the sensors 60, may be uεed to adjuεt exertion and protect a uεer.
Likewiεe, the activation center 220 may control other similarly placed pairs of electrodes 226, 228. If wires are used, certain bandwidth limitations may be relaxed, but each sensor 60, actuator 90, or other device may have a procesεor and memory organic or inherent to itself. Thus, all data that is not likely to change rapidly may be downloaded, including applications, and session data to a lowest level of use. In many cases data may be stored in the controller 12. Session data may be information corresponding to positions, motion, condition, and so forth of an opponent. Thus, much of the session data in the databaεeε 160, 162 and tables 166, 168 may be provided to the user and controller 12 asεociated with the databases 150, 152 and tables 156, 158 for use during a contest, competition, or the like. Thus, the necessary data traffic passed through the transceiver 219 of each of two or more remotely interacting participants (contestantε, opponents, teammateε, etc.) may be minimized to improve real time performance of the system 10, and the wireless communications of the transceiver.
An environmental suit 232 may provide heating or cooling to create an environment, or to protect a user from the effects of exertion. Actuation of the εuit 232 may be provided by the εenεory interface device 16 through hard connections or wirelesεly through the tranεceiver 219. Thus, for example, a user cycling indoors may obtain needed additional body cooling to facilitate personal performance εimilar to that available on an open road at 30 mile-per- hour εpeeds. The environment suit may also be provided with other sensors 60 and actuators 90.
An apparatus in accordance with the invention may be used to create a duplicated reality, rather than a virtual reality. That is, two remote users may experience interaction based upon tracking of the activities of each. Thus, the apparatus 10 may track the movements of a first user and transmit to a second user sufficient data to provide an interactive environment for the second user. Meanwhile, another apparatuε 10 may do the equivalent εervice for certain activitieε of the second user. Feedback on each user may be provided to the other user. Thus, rather than a synthesized environment, a real environment may be properly duplicated.
For example, two users may engage in mutual combat in the martial arts. Each user may be faced with an opponent represented by an image moving through the motions of the opponent. The opponent, meanwhile, may be tracked by an apparatus 10 in order to provide the information for creating the image to be viewed by the user.
In one embodiment of an apparatuε 10 made in accordance with the invention, for example, two competitors may run a bicycle course that is a camera-digitized, actual course. Each competitor may experience reεiεtance to motion, apparent wind εpeed, and orientation of a bicycle determined by actual conditionε on an actual courεe. Thuε, a duplicated reality may be presented to each user, based on the actual reality experienced by the other user. Effectively, a hybrid actual/duplicate reality exists for each user.
Two users, in this example, may compete on a course not experienced by either. Each may experience the sensations of speed, grade, resistance, and external environment. Each sensation may be exactly as though the user were positioned on the course moving at the user's developed rate of speed. Each user may see the εurrounding countryεide pass by at the appropriate speed.
Moreover, the two racers could be removed great distances from one another, and yet compete on the course, each seeing the image of the competitor. The opposing competitor's location, relative to the speed of each user, may be reflected by each reεpective image of the courεe diεplayed to the users.
Electromuscular stimulation apparatus 100 may be worn to assist a user to exercise at a speed, or at an exertion level above that normally experienced. Alternatively, the EMS may be worn to ensure that muscles do experience total exertion in a limited time. Thus, for example, a user may obtain a one hour workout from 30 minutes of activity. Likewise, in the above examples of two competitors, one competitor may be handicapped. That is one user may receive greater exertion, a more difficult workout, against a lesεer opponent, without being credited with the exertion by the system. A cyclist may have to exert, for example, ten percent more energy that would actually be required by an actual courεe. The motivation of having a competitor close by could then remain, while the better competitor would receive a more appropriate workout. Speed, energy, and so forth may also be similarly handicapped for martial arts contestants in the above example.
In another example, a skilled mechanic may direct another mechanic at a remote location. Thus, for example, a εkilled mechanic may better recognize the nature of an environment or a machine, or may simply not be available to travel to numerous locationε in real time. Thus, a principal mechanic on a site may be equipped with cameras. Also, a subject machine may be instrumented.
Then, certain information needed by a consulting mechanic located a diεtance away from the principal mechanic may be readily provided in real time. Data may be tranεmitted dynamically as the machine or equipment operates. Thus, for example, a location or velocity in space may be represented by an image, based upon tracking information provided from the actual device at a remote location.
Thus, one physical object may be positioned in space relative to another physical object, although one of the objects may be a re-creation or duplication of its real object at a remote location. Rather than syntheεiε (a creation of an imaginary environment by uεe of computed imageε) , an environment is duplicated (represented by the best available data to duplicate an actual but remote environment) . One advantage of a duplicated environment rather than a synthesized environment is that certain information may be provided in advance to an apparatus 10 controlled by a uεer. Some leεεer, required amount of necessary operational data may be passed from a remote site. A machine, for example, may be represented by images and operational data downloaded into a file stored on a user's computer.
During operation of the machine, the user's computer may provide most of the information needed to re-create an image of the diεtant machinery. Nevertheleεε, the actual speeds, positioning, and the like, corresponding to the machine, may be provided with a limited amount of required data. Such operation may require less data and a far lower bandwidth for transmission.
In one embodiment, the invention may include a presentation of multiple stimuli to a user, the stimuli including an image preεented viεually. The apparatuε 10 may then include control of actuatorε 90 by a combination of pre-inputε provided aε an open loop control contribution by an application, data file, hardware module, or the like. Thuε, pre-inputε may include open-loop controlε and commands.
Similarly, user-selected inputs may be provided. A user, for example, may εelect optionε or set up a seεεion through a programming interface module 124. Alternatively, a uεer may interact with another input device connected to provide inputε through the input module 116. The apparatuε 10 may obtain a performance of the εystem 10 in accordance with the user-selected inputs. Thuε, a "man-in-the-loop" may exert a certain amount of control.
In addition to theεe control functionε, the εensors 60 of the tracker device 14 may provide feedback from a user. The feedback, in combination with the user-selected data and the pre-inputs, may control actuatorε 90 of the sensory interface device 16. The apparatus 10 may provide stimuli to a user at an appropriate level based on all three different types of inputs. The condition of a user as indicated by feedback from a sensor 60 may be programmed to override a pre-input from the controller 12, or an input from a user through the programming interface module 124.
The present invention may be embodied in other specific forms without departing from its spirit or essential characteristicε. The deεcribed embodiments are to be considered in all respects only as illustrative, and not restrictive. The scope of the invention is, therefore, indicated by the appended claims, rather than by the foregoing description. All changes that come within the meaning and range of equivalency of the claims are to be embraced within their scope.
What is claimed and desired to be secured by Letters Patent is:

Claims

1. An apparatus for training a user, the apparatus comprising: an actuation device for presenting to a user a stimuluε εensible by a user; a controller operably connected to the actuation device and comprising a processor for procesεing data, a memory device for storing data, an input device for receiving feedback data corresponding to a condition of a user, and an output device for sending control signalε for controlling the actuation device; a tracking device operably connected to communicate the feedback data to the input device of the controller and including a sensor for detecting a condition of a user; the controller further programmed to operate on input data provided independently from the user by a program executable by the proceεεor, user data correεponding to inputs selected by a user, and feedback data corresponding to a user's condition and provided from the tracking device; and the actuation device operably connected to the controller and tracking device for providing stimuli according to a control εignal correεponding to the feedback data, uεer data, and input data.
2. The apparatuε of claim 1 wherein the actuation device further comprises an electromuscular stimulation device comprising a receiver for receiving input signals corresponding to the user data and feedback data.
3. The apparatus of claim 1 wherein the tracking device further compriseε a εensor selected from a position detector, motion senεor, accelerometer, radar receiver, force tranεducer, preεεure tranεducer, temperature εenεor, heart rate detector, humidity εensor, and imaging sensor.
4. The apparatus of claim 3 wherein the imaging senεor is εelected from a magnetic resonance imaging device, a sonar imaging device, an ultrasonic imaging device, an x-ray imaging device, an imaging device operating in the infrared imaging spectrum, an imaging device operating in the ultraviolet εpectrum, an imaging device operating in the visible light εpectrum, a radar imaging device, and a tomographic imaging device.
5. The apparatuε of claim 1 wherein the εensor of the tracking device includes a transducer for detecting a condition of a user, the transducer being selected from detectors for detecting spatial position, a relative displacement, a velocity, a speed, a force, a presεure, an environmental temperature, and a pulεe rate correεponding to a bodily member of a uεer.
6. The apparatuε of claim 1 wherein the εensor is adapted to detect a position of a bodily member of a user, the senεor being εelected from a radar receiver, a gyroεcopic device for establishing spatial position, a global positioning εystem detecting a target positioned on the bodily member from three sensorε εpaced from one another and from the bodily member, and an imaging system adapted for detecting, recording, and interpreting positionε of bodily memberε of a uεer and proceεεing data correεponding to the poεitionε to provide outputs from the tracking device to the controller.
7. The apparatus of claim 1 wherein the tracking device includes an instrumented, movable member incorporated into an article of body wear placeable proximate a bodily member of the user.
8. The apparatus of claim 7 wherein the article of body wear. is selected from a sleeve fittable to an arm of a user, a glove, a hat, a helmet, a sleeve fittable to a torso of a user, a sleeve fittable to a leg of a user, a stocking fittable to a foot of a user, a boot, and a suit fittable to arms, torso and legs of a uεer.
9. A method of exerciεing comprising: inputting a procesε parameter signal into an input device for operating an executable program in a procesεor of a controller, the proceεε parameter εignal correεponding to data required by the executable program; inputting a uεer εelection εignal into the input device, the uεer selections corresponding to optional data selectable by a user and uεeable by the executable program; tracking a condition of a uεer by a tracking device, the condition being εelected from a εpatial position, a relative displacement, a velocity, a speed, a force, a presεure, an environmental temperature, and a pulse rate corresponding to a bodily member of a user, and the tracking device comprising a sensor selected from a position detector, motion sensor, accelerometer, radar receiver, force transducer, pressure transducer, temperature sensor, heart rate detector, humidity sensor, and imaging sensor; procesεing the proceεε parameter εignal, the user selection εignal, and a εenεor εignal from the tracking device, the εenεor εignal being received by the controller operably connected to the tracking device, to provide an actuator signal to a senεory interface device operably connected to the controller to control an actuator; and providing to a bodily member of a user a stimulus corresponding to the process parameter signal, the user selection signal, and the sensor signal.
10. The method of claim 9 further comprising setting a control of an electromuscular stimulation device to deliver senεory impact to muεcles of a user at interactively determined times, the electromuscular stimulation device comprising a power supply, a voltage source connected to the power supply, a timing control connected between the voltage source and a plurality of electrodes secured to the body of a user to actuate selected muscles, the timing control being controlled by the controller in accordance with settings input by a user, pre-programmed control parameters, and feedback signals correεponding to a selected condition of a user provided from the tracking device.
ll. An apparatus for training a user, the apparatuε compriεing: an actuation device for preεenting to a user a stimulus sensible by a user; a controller operably connected to the actuation device and compriεing a processor, an input device for receiving feedback data corresponding to a condition of a user, and an output device for sending control signals for controlling the actuation device,- a tracking device operably connected to communicate the feedback data to the controller and including a sensor for detecting a condition of a user; and an electromuscular stimulation device comprising a receiver for receiving input signals corresponding to user inputs selected by a user and to the feedback data, the electromuscular stimulation device being operably connected to the controller to provide stimulation directly to a uεer from the controller.
12. The apparatus of claim 11 wherein the tracking device further compriseε a εenεor εelected from a poεition detector, motion εensor, accelerometer, radar receiver, force tranεducer, preεεure tranεducer, temperature εenεor, heart rate detector, humidity sensor, and imaging sensor
13. The apparatus of claim 12 wherein the sensor is an imaging senεor and iε εelected from a magnetic reεonance imaging device, a sonar imaging device, an ultrasonic imaging device, an x-ray imaging device, an imaging device operating in the infrared imaging spectrum, an imaging device operating in the ultraviolet spectrum, an imaging device operating in the visible light spectrum, a radar imaging device, and a tomographic imaging device.
14. The apparatus of claim 11 wherein the senεor of the tracking device includes a transducer for detecting a condition of a uεer, the condition being selected from a spatial position, a relative displacement, a velocity, a speed, a force, a presεure, an environmental temperature, and a pulεe rate corresponding to a bodily member of a user.
15. The apparatus of claim 11 wherein the sensor is adapted to detect a position of a bodily member of a uεer, the εenεor being εelected from a radar receiver, a gyroscopic device for establishing spatial position, a global positioning syεtem detecting a target positioned on the bodily member from three senεorε εpaced from one another and from the bodily member, and an imaging εyεtem adapted for detecting, recording, and interpreting poεitionε of bodily memberε of a uεer and proceεεing data correεponding to the poεitionε to provide outputs from the tracking device to the controller.
16. The apparatus of claim 15 wherein the tracking device includes an instrumented, movable member incorporated into an article of body wear placeable over a bodily member of the user to move therewith.
17. The apparatus of claim 16 wherein the article of body wear is selected from a εleeve fittable to an arm of a uεer, a glove, a hat, a helmet, a εleeve fittable to a torso of a user, a sleeve fittable to a leg of a uεer, a εtocking fittable to a foot of a uεer, a boot, and a εuit fittable to armε, torεo and legε of a uεer.
PCT/US1996/011885 1995-07-26 1996-07-19 Electronic exercise enhancer WO1997004840A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP96925358A EP0840638B1 (en) 1995-07-26 1996-07-19 Electronic exercise enhancer
AU65482/96A AU6548296A (en) 1995-07-26 1996-07-19 Electronic exercise enhancer
DE69634915T DE69634915D1 (en) 1995-07-26 1996-07-19 ELECTRONIC DEVICE FOR EXERCISE OPTIMIZATION

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/507,550 US5702323A (en) 1995-07-26 1995-07-26 Electronic exercise enhancer
US08/507,550 1995-07-26

Publications (1)

Publication Number Publication Date
WO1997004840A1 true WO1997004840A1 (en) 1997-02-13

Family

ID=24019086

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1996/011885 WO1997004840A1 (en) 1995-07-26 1996-07-19 Electronic exercise enhancer

Country Status (5)

Country Link
US (2) US5702323A (en)
EP (1) EP0840638B1 (en)
AU (1) AU6548296A (en)
DE (1) DE69634915D1 (en)
WO (1) WO1997004840A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008055949A1 (en) * 2006-11-10 2008-05-15 Khymeia S.R.L. Apparatus for motor training and exercise of the human body

Families Citing this family (567)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8352400B2 (en) 1991-12-23 2013-01-08 Hoffberg Steven M Adaptive pattern recognition based controller apparatus and method and human-factored interface therefore
US7084859B1 (en) * 1992-09-18 2006-08-01 Pryor Timothy R Programmable tactile touch screen displays and man-machine interfaces for improved vehicle instrumentation and telematics
US9513744B2 (en) * 1994-08-15 2016-12-06 Apple Inc. Control systems employing novel physical controls and touch screens
US7386401B2 (en) * 1994-11-21 2008-06-10 Phatrat Technology, Llc Helmet that reports impact information, and associated methods
US7739076B1 (en) 1999-06-30 2010-06-15 Nike, Inc. Event and sport performance methods and systems
US6539336B1 (en) * 1996-12-12 2003-03-25 Phatrat Technologies, Inc. Sport monitoring system for determining airtime, speed, power absorbed and other factors such as drop distance
US7678023B1 (en) 1995-06-22 2010-03-16 Shea Michael J Method for providing mental activity for an exerciser
US8228305B2 (en) 1995-06-29 2012-07-24 Apple Inc. Method for providing human input to a computer
US20090322499A1 (en) * 1995-06-29 2009-12-31 Pryor Timothy R Programmable tactile touch screen displays and man-machine interfaces for improved vehicle instrumentation and telematics
US6430997B1 (en) * 1995-11-06 2002-08-13 Trazer Technologies, Inc. System and method for tracking and assessing movement skills in multidimensional space
US6059692A (en) * 1996-12-13 2000-05-09 Hickman; Paul L. Apparatus for remote interactive exercise and health equipment
US6808472B1 (en) 1995-12-14 2004-10-26 Paul L. Hickman Method and apparatus for remote interactive exercise and health equipment
US5890996A (en) * 1996-05-30 1999-04-06 Interactive Performance Monitoring, Inc. Exerciser and physical performance monitoring system
US7146408B1 (en) * 1996-05-30 2006-12-05 Schneider Automation Inc. Method and system for monitoring a controller and displaying data from the controller in a format provided by the controller
US6758755B2 (en) 1996-11-14 2004-07-06 Arcade Planet, Inc. Prize redemption system for games executed over a wide area network
JP3469410B2 (en) * 1996-11-25 2003-11-25 三菱電機株式会社 Wellness system
IL120507A (en) * 1997-03-24 2001-06-14 Keytron Electronics & Technolo Exercise monitoring system
US6345232B1 (en) 1997-04-10 2002-02-05 Urban H. D. Lynch Determining aircraft position and attitude using GPS position data
US6050924A (en) 1997-04-28 2000-04-18 Shea; Michael J. Exercise system
US5947868A (en) * 1997-06-27 1999-09-07 Dugan; Brian M. System and method for improving fitness equipment and exercise
JP3103045B2 (en) * 1997-07-10 2000-10-23 三菱電機株式会社 Image capturing / reproducing method and method, and recording medium recording image reproducing program
GB9722766D0 (en) 1997-10-28 1997-12-24 British Telecomm Portable computers
US7856750B2 (en) 1997-12-08 2010-12-28 Horus Vision Llc Apparatus and method for calculating aiming point information
US5989120A (en) * 1998-01-07 1999-11-23 Pragmatic Designs, Inc. Electronic counting apparatus for a child's game and method therefor
US8479122B2 (en) 2004-07-30 2013-07-02 Apple Inc. Gestures for touch sensitive input devices
US7614008B2 (en) 2004-07-30 2009-11-03 Apple Inc. Operation of a computer with touch screen interface
US9292111B2 (en) 1998-01-26 2016-03-22 Apple Inc. Gesturing with a multipoint sensing device
US9239673B2 (en) 1998-01-26 2016-01-19 Apple Inc. Gesturing with a multipoint sensing device
US6217027B1 (en) * 1998-03-02 2001-04-17 United States Of America Computerized portable pneumatic target apparatus
DE19819257C2 (en) * 1998-04-29 2000-05-18 Wolfgang Siegfried Obesity treatment device
JP3120065B2 (en) * 1998-05-27 2000-12-25 科学技術振興事業団 Feedforward exercise training device and feedforward exercise evaluation system
US6483484B1 (en) * 1998-12-18 2002-11-19 Semiconductor Energy Laboratory Co., Ltd. Goggle type display system
US7904187B2 (en) 1999-02-01 2011-03-08 Hoffberg Steven M Internet appliance system and method
US7749089B1 (en) 1999-02-26 2010-07-06 Creative Kingdoms, Llc Multi-media interactive play system
US6307952B1 (en) * 1999-03-03 2001-10-23 Disney Enterprises, Inc. Apparatus for detecting guest interactions and method therefore
US6243624B1 (en) * 1999-03-19 2001-06-05 Northwestern University Non-Linear muscle-like compliant controller
GB9909254D0 (en) * 1999-04-23 1999-06-16 Interex Ltd A Computer Interface
US7789742B1 (en) * 1999-05-12 2010-09-07 Wilbert Q. Murdock Smart golf club multiplayer system for the internet
JP2000350865A (en) * 1999-06-11 2000-12-19 Mr System Kenkyusho:Kk Game device for composite real space, image processing method therefor and program storage medium
WO2001001706A1 (en) * 1999-06-30 2001-01-04 Phatrat Technology, Inc. Event and sport performance methods and systems
US6997852B2 (en) 1999-07-08 2006-02-14 Icon Ip, Inc. Methods and systems for controlling an exercise apparatus using a portable remote device
US6312363B1 (en) 1999-07-08 2001-11-06 Icon Health & Fitness, Inc. Systems and methods for providing an improved exercise device with motivational programming
US6458060B1 (en) 1999-07-08 2002-10-01 Icon Ip, Inc. Systems and methods for interaction with exercise device
US20080051256A1 (en) * 1999-07-08 2008-02-28 Icon Ip, Inc. Exercise device with on board personal trainer
US6918858B2 (en) 1999-07-08 2005-07-19 Icon Ip, Inc. Systems and methods for providing an improved exercise device with access to motivational programming over telephone communication connection lines
US8029415B2 (en) 1999-07-08 2011-10-04 Icon Ip, Inc. Systems, methods, and devices for simulating real world terrain on an exercise device
US7985164B2 (en) 1999-07-08 2011-07-26 Icon Ip, Inc. Methods and systems for controlling an exercise apparatus using a portable data storage device
US7166062B1 (en) 1999-07-08 2007-01-23 Icon Ip, Inc. System for interaction with exercise device
US7628730B1 (en) 1999-07-08 2009-12-08 Icon Ip, Inc. Methods and systems for controlling an exercise apparatus using a USB compatible portable remote device
US7537546B2 (en) 1999-07-08 2009-05-26 Icon Ip, Inc. Systems and methods for controlling the operation of one or more exercise devices and providing motivational programming
US6447424B1 (en) * 2000-02-02 2002-09-10 Icon Health & Fitness Inc System and method for selective adjustment of exercise apparatus
US7166064B2 (en) * 1999-07-08 2007-01-23 Icon Ip, Inc. Systems and methods for enabling two-way communication between one or more exercise devices and computer devices and for enabling users of the one or more exercise devices to competitively exercise
US7060006B1 (en) 1999-07-08 2006-06-13 Icon Ip, Inc. Computer systems and methods for interaction with exercise device
US6922615B2 (en) * 1999-07-30 2005-07-26 Oshkosh Truck Corporation Turret envelope control system and method for a fire fighting vehicle
US7162332B2 (en) 1999-07-30 2007-01-09 Oshkosh Truck Corporation Turret deployment system and method for a fire fighting vehicle
US7729831B2 (en) 1999-07-30 2010-06-01 Oshkosh Corporation Concrete placement vehicle control system and method
US7006902B2 (en) * 1999-07-30 2006-02-28 Oshkosh Truck Corporation Control system and method for an equipment service vehicle
US7107129B2 (en) * 2002-02-28 2006-09-12 Oshkosh Truck Corporation Turret positioning system and method for a fire fighting vehicle
US7184862B2 (en) * 1999-07-30 2007-02-27 Oshkosh Truck Corporation Turret targeting system and method for a fire fighting vehicle
US6749432B2 (en) * 1999-10-20 2004-06-15 Impulse Technology Ltd Education system challenging a subject's physiologic and kinesthetic systems to synergistically enhance cognitive function
US6811516B1 (en) 1999-10-29 2004-11-02 Brian M. Dugan Methods and apparatus for monitoring and encouraging health and fitness
US8482535B2 (en) * 1999-11-08 2013-07-09 Apple Inc. Programmable tactile touch screen displays and man-machine interfaces for improved vehicle instrumentation and telematics
US8956228B2 (en) * 1999-12-03 2015-02-17 Nike, Inc. Game pod
US6585622B1 (en) 1999-12-03 2003-07-01 Nike, Inc. Interactive use an athletic performance monitoring and reward method, system, and computer program product
US20020091843A1 (en) * 1999-12-21 2002-07-11 Vaid Rahul R. Wireless network adapter
TWM244584U (en) 2000-01-17 2004-09-21 Semiconductor Energy Lab Display system and electrical appliance
US8290809B1 (en) 2000-02-14 2012-10-16 Ebay Inc. Determining a community rating for a user using feedback ratings of related users in an electronic environment
US8576199B1 (en) 2000-02-22 2013-11-05 Apple Inc. Computer control systems
US20080122799A1 (en) * 2001-02-22 2008-05-29 Pryor Timothy R Human interfaces for vehicles, homes, and other applications
US6761637B2 (en) 2000-02-22 2004-07-13 Creative Kingdoms, Llc Method of game play using RFID tracking device
US7445550B2 (en) 2000-02-22 2008-11-04 Creative Kingdoms, Llc Magical wand and interactive play experience
US7878905B2 (en) 2000-02-22 2011-02-01 Creative Kingdoms, Llc Multi-layered interactive play experience
US9614934B2 (en) 2000-02-29 2017-04-04 Paypal, Inc. Methods and systems for harvesting comments regarding users on a network-based facility
US7428505B1 (en) * 2000-02-29 2008-09-23 Ebay, Inc. Method and system for harvesting feedback and comments regarding multiple items from users of a network-based transaction facility
IT1321144B1 (en) * 2000-03-01 2003-12-30 Technogym Srl EXPERT SYSTEM FOR INTERACTIVE EXCHANGE OF INFORMATION BETWEEN A USER AND A DEDICATED INFORMATION SYSTEM.
US8103517B2 (en) 2000-04-12 2012-01-24 Michael Hinnebusch System and method to improve fitness training
US6746371B1 (en) 2000-04-28 2004-06-08 International Business Machines Corporation Managing fitness activity across diverse exercise machines utilizing a portable computer system
US6601016B1 (en) 2000-04-28 2003-07-29 International Business Machines Corporation Monitoring fitness activity across diverse exercise machines utilizing a universally accessible server system
US6702719B1 (en) 2000-04-28 2004-03-09 International Business Machines Corporation Exercise machine
JP2001356849A (en) * 2000-05-08 2001-12-26 Ken Tamada Business model for human interface and hardware
EP1159989A1 (en) * 2000-05-24 2001-12-05 In2Sports B.V. A method of generating and/or adjusting a training schedule
US6995753B2 (en) 2000-06-06 2006-02-07 Semiconductor Energy Laboratory Co., Ltd. Display device and method of manufacturing the same
JP2002072963A (en) * 2000-06-12 2002-03-12 Semiconductor Energy Lab Co Ltd Light-emitting module and driving method therefor, and optical sensor
GB0021327D0 (en) * 2000-08-31 2000-10-18 Smith & Nephew Rehabilitation device
US7066781B2 (en) 2000-10-20 2006-06-27 Denise Chapman Weston Children's toy with wireless tag/transponder
US7171331B2 (en) * 2001-12-17 2007-01-30 Phatrat Technology, Llc Shoes employing monitoring devices, and associated methods
US20020078152A1 (en) * 2000-12-19 2002-06-20 Barry Boone Method and apparatus for providing predefined feedback
US7277782B2 (en) 2001-01-31 2007-10-02 Oshkosh Truck Corporation Control system and method for electric vehicle
AU2002255568B8 (en) 2001-02-20 2014-01-09 Adidas Ag Modular personal network systems and methods
US20080088587A1 (en) * 2001-02-22 2008-04-17 Timothy Pryor Compact rtd instrument panels and computer interfaces
US20080024463A1 (en) * 2001-02-22 2008-01-31 Timothy Pryor Reconfigurable tactile control display applications
US6834436B2 (en) * 2001-02-23 2004-12-28 Microstrain, Inc. Posture and body movement measuring system
US8939831B2 (en) 2001-03-08 2015-01-27 Brian M. Dugan Systems and methods for improving fitness equipment and exercise
US20020160883A1 (en) 2001-03-08 2002-10-31 Dugan Brian M. System and method for improving fitness equipment and exercise
FR2822385B1 (en) * 2001-03-23 2003-09-26 Euro Gem Gmbh TRAINING APPARATUS FOR ABDOMINAL MUSCLE
EP1256316A1 (en) * 2001-05-07 2002-11-13 Move2Health B.V. Portable device comprising an acceleration sensor and method of generating instructions or advice
US20030059754A1 (en) * 2001-09-27 2003-03-27 Jackson Jeff Wayne Routine machine
NL1019154C2 (en) * 2001-10-11 2003-04-14 Tech Ind Tacx B V Home trainer comprises frame in which bicycle is fixed and which is provided with adjustable brake component in frictional contact with driven wheel of bicycle
DE10151152A1 (en) * 2001-10-19 2003-08-07 Wolfgang Kloeckner Device for using a biofeedback method, and method for generating and displaying data when using a biofeedback method
US6921351B1 (en) 2001-10-19 2005-07-26 Cybergym, Inc. Method and apparatus for remote interactive exercise and health equipment
US7593775B2 (en) * 2002-01-15 2009-09-22 Therapeutic Innovations Sports equipment with resonant muscle stimulator for developing muscle strength
US6955542B2 (en) * 2002-01-23 2005-10-18 Aquatech Fitness Corp. System for monitoring repetitive movement
US6925851B2 (en) * 2002-01-24 2005-08-09 Sensorpad Systems Inc. Method and system for detecting and displaying the impact of a blow
US6990639B2 (en) * 2002-02-07 2006-01-24 Microsoft Corporation System and process for controlling electronic components in a ubiquitous computing environment using multimodal integration
DE20221083U1 (en) * 2002-04-03 2005-01-27 Proxomed Medizintechnik Gmbh Measuring device for training devices
US20070066396A1 (en) 2002-04-05 2007-03-22 Denise Chapman Weston Retail methods for providing an interactive product to a consumer
US6836711B2 (en) * 2002-04-05 2004-12-28 Michael Leonard Gentilcore Bicycle data acquisition
US6967566B2 (en) 2002-04-05 2005-11-22 Creative Kingdoms, Llc Live-action interactive adventure game
GB2387685B (en) * 2002-04-19 2005-07-20 Thales Plc Apparatus and method for vehicle simulation
US20040002843A1 (en) * 2002-05-13 2004-01-01 Consolidated Global Fun Unlimited, Llc Method and system for interacting with simulated phenomena
US20050009608A1 (en) * 2002-05-13 2005-01-13 Consolidated Global Fun Unlimited Commerce-enabled environment for interacting with simulated phenomena
US7946959B2 (en) * 2002-05-30 2011-05-24 Nike, Inc. Training scripts
US20040015318A1 (en) * 2002-07-11 2004-01-22 Heller Alan C. Automatic sensory logger
US7674184B2 (en) 2002-08-01 2010-03-09 Creative Kingdoms, Llc Interactive water attraction and quest game
US7651442B2 (en) * 2002-08-15 2010-01-26 Alan Carlson Universal system for monitoring and controlling exercise parameters
US6840892B1 (en) * 2002-08-22 2005-01-11 Tonic Fitness Technology, Inc. Recuperating machine
US7358963B2 (en) 2002-09-09 2008-04-15 Apple Inc. Mouse having an optically-based scrolling feature
JP2004110628A (en) * 2002-09-20 2004-04-08 Shimano Inc Bicycle user's information management device and cycle computer
ATE454195T1 (en) 2002-10-30 2010-01-15 Nike International Ltd GARMENTS WITH MOTION DETECTION MARKERS FOR VIDEO GAMES
US8206219B2 (en) 2002-10-30 2012-06-26 Nike, Inc. Interactive gaming apparel for interactive gaming
US7805149B2 (en) 2004-01-16 2010-09-28 Adidas Ag Location-aware fitness training device, methods, and program products that support real-time interactive communication and automated route generation
US8157706B2 (en) * 2009-10-19 2012-04-17 Precor Incorporated Fitness facility equipment usage control system and method
JP2006517830A (en) * 2003-01-26 2006-08-03 プレコ−ル インコ−ポレイテッド Fitness equipment maintenance tracking and alarm system
US8882637B2 (en) 2003-01-26 2014-11-11 Precor Incorporated Fitness facility equipment distribution management
ES2234384B1 (en) * 2003-02-05 2006-10-16 Jesus Ciudad Colado INFORMATIZED TEAM TO VIRTUALLY SIMULATE THE SALON TOREO.
US7621850B2 (en) 2003-02-28 2009-11-24 Nautilus, Inc. Dual deck exercise device
US7815549B2 (en) * 2003-02-28 2010-10-19 Nautilus, Inc. Control system and method for an exercise apparatus
US9446319B2 (en) 2003-03-25 2016-09-20 Mq Gaming, Llc Interactive gaming toy
US8745541B2 (en) 2003-03-25 2014-06-03 Microsoft Corporation Architecture for controlling a computer using hand gestures
US7665041B2 (en) 2003-03-25 2010-02-16 Microsoft Corporation Architecture for controlling a computer using hand gestures
US7182738B2 (en) * 2003-04-23 2007-02-27 Marctec, Llc Patient monitoring apparatus and method for orthosis and other devices
CN1247284C (en) * 2003-05-27 2006-03-29 李明 Method and apparatus for realizing virtual riding bicycle
US6837827B1 (en) 2003-06-17 2005-01-04 Garmin Ltd. Personal training device using GPS data
US7497807B2 (en) * 2003-07-15 2009-03-03 Cube X Incorporated Interactive computer simulation enhanced exercise machine
US7497812B2 (en) * 2003-07-15 2009-03-03 Cube X, Incorporated Interactive computer simulation enhanced exercise machine
US20050054492A1 (en) * 2003-07-15 2005-03-10 Neff John D. Exercise device for under a desk
US6955094B1 (en) * 2003-07-18 2005-10-18 Cleveland Medical Devices Inc. Sensor for measuring shear forces
US7217224B2 (en) * 2003-08-14 2007-05-15 Tom Thomas Virtual exercise system and method
US20060252602A1 (en) * 2003-10-14 2006-11-09 Brown Michael W Program and system for managing fitness activity across diverse exercise machines utilizing a portable computer system
US7716079B2 (en) * 2003-11-20 2010-05-11 Ebay Inc. Feedback cancellation in a network-based transaction facility
WO2005069251A2 (en) * 2004-01-06 2005-07-28 Mayo Foundation For Medical Education And Research Hypoxia awareness training system
US7308818B2 (en) * 2004-02-09 2007-12-18 Garri Productions, Inc. Impact-sensing and measurement systems, methods for using same, and related business methods
US7398151B1 (en) 2004-02-25 2008-07-08 Garmin Ltd. Wearable electronic device
US7507187B2 (en) 2004-04-06 2009-03-24 Precor Incorporated Parameter sensing system for an exercise device
US7901292B1 (en) * 2004-04-15 2011-03-08 Navteq North America, Llc Method for comparing performances on remotely located courses
EP1749290A1 (en) * 2004-05-24 2007-02-07 Nederlandse Organisatie voor Toegepast-Natuuurwetenschappelijk Onderzoek TNO System, use of said system and method for monitoring and optimising a performance of at least one human operator
EP1600911A1 (en) * 2004-05-24 2005-11-30 Nederlandse Organisatie voor toegepast-natuurwetenschappelijk onderzoek TNO System, use of said system and method for monitoring and optimising a performance of at least one human operator
WO2006014810A2 (en) * 2004-07-29 2006-02-09 Kevin Ferguson A human movement measurement system
US8381135B2 (en) 2004-07-30 2013-02-19 Apple Inc. Proximity detector in handheld device
US20100231506A1 (en) * 2004-09-07 2010-09-16 Timothy Pryor Control of appliances, kitchen and home
US8628333B2 (en) * 2004-09-10 2014-01-14 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and apparatus for performance optimization through physical perturbation of task elements
US7591764B2 (en) * 2004-09-24 2009-09-22 Swimworks, Inc. Exercise apparatus
US20060085253A1 (en) * 2004-10-18 2006-04-20 Matthew Mengerink Method and system to utilize a user network within a network-based commerce platform
US7373820B1 (en) 2004-11-23 2008-05-20 James Terry L Accelerometer for data collection and communication
US8021277B2 (en) * 2005-02-02 2011-09-20 Mad Dogg Athletics, Inc. Programmed exercise bicycle with computer aided guidance
US20060240947A1 (en) * 2005-03-16 2006-10-26 Nautilus, Inc. Apparatus and methods for transmitting programming, receiving and displaying programming, communicating with exercise equipment, and accessing and passing data to and from applications
US20060234840A1 (en) * 2005-03-23 2006-10-19 Watson Edward M Closed loop control of resistance in a resistance-type exercise system
DE102005022005B4 (en) * 2005-05-09 2014-10-30 Anna Gutmann Method and device for facilitating the movement control of body parts
US7864168B2 (en) * 2005-05-25 2011-01-04 Impulse Technology Ltd. Virtual reality movement system
WO2007020663A1 (en) * 2005-08-12 2007-02-22 Vupiesse Italia S.R.L. Self-coaching portable device for abdominal muscles
WO2007027706A2 (en) * 2005-08-29 2007-03-08 Blanarovich Adrian M Apparatus and system for measuring and communicating physical activity data
US7591795B2 (en) * 2005-09-28 2009-09-22 Alterg, Inc. System, method and apparatus for applying air pressure on a portion of the body of an individual
US20070088469A1 (en) * 2005-10-04 2007-04-19 Oshkosh Truck Corporation Vehicle control system and method
US7697827B2 (en) 2005-10-17 2010-04-13 Konicek Jeffrey C User-friendlier interfaces for a camera
EP1971404A2 (en) * 2005-12-22 2008-09-24 Radow, Scott B. Exercise device
TWI291889B (en) * 2005-12-22 2008-01-01 Ind Tech Res Inst Interactive control system
US8077147B2 (en) 2005-12-30 2011-12-13 Apple Inc. Mouse with optical sensing surface
US11826652B2 (en) 2006-01-04 2023-11-28 Dugan Health, Llc Systems and methods for improving fitness equipment and exercise
US20100201512A1 (en) 2006-01-09 2010-08-12 Harold Dan Stirling Apparatus, systems, and methods for evaluating body movements
KR101230536B1 (en) * 2006-02-13 2013-02-06 한로봇 주식회사 Arm-wrestling robot and the control method
US20070208392A1 (en) * 2006-02-17 2007-09-06 Alfred E. Mann Foundation For Scientific Research System for functional electrical stimulation
JP4151982B2 (en) * 2006-03-10 2008-09-17 任天堂株式会社 Motion discrimination device and motion discrimination program
JP4684147B2 (en) 2006-03-28 2011-05-18 任天堂株式会社 Inclination calculation device, inclination calculation program, game device, and game program
US7735230B2 (en) * 2006-03-29 2010-06-15 Novatac, Inc. Head-mounted navigation system
US8706560B2 (en) 2011-07-27 2014-04-22 Ebay Inc. Community based network shopping
JP4757089B2 (en) * 2006-04-25 2011-08-24 任天堂株式会社 Music performance program and music performance apparatus
JP4916762B2 (en) * 2006-05-02 2012-04-18 任天堂株式会社 GAME PROGRAM AND GAME DEVICE
US8152693B2 (en) * 2006-05-08 2012-04-10 Nokia Corporation Exercise data device, server, system and method
JP5199246B2 (en) 2006-05-22 2013-05-15 ナイキ インターナショナル リミテッド Clock display with translucent cover using light source
US7787857B2 (en) * 2006-06-12 2010-08-31 Garmin Ltd. Method and apparatus for providing an alert utilizing geographic locations
US8781568B2 (en) 2006-06-23 2014-07-15 Brian M. Dugan Systems and methods for heart rate monitoring, data transmission, and use
DE202006011058U1 (en) * 2006-07-18 2006-10-12 Jerichow, Ulrich, Dr. Generation of a virtual environment for fitness training uses computer system control of wind and temperature and noise simulations
US8430770B2 (en) 2006-10-07 2013-04-30 Brian M. Dugan Systems and methods for measuring and/or analyzing swing information
US8337335B2 (en) 2006-10-07 2012-12-25 Dugan Brian M Systems and methods for measuring and/or analyzing swing information
US20080110115A1 (en) * 2006-11-13 2008-05-15 French Barry J Exercise facility and method
US20080161731A1 (en) * 2006-12-27 2008-07-03 Woods Sherrod A Apparatus, system, and method for monitoring the range of motion of a patient's joint
US7844915B2 (en) 2007-01-07 2010-11-30 Apple Inc. Application programming interfaces for scrolling operations
DE102007002614B4 (en) * 2007-01-12 2009-04-02 Jerichow, Ulrich, Dr. Device for monitoring, controlling and / or regulating a gas composition
JP5127242B2 (en) * 2007-01-19 2013-01-23 任天堂株式会社 Acceleration data processing program and game program
US20080182724A1 (en) * 2007-01-25 2008-07-31 Nicole Lee Guthrie Activity Monitor with Incentive Features
US7913178B2 (en) 2007-01-31 2011-03-22 Ebay Inc. Method and system for collaborative and private sessions
DE202007019152U1 (en) * 2007-03-01 2011-03-17 Telefonaktiebolaget L M Ericsson (Publ) Mobile service to stay up-to-date during a competitor race
US7931563B2 (en) * 2007-03-08 2011-04-26 Health Hero Network, Inc. Virtual trainer system and method
US8005238B2 (en) 2007-03-22 2011-08-23 Microsoft Corporation Robust adaptive beamforming with enhanced noise suppression
US7909741B2 (en) 2007-03-27 2011-03-22 Dhkl, Inc. Devices, systems and methods for receiving, recording and displaying information relating to physical exercise
US20090029831A1 (en) 2007-03-30 2009-01-29 Nautilus, Inc. Device and method for limiting travel in an exercise device, and an exercise device including such a limiting device
US8005237B2 (en) 2007-05-17 2011-08-23 Microsoft Corp. Sensor array beamformer post-processor
US20080310707A1 (en) * 2007-06-15 2008-12-18 Microsoft Corporation Virtual reality enhancement using real world data
US8430752B2 (en) 2007-06-20 2013-04-30 The Nielsen Company (Us), Llc Methods and apparatus to meter video game play
WO2009003170A1 (en) * 2007-06-27 2008-12-31 Radow Scott B Stationary exercise equipment
TWI343031B (en) * 2007-07-16 2011-06-01 Ind Tech Res Inst Method and device for keyboard instrument learning
US20090048493A1 (en) * 2007-08-17 2009-02-19 James Terry L Health and Entertainment Device for Collecting, Converting, Displaying and Communicating Data
US8221290B2 (en) 2007-08-17 2012-07-17 Adidas International Marketing B.V. Sports electronic training system with electronic gaming features, and applications thereof
EP2195099B1 (en) * 2007-09-10 2011-08-24 Trixter PLC Sensing apparatus for use with exercise bicycles
US8265949B2 (en) 2007-09-27 2012-09-11 Depuy Products, Inc. Customized patient surgical plan
US8357111B2 (en) 2007-09-30 2013-01-22 Depuy Products, Inc. Method and system for designing patient-specific orthopaedic surgical instruments
US8361076B2 (en) 2007-09-30 2013-01-29 Depuy Products, Inc. Patient-customizable device and system for performing an orthopaedic surgical procedure
US8629976B2 (en) * 2007-10-02 2014-01-14 Microsoft Corporation Methods and systems for hierarchical de-aliasing time-of-flight (TOF) systems
WO2014153201A1 (en) 2013-03-14 2014-09-25 Alterg, Inc. Method of gait evaluation and training with differential pressure system
US20120238921A1 (en) 2011-03-18 2012-09-20 Eric Richard Kuehne Differential air pressure systems and methods of using and calibrating such systems for mobility impaired users
US10342461B2 (en) 2007-10-15 2019-07-09 Alterg, Inc. Method of gait evaluation and training with differential pressure system
WO2009051750A1 (en) 2007-10-15 2009-04-23 Alterg, Inc. Systems, methods and apparatus for calibrating differential air pressure devices
US20090111656A1 (en) * 2007-10-26 2009-04-30 At&T Knowledge Ventures, L.P. Networked exercise machine
AU2007362978B2 (en) * 2007-12-21 2012-04-05 Tonic Fitness Technology, Inc. Exercise apparratus adapting individual physical ability and control method thereof
US20090166684A1 (en) * 2007-12-26 2009-07-02 3Dv Systems Ltd. Photogate cmos pixel for 3d cameras having reduced intra-pixel cross talk
US7676332B2 (en) 2007-12-27 2010-03-09 Kersh Risk Management, Inc. System and method for processing raw activity energy expenditure data
US8264505B2 (en) 2007-12-28 2012-09-11 Microsoft Corporation Augmented reality and filtering
US8047966B2 (en) * 2008-02-29 2011-11-01 Apple Inc. Interfacing portable media devices and sports equipment
US8951168B2 (en) 2008-03-05 2015-02-10 Mad Dogg Athletics, Inc. Programmable exercise bicycle
TWI378781B (en) * 2008-03-12 2012-12-11 Jung Tang Huang A belt integrated with stress sensing and output reaction
US20090270743A1 (en) * 2008-04-17 2009-10-29 Dugan Brian M Systems and methods for providing authenticated biofeedback information to a mobile device and for using such information
US8976007B2 (en) 2008-08-09 2015-03-10 Brian M. Dugan Systems and methods for providing biofeedback information to a cellular telephone and for using such information
US9549585B2 (en) 2008-06-13 2017-01-24 Nike, Inc. Footwear having sensor system
EP3087858B1 (en) 2008-06-13 2021-04-28 NIKE Innovate C.V. Footwear having sensor system
US9002680B2 (en) * 2008-06-13 2015-04-07 Nike, Inc. Foot gestures for computer input and interface control
US10070680B2 (en) 2008-06-13 2018-09-11 Nike, Inc. Footwear having sensor system
US8385557B2 (en) 2008-06-19 2013-02-26 Microsoft Corporation Multichannel acoustic echo reduction
US8325909B2 (en) 2008-06-25 2012-12-04 Microsoft Corporation Acoustic echo suppression
US20100009809A1 (en) * 2008-06-26 2010-01-14 Janice Carrington System for simulating a tour of or being in a remote location while exercising
US8203699B2 (en) 2008-06-30 2012-06-19 Microsoft Corporation System architecture design for time-of-flight system having reduced differential pixel size, and time-of-flight systems so designed
US8892210B2 (en) 2008-07-02 2014-11-18 Niveus Medical, Inc. Devices, systems, and methods for automated optimization of energy delivery
WO2010003106A2 (en) * 2008-07-02 2010-01-07 Niveus Medical Inc. Systems and methods for automated muscle stimulation
US8625899B2 (en) * 2008-07-10 2014-01-07 Samsung Electronics Co., Ltd. Method for recognizing and translating characters in camera-based image
US20100022354A1 (en) * 2008-07-25 2010-01-28 Expresso Fitness Corp. Exercise equipment with movable handle bars to simulate steering motion in a simulated environment and methods therefor
US9149386B2 (en) 2008-08-19 2015-10-06 Niveus Medical, Inc. Devices and systems for stimulation of tissues
WO2010027874A2 (en) * 2008-08-26 2010-03-11 Niveus Medical, Inc. Device, system, and method to improve powered muscle stimulation performance in the presence of tissue edema
US7713172B2 (en) * 2008-10-14 2010-05-11 Icon Ip, Inc. Exercise device with proximity sensor
ES2781331T3 (en) * 2008-11-06 2020-09-01 Jin Y Song Apparatus for monitoring and recording the location and intensity of impacts in sports
US8681321B2 (en) 2009-01-04 2014-03-25 Microsoft International Holdings B.V. Gated 3D camera
US8577084B2 (en) * 2009-01-30 2013-11-05 Microsoft Corporation Visual target tracking
US8448094B2 (en) * 2009-01-30 2013-05-21 Microsoft Corporation Mapping a natural input device to a legacy system
US8577085B2 (en) * 2009-01-30 2013-11-05 Microsoft Corporation Visual target tracking
US8682028B2 (en) * 2009-01-30 2014-03-25 Microsoft Corporation Visual target tracking
US8565477B2 (en) 2009-01-30 2013-10-22 Microsoft Corporation Visual target tracking
US20100199228A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Gesture Keyboarding
US8267781B2 (en) * 2009-01-30 2012-09-18 Microsoft Corporation Visual target tracking
US20100199231A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Predictive determination
US7996793B2 (en) 2009-01-30 2011-08-09 Microsoft Corporation Gesture recognizer system architecture
US8565476B2 (en) 2009-01-30 2013-10-22 Microsoft Corporation Visual target tracking
US8294767B2 (en) 2009-01-30 2012-10-23 Microsoft Corporation Body scan
US8487938B2 (en) * 2009-01-30 2013-07-16 Microsoft Corporation Standard Gestures
US8295546B2 (en) 2009-01-30 2012-10-23 Microsoft Corporation Pose tracking pipeline
US8588465B2 (en) 2009-01-30 2013-11-19 Microsoft Corporation Visual target tracking
WO2010096776A2 (en) 2009-02-20 2010-08-26 Niveus Medical, Inc. Systems and methods of powered muscle stimulation using an energy guidance field
CA2702319C (en) 2009-03-11 2011-06-14 Mytrak Health System Inc. Ergonomic/physiotherapy programme monitoring system and method of using same
US8773355B2 (en) 2009-03-16 2014-07-08 Microsoft Corporation Adaptive cursor sizing
US9256282B2 (en) * 2009-03-20 2016-02-09 Microsoft Technology Licensing, Llc Virtual object manipulation
US8988437B2 (en) 2009-03-20 2015-03-24 Microsoft Technology Licensing, Llc Chaining animations
US8251874B2 (en) 2009-03-27 2012-08-28 Icon Health & Fitness, Inc. Exercise systems for simulating real world terrain
US9313376B1 (en) 2009-04-01 2016-04-12 Microsoft Technology Licensing, Llc Dynamic depth power equalization
US8454437B2 (en) 2009-07-17 2013-06-04 Brian M. Dugan Systems and methods for portable exergaming
US8181123B2 (en) 2009-05-01 2012-05-15 Microsoft Corporation Managing virtual port associations to users in a gesture-based computing environment
US8340432B2 (en) 2009-05-01 2012-12-25 Microsoft Corporation Systems and methods for detecting a tilt angle from a depth image
US9498718B2 (en) 2009-05-01 2016-11-22 Microsoft Technology Licensing, Llc Altering a view perspective within a display environment
US9377857B2 (en) 2009-05-01 2016-06-28 Microsoft Technology Licensing, Llc Show body position
US8253746B2 (en) 2009-05-01 2012-08-28 Microsoft Corporation Determine intended motions
US8503720B2 (en) 2009-05-01 2013-08-06 Microsoft Corporation Human body pose estimation
US8660303B2 (en) * 2009-05-01 2014-02-25 Microsoft Corporation Detection of body and props
US8649554B2 (en) 2009-05-01 2014-02-11 Microsoft Corporation Method to control perspective for a camera-controlled computer
US8638985B2 (en) 2009-05-01 2014-01-28 Microsoft Corporation Human body pose estimation
US9015638B2 (en) 2009-05-01 2015-04-21 Microsoft Technology Licensing, Llc Binding users to a gesture based system and providing feedback to the users
US8942428B2 (en) * 2009-05-01 2015-01-27 Microsoft Corporation Isolate extraneous motions
US9898675B2 (en) * 2009-05-01 2018-02-20 Microsoft Technology Licensing, Llc User movement tracking feedback to improve tracking
ES2709512T3 (en) * 2009-05-15 2019-04-16 Alterg Inc Differential air pressure systems
US8033959B2 (en) 2009-05-18 2011-10-11 Adidas Ag Portable fitness monitoring systems, and applications thereof
US8744121B2 (en) 2009-05-29 2014-06-03 Microsoft Corporation Device for identifying and tracking multiple humans over time
US8418085B2 (en) 2009-05-29 2013-04-09 Microsoft Corporation Gesture coach
US8379101B2 (en) 2009-05-29 2013-02-19 Microsoft Corporation Environment and/or target segmentation
US8693724B2 (en) 2009-05-29 2014-04-08 Microsoft Corporation Method and system implementing user-centric gesture control
US20100302365A1 (en) * 2009-05-29 2010-12-02 Microsoft Corporation Depth Image Noise Reduction
US8856691B2 (en) 2009-05-29 2014-10-07 Microsoft Corporation Gesture tool
US9182814B2 (en) 2009-05-29 2015-11-10 Microsoft Technology Licensing, Llc Systems and methods for estimating a non-visible or occluded body part
US8625837B2 (en) 2009-05-29 2014-01-07 Microsoft Corporation Protocol and format for communicating an image from a camera to a computing environment
US8542252B2 (en) * 2009-05-29 2013-09-24 Microsoft Corporation Target digitization, extraction, and tracking
US8320619B2 (en) 2009-05-29 2012-11-27 Microsoft Corporation Systems and methods for tracking a model
US9383823B2 (en) 2009-05-29 2016-07-05 Microsoft Technology Licensing, Llc Combining gestures beyond skeletal
US9400559B2 (en) * 2009-05-29 2016-07-26 Microsoft Technology Licensing, Llc Gesture shortcuts
US8509479B2 (en) * 2009-05-29 2013-08-13 Microsoft Corporation Virtual object
US8487871B2 (en) * 2009-06-01 2013-07-16 Microsoft Corporation Virtual desktop coordinate transformation
US8390680B2 (en) 2009-07-09 2013-03-05 Microsoft Corporation Visual representation expression based on player expression
US9159151B2 (en) 2009-07-13 2015-10-13 Microsoft Technology Licensing, Llc Bringing a visual representation to life via learned input from the user
US8264536B2 (en) * 2009-08-25 2012-09-11 Microsoft Corporation Depth-sensitive imaging via polarization-state mapping
US9141193B2 (en) 2009-08-31 2015-09-22 Microsoft Technology Licensing, Llc Techniques for using human gestures to control gesture unaware programs
US8508919B2 (en) * 2009-09-14 2013-08-13 Microsoft Corporation Separation of electrical and optical components
US8330134B2 (en) 2009-09-14 2012-12-11 Microsoft Corporation Optical fault monitoring
US8428340B2 (en) * 2009-09-21 2013-04-23 Microsoft Corporation Screen space plane identification
US8976986B2 (en) * 2009-09-21 2015-03-10 Microsoft Technology Licensing, Llc Volume adjustment based on listener position
US8760571B2 (en) * 2009-09-21 2014-06-24 Microsoft Corporation Alignment of lens and image sensor
US9014546B2 (en) 2009-09-23 2015-04-21 Rovi Guides, Inc. Systems and methods for automatically detecting users within detection regions of media devices
US8452087B2 (en) * 2009-09-30 2013-05-28 Microsoft Corporation Image selection techniques
US8723118B2 (en) * 2009-10-01 2014-05-13 Microsoft Corporation Imager for constructing color and depth images
US7955219B2 (en) * 2009-10-02 2011-06-07 Precor Incorporated Exercise community system
US8613689B2 (en) 2010-09-23 2013-12-24 Precor Incorporated Universal exercise guidance system
US8827870B2 (en) * 2009-10-02 2014-09-09 Precor Incorporated Exercise guidance system
US20110083108A1 (en) * 2009-10-05 2011-04-07 Microsoft Corporation Providing user interface feedback regarding cursor position on a display screen
US8867820B2 (en) 2009-10-07 2014-10-21 Microsoft Corporation Systems and methods for removing a background of an image
US8564534B2 (en) 2009-10-07 2013-10-22 Microsoft Corporation Human tracking system
US7961910B2 (en) 2009-10-07 2011-06-14 Microsoft Corporation Systems and methods for tracking a model
US8963829B2 (en) 2009-10-07 2015-02-24 Microsoft Corporation Methods and systems for determining and tracking extremities of a target
US9400548B2 (en) * 2009-10-19 2016-07-26 Microsoft Technology Licensing, Llc Gesture personalization and profile roaming
US20110099476A1 (en) * 2009-10-23 2011-04-28 Microsoft Corporation Decorating a display environment
US8988432B2 (en) * 2009-11-05 2015-03-24 Microsoft Technology Licensing, Llc Systems and methods for processing an image for target tracking
WO2011060056A2 (en) 2009-11-11 2011-05-19 Niveus Mediacl, Inc. Synergistic muscle activation device
US8843857B2 (en) 2009-11-19 2014-09-23 Microsoft Corporation Distance scalable no touch computing
US9244533B2 (en) 2009-12-17 2016-01-26 Microsoft Technology Licensing, Llc Camera navigation for presentations
US20110151974A1 (en) * 2009-12-18 2011-06-23 Microsoft Corporation Gesture style recognition and reward
US20110150271A1 (en) 2009-12-18 2011-06-23 Microsoft Corporation Motion detection using depth images
US8320621B2 (en) 2009-12-21 2012-11-27 Microsoft Corporation Depth projector system with integrated VCSEL array
SG182780A1 (en) * 2009-12-30 2012-09-27 Fundacion Tecnalia Res & Innovation Apparatus for external activation of paralyzed body parts by stimulation of peripheral nerves
IT1397157B1 (en) * 2010-01-07 2013-01-04 Camerota MACHINE FOR THE PHYSICAL EXERCISE OF A USER.
US9019201B2 (en) * 2010-01-08 2015-04-28 Microsoft Technology Licensing, Llc Evolving universal gesture sets
US8631355B2 (en) * 2010-01-08 2014-01-14 Microsoft Corporation Assigning gesture dictionaries
US9268404B2 (en) * 2010-01-08 2016-02-23 Microsoft Technology Licensing, Llc Application gesture interpretation
US8933884B2 (en) * 2010-01-15 2015-01-13 Microsoft Corporation Tracking groups of users in motion capture system
US8334842B2 (en) 2010-01-15 2012-12-18 Microsoft Corporation Recognizing user intent in motion capture system
US8676581B2 (en) 2010-01-22 2014-03-18 Microsoft Corporation Speech recognition analysis via identification information
US8221292B2 (en) * 2010-01-25 2012-07-17 Precor Incorporated User status notification system
US8265341B2 (en) 2010-01-25 2012-09-11 Microsoft Corporation Voice-body identity correlation
US8864581B2 (en) 2010-01-29 2014-10-21 Microsoft Corporation Visual based identitiy tracking
US8891067B2 (en) * 2010-02-01 2014-11-18 Microsoft Corporation Multiple synchronized optical sources for time-of-flight range finding systems
US8619122B2 (en) * 2010-02-02 2013-12-31 Microsoft Corporation Depth camera compatibility
US8687044B2 (en) * 2010-02-02 2014-04-01 Microsoft Corporation Depth camera compatibility
US8717469B2 (en) * 2010-02-03 2014-05-06 Microsoft Corporation Fast gating photosurface
US8659658B2 (en) * 2010-02-09 2014-02-25 Microsoft Corporation Physical interaction zone for gesture-based user interfaces
US8499257B2 (en) * 2010-02-09 2013-07-30 Microsoft Corporation Handles interactions for human—computer interface
US8633890B2 (en) * 2010-02-16 2014-01-21 Microsoft Corporation Gesture detection based on joint skipping
US20110199302A1 (en) * 2010-02-16 2011-08-18 Microsoft Corporation Capturing screen objects using a collision volume
US8928579B2 (en) * 2010-02-22 2015-01-06 Andrew David Wilson Interacting with an omni-directionally projected display
US8422769B2 (en) * 2010-03-05 2013-04-16 Microsoft Corporation Image segmentation using reduced foreground training data
US8411948B2 (en) 2010-03-05 2013-04-02 Microsoft Corporation Up-sampling binary images for segmentation
US8655069B2 (en) 2010-03-05 2014-02-18 Microsoft Corporation Updating image segmentation following user input
US20110221755A1 (en) * 2010-03-12 2011-09-15 Kevin Geisner Bionic motion
US20110223995A1 (en) 2010-03-12 2011-09-15 Kevin Geisner Interacting with a computer based application
US8279418B2 (en) * 2010-03-17 2012-10-02 Microsoft Corporation Raster scanning for depth detection
US8213680B2 (en) * 2010-03-19 2012-07-03 Microsoft Corporation Proxy training data for human body tracking
US8514269B2 (en) * 2010-03-26 2013-08-20 Microsoft Corporation De-aliasing depth images
US20110234481A1 (en) * 2010-03-26 2011-09-29 Sagi Katz Enhancing presentations using depth sensing cameras
US8523667B2 (en) * 2010-03-29 2013-09-03 Microsoft Corporation Parental control settings based on body dimensions
US8605763B2 (en) 2010-03-31 2013-12-10 Microsoft Corporation Temperature measurement and control for laser and light-emitting diodes
US9646340B2 (en) 2010-04-01 2017-05-09 Microsoft Technology Licensing, Llc Avatar-based virtual dressing room
US9098873B2 (en) 2010-04-01 2015-08-04 Microsoft Technology Licensing, Llc Motion-based interactive shopping environment
US8351651B2 (en) 2010-04-26 2013-01-08 Microsoft Corporation Hand-location post-process refinement in a tracking system
US11117033B2 (en) 2010-04-26 2021-09-14 Wilbert Quinc Murdock Smart system for display of dynamic movement parameters in sports and training
US8379919B2 (en) 2010-04-29 2013-02-19 Microsoft Corporation Multiple centroid condensation of probability distribution clouds
US8284847B2 (en) 2010-05-03 2012-10-09 Microsoft Corporation Detecting motion for a multifunction sensor device
US8885890B2 (en) 2010-05-07 2014-11-11 Microsoft Corporation Depth map confidence filtering
US8498481B2 (en) 2010-05-07 2013-07-30 Microsoft Corporation Image segmentation using star-convexity constraints
US8457353B2 (en) 2010-05-18 2013-06-04 Microsoft Corporation Gestures and gesture modifiers for manipulating a user-interface
US8803888B2 (en) 2010-06-02 2014-08-12 Microsoft Corporation Recognition system for sharing information
US8751215B2 (en) 2010-06-04 2014-06-10 Microsoft Corporation Machine based sign language interpreter
US9008355B2 (en) 2010-06-04 2015-04-14 Microsoft Technology Licensing, Llc Automatic depth camera aiming
US9557574B2 (en) 2010-06-08 2017-01-31 Microsoft Technology Licensing, Llc Depth illumination and detection optics
US8330822B2 (en) 2010-06-09 2012-12-11 Microsoft Corporation Thermally-tuned depth camera light source
US8749557B2 (en) 2010-06-11 2014-06-10 Microsoft Corporation Interacting with user interface via avatar
US9384329B2 (en) 2010-06-11 2016-07-05 Microsoft Technology Licensing, Llc Caloric burn determination from body movement
US8675981B2 (en) 2010-06-11 2014-03-18 Microsoft Corporation Multi-modal gender recognition including depth data
US8982151B2 (en) 2010-06-14 2015-03-17 Microsoft Technology Licensing, Llc Independently processing planes of display data
US8670029B2 (en) 2010-06-16 2014-03-11 Microsoft Corporation Depth camera illuminator with superluminescent light-emitting diode
US8558873B2 (en) 2010-06-16 2013-10-15 Microsoft Corporation Use of wavefront coding to create a depth image
US8296151B2 (en) 2010-06-18 2012-10-23 Microsoft Corporation Compound gesture-speech commands
US8381108B2 (en) 2010-06-21 2013-02-19 Microsoft Corporation Natural user input for driving interactive stories
US8416187B2 (en) 2010-06-22 2013-04-09 Microsoft Corporation Item navigation using motion-capture data
US9272139B2 (en) 2010-07-01 2016-03-01 Marilyn J. Hamilton Universal closed-loop electrical stimulation system
US8827717B2 (en) 2010-07-02 2014-09-09 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Physiologically modulating videogames or simulations which use motion-sensing input devices
US10192173B2 (en) 2010-07-02 2019-01-29 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration System and method for training of state-classifiers
US8493822B2 (en) 2010-07-14 2013-07-23 Adidas Ag Methods, systems, and program products for controlling the playback of music
US10039970B2 (en) 2010-07-14 2018-08-07 Adidas Ag Location-aware fitness monitoring methods, systems, and program products, and applications thereof
US9392941B2 (en) 2010-07-14 2016-07-19 Adidas Ag Fitness monitoring methods, systems, and program products, and applications thereof
US9075434B2 (en) 2010-08-20 2015-07-07 Microsoft Technology Licensing, Llc Translating user motion into multiple object responses
US8613666B2 (en) 2010-08-31 2013-12-24 Microsoft Corporation User selection and navigation based on looped motions
US8437506B2 (en) 2010-09-07 2013-05-07 Microsoft Corporation System for fast, probabilistic skeletal tracking
US20120058824A1 (en) 2010-09-07 2012-03-08 Microsoft Corporation Scalable real-time motion recognition
US8988508B2 (en) 2010-09-24 2015-03-24 Microsoft Technology Licensing, Llc. Wide angle field of view active illumination imaging system
US8681255B2 (en) 2010-09-28 2014-03-25 Microsoft Corporation Integrated low power depth camera and projection device
US8548270B2 (en) 2010-10-04 2013-10-01 Microsoft Corporation Time-of-flight depth imaging
US9484065B2 (en) 2010-10-15 2016-11-01 Microsoft Technology Licensing, Llc Intelligent determination of replays based on event identification
US8592739B2 (en) 2010-11-02 2013-11-26 Microsoft Corporation Detection of configuration changes of an optical element in an illumination system
US8866889B2 (en) 2010-11-03 2014-10-21 Microsoft Corporation In-home depth camera calibration
KR101818092B1 (en) 2010-11-10 2018-01-12 나이키 이노베이트 씨.브이. Systems and methods for time-based athletic activity measurement and display
US8667519B2 (en) 2010-11-12 2014-03-04 Microsoft Corporation Automatic passive and anonymous feedback system
US10726861B2 (en) 2010-11-15 2020-07-28 Microsoft Technology Licensing, Llc Semi-private communication in open environments
US9349040B2 (en) 2010-11-19 2016-05-24 Microsoft Technology Licensing, Llc Bi-modal depth-image analysis
US10234545B2 (en) 2010-12-01 2019-03-19 Microsoft Technology Licensing, Llc Light source module
US8553934B2 (en) 2010-12-08 2013-10-08 Microsoft Corporation Orienting the position of a sensor
US11064910B2 (en) 2010-12-08 2021-07-20 Activbody, Inc. Physical activity monitoring system
US8618405B2 (en) 2010-12-09 2013-12-31 Microsoft Corp. Free-space gesture musical instrument digital interface (MIDI) controller
US8408706B2 (en) 2010-12-13 2013-04-02 Microsoft Corporation 3D gaze tracker
US9171264B2 (en) 2010-12-15 2015-10-27 Microsoft Technology Licensing, Llc Parallel processing machine learning decision tree training
US8884968B2 (en) 2010-12-15 2014-11-11 Microsoft Corporation Modeling an object from image data
US8920241B2 (en) 2010-12-15 2014-12-30 Microsoft Corporation Gesture controlled persistent handles for interface guides
US8448056B2 (en) 2010-12-17 2013-05-21 Microsoft Corporation Validation analysis of human target
US8803952B2 (en) 2010-12-20 2014-08-12 Microsoft Corporation Plural detector time-of-flight depth mapping
US9848106B2 (en) 2010-12-21 2017-12-19 Microsoft Technology Licensing, Llc Intelligent gameplay photo capture
US9823339B2 (en) 2010-12-21 2017-11-21 Microsoft Technology Licensing, Llc Plural anode time-of-flight sensor
US9821224B2 (en) 2010-12-21 2017-11-21 Microsoft Technology Licensing, Llc Driving simulator control with virtual skeleton
US8994718B2 (en) 2010-12-21 2015-03-31 Microsoft Technology Licensing, Llc Skeletal control of three-dimensional virtual world
US8385596B2 (en) 2010-12-21 2013-02-26 Microsoft Corporation First person shooter control with virtual skeleton
US9123316B2 (en) 2010-12-27 2015-09-01 Microsoft Technology Licensing, Llc Interactive content creation
US8488888B2 (en) 2010-12-28 2013-07-16 Microsoft Corporation Classification of posture states
US9256281B2 (en) 2011-01-28 2016-02-09 Empire Technology Development Llc Remote movement guidance
US9349301B2 (en) * 2011-01-28 2016-05-24 Empire Technology Development Llc Sensor-based movement guidance
US8401225B2 (en) 2011-01-31 2013-03-19 Microsoft Corporation Moving object segmentation using depth images
US8401242B2 (en) 2011-01-31 2013-03-19 Microsoft Corporation Real-time camera tracking using depth maps
US8587583B2 (en) 2011-01-31 2013-11-19 Microsoft Corporation Three-dimensional environment reconstruction
US9247238B2 (en) 2011-01-31 2016-01-26 Microsoft Technology Licensing, Llc Reducing interference between multiple infra-red depth cameras
US8724887B2 (en) 2011-02-03 2014-05-13 Microsoft Corporation Environmental modifications to mitigate environmental factors
US8942917B2 (en) 2011-02-14 2015-01-27 Microsoft Corporation Change invariant scene recognition by an agent
US8497838B2 (en) 2011-02-16 2013-07-30 Microsoft Corporation Push actuation of interface controls
US20120323496A1 (en) 2011-02-17 2012-12-20 Nike, Inc. Tracking of User Performance Metrics During a Workout Session
JP5805218B2 (en) 2011-02-17 2015-11-04 ナイキ イノベイト シーブイ Footwear with sensor system
BR112013021141A2 (en) 2011-02-17 2019-12-10 Nike Int Ltd footwear with sensor system
US9381420B2 (en) 2011-02-17 2016-07-05 Nike, Inc. Workout user experience
US9551914B2 (en) 2011-03-07 2017-01-24 Microsoft Technology Licensing, Llc Illuminator with refractive optical element
US9067136B2 (en) 2011-03-10 2015-06-30 Microsoft Technology Licensing, Llc Push personalization of interface controls
US8571263B2 (en) 2011-03-17 2013-10-29 Microsoft Corporation Predicting joint positions
US9533228B2 (en) 2011-03-28 2017-01-03 Brian M. Dugan Systems and methods for fitness and video games
US20120253489A1 (en) 2011-03-28 2012-10-04 Dugan Brian M Systems and methods for fitness and video games
US9610506B2 (en) 2011-03-28 2017-04-04 Brian M. Dugan Systems and methods for fitness and video games
US9470778B2 (en) 2011-03-29 2016-10-18 Microsoft Technology Licensing, Llc Learning from high quality depth measurements
US9298287B2 (en) 2011-03-31 2016-03-29 Microsoft Technology Licensing, Llc Combined activation for natural user interface systems
US9842168B2 (en) 2011-03-31 2017-12-12 Microsoft Technology Licensing, Llc Task driven user intents
US10642934B2 (en) 2011-03-31 2020-05-05 Microsoft Technology Licensing, Llc Augmented conversational understanding architecture
US9760566B2 (en) 2011-03-31 2017-09-12 Microsoft Technology Licensing, Llc Augmented conversational understanding agent to identify conversation context between two humans and taking an agent action thereof
US8503494B2 (en) 2011-04-05 2013-08-06 Microsoft Corporation Thermal management system
US20120258433A1 (en) 2011-04-05 2012-10-11 Adidas Ag Fitness Monitoring Methods, Systems, And Program Products, And Applications Thereof
US8824749B2 (en) 2011-04-05 2014-09-02 Microsoft Corporation Biometric recognition
US8620113B2 (en) 2011-04-25 2013-12-31 Microsoft Corporation Laser diode modes
US9259643B2 (en) 2011-04-28 2016-02-16 Microsoft Technology Licensing, Llc Control of separate computer game elements
US8702507B2 (en) 2011-04-28 2014-04-22 Microsoft Corporation Manual and camera-based avatar control
US10671841B2 (en) 2011-05-02 2020-06-02 Microsoft Technology Licensing, Llc Attribute state classification
US8888331B2 (en) 2011-05-09 2014-11-18 Microsoft Corporation Low inductance light source module
US9064006B2 (en) 2012-08-23 2015-06-23 Microsoft Technology Licensing, Llc Translating natural language utterances to keyword search queries
US9137463B2 (en) 2011-05-12 2015-09-15 Microsoft Technology Licensing, Llc Adaptive high dynamic range camera
US8788973B2 (en) 2011-05-23 2014-07-22 Microsoft Corporation Three-dimensional gesture controlled avatar configuration interface
US8760395B2 (en) 2011-05-31 2014-06-24 Microsoft Corporation Gesture recognition techniques
US9594430B2 (en) 2011-06-01 2017-03-14 Microsoft Technology Licensing, Llc Three-dimensional foreground selection for vision system
US8526734B2 (en) 2011-06-01 2013-09-03 Microsoft Corporation Three-dimensional background removal for vision system
US8947226B2 (en) 2011-06-03 2015-02-03 Brian M. Dugan Bands for measuring biometric information
US10796494B2 (en) 2011-06-06 2020-10-06 Microsoft Technology Licensing, Llc Adding attributes to virtual representations of real-world objects
US9208571B2 (en) 2011-06-06 2015-12-08 Microsoft Technology Licensing, Llc Object digitization
US8597142B2 (en) 2011-06-06 2013-12-03 Microsoft Corporation Dynamic camera based practice mode
US9724600B2 (en) 2011-06-06 2017-08-08 Microsoft Technology Licensing, Llc Controlling objects in a virtual environment
US8929612B2 (en) 2011-06-06 2015-01-06 Microsoft Corporation System for recognizing an open or closed hand
US9098110B2 (en) 2011-06-06 2015-08-04 Microsoft Technology Licensing, Llc Head rotation tracking from depth-based center of mass
US9013489B2 (en) 2011-06-06 2015-04-21 Microsoft Technology Licensing, Llc Generation of avatar reflecting player appearance
US8897491B2 (en) 2011-06-06 2014-11-25 Microsoft Corporation System for finger recognition and tracking
US9597587B2 (en) 2011-06-08 2017-03-21 Microsoft Technology Licensing, Llc Locational node device
US8786730B2 (en) 2011-08-18 2014-07-22 Microsoft Corporation Image exposure using exclusion regions
US20130089843A1 (en) * 2011-10-07 2013-04-11 Ikkos, Llc Training motion with observation-based feedback
US9557836B2 (en) 2011-11-01 2017-01-31 Microsoft Technology Licensing, Llc Depth image compression
US9117281B2 (en) 2011-11-02 2015-08-25 Microsoft Corporation Surface segmentation from RGB and depth images
US8854426B2 (en) 2011-11-07 2014-10-07 Microsoft Corporation Time-of-flight camera with guided light
US8724906B2 (en) 2011-11-18 2014-05-13 Microsoft Corporation Computing pose and/or shape of modifiable entities
US20130130870A1 (en) * 2011-11-23 2013-05-23 Ian McCranor Martial arts training and scoring gear
US8509545B2 (en) 2011-11-29 2013-08-13 Microsoft Corporation Foreground subject detection
US8803800B2 (en) 2011-12-02 2014-08-12 Microsoft Corporation User interface control based on head orientation
US8635637B2 (en) 2011-12-02 2014-01-21 Microsoft Corporation User interface presenting an animated avatar performing a media reaction
US9100685B2 (en) 2011-12-09 2015-08-04 Microsoft Technology Licensing, Llc Determining audience state or interest using passive sensor data
US8971612B2 (en) 2011-12-15 2015-03-03 Microsoft Corporation Learning image processing tasks from scene reconstructions
US8879831B2 (en) 2011-12-15 2014-11-04 Microsoft Corporation Using high-level attributes to guide image processing
US8630457B2 (en) 2011-12-15 2014-01-14 Microsoft Corporation Problem states for pose tracking pipeline
US8811938B2 (en) 2011-12-16 2014-08-19 Microsoft Corporation Providing a user interface experience based on inferred vehicle state
US9342139B2 (en) 2011-12-19 2016-05-17 Microsoft Technology Licensing, Llc Pairing a computing device to a user
US9339691B2 (en) 2012-01-05 2016-05-17 Icon Health & Fitness, Inc. System and method for controlling an exercise device
US9720089B2 (en) 2012-01-23 2017-08-01 Microsoft Technology Licensing, Llc 3D zoom imager
US20130213146A1 (en) 2012-02-22 2013-08-22 Nike, Inc. Footwear Having Sensor System
US8739639B2 (en) 2012-02-22 2014-06-03 Nike, Inc. Footwear having sensor system
US11071344B2 (en) 2012-02-22 2021-07-27 Nike, Inc. Motorized shoe with gesture control
US11684111B2 (en) 2012-02-22 2023-06-27 Nike, Inc. Motorized shoe with gesture control
US20130213147A1 (en) 2012-02-22 2013-08-22 Nike, Inc. Footwear Having Sensor System
US9367668B2 (en) 2012-02-28 2016-06-14 Precor Incorporated Dynamic fitness equipment user interface adjustment
US8898687B2 (en) 2012-04-04 2014-11-25 Microsoft Corporation Controlling a media program based on a media reaction
US9681836B2 (en) 2012-04-23 2017-06-20 Cyberonics, Inc. Methods, systems and apparatuses for detecting seizure and non-seizure states
US9210401B2 (en) 2012-05-03 2015-12-08 Microsoft Technology Licensing, Llc Projected visual cues for guiding physical movement
CA2775700C (en) 2012-05-04 2013-07-23 Microsoft Corporation Determining a future portion of a currently presented media program
US10102345B2 (en) 2012-06-19 2018-10-16 Activbody, Inc. Personal wellness management platform
US10133849B2 (en) 2012-06-19 2018-11-20 Activbody, Inc. Merchandizing, socializing, and/or gaming via a personal wellness device and/or a personal wellness platform
US9230064B2 (en) 2012-06-19 2016-01-05 EZ as a Drink Productions, Inc. Personal wellness device
JP6018707B2 (en) 2012-06-21 2016-11-02 マイクロソフト コーポレーション Building an avatar using a depth camera
US9630093B2 (en) 2012-06-22 2017-04-25 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and system for physiologically modulating videogames and simulations which use gesture and body image sensing control input devices
US9836590B2 (en) 2012-06-22 2017-12-05 Microsoft Technology Licensing, Llc Enhanced accuracy of user presence status determination
US9696427B2 (en) 2012-08-14 2017-07-04 Microsoft Technology Licensing, Llc Wide angle depth detection
US8882310B2 (en) 2012-12-10 2014-11-11 Microsoft Corporation Laser die light source module with low inductance
US9043004B2 (en) 2012-12-13 2015-05-26 Nike, Inc. Apparel having sensor system
US9857470B2 (en) 2012-12-28 2018-01-02 Microsoft Technology Licensing, Llc Using photometric stereo for 3D environment modeling
US20140200116A1 (en) * 2013-01-17 2014-07-17 Alex Aquatics Real Time Feedback Swim Training System and Method Based on Instantaneous Speed
US9251590B2 (en) 2013-01-24 2016-02-02 Microsoft Technology Licensing, Llc Camera pose estimation for 3D reconstruction
US9743861B2 (en) 2013-02-01 2017-08-29 Nike, Inc. System and method for analyzing athletic activity
US10926133B2 (en) 2013-02-01 2021-02-23 Nike, Inc. System and method for analyzing athletic activity
US11006690B2 (en) 2013-02-01 2021-05-18 Nike, Inc. System and method for analyzing athletic activity
US9052746B2 (en) 2013-02-15 2015-06-09 Microsoft Technology Licensing, Llc User center-of-mass and mass distribution extraction using depth images
US9940553B2 (en) 2013-02-22 2018-04-10 Microsoft Technology Licensing, Llc Camera/object pose from predicted coordinates
US9135516B2 (en) 2013-03-08 2015-09-15 Microsoft Technology Licensing, Llc User body angle, curvature and average extremity positions extraction using depth images
US9092657B2 (en) 2013-03-13 2015-07-28 Microsoft Technology Licensing, Llc Depth image processing
US9274606B2 (en) 2013-03-14 2016-03-01 Microsoft Technology Licensing, Llc NUI video conference controls
EP2969058B1 (en) 2013-03-14 2020-05-13 Icon Health & Fitness, Inc. Strength training apparatus with flywheel and related methods
US9279734B2 (en) 2013-03-15 2016-03-08 Nike, Inc. System and method for analyzing athletic activity
US9720443B2 (en) 2013-03-15 2017-08-01 Nike, Inc. Wearable device assembly having athletic functionality
US9953213B2 (en) 2013-03-27 2018-04-24 Microsoft Technology Licensing, Llc Self discovery of autonomous NUI devices
US9229476B2 (en) 2013-05-08 2016-01-05 EZ as a Drink Productions, Inc. Personal handheld electronic device with a touchscreen on a peripheral surface
US9442186B2 (en) 2013-05-13 2016-09-13 Microsoft Technology Licensing, Llc Interference reduction for TOF systems
US9262064B2 (en) 2013-07-09 2016-02-16 EZ as a Drink Productions, Inc. Handheld computing platform with integrated pressure sensor and associated methods of use
US9462253B2 (en) 2013-09-23 2016-10-04 Microsoft Technology Licensing, Llc Optical modules that reduce speckle contrast and diffraction artifacts
US9443310B2 (en) 2013-10-09 2016-09-13 Microsoft Technology Licensing, Llc Illumination modules that emit structured light
US9674563B2 (en) 2013-11-04 2017-06-06 Rovi Guides, Inc. Systems and methods for recommending content
US9769459B2 (en) 2013-11-12 2017-09-19 Microsoft Technology Licensing, Llc Power efficient laser diode driver circuit and method
US9508385B2 (en) 2013-11-21 2016-11-29 Microsoft Technology Licensing, Llc Audio-visual project generator
EP3623020A1 (en) 2013-12-26 2020-03-18 Icon Health & Fitness, Inc. Magnetic resistance mechanism in a cable machine
US9971491B2 (en) 2014-01-09 2018-05-15 Microsoft Technology Licensing, Llc Gesture library for natural user input
US20150237927A1 (en) * 2014-02-22 2015-08-27 Jan Nelson Temperature Controlled Personal Environment
US9739883B2 (en) 2014-05-16 2017-08-22 Elwha Llc Systems and methods for ultrasonic velocity and acceleration detection
WO2015138339A1 (en) 2014-03-10 2015-09-17 Icon Health & Fitness, Inc. Pressure sensor to quantify work
US9618618B2 (en) 2014-03-10 2017-04-11 Elwha Llc Systems and methods for ultrasonic position and motion detection
US9437002B2 (en) 2014-09-25 2016-09-06 Elwha Llc Systems and methods for a dual modality sensor system
US10124246B2 (en) 2014-04-21 2018-11-13 Activbody, Inc. Pressure sensitive peripheral devices, and associated methods of use
WO2015191445A1 (en) 2014-06-09 2015-12-17 Icon Health & Fitness, Inc. Cable system incorporated into a treadmill
WO2015195965A1 (en) 2014-06-20 2015-12-23 Icon Health & Fitness, Inc. Post workout massage device
GB2546194A (en) 2014-09-29 2017-07-12 Equinox Holdings Inc Exercise class apparatus and method
US10258828B2 (en) 2015-01-16 2019-04-16 Icon Health & Fitness, Inc. Controls for an exercise device
US9937402B2 (en) 2015-01-30 2018-04-10 Eras Roy Noel, III Speedbag performance monitor
US10391361B2 (en) 2015-02-27 2019-08-27 Icon Health & Fitness, Inc. Simulating real-world terrain on an exercise device
US10380656B2 (en) 2015-02-27 2019-08-13 Ebay Inc. Dynamic predefined product reviews
CA2989417A1 (en) 2015-04-20 2016-10-27 Michael V. SCHAEFER Apparatus and method for increased realism of training on exercise machines
US9995823B2 (en) 2015-07-31 2018-06-12 Elwha Llc Systems and methods for utilizing compressed sensing in an entertainment system
US10953305B2 (en) 2015-08-26 2021-03-23 Icon Health & Fitness, Inc. Strength exercise mechanisms
US10412280B2 (en) 2016-02-10 2019-09-10 Microsoft Technology Licensing, Llc Camera with light valve over sensor array
US10257932B2 (en) 2016-02-16 2019-04-09 Microsoft Technology Licensing, Llc. Laser diode chip on printed circuit board
US10462452B2 (en) 2016-03-16 2019-10-29 Microsoft Technology Licensing, Llc Synchronizing active illumination cameras
US10625137B2 (en) 2016-03-18 2020-04-21 Icon Health & Fitness, Inc. Coordinated displays in an exercise device
US10272317B2 (en) 2016-03-18 2019-04-30 Icon Health & Fitness, Inc. Lighted pace feature in a treadmill
US10493349B2 (en) 2016-03-18 2019-12-03 Icon Health & Fitness, Inc. Display on exercise device
US10561894B2 (en) 2016-03-18 2020-02-18 Icon Health & Fitness, Inc. Treadmill with removable supports
US10293211B2 (en) 2016-03-18 2019-05-21 Icon Health & Fitness, Inc. Coordinated weight selection
US10252109B2 (en) 2016-05-13 2019-04-09 Icon Health & Fitness, Inc. Weight platform treadmill
US10441844B2 (en) 2016-07-01 2019-10-15 Icon Health & Fitness, Inc. Cooling systems and methods for exercise equipment
US10471299B2 (en) 2016-07-01 2019-11-12 Icon Health & Fitness, Inc. Systems and methods for cooling internal exercise equipment components
KR102196380B1 (en) 2016-07-21 2020-12-30 매직 립, 인코포레이티드 Technology for controlling a virtual image generation system using user's emotional states
US10671705B2 (en) 2016-09-28 2020-06-02 Icon Health & Fitness, Inc. Customizing recipe recommendations
US10500473B2 (en) 2016-10-10 2019-12-10 Icon Health & Fitness, Inc. Console positioning
US10376736B2 (en) 2016-10-12 2019-08-13 Icon Health & Fitness, Inc. Cooling an exercise device during a dive motor runway condition
US11141092B2 (en) 2016-10-19 2021-10-12 United States Of America As Represented By The Administrator Of Nasa Method and system for incorporating physiological self-regulation challenge into geospatial scenario games and/or simulations
US10661114B2 (en) 2016-11-01 2020-05-26 Icon Health & Fitness, Inc. Body weight lift mechanism on treadmill
TWI646997B (en) 2016-11-01 2019-01-11 美商愛康運動與健康公司 Distance sensor for console positioning
TWI680782B (en) 2016-12-05 2020-01-01 美商愛康運動與健康公司 Offsetting treadmill deck weight during operation
USD1010028S1 (en) 2017-06-22 2024-01-02 Boost Treadmills, LLC Unweighting exercise treadmill
TWI722450B (en) 2017-08-16 2021-03-21 美商愛康運動與健康公司 System for opposing axial impact loading in a motor
KR101866483B1 (en) * 2017-09-26 2018-07-19 주식회사 와이앤제이바이오 Electric stimulated indoor bike
US11654327B2 (en) 2017-10-31 2023-05-23 Alterg, Inc. System for unweighting a user and related methods of exercise
US10729965B2 (en) 2017-12-22 2020-08-04 Icon Health & Fitness, Inc. Audible belt guide in a treadmill
US11040246B2 (en) 2018-02-06 2021-06-22 Adidas Ag Increasing accuracy in workout autodetection systems and methods
US11051829B2 (en) 2018-06-26 2021-07-06 DePuy Synthes Products, Inc. Customized patient-specific orthopaedic surgical instrument
AU2019388605A1 (en) 2018-09-04 2021-02-18 Hvrt Corp. Reticles, methods of use and manufacture
US11364419B2 (en) 2019-02-21 2022-06-21 Scott B. Radow Exercise equipment with music synchronization
US11872433B2 (en) 2020-12-01 2024-01-16 Boost Treadmills, LLC Unweighting enclosure, system and method for an exercise device
CN113398468A (en) * 2021-06-16 2021-09-17 北京金林高科科技有限公司 Control system and method of EMS wearable device and wearable device
US20230115873A1 (en) 2021-10-12 2023-04-13 Boost Treadmills, LLC DAP Platform, Integrated Lifts, System and Related Devices and Methods
US11806577B1 (en) 2023-02-17 2023-11-07 Mad Dogg Athletics, Inc. Programmed exercise bicycle with computer aided guidance

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5489249A (en) * 1991-07-02 1996-02-06 Proform Fitness Products, Inc. Video exercise control system
US5492514A (en) * 1993-05-14 1996-02-20 Daum Electronic Gmbh Home trainer with biofeedback
US5527239A (en) * 1993-02-04 1996-06-18 Abbondanza; James M. Pulse rate controlled exercise system

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4148321A (en) * 1973-11-26 1979-04-10 Wyss Oscar A M Apparatuses and methods for therapeutic treatment and active massages of muscles
US4724842A (en) * 1982-05-19 1988-02-16 Charters Thomas H Method and apparatus for muscle stimulation
US4569352A (en) * 1983-05-13 1986-02-11 Wright State University Feedback control system for walking
US4665928A (en) * 1983-08-10 1987-05-19 Orthotronics, Inc. Range of motion measuring and displaying device
US4620543A (en) * 1984-06-15 1986-11-04 Richards Medical Company Enhanced fracture healing and muscle exercise through defined cycles of electric stimulation
US5277197A (en) * 1986-12-08 1994-01-11 Physical Health Device, Inc. Microprocessor controlled system for unsupervised EMG feedback and exercise training
US4838272A (en) * 1987-08-19 1989-06-13 The Regents Of The University Of California Method and apparatus for adaptive closed loop electrical stimulation of muscles
US4947836A (en) * 1987-09-21 1990-08-14 Hillcrest Medical Center Exerciser with muscle stimulation
US4809696A (en) * 1987-09-21 1989-03-07 Hillcrest Medical Center Functional electrical stimulation synchronizer switch
US4863157A (en) * 1988-04-29 1989-09-05 State University Of New York Method and apparatus for exercising a paralyzed limb
US5503149A (en) * 1990-07-09 1996-04-02 Beavin; William C. Computer simulation of live organ using arthroscopic and/or laparoscopic data
US5273038A (en) * 1990-07-09 1993-12-28 Beavin William C Computer simulation of live organ
US5328424A (en) * 1993-03-19 1994-07-12 Greco Bruce C Upper and lower body exerciser that can be used by people with lower body paralysis
US5549656A (en) * 1993-08-16 1996-08-27 Med Serve Group, Inc. Combination neuromuscular stimulator and electromyograph system
US5549646A (en) * 1994-12-06 1996-08-27 Pacesetter, Inc. Periodic electrical lead intergrity testing system and method for implantable cardiac stimulating devices

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5489249A (en) * 1991-07-02 1996-02-06 Proform Fitness Products, Inc. Video exercise control system
US5527239A (en) * 1993-02-04 1996-06-18 Abbondanza; James M. Pulse rate controlled exercise system
US5492514A (en) * 1993-05-14 1996-02-20 Daum Electronic Gmbh Home trainer with biofeedback

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP0840638A4 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008055949A1 (en) * 2006-11-10 2008-05-15 Khymeia S.R.L. Apparatus for motor training and exercise of the human body

Also Published As

Publication number Publication date
EP0840638A4 (en) 2002-10-09
US6066075A (en) 2000-05-23
US5702323A (en) 1997-12-30
AU6548296A (en) 1997-02-26
EP0840638A1 (en) 1998-05-13
DE69634915D1 (en) 2005-08-11
EP0840638B1 (en) 2005-07-06

Similar Documents

Publication Publication Date Title
US5702323A (en) Electronic exercise enhancer
US8503086B2 (en) System and method for tracking and assessing movement skills in multidimensional space
US6308565B1 (en) System and method for tracking and assessing movement skills in multidimensional space
US5913727A (en) Interactive movement and contact simulation game
US6749432B2 (en) Education system challenging a subject's physiologic and kinesthetic systems to synergistically enhance cognitive function
US9046919B2 (en) Wearable user interface device, system, and method of use
US6073489A (en) Testing and training system for assessing the ability of a player to complete a task
EP1059970A2 (en) System and method for tracking and assessing movement skills in multidimensional space
US8620146B1 (en) Picture-in-picture video system for virtual exercise, instruction and entertainment
US10007332B2 (en) Electronic gaming machine in communicative control with avatar display from motion-capture system
KR20070095407A (en) Method and system for athletic motion analysis and instruction
US11083967B1 (en) Virtual reality haptic system and apparatus
CN112642133B (en) Rehabilitation training system based on virtual reality
JP2003180896A5 (en)
WO1998003110A2 (en) Motion measurement apparatus
Liebermann et al. The use of feedback-based technologies
WO2000057387A1 (en) Virtual-reality exercise system and method
RU104852U1 (en) SPORT LOAD CONTROL SYSTEM AND SPORTS SIMULATOR FOR TRAINING OR COMPETITIONS
CN113017615A (en) Virtual interactive motion auxiliary system and method based on inertial motion capture equipment
Katz et al. Virtual reality
RU2106695C1 (en) Method for representation of virtual space for user and device which implements said method
WO2001029799A2 (en) Education system challenging a subject's physiologic and kinesthetic systems to synergistically enhance cognitive function
Emanuele et al. Skills Accelerators Using Multimodal Systems for Boxing Training
Katz et al. Sport Technology Research Laboratory, University of Calgary
WO2006008704A1 (en) Exercise system and method

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AU AZ BB BG BR BY CA CH CN CZ DE DK EE ES FI GB GE HU IL IS JP KE KG KP KR KZ LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK TJ TM TR TT UA UG US UZ VN AM AZ BY KG KZ MD RU TJ TM

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): KE LS MW SD SZ UG AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 1996925358

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1996925358

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

NENP Non-entry into the national phase

Ref country code: CA

WWG Wipo information: grant in national office

Ref document number: 1996925358

Country of ref document: EP