[go: up one dir, main page]

US20190160684A1 - Massage Robot Using Machine Vision - Google Patents

Massage Robot Using Machine Vision Download PDF

Info

Publication number
US20190160684A1
US20190160684A1 US15/826,569 US201715826569A US2019160684A1 US 20190160684 A1 US20190160684 A1 US 20190160684A1 US 201715826569 A US201715826569 A US 201715826569A US 2019160684 A1 US2019160684 A1 US 2019160684A1
Authority
US
United States
Prior art keywords
massage
user
robot
massage robot
robotic arm
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/826,569
Other versions
US10434658B2 (en
Inventor
Haisong Gu
Dongyan Wang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Midea Group Co Ltd
Original Assignee
Midea Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Midea Group Co Ltd filed Critical Midea Group Co Ltd
Priority to US15/826,569 priority Critical patent/US10434658B2/en
Assigned to MIDEA GROUP CO., LTD. reassignment MIDEA GROUP CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GU, HAISONG, WANG, DONGYAN
Priority to CN201880073575.2A priority patent/CN111343958A/en
Priority to PCT/CN2018/114024 priority patent/WO2019105183A1/en
Publication of US20190160684A1 publication Critical patent/US20190160684A1/en
Application granted granted Critical
Publication of US10434658B2 publication Critical patent/US10434658B2/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/008Manipulators for service tasks
    • B25J11/009Nursing, e.g. carrying sick persons, pushing wheelchairs, distributing drugs
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J5/00Manipulators mounted on wheels or on carriages
    • B25J5/007Manipulators mounted on wheels or on carriages mounted on wheels
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H23/00Percussion or vibration massage, e.g. using supersonic vibration; Suction-vibration massage; Massage with moving diaphragms
    • A61H23/006Percussion or tapping massage
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H39/00Devices for locating or stimulating specific reflex points of the body for physical therapy, e.g. acupuncture
    • A61H39/02Devices for locating such points
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H7/00Devices for suction-kneading massage; Devices for massaging the skin by rubbing or brushing not otherwise provided for
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/0005Manipulators having means for high-level communication with users, e.g. speech generator, face recognition means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/008Manipulators for service tasks
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J13/00Controls for manipulators
    • B25J13/006Controls for manipulators by means of a wireless system for controlling one or several manipulators
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J13/00Controls for manipulators
    • B25J13/08Controls for manipulators by means of sensing devices, e.g. viewing or touching devices
    • B25J13/081Touching devices, e.g. pressure-sensitive
    • B25J13/084Tactile sensors
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J13/00Controls for manipulators
    • B25J13/08Controls for manipulators by means of sensing devices, e.g. viewing or touching devices
    • B25J13/085Force or torque sensors
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J13/00Controls for manipulators
    • B25J13/08Controls for manipulators by means of sensing devices, e.g. viewing or touching devices
    • B25J13/087Controls for manipulators by means of sensing devices, e.g. viewing or touching devices for sensing other physical parameters, e.g. electrical or chemical properties
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J19/00Accessories fitted to manipulators, e.g. for monitoring, for viewing; Safety devices combined with or specially adapted for use in connection with manipulators
    • B25J19/02Sensing devices
    • B25J19/021Optical sensing devices
    • B25J19/023Optical sensing devices including video camera means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/0003Home robots, i.e. small robots for domestic use
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1694Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
    • B25J9/1697Vision controlled systems
    • G06K9/00255
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/166Detection; Localisation; Normalisation using acquisition arrangements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/16Physical interface with patient
    • A61H2201/1657Movement of interface, i.e. force application means
    • A61H2201/1659Free spatial automatic movement of interface within a working area, e.g. Robot
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5002Means for controlling a set of similar massage devices acting in sequence at different locations on a patient
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5007Control means thereof computer controlled
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5023Interfaces to the user
    • A61H2201/5038Interfaces to the user freely programmable by the user
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5023Interfaces to the user
    • A61H2201/5048Audio interfaces, e.g. voice or music controlled
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5058Sensors or detectors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5058Sensors or detectors
    • A61H2201/5064Position sensors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5058Sensors or detectors
    • A61H2201/5071Pressure sensors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5058Sensors or detectors
    • A61H2201/5082Temperature sensors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H2201/00Characteristics of apparatus not provided for in the preceding codes
    • A61H2201/50Control means thereof
    • A61H2201/5097Control means thereof wireless
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S901/00Robots
    • Y10S901/02Arm motion controller
    • Y10S901/09Closed loop, sensor feedback controls arm movement
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S901/00Robots
    • Y10S901/02Arm motion controller
    • Y10S901/09Closed loop, sensor feedback controls arm movement
    • Y10S901/10Sensor physically contacts and follows work contour

Definitions

  • This disclosure relates generally to massage robots.
  • the masseuse may not perform massage according to the person's personal preferences or to address the person's specific idiosyncrasies or problem spots.
  • a massage robot that uses machine vision to locate treatment spots for massage.
  • a massage robot includes one or more robotic arms, an image sensor and a control system.
  • the image sensor captures images of the user.
  • the control system includes an image processing module and a motion controller.
  • the image processing module processes the images to locate a treatment spot on the user.
  • the motion controller controls the robotic arm to perform a massage procedure on the identified treatment spot.
  • the massage can be personalized by selecting the massage procedure based on the identity of the user, for example based on the user's past history of massages, on the user's personal preferences or on instructions from a massage professional treating the user.
  • Sensors other than machine vision can also be used. Examples include ultrasound sensors, touch sensors, pressure sensors, and temperature sensors.
  • the robotic arm can be fitted with different end effectors, which are selected based on the desired massage procedure.
  • FIG. 1 is a block diagram of a massage robot, according to an embodiment.
  • FIG. 2 shows a user receiving a massage from a mobile, home assistant robot.
  • FIG. 3 shows a user receiving a massage from a robot fixed to a massage table.
  • FIG. 4 is a block diagram of a residential environment including a massage robot, according to an embodiment.
  • FIG. 1 is a block diagram of a massage robot 100 , according to an embodiment.
  • the robot 100 has one or more robotic arms 110 which are guided by a control system 150 that uses machine vision to identify a treatment spot for the massage procedure.
  • the massage robot 100 includes an image sensor 130 (e.g., a camera), an image processing module 132 and a motion controller 155 .
  • the image sensor 130 captures images of the user—the person receiving the massage.
  • the image processing module 132 processes these images to locate the treatment spot, and the motion controller 155 controls the robotic arm 110 to perform the desired massage procedure on the identified treatment spot.
  • robotic arms can be used, with different degrees of freedom.
  • the robotic arm will have from 3-7 degrees of freedom.
  • Example degrees of freedom include shoulder, elbow, wrist, finger/hand manipulators, and telescoping upper arm or forearm.
  • the control system 150 may also select the appropriate end effector and control the robotic arm to automatically change to the correct end effector.
  • end effectors include those with fingers (typically from one to five, with or without thumb) or pad without fingers.
  • the massage robot 100 may also include other sensors 135 .
  • sensors 135 include an ultrasound sensor to capture ultrasound images, a pressure sensor to sense the pressure exerted by the robotic arm on the treatment spot, a touch sensor to detect the hardness or firmness of the treatment spot, and a temperature sensor to detect the temperature of the treatment spot. Information from these sensors can also be processed 137 by the control system 150 and used by the motion controller 155 to control the robotic arm.
  • a personalization module 140 is used to personalize the massage procedure for the user. It first identifies the user. For example, the user may be identified when he logs into his user account for the massage robot or for a home network connected to the robot. Alternately, the user may be identified by face recognition or other techniques. Once the user is identified, the personalization module 140 determines the appropriate massage procedure based on the user's identity. For example, the massage procedure may be selected based on the user's profile 142 . Perhaps the user has specified certain types of massage on specific days of the week. Alternately, the massage procedure may be defined by a massage professional treating the user. The professional's instructions specifying certain types of massage for the user may be stored somewhere and retrieved by the massage robot 100 .
  • the massage robot 100 may determine or suggest a massage procedure based on the user's prior history 143 of massages. If in the past the user has received upper back massage on Tuesday and Thursday and whole body massage on Saturday, and today is Saturday, then the massage robot 100 may assume whole body massage or may ask the user whether he would like a whole body massage.
  • the massage procedure may also be selected based on the user's medical record, such as a record of back pain at certain locations.
  • the massage procedure may also be determined based on the user's pose. If the user is lying face down with exposed back, legs and arms, the massage robot 100 may determine that the user would like a whole body massage. If the user is sitting with back facing the robot 100 , the robot may determine that some sort of back massage is appropriate.
  • the massage robot 100 also includes a user interface 160 , which in this example includes speech synthesis 162 and voice recognition 164 .
  • the speech synthesis module 162 is used to give instructions and otherwise communicate to the user.
  • the voice recognition module 164 is used to receive voice commands or other communications.
  • FIG. 2 shows an example of a massage robot 200 preparing to deliver a massage to a user.
  • Jane User comes home from work on Wednesday, sits down to watch TV and asks her home robot assistant for a beer.
  • the massage robot is a mobile, more general purpose home assistant robot.
  • the robot can also perform other household tasks, such as cleaning, retrieving items, organizing and putting away items.
  • Assistant retrieves a beer from the refrigerator and delivers it to User.
  • Assistant's camera captures images of User, and determines that she is sitting in her favorite chair and watching TV.
  • Assistant also knows that User has requested a beer. Based on this pattern and User's past history, Assistant asks, “I see today is Wednesday and you look tired. Would you like a back massage?”
  • Assistant determines from its captured images that User is sitting in the massage chair and that the back is exposed ready for massage, but the pose is not quite right. Assistant instructs, “Please sit straight and move slightly to the left.” Assistant also determines that User is still drinking beer. Assistant warns, “Drinking while receiving massage is not recommended. Please put down your beer.”
  • Assistant determines that User's pose is now ready for massage. From User's profile and past history, Assistant selects a general 15-minute massage of the upper back and also notices that User recently has had an especially tight right trapezius. Assistant: “I will perform your usual upper back massage. Is your right side still bothering you?”
  • the vision system guides the end effector of the robotic arm to carry out a specific pattern of massage and/or to address specific problem areas.
  • the image processing module 132 determines the location of the upper back, possibly including anatomical landmarks such as the spinal column, neck and trapezius muscles. Massage is applied to different treatment spots based on the captured images. The massage may be applied with different size pressure areas and different amounts of pressure.
  • the motion controller 155 preferably controls the location of the massage, the pressure of the massage, and the speed and/or length of massage strokes.
  • Assistant changes the end effector of the robotic arm to a long broad end effector and begins the massage by applying massage oil to User's back and stroking from the spine outwards in order to emulate a general back massage by a masseuse's forearm.
  • Assistant then changes to a tennis-ball sized end effector with more of a pressuring motion rather than a stroking motion to emulate pressure by a masseuse's palm.
  • the amount of pressure is based on the pressures recorded for User's last massage.
  • Assistant works a pattern of palm pressures across User's back. During this pattern, sensors in the end effector detect a small knot on the left side and note the location for later treatment.
  • the image sensor is a thermal camera
  • the image processing module 132 may process the thermal images to identify unusually hot spots or temperature peaks for treatment.
  • Assistant then changes to a smaller size end effector to emulate a more localized pressure massage by fingers and thumbs.
  • Assistant starts by applying pressure to both sides of the spinal column, working up and down the spinal column. It then massages specific spots, including the identified knot.
  • Assistant applies concentrated pressure to the knot area. User yells, “Ouch. That hurts.”
  • the Assistant increases the temperature of the end effector and finishes treatment of the knot in this manner.
  • the massage time may be a predetermined time period: five minutes for example.
  • the Assistant may massage the knot until sensors indicate the knot has been released or until User instructs the Assistant to stop.
  • Assistant then moves to the right trapezius.
  • Assistant changes to opposing end effectors.
  • the opposing end effectors may be either on the same robotic arm or on different robotic arms.
  • Opposing end effectors may be used to emulate pinching or squeezing motions.
  • Assistant uses the opposing end effectors to massage the right trapezius.
  • FIG. 3 shows another example.
  • the massage robot 300 is stationary. It includes a set of robotic arms that are stationary relative to a massage table. The user lies on the massage table and the robotic arms perform the massage. The arms may work simultaneously, with different arms performing different procedures. For example, some arms may provide massage for the back while others provide massage of the legs or arms. Alternately, some arms may provide general massage of an area while others massage specific treatment spots.
  • the massage robot 300 may be used in professional settings, such as at a massage studio or doctor's office, in addition to providing home massage.
  • FIG. 4 is a block diagram of a residential environment that includes a massage robot, according to an embodiment.
  • the residential environment 400 is an environment designed for people to live in.
  • the residential environment 400 can be a dwelling, such as a house, a condo, an apartment, or a dormitory.
  • the residential environment 400 includes home devices 410 A-N, including the massage robot described above. It also includes a home device network 420 connecting the home devices 410 , and a resident profiles database 430 that contains residents' preferences for the home devices.
  • the components in FIG. 4 are shown as separate blocks but they may be combined depending on the implementation.
  • the resident profiles 430 may be part of the home devices 410 .
  • the residential environment 400 may include a hub for the network 420 .
  • the hub may also control the home devices 410 .
  • the network 420 may also provide access to external devices, such as cloud-based services.
  • the home devices 410 are household devices that are made available to the different persons associated with the residential environment 400 .
  • Examples of other home devices 410 include HVAC devices (e.g., air conditioner, heater, air venting), lighting, powered window and door treatments (e.g., door locks, power blinds and shades), powered furniture or furnishings (e.g., standing desk, recliner chair), audio devices (e.g., music player), video device (e.g., television, home theater), environmental controls (e.g., air filter, air freshener), kitchen appliances (e.g., rice cooker, coffee machine, refrigerator), bathroom appliances, and household robotic devices (e.g., vacuum robot, robot butler).
  • the home devices 410 can include other types of devices that can be used in a household.
  • the resident profiles 430 typically include information about the different residents, such as name, an identifier used by the system, age, gender, and health information.
  • the resident profiles 430 can also include settings and other preferences of the home devices 410 selected by the different residents.
  • the network 420 provides connectivity between the different components of the residential environment 400 and allows the components to exchange data with each other.
  • the term “network” is intended to be interpreted broadly. It can include formal networks with standard defined protocols, such as Ethernet and InfiniBand.
  • the network 420 is a local area network that has its network equipment and interconnects managed within the residential environment 400 .
  • the network 420 can also combine different types of connectivity. It may include a combination of local area and/or wide area networks, using both wired and/or wireless links. Data exchanged between the components may be represented using any suitable format. In some embodiments, all or some of the data and communications may be encrypted.
  • the functionality described above can be physically implemented in the individual massage robot (one of the home devices 410 ), in a central hub for the home network, in a cloud-based service or elsewhere accessible by the massage robot via the network 420 , or in combinations of any of the above.
  • Alternate embodiments are implemented in computer hardware, firmware, software, and/or combinations thereof. Implementations can be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a programmable processor; and method steps can be performed by a programmable processor executing a program of instructions to perform functions by operating on input data and generating output. Embodiments can be implemented advantageously in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device.
  • Each computer program can be implemented in a high-level procedural or object-oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language.
  • Suitable processors include, by way of example, both general and special purpose microprocessors.
  • a processor will receive instructions and data from a read-only memory and/or a random access memory.
  • a computer will include one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and optical disks.
  • Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits) and other forms of hardware.
  • ASICs application-specific integrated circuits
  • module is not meant to be limited to a specific physical form. Depending on the specific application, modules can be implemented as hardware, firmware, software, and/or combinations of these. Furthermore, different modules can share common components or even be implemented by the same components. There may or may not be a clear boundary between different modules, even if drawn as separate elements in the figures.

Landscapes

  • Engineering & Computer Science (AREA)
  • Robotics (AREA)
  • Mechanical Engineering (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Rehabilitation Therapy (AREA)
  • Pain & Pain Management (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Veterinary Medicine (AREA)
  • Epidemiology (AREA)
  • Public Health (AREA)
  • Physical Education & Sports Medicine (AREA)
  • Animal Behavior & Ethology (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Dermatology (AREA)
  • Nursing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Manipulator (AREA)
  • Massaging Devices (AREA)

Abstract

A massage robot uses machine vision to locate treatment spots for massage. In one approach, a massage robot includes one or more robotic arms, an image sensor and a control system. The image sensor captures images of the user. The control system includes an image processing module and a motion controller. The image processing module processes the images to locate a treatment spot on the user. The motion controller controls the robotic arm to perform a massage procedure on the identified treatment spot.

Description

    BACKGROUND 1. Technical Field
  • This disclosure relates generally to massage robots.
  • 2. Description of Related Art
  • Many people enjoy receiving massage and enjoy the health benefits of massage. However, in most cases, people receive massage at a massage parlor or other type of studio. To do this, people typically must make appointments in advance, must leave their homes and travel to the studio and must spend time to find a good studio and masseuse. If the studio closes or the masseuse leaves, the person must start over and find another acceptable provider of massage treatment. Treatments and their quality may also vary from one studio to the next, and even between masseuses within a studio. In addition, if the person is not using the same masseuse and studio on a regular basis, there may be no past history about the person and his massage treatments, and the masseuse may not perform massage according to the person's personal preferences or to address the person's specific idiosyncrasies or problem spots.
  • Thus, there is a need for better and more convenient approaches to provide massage.
  • SUMMARY
  • The present disclosure overcomes the limitations of the prior art by providing a massage robot that uses machine vision to locate treatment spots for massage. In one approach, a massage robot includes one or more robotic arms, an image sensor and a control system. The image sensor captures images of the user. The control system includes an image processing module and a motion controller. The image processing module processes the images to locate a treatment spot on the user. The motion controller controls the robotic arm to perform a massage procedure on the identified treatment spot.
  • The massage can be personalized by selecting the massage procedure based on the identity of the user, for example based on the user's past history of massages, on the user's personal preferences or on instructions from a massage professional treating the user.
  • Sensors other than machine vision can also be used. Examples include ultrasound sensors, touch sensors, pressure sensors, and temperature sensors. In some designs, the robotic arm can be fitted with different end effectors, which are selected based on the desired massage procedure.
  • Other aspects include components, devices, systems, improvements, methods, processes, applications, computer readable mediums, and other technologies related to any of the above.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Embodiments of the disclosure have other advantages and features which will be more readily apparent from the following detailed description and the appended claims, when taken in conjunction with the examples in the accompanying drawings, in which:
  • FIG. 1 is a block diagram of a massage robot, according to an embodiment.
  • FIG. 2 shows a user receiving a massage from a mobile, home assistant robot.
  • FIG. 3 shows a user receiving a massage from a robot fixed to a massage table.
  • FIG. 4 is a block diagram of a residential environment including a massage robot, according to an embodiment.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • The figures and the following description relate to preferred embodiments by way of illustration only. It should be noted that from the following discussion, alternative embodiments of the structures and methods disclosed herein will be readily recognized as viable alternatives that may be employed without departing from the principles of what is claimed.
  • FIG. 1 is a block diagram of a massage robot 100, according to an embodiment. The robot 100 has one or more robotic arms 110 which are guided by a control system 150 that uses machine vision to identify a treatment spot for the massage procedure. In FIG. 1, the massage robot 100 includes an image sensor 130 (e.g., a camera), an image processing module 132 and a motion controller 155. The image sensor 130 captures images of the user—the person receiving the massage. The image processing module 132 processes these images to locate the treatment spot, and the motion controller 155 controls the robotic arm 110 to perform the desired massage procedure on the identified treatment spot.
  • Different types of robotic arms can be used, with different degrees of freedom. Typically, the robotic arm will have from 3-7 degrees of freedom. Example degrees of freedom include shoulder, elbow, wrist, finger/hand manipulators, and telescoping upper arm or forearm. If the robotic arm can be fitted with different end effectors, the control system 150 may also select the appropriate end effector and control the robotic arm to automatically change to the correct end effector. Examples of end effectors include those with fingers (typically from one to five, with or without thumb) or pad without fingers.
  • The massage robot 100 may also include other sensors 135. Examples include an ultrasound sensor to capture ultrasound images, a pressure sensor to sense the pressure exerted by the robotic arm on the treatment spot, a touch sensor to detect the hardness or firmness of the treatment spot, and a temperature sensor to detect the temperature of the treatment spot. Information from these sensors can also be processed 137 by the control system 150 and used by the motion controller 155 to control the robotic arm.
  • A personalization module 140 is used to personalize the massage procedure for the user. It first identifies the user. For example, the user may be identified when he logs into his user account for the massage robot or for a home network connected to the robot. Alternately, the user may be identified by face recognition or other techniques. Once the user is identified, the personalization module 140 determines the appropriate massage procedure based on the user's identity. For example, the massage procedure may be selected based on the user's profile 142. Perhaps the user has specified certain types of massage on specific days of the week. Alternately, the massage procedure may be defined by a massage professional treating the user. The professional's instructions specifying certain types of massage for the user may be stored somewhere and retrieved by the massage robot 100.
  • Alternately, the massage robot 100 may determine or suggest a massage procedure based on the user's prior history 143 of massages. If in the past the user has received upper back massage on Tuesday and Thursday and whole body massage on Saturday, and today is Saturday, then the massage robot 100 may assume whole body massage or may ask the user whether he would like a whole body massage. The massage procedure may also be selected based on the user's medical record, such as a record of back pain at certain locations.
  • The massage procedure may also be determined based on the user's pose. If the user is lying face down with exposed back, legs and arms, the massage robot 100 may determine that the user would like a whole body massage. If the user is sitting with back facing the robot 100, the robot may determine that some sort of back massage is appropriate.
  • Returning to FIG. 1, the massage robot 100 also includes a user interface 160, which in this example includes speech synthesis 162 and voice recognition 164. The speech synthesis module 162 is used to give instructions and otherwise communicate to the user. The voice recognition module 164 is used to receive voice commands or other communications.
  • FIG. 2 shows an example of a massage robot 200 preparing to deliver a massage to a user. Jane User comes home from work on Wednesday, sits down to watch TV and asks her home robot assistant for a beer. In this example, the massage robot is a mobile, more general purpose home assistant robot. In addition to providing massage, the robot can also perform other household tasks, such as cleaning, retrieving items, organizing and putting away items.
  • User: “Assistant, may I please have a beer?”
  • Assistant retrieves a beer from the refrigerator and delivers it to User. Assistant's camera captures images of User, and determines that she is sitting in her favorite chair and watching TV. Assistant also knows that User has requested a beer. Based on this pattern and User's past history, Assistant asks, “I see today is Wednesday and you look tired. Would you like a back massage?”
  • User: “Yes. That would be good.”
  • Assistant: “Please move to the massage chair and prepare for back massage.”
  • User moves to the massage chair, where her face is facing the chair back and her back is exposed.
  • Assistant determines from its captured images that User is sitting in the massage chair and that the back is exposed ready for massage, but the pose is not quite right. Assistant instructs, “Please sit straight and move slightly to the left.” Assistant also determines that User is still drinking beer. Assistant warns, “Drinking while receiving massage is not recommended. Please put down your beer.”
  • User follows Assistant's instructions. Assistant determines that User's pose is now ready for massage. From User's profile and past history, Assistant selects a general 15-minute massage of the upper back and also notices that User recently has had an especially tight right trapezius. Assistant: “I will perform your usual upper back massage. Is your right side still bothering you?”
  • User: “15-minute back massage is fine. My right side is not painful, but still tight. Better than last time.”
  • Assistant begins the massage of the upper back. The vision system guides the end effector of the robotic arm to carry out a specific pattern of massage and/or to address specific problem areas. The image processing module 132 determines the location of the upper back, possibly including anatomical landmarks such as the spinal column, neck and trapezius muscles. Massage is applied to different treatment spots based on the captured images. The massage may be applied with different size pressure areas and different amounts of pressure. The motion controller 155 preferably controls the location of the massage, the pressure of the massage, and the speed and/or length of massage strokes.
  • In this example, Assistant changes the end effector of the robotic arm to a long broad end effector and begins the massage by applying massage oil to User's back and stroking from the spine outwards in order to emulate a general back massage by a masseuse's forearm.
  • Assistant then changes to a tennis-ball sized end effector with more of a pressuring motion rather than a stroking motion to emulate pressure by a masseuse's palm. The amount of pressure is based on the pressures recorded for User's last massage. Assistant works a pattern of palm pressures across User's back. During this pattern, sensors in the end effector detect a small knot on the left side and note the location for later treatment. As another example where the massage robot itself may detect problem areas, if the image sensor is a thermal camera, the image processing module 132 may process the thermal images to identify unusually hot spots or temperature peaks for treatment.
  • Assistant then changes to a smaller size end effector to emulate a more localized pressure massage by fingers and thumbs. Assistant starts by applying pressure to both sides of the spinal column, working up and down the spinal column. It then massages specific spots, including the identified knot. Assistant applies concentrated pressure to the knot area. User yells, “Ouch. That hurts.”
  • Assistant stops and asks, “Is that too much pressure?”
  • User: “No. It is okay. Just a little sensitive.”
  • Assistant continues with the same amount of pressure, but with more of a kneading motion rather than just direct pressure.
  • Assistant: “How is that?”
  • User: “Better. Thank you.”
  • Assistant: “I am going to add some heat. OK?”
  • User: “OK.”
  • Assistant increases the temperature of the end effector and finishes treatment of the knot in this manner. In some cases, the massage time may be a predetermined time period: five minutes for example. Alternately, the Assistant may massage the knot until sensors indicate the knot has been released or until User instructs the Assistant to stop.
  • Assistant then moves to the right trapezius. Assistant changes to opposing end effectors. The opposing end effectors may be either on the same robotic arm or on different robotic arms. Opposing end effectors may be used to emulate pinching or squeezing motions. Assistant uses the opposing end effectors to massage the right trapezius.
  • When this part of the procedure is completed, Assistant states, “Massage is completed.” User replies, “Good. Thanks.”
  • FIG. 3 shows another example. In this example, the massage robot 300 is stationary. It includes a set of robotic arms that are stationary relative to a massage table. The user lies on the massage table and the robotic arms perform the massage. The arms may work simultaneously, with different arms performing different procedures. For example, some arms may provide massage for the back while others provide massage of the legs or arms. Alternately, some arms may provide general massage of an area while others massage specific treatment spots. The massage robot 300 may be used in professional settings, such as at a massage studio or doctor's office, in addition to providing home massage.
  • In another aspect, the massage robot may be part of a home network. FIG. 4 is a block diagram of a residential environment that includes a massage robot, according to an embodiment. The residential environment 400 is an environment designed for people to live in. The residential environment 400 can be a dwelling, such as a house, a condo, an apartment, or a dormitory. The residential environment 400 includes home devices 410A-N, including the massage robot described above. It also includes a home device network 420 connecting the home devices 410, and a resident profiles database 430 that contains residents' preferences for the home devices. The components in FIG. 4 are shown as separate blocks but they may be combined depending on the implementation. For example, the resident profiles 430 may be part of the home devices 410. Also, the residential environment 400 may include a hub for the network 420. The hub may also control the home devices 410. The network 420 may also provide access to external devices, such as cloud-based services.
  • The home devices 410 are household devices that are made available to the different persons associated with the residential environment 400. Examples of other home devices 410 include HVAC devices (e.g., air conditioner, heater, air venting), lighting, powered window and door treatments (e.g., door locks, power blinds and shades), powered furniture or furnishings (e.g., standing desk, recliner chair), audio devices (e.g., music player), video device (e.g., television, home theater), environmental controls (e.g., air filter, air freshener), kitchen appliances (e.g., rice cooker, coffee machine, refrigerator), bathroom appliances, and household robotic devices (e.g., vacuum robot, robot butler). The home devices 410 can include other types of devices that can be used in a household.
  • The resident profiles 430 typically include information about the different residents, such as name, an identifier used by the system, age, gender, and health information. The resident profiles 430 can also include settings and other preferences of the home devices 410 selected by the different residents.
  • The network 420 provides connectivity between the different components of the residential environment 400 and allows the components to exchange data with each other. The term “network” is intended to be interpreted broadly. It can include formal networks with standard defined protocols, such as Ethernet and InfiniBand. In one embodiment, the network 420 is a local area network that has its network equipment and interconnects managed within the residential environment 400. The network 420 can also combine different types of connectivity. It may include a combination of local area and/or wide area networks, using both wired and/or wireless links. Data exchanged between the components may be represented using any suitable format. In some embodiments, all or some of the data and communications may be encrypted.
  • The functionality described above can be physically implemented in the individual massage robot (one of the home devices 410), in a central hub for the home network, in a cloud-based service or elsewhere accessible by the massage robot via the network 420, or in combinations of any of the above.
  • Although the detailed description contains many specifics, these should not be construed as limiting the scope of the invention but merely as illustrating different examples. It should be appreciated that the scope of the disclosure includes other embodiments not discussed in detail above. Various other modifications, changes and variations which will be apparent to those skilled in the art may be made in the arrangement, operation and details of the method and apparatus disclosed herein without departing from the spirit and scope as defined in the appended claims. Therefore, the scope of the invention should be determined by the appended claims and their legal equivalents.
  • Alternate embodiments are implemented in computer hardware, firmware, software, and/or combinations thereof. Implementations can be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a programmable processor; and method steps can be performed by a programmable processor executing a program of instructions to perform functions by operating on input data and generating output. Embodiments can be implemented advantageously in one or more computer programs that are executable on a programmable system including at least one programmable processor coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. Each computer program can be implemented in a high-level procedural or object-oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language. Suitable processors include, by way of example, both general and special purpose microprocessors. Generally, a processor will receive instructions and data from a read-only memory and/or a random access memory. Generally, a computer will include one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and optical disks. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits) and other forms of hardware.
  • The term “module” is not meant to be limited to a specific physical form. Depending on the specific application, modules can be implemented as hardware, firmware, software, and/or combinations of these. Furthermore, different modules can share common components or even be implemented by the same components. There may or may not be a clear boundary between different modules, even if drawn as separate elements in the figures.

Claims (20)

What is claimed is:
1. A massage robot for providing massage to a user, the massage robot comprising:
a robotic arm;
an image sensor that captures images of the user; and
a control system comprising:
an image processing module that processes the images to locate a treatment spot on the user; and
a motion controller that controls the robotic arm to perform a massage procedure on the identified treatment spot.
2. The massage robot of claim 1 further comprising:
a personalization module that identifies the user and determines the massage procedure based on the identity of the user.
3. The massage robot of claim 2 wherein the personalization module determines the massage procedure based on a history of past massage procedures performed on the identified user.
4. The massage robot of claim 2 wherein the personalization module determines the massage procedure based on stored instructions from a massage professional treating the identified user.
5. The massage robot of claim 1 wherein the image processing module further determines a pose of the user, and the motion controller controls the massage robot according to the pose.
6. The massage robot of claim 5 wherein the image processing module determines whether the user is ready for the massage procedure based on the user's pose, and the control system determines whether to begin the massage procedure depending on whether the user is ready for the massage procedure.
7. The massage robot of claim 5 further comprising:
a speech synthesis module, wherein the control system controls the speech synthesis module to produce voice instructions instructing the user to position himself for the massage procedure.
8. The massage robot of claim 1 wherein the image processing module processes the images to identify body parts of the user and identifies the treatment spot on the relevant body part.
9. The massage robot of claim 1 wherein the image sensor is a thermal camera, and the image processing module processes thermal images captured by the thermal camera to identify a hot spot on the user as the treatment spot.
10. The massage robot of claim 1 wherein the robotic arm is compatible with different end effectors, and the control system selects the end effector appropriate for the massage procedure.
11. The massage robot of claim 1 further comprising:
an ultrasound sensor, the motion controller further controlling the robotic arm according to signals from the ultrasound sensor.
12. The massage robot of claim 1 further comprising:
a touch sensor, the motion controller further controlling the robotic arm according to a hardness of the treatment spot detected by the touch sensor.
13. The massage robot of claim 1 further comprising:
a pressure sensor, the motion controller further controlling the robotic arm according to a pressure exerted by the robotic arm on the treatment spot detected by the pressure sensor.
14. The massage robot of claim 1 further comprising:
a temperature sensor, the motion controller further controlling the robotic arm according to a temperature of the treatment spot detected by the temperature sensor.
15. The massage robot of claim 1 further comprising:
a voice recognition module that receives voice commands, the motion controller further controlling the robotic arm according to the voice commands.
16. The massage robot of claim 1 wherein the motion controller further controls a pressure exerted by the robotic arm on the treatment spot.
17. The massage robot of claim 1 wherein the motion controller further controls a speed and/or a length of a stroke of the robotic arm across the treatment spot.
18. The massage robot of claim 1 further comprising:
a second robotic arm, the motion controller controlling both robotic arms to perform the massage procedure on the identified treatment spot.
19. The massage robot of claim 1 wherein the massage robot is mobile and also has a capability to perform household chores.
20. The massage robot of claim 1 wherein at least a portion of the control system is implemented by one of a cloud service or a home network hub.
US15/826,569 2017-11-29 2017-11-29 Massage robot using machine vision Expired - Fee Related US10434658B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US15/826,569 US10434658B2 (en) 2017-11-29 2017-11-29 Massage robot using machine vision
CN201880073575.2A CN111343958A (en) 2017-11-29 2018-11-05 Massage robot using machine vision
PCT/CN2018/114024 WO2019105183A1 (en) 2017-11-29 2018-11-05 Massage robot using machine vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/826,569 US10434658B2 (en) 2017-11-29 2017-11-29 Massage robot using machine vision

Publications (2)

Publication Number Publication Date
US20190160684A1 true US20190160684A1 (en) 2019-05-30
US10434658B2 US10434658B2 (en) 2019-10-08

Family

ID=66634225

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/826,569 Expired - Fee Related US10434658B2 (en) 2017-11-29 2017-11-29 Massage robot using machine vision

Country Status (3)

Country Link
US (1) US10434658B2 (en)
CN (1) CN111343958A (en)
WO (1) WO2019105183A1 (en)

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200077915A1 (en) * 2018-09-06 2020-03-12 Fuji Medical Instruments Mfg. Co., Ltd. Massage Machine
CN110960414A (en) * 2020-01-10 2020-04-07 天寿健康科技(济南)有限公司 Internet of things health robot
CN111010536A (en) * 2019-12-04 2020-04-14 武汉市海沁医疗科技有限公司 Video monitoring data storage method and device for massage robot
CN111571611A (en) * 2020-05-26 2020-08-25 广州纳丽生物科技有限公司 Facial operation robot track planning method based on facial and skin features
US20200390644A1 (en) * 2019-06-11 2020-12-17 Addaday, Inc. Programmable therapeutic device, method, and system
US20210085558A1 (en) * 2019-09-24 2021-03-25 Lg Electronics Inc. Artificial intelligence massage apparatus and method for controlling massage operation in consideration of facial expression or utterance of user
CN112809697A (en) * 2021-01-11 2021-05-18 福建思特电子有限公司 5G intelligent entrance guard robot
CN113813170A (en) * 2021-08-30 2021-12-21 中科尚易健康科技(北京)有限公司 Target point conversion method between cameras in multi-camera physiotherapy system
US20220161441A1 (en) * 2019-07-17 2022-05-26 Riccoh Co., Ltd. Robot body care system, robot body care method, and robot body care program
CN115054504A (en) * 2022-06-27 2022-09-16 浙江师范大学 Intelligent finger massager
US20220387118A1 (en) * 2020-05-12 2022-12-08 Aescape, Inc. Method and system for autonomous therapy
US20230064229A1 (en) * 2021-08-30 2023-03-02 Sayed H. Rohani Relaxing, Massaging Hands
WO2023110489A1 (en) * 2021-12-17 2023-06-22 Kuka Deutschland Gmbh Method for planning movements to be performed automatically of a robot arm guiding a massage appliance, and associated robot system
CN116999305A (en) * 2023-09-08 2023-11-07 遨博(北京)智能科技股份有限公司 A physical therapy robot system
US11858144B2 (en) 2020-05-12 2024-01-02 Aescape, Inc. Method and system for autonomous body interaction
CN118061180A (en) * 2024-02-29 2024-05-24 河南翔宇医疗设备股份有限公司 Motion control method and system for rehabilitation robot
US11999061B2 (en) 2020-05-12 2024-06-04 Aescape, Inc. Method and system for autonomous object manipulation
WO2024192220A1 (en) * 2023-03-16 2024-09-19 Aescape, Inc. Robot system to perform coordinated body work
US12122050B2 (en) 2020-09-09 2024-10-22 Aescape, Inc. Method and system for generating a therapeutic massage plan
US20250010486A1 (en) * 2023-07-07 2025-01-09 Denis LEDENKOV Automatic massage method
US12214509B2 (en) * 2023-03-16 2025-02-04 Aescape, Inc. Robot system to perform coordinated body work
DE112022004212B4 (en) * 2021-08-27 2025-07-10 Andrew Flessas System and method for precise multidimensional movement of a haptic stimulator
US20250255772A1 (en) * 2024-02-13 2025-08-14 Aescape, Inc. Therapeutic content personalization

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110882152A (en) * 2019-12-04 2020-03-17 武汉市海沁医疗科技有限公司 Self-rotating massage mechanism and massage system
WO2022073467A1 (en) * 2020-10-09 2022-04-14 谈斯聪 Dual-arm multitask parallel processing robot device for caregiving massages
CN113648215B (en) * 2021-07-08 2024-06-04 康民智美(成都)健康科技有限公司 Physiotherapy equipment
CN114404256A (en) * 2022-01-24 2022-04-29 辽宁师范大学 Intelligent sojourn bed suitable for health maintenance of middle-aged and elderly people
CN115972202B (en) * 2022-12-09 2024-10-11 上海非夕机器人科技有限公司 Method, robot, device, medium and product for controlling operation of a robot arm
CN119238547B (en) * 2024-09-14 2025-06-24 湖南中医药大学 Massage robot control method and system

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06209A (en) 1992-06-19 1994-01-11 Matsushita Electric Works Ltd Relax/refresh system
US5876325A (en) * 1993-11-02 1999-03-02 Olympus Optical Co., Ltd. Surgical manipulation system
DE10080012B4 (en) 1999-03-19 2005-04-14 Matsushita Electric Works, Ltd., Kadoma Three-dimensional method of detecting objects and system for picking up an object from a container using the method
IL133551A0 (en) * 1999-12-16 2001-04-30 Nissim Elias Human touch massager
TW453875B (en) * 2000-01-31 2001-09-11 Sanyo Electric Co Massaging machine and method for controlling the therapeutic force of a massaging machine
CN1759822A (en) * 2004-10-15 2006-04-19 严红缨 Massage method of intelligent initiative type appliance and massage robot
KR100830214B1 (en) * 2006-05-10 2008-05-16 최형근 Hair growth promoting device for both scalp massage
CN101884585A (en) 2010-07-30 2010-11-17 山东建筑大学 A Chinese medicine massage robot control system
JP2014094022A (en) 2012-11-07 2014-05-22 Panasonic Corp Image display program, electronic device, massage device, and massage system
JP6213762B2 (en) 2013-01-08 2017-10-18 パナソニックIpマネジメント株式会社 Massage machine operation program, massage machine operator, massage machine, and massage system
CN103753549A (en) * 2014-01-24 2014-04-30 成都万先自动化科技有限责任公司 Robot for hand massage service
US10492861B2 (en) * 2015-04-02 2019-12-03 Mark John McDonald Rogers Method and apparatus for treating soft tissue injury
US20170266077A1 (en) * 2016-03-21 2017-09-21 Christian Campbell Mackin Robotic massage machine and method of use
CN206007535U (en) * 2016-05-10 2017-03-15 深圳光启合众科技有限公司 Massage apparatus
CN106901941A (en) 2017-02-21 2017-06-30 哈尔滨医科大学 A kind of joint of vertebral column depressurized system
US10733004B2 (en) * 2017-04-26 2020-08-04 At&T Intellectual Property I, L.P. Intelligent service on-demand robot virtualization
CN107095510A (en) 2017-06-14 2017-08-29 海宁市金郑家具有限公司 Sofa with treatment cervical vertebra function

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200077915A1 (en) * 2018-09-06 2020-03-12 Fuji Medical Instruments Mfg. Co., Ltd. Massage Machine
US20200390644A1 (en) * 2019-06-11 2020-12-17 Addaday, Inc. Programmable therapeutic device, method, and system
US20220161441A1 (en) * 2019-07-17 2022-05-26 Riccoh Co., Ltd. Robot body care system, robot body care method, and robot body care program
US20210085558A1 (en) * 2019-09-24 2021-03-25 Lg Electronics Inc. Artificial intelligence massage apparatus and method for controlling massage operation in consideration of facial expression or utterance of user
CN111010536A (en) * 2019-12-04 2020-04-14 武汉市海沁医疗科技有限公司 Video monitoring data storage method and device for massage robot
CN110960414A (en) * 2020-01-10 2020-04-07 天寿健康科技(济南)有限公司 Internet of things health robot
US12440974B2 (en) * 2020-05-12 2025-10-14 Aescape, Inc. Method and system for autonomous object manipulation
US12296487B2 (en) * 2020-05-12 2025-05-13 Aescape, Inc. Method and system for autonomous body interaction
US11998289B2 (en) * 2020-05-12 2024-06-04 Aescape, Inc. Method and system for autonomous therapy
US20220387118A1 (en) * 2020-05-12 2022-12-08 Aescape, Inc. Method and system for autonomous therapy
US20240342903A1 (en) * 2020-05-12 2024-10-17 Aescape, Inc. Method and system for autonomous object manipulation
US20240208057A1 (en) * 2020-05-12 2024-06-27 Aescape, Inc. Method and system for autonomous body interaction
US11999061B2 (en) 2020-05-12 2024-06-04 Aescape, Inc. Method and system for autonomous object manipulation
US11858144B2 (en) 2020-05-12 2024-01-02 Aescape, Inc. Method and system for autonomous body interaction
CN111571611A (en) * 2020-05-26 2020-08-25 广州纳丽生物科技有限公司 Facial operation robot track planning method based on facial and skin features
US12122050B2 (en) 2020-09-09 2024-10-22 Aescape, Inc. Method and system for generating a therapeutic massage plan
CN112809697A (en) * 2021-01-11 2021-05-18 福建思特电子有限公司 5G intelligent entrance guard robot
DE112022004212B4 (en) * 2021-08-27 2025-07-10 Andrew Flessas System and method for precise multidimensional movement of a haptic stimulator
CN113813170A (en) * 2021-08-30 2021-12-21 中科尚易健康科技(北京)有限公司 Target point conversion method between cameras in multi-camera physiotherapy system
US20230064229A1 (en) * 2021-08-30 2023-03-02 Sayed H. Rohani Relaxing, Massaging Hands
WO2023110489A1 (en) * 2021-12-17 2023-06-22 Kuka Deutschland Gmbh Method for planning movements to be performed automatically of a robot arm guiding a massage appliance, and associated robot system
CN115054504A (en) * 2022-06-27 2022-09-16 浙江师范大学 Intelligent finger massager
WO2024192220A1 (en) * 2023-03-16 2024-09-19 Aescape, Inc. Robot system to perform coordinated body work
US12214509B2 (en) * 2023-03-16 2025-02-04 Aescape, Inc. Robot system to perform coordinated body work
US12533281B2 (en) 2023-03-16 2026-01-27 Aescape, Inc. Robotic massage end effector
US20250010486A1 (en) * 2023-07-07 2025-01-09 Denis LEDENKOV Automatic massage method
CN116999305A (en) * 2023-09-08 2023-11-07 遨博(北京)智能科技股份有限公司 A physical therapy robot system
US20250255772A1 (en) * 2024-02-13 2025-08-14 Aescape, Inc. Therapeutic content personalization
US12458560B2 (en) * 2024-02-13 2025-11-04 Aescape, Inc. Therapeutic content personalization
CN118061180A (en) * 2024-02-29 2024-05-24 河南翔宇医疗设备股份有限公司 Motion control method and system for rehabilitation robot

Also Published As

Publication number Publication date
CN111343958A (en) 2020-06-26
US10434658B2 (en) 2019-10-08
WO2019105183A1 (en) 2019-06-06

Similar Documents

Publication Publication Date Title
US10434658B2 (en) Massage robot using machine vision
WO2022227550A1 (en) Control method and control apparatus for air conditioner, and air conditioner
US11825953B1 (en) Multi-zone adjustable bed with smart adjustment mechanism
CN113251611B (en) Control method and device for environment purification equipment and environment purification equipment
CN110726212B (en) Control method and device of air conditioner and air conditioner equipment
CN105423558A (en) Water heater, water heater system and control method for water heater
EP2726164A1 (en) Augmented-reality range-of-motion therapy system and method of operation thereof
WO2022252673A1 (en) Control method and apparatus for household appliance for adjusting fatigue degree, and household appliance
CN114675553A (en) Control method, device, equipment and medium for intelligent home system
CN107949316A (en) User terminal and dormancy management method
US20180192779A1 (en) Tv bed, tv, bed, and method for operating the same
WO2007012253A1 (en) An intellective blowing air conditioner and an intellective blowing method thereof
CN115128964A (en) Smart mattress based household equipment linkage method and related products
CN113168641B (en) Information processing apparatus and information processing method
CN106196617A (en) Control device of water heater, water heater and control method of water heater
CN107132787B (en) Kitchen housekeeper robot cooking control method and kitchen housekeeper robot
US20220008607A1 (en) Room Sterilization System
Ding et al. AToM-Bot: Embodied Fulfillment of Unspoken Human Needs with Affective Theory of Mind
CN111306729A (en) Environmental adjustment method and adjustment system for bathroom environment
CN107714457B (en) Moxibustion equipment
CN112957240A (en) Voice recognition control massage chair
JP7225659B2 (en) Information processing device, information processing method and information processing program
CN115670170B (en) Intelligent mattress adjusting method and related products
CN106802571A (en) Working method, user terminal and control center of a smart home device
CN118512092A (en) Control method and device of zero gravity sofa, zero gravity sofa and storage medium

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: MIDEA GROUP CO., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GU, HAISONG;WANG, DONGYAN;REEL/FRAME:044262/0089

Effective date: 20171129

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20231008