Samsung Patent | Method and wearable device for providing a virtual input interface

Patent: Method and wearable device for providing a virtual input interface

Publication Number: 10168792

Publication Date: 2019-01-01

Applicants: Samsung

Abstract

Provided is a wearable device including: an image sensor configured to sense a gesture image of a user setting a user input region; and a display configured to provide a virtual input interface corresponding to the set user input region.

Background

The real world is a space consisting of 3-dimensional (3D) coordinates. People are able to recognize 3D space by combining visual information obtained using two eyes. However, a photograph or a moving image captured by a general digital device is expressed in 2D coordinates, and thus does not include information about space. In order to give a feeling of space, 3D cameras or display products that capture and display 3D images by using two cameras have been introduced.

Meanwhile, a current input method of smart glasses is limited. A user basically controls the smart glasses by using a voice command. However, it is difficult for the user to control the smart glasses by using only a voice command if a text input is required. Thus, a wearable system that provides various input interaction methods is required.

Summary

Methods and apparatuses consistent with exemplary embodiments include a method and wearable device for setting an input region in the air or on an actual object based on a user motion, and providing a virtual input interface in the set input region.

Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented exemplary embodiments.

According to one or more exemplary embodiments, a wearable device includes: an image sensor configured to sense a gesture image of a user setting a user input region; and a display configured to provide a virtual input interface corresponding to the user input region set by using the sensed gesture image.

The sensed gesture image may correspond to a figure drawn by the user, and the virtual input interface may be displayed to correspond to the sensed figure.

The virtual input interface may be displayed to correspond to a size of the user input region.

The virtual input interface may be determined based on a type of an application being executed by the glasses type wearable device.

The display may include a transparent display configured to provide the virtual input interface on a region of the transparent display corresponding to the user input region as observed through the transparent display.

The image sensor may be configured to capture a first image of the user input region, and the display may be configured to display a second image of the virtual input interface over the user input region of the first image.

The glasses type wearable device may further include: a depth sensor configured to sense a first depth value corresponding to a distance from the wearable device to the user input region, and a second depth value corresponding to a distance from the wearable device to an input tool; and a controller configured to determine whether an input is generated through the virtual input interface based on the first depth value and the second depth value.

The displayed size of the virtual input interface may be determined based the first depth value.

The controller may be configured to determine that an input is generated through the virtual input interface when a difference between the first and second depth values is less than a threshold value.

The controller may be configured to determine that an input is generated through the virtual input interface when the second depth value is greater than the first depth value.

According to one or more exemplary embodiments, a method of providing, by a wearable device, a virtual input interface, includes: obtaining a gesture image of a user for setting a user input region; and providing a virtual input interface corresponding to the user input region such that the virtual input interface corresponds to a size of the user input region.

The obtaining of the gesture image may include: obtaining the gesture image by recognizing a figure drawn by the user; and setting a region corresponding to the figure as the user input region.

The virtual input interface may be determined based on a size of the user input region.

The method may further include determining the virtual input interface based on a type of object where the user input region is set.

The method may further include determining the virtual input interface based on a type of an application being executed by the wearable device.

The virtual input interface may be provided on a transparent display such that the virtual input interface corresponds to the user input region as observed through the transparent display.

The providing of the virtual input interface may include: capturing a first image of the user input region by using an image sensor; generating a second image of the virtual input interface; and displaying the second image over the user input region of the first image.

The method may further include: obtaining a first depth value corresponding to a distance from the wearable device to the user input region, and a second depth value corresponding to a distance from the wearable device to an input tool; and determining whether an input is generated through the virtual input interface based on the first depth value and the second depth value.

A displayed size of the virtual input interface may be determined based on a size of the user input region.

The determining of whether the input is generated may include determining that a difference between the first and second depth values is less than a threshold value.

The determining of whether the input is generated may include determining the second depth value is greater than the first depth value.

According to one or more exemplary embodiments, a wearable input device includes: a sensor configured to sense a plurality of gestures and a real world image; a display configured to display a graphic user interface; and a controller configured to determine an input region of the real world image, control the display to display the graphic user interface on an area corresponding to the determined input region, and determine an input based on an input gesture of the plurality of gestures.

The wearable input device may include a communicator configured to receive a touch signal from an external device. The controller may be further configured to determine the input based on the touch signal.

The may be further determined based on an input region defining gesture of the plurality of gestures.

The sensor may be further configured to determine a distance between the wearable input device and the input region.

The controller may be further configured to continuously update a display region of the graphic user interface based on the real world image.

发表评论

电子邮件地址不会被公开。 必填项已用*标注