Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

301 results about "Natural user interface" patented technology

In computing, a natural user interface, or NUI, or natural interface is a user interface that is effectively invisible, and remains invisible as the user continuously learns increasingly complex interactions. The word natural is used because most computer interfaces use artificial control devices whose operation has to be learned.

Natural language interface using constrained intermediate dictionary of results

A method for processing a natural language input provided by a user includes: providing a natural language query input to the user; performing, based on the input, a search of one or more language-based databases; providing, through a user interface, a result of the search to the user; identifying, for the one or more language-based databases, a finite number of database objects; and determining a plurality of combinations of the finite number of database objects. The one or more language-based databases include at least one metadata database including at least one of a group of information types including case information, keywords, information models, and database values.
Owner:RENESSELAER POLYTECHNIC INST

Hand-ear user interface for hand-held device

A hand-held information processing device, such as a mobile phone, has a thumb wheel that lets the user scan a circular array of options. Each respective one of the options is represented by a respective audio output that gets played out when the wheel is turned a notch up or down. This enables the user to select an option with one hand and without having to look at the device. It also allows for a form factor smaller than that of a conventional mobile phones since a keypad is not needed for entering digits to make a call from a personalized directory.
Owner:KONINKLIJKE PHILIPS ELECTRONICS NV

Synonyms mechanism for natural language systems

Roughly described, a natural language interface to a back-end application incorporates synonyms automatically added to user input to enhance the natural language interpretation. Synonyms can be learned from user input and written into a synonyms database. Their selection can be based on tokens identified in user input. Natural language interpretation can be performed by agents arranged in a network, which parse the user input in a distributed manner. In an embodiment, a particular agent of the natural language interpreter receives a first message that includes the user input, returns a message claiming at least a portion of the user input, and subsequently receives a second message delegating actuation of at least that portion to the particular agent.
Owner:IANYWHERE SOLUTIONS

Method and system enabling natural user interface gestures with an electronic system

An electronic device coupleable to a display screen includes a camera system that acquires optical data of a user comfortably gesturing in a user-customizable interaction zone having a z0 plane, while controlling operation of the device. Subtle gestures include hand movements commenced in a dynamically resizable and relocatable interaction zone. Preferably (x,y,z) locations in the interaction zone are mapped to two-dimensional display screen locations. Detected user hand movements can signal the device that an interaction is occurring in gesture mode. Device response includes presenting GUI on the display screen, creating user feedback including haptic feedback. User three-dimensional interaction can manipulate displayed virtual objects, including releasing such objects. User hand gesture trajectory clues enable the device to anticipate probable user intent and to appropriately update display screen renderings.
Owner:KAYA DYNAMICS LLC

Adaptive user interface input device

The present invention relates to an adaptable user interface. The adaptable user interface provides a more error free input function as well as greater ease of use when being used during certain events such as while moving. A user interface in accordance with the principles of the present invention comprises a user input, at least one sensor, and a display unit functionally in communication with the at least one sensor and adapted to change its user interface input mode. The user interface is capable of adapting its user interface input in response to a stimulus sensed by the sensor.
Owner:NOKIA TECHNOLOGLES OY

Method and system enabling natural user interface gestures with user wearable glasses

User wearable eye glasses include a pair of two-dimensional cameras that optically acquire information for user gestures made with an unadorned user object in an interaction zone responsive to viewing displayed imagery, with which the user can interact. Glasses systems intelligently signal process and map acquired optical information to rapidly ascertain a sparse (x,y,z) set of locations adequate to identify user gestures. The displayed imagery can be created by glasses systems and presented with a virtual on-glasses display, or can be created and / or viewed off-glasses. In some embodiments the user can see local views directly, but augmented with imagery showing internet provided tags identifying and / or providing information as to viewed objects. On-glasses systems can communicate wirelessly with cloud servers and with off-glasses systems that the user can carry in a pocket or purse.
Owner:KAYA DYNAMICS LLC

Audio user interface

An audio user interface that provides audio prompts that help a user interact with a user interface of an electronic device is disclosed. The audio prompts can provide audio indicators that allow a user to focus his or her visual attention upon other tasks such as driving an automobile, exercising, or crossing a street, yet still enable the user to interact with the user interface. An intelligent path can provide access to different types of audio prompts from a variety of different sources. The different types of audio prompts may be presented based on availability of a particular type of audio prompt. As examples, the audio prompts may include pre-recorded voice audio, such as celebrity voices or cartoon characters, obtained from a dedicate voice server. Absent availability of pre-recorded or synthesized audio data, non-voice audio prompts may be provided.
Owner:APPLE INC

Audio user interface

An audio user interface that provides audio prompts that help a user interact with a user interface of an electronic device is disclosed. The audio prompts can provide audio indicators that allow a user to focus his or her visual attention upon other tasks such as driving an automobile, exercising, or crossing a street, yet still enable the user to interact with the user interface. An intelligent path can provide access to different types of audio prompts from a variety of different sources. The different types of audio prompts may be presented based on availability of a particular type of audio prompt. As examples, the audio prompts may include pre-recorded voice audio, such as celebrity voices or cartoon characters, obtained from a dedicate voice server. Absent availability of pre-recorded or synthesized audio data, non-voice audio prompts may be provided.
Owner:APPLE INC

Private data entry

A device and method for creating a private entry display on a touchscreen is provided, wherein the touchscreen includes a plurality of touch cells corresponding to spatial locations on the touchscreen. A graphical user interface is generated for display on the touchscreen for a predefined operation, the graphical user interface including a plurality of input zones. A characteristic of the graphical user interface as displayed on the touchscreen for the predefined operation is alter in order to change the touch cells associated with the graphical user interface. The altered user interface then is displayed on the touchscreen.
Owner:SHARP KK

Sensor-based distributed tangible user interface

A distributed tangible user interface comprises compact, self-powered, tangible user interface manipulative devices having sensing, display, and wireless communication capabilities, along with one or more associated digital content or other interactive software management applications. The manipulative devices display visual representations of digital content or program controls and can be physically manipulated as a group by a user for interaction with the digital information or software application. A controller on each manipulative device receives and processes data from a movement sensor, initiating behavior on the manipulative and / or forwarding the results to a management application that uses the information to manage the digital content, software application, and / or the manipulative devices. The manipulative devices may also detect the proximity and identity of other manipulative devices, responding to and / or forwarding that information to the management application, and may have feedback devices for presenting responsive information to the user.
Owner:MASSACHUSETTS INST OF TECH

Interactive video programming methods

An entertainment head-end provides broadcast programming, video-on-demand services, and HTML-based interactive programming through a distribution network to client terminals in subscribers' homes. A number of different features are provided, including novel user interfaces, enhanced video-on-demand controls, a variety of interactive services (personalized news, jukebox, games, celebrity chat), and techniques that combine to provide user experiences evocative of conventional television.
Owner:MICROSOFT TECH LICENSING LLC

Display terminal user interface with ability to select remotely stored surface finish for mapping onto displayed 3-D surface

A user interface on a display terminal, such as a personal computer, includes a 3D display region which shows a scene incorporating a number of objects, such as items of furniture. A surface finish selector is also displayed and is used to select a surface finish from a number of alternatives. In the case of items of furniture, these finishes may correspond to different fabrics for upholstery. A surface texture data for a selected finish is automatically downloaded from a remote source and mapped onto the object in the 3D scene. In a preferred implementation, the surface finish selector is a frame of a web page and generates control data which is passed to another frame containing the 3D scene together with movement controls for changing the viewpoint in the scene.
Owner:BRITISH TELECOMM PLC

System and method for an intelligent multi-modal user interface for route drawing

InactiveUS7216034B2Facilitate route drawingFacilitate the route drawingInstruments for road network navigationRoad vehicles traffic controlGraphicsDisplay device
A system and method for providing an interactive, multi-modal interface used to generate a route drawing. A mobile terminal display initially displays map data indicative of a user's region of interest. Through a series of audible, visual, and tactile excitations, the user interacts with the route drawing system to generate a desired route to be superimposed upon the user's region of interest. User preferences are used by the system to intelligently aid the user in selection of each segment of the route to be drawn. An analysis of each segment, as well as an analysis of the final route, is performed so that the attributes of the route, along with the graphical representation of the route itself, may be stored either locally within the mobile terminal or remotely within a map server.
Owner:NOKIA TECHNOLOGLES OY

Query generation from displayed text documents using virtual magnets

A system and method are provided for dynamically generating a query using touch gestures. A virtual magnet is movable on a display device of a tactile user interface in response to touch. A user selects one of a set of text documents for review, which is displayed on the display. The system is configured for recognizing a highlighting gesture on the tactile user interface over the displayed document as a selection of a text fragment from the document text. The virtual magnet is populated with a query which is based on the text fragment selected with the highlighting gesture. The populated magnet is able to cause a subset of displayed graphic objects to exhibit a response to the magnet as a function of the query and the text content of the respective documents which the objects represent and / or to cause responsive instances in a text document to be displayed.
Owner:XEROX CORP

External user interface for head worn computing

Aspects of the present invention relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in the form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC

External user interface for head worn computing

Aspects of the present disclosure relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in the form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC

External user interface for head worn computing

Aspects of the present disclosure relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in the form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC

External user interface for head worn computing

Aspects of the present disclosure relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in a form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC

Image processing device, information processing method and computer-readable storage medium storing a control program for performing an operation based on whether a function is being set or requested to be set

For example, an image processing device which can be operated from both remote user interface such as a Web browser and local interface by an operation panel or a card reader, and a computer system communicating with the device are embodied. A system of integrating authenticating processes by each of the remote and local user interfaces can be provided. In this case, an adjustment is made not to cause inconsistency by the conflict between the operation from the remote user interface and the operation from the local user interface so as to practically utilize each user interface.
Owner:CANON KK

External user interface for head worn computing

Aspects of the present disclosure relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in the form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC

External user interface for head worn computing

Aspects of the present disclosure relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in the form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC

External user interface for head worn computing

Aspects of the present invention relate to external user interfaces used in connection with head worn computers (HWC). Embodiments relate to an external user interface that has a physical form intended to be hand held. The hand held user interface may be in the form similar to that of a writing instrument, such as a pen. In embodiments, the hand held user interface includes technologies relating to writing surface tip pressure monitoring, lens configurations setting a predetermined imaging distance, user interface software mode selection, quick software application launching, and other interface technologies.
Owner:OSTERHOUT GROUP INC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products