US20150030144A1 - Communications using different modalities - Google Patents
Communications using different modalities Download PDFInfo
- Publication number
- US20150030144A1 US20150030144A1 US14/456,008 US201414456008A US2015030144A1 US 20150030144 A1 US20150030144 A1 US 20150030144A1 US 201414456008 A US201414456008 A US 201414456008A US 2015030144 A1 US2015030144 A1 US 2015030144A1
- Authority
- US
- United States
- Prior art keywords
- mode
- pva
- call
- caller
- voice
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/04—Real-time or near real-time messaging, e.g. instant messaging [IM]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42382—Text-based messaging services in telephone networks such as PSTN/ISDN, e.g. User-to-User Signalling or Short Message Service for fixed networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/06—Message adaptation to terminal or network requirements
- H04L51/066—Format adaptation, e.g. format conversion or compression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/006—Call diverting means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/247—Telephone sets including user guidance or feature selection means facilitating their use
- H04M1/2478—Telephone terminals specially adapted for non-voice services, e.g. email, internet access
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/253—Telephone sets using digital voice transmission
- H04M1/2535—Telephone sets using digital voice transmission adapted for voice communication over an Internet Protocol [IP] network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/57—Arrangements for indicating or recording the number of the calling subscriber at the called subscriber's set
- H04M1/575—Means for retrieving and displaying personal data about calling party
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/64—Automatic arrangements for answering calls; Automatic arrangements for recording messages for absent subscribers; Arrangements for recording conversations
- H04M1/65—Recording arrangements for recording a message from the calling party
- H04M1/6505—Recording arrangements for recording a message from the calling party storing speech in digital form
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M7/00—Arrangements for interconnection between switching centres
- H04M7/0024—Services and arrangements where telephone services are combined with data services
- H04M7/0042—Services and arrangements where telephone services are combined with data services where the data service is a text-based messaging service
- H04M7/0045—Services and arrangements where telephone services are combined with data services where the data service is a text-based messaging service where the text-based messaging service is an instant messaging service
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/60—Medium conversion
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42391—Systems providing special services or facilities to subscribers where the subscribers are hearing-impaired persons, e.g. telephone devices for the deaf
Definitions
- POTS Plain old telephone service
- POTS is the most basic form of residential and small business telephone service. POTS became available soon after the telephone system was introduced in the late 19th century and, from the standpoint of the user, has existed more or less unchanged ever since. POTS services include bi-directional or full duplex voice path with a limited frequency range, a dial tone and ringing signal, subscriber dialing, operator services such as directory assistance, and long distance and conference calling.
- new network services became available due to the creation of electronic telephone exchanges and computerization. New services included voice mail, caller identification, call waiting, reminder calls and other similar services.
- IP Telephony is a form of telephony which uses the TCP/IP protocol popularized by the Internet to transmit digitized voice data.
- the routing of voice conversations over the Internet or through other IP networks is also called VoIP (Voice over Internet Protocol).
- Digital telephony was introduced to improve voice services, but was subsequently found to be very useful in the creation of new network services because it can transfer data quickly over telephone lines.
- Computer Telephony Integration enables a computer to control phone functions such as making and receiving voice, fax, and data calls.
- Session Initiation Protocol is a signaling protocol used for creating, modifying and terminating sessions (voice or video calls) with one or more participants. Sessions include Internet telephone calls, multimedia distribution and multimedia conferences. Development of the SIP protocol was motivated by the need for a signaling and call setup protocol for IP-based communications that could support the call processing functions and features present in the public switched telephone network (PSTN) using proxy servers and user agents.
- PSTN public switched telephone network
- Interactive text messaging is a form of real-time communication between two or more people based on typed text.
- Interactive text messaging differs from email in that text conversations are conducted in real-time using computers connected over a network such as the Internet.
- a client program on a text messaging user's computer connects to an interactive text messaging service. The user types a message and the service sends the text message to the recipient.
- Popular interactive text messaging services on the Internet include .NET Messenger Service, AOL Instant Messenger, Excite/Pal, Gadu-Gadu, Google Talk, iChat, ICQ, Jabber, Qnext, QQ, Meetro, Skype, Trillian, Yahoo! Messenger and Rediff Bol Instant Messenger.
- Communications between users of different modalities is enabled by a single integrated platform that allows both the input of voice (from a telephone, for example) to be realized as text (as an interactive text message) and allows the input of text (from the interactive text messaging application) to be realized as voice (on the telephone).
- Real-time communication may be enabled between any permutation of any number of text devices (desktop, PDA, mobile telephone) and voice devices (mobile telephone, regular telephone, etc.).
- a call to a text device user may be initiated by a voice device user or vice versa.
- FIG. 1 is a block diagram illustrating an exemplary computing environment in which aspects of the invention may be implemented
- FIG. 2 a is a block diagram of a system for cross-modal communication in accordance with some embodiments of the invention.
- FIG. 2 b is another block diagram of a system for cross-modal communication in accordance with embodiments of the invention.
- FIG. 3-5 are flow diagrams of methods for cross-modal communication in accordance with some embodiments of the invention.
- a VoIM Gateway (“Voice over IM”) is a platform which enables real-time communication between users of different modalities. The platform allows both the input of voice (from the telephone) to be realized as text (in the interactive text messaging client) and the input of text (from the text messaging client) to be realized as voice (on the telephone).
- Embodiments of the invention may permit multiple-user scenarios (i.e., more than two users), different originating devices, different speech recognition grammars and aspects of Voice User Interface (VUI) (interacting with a computer through a voice/speech platform in order to initiate an automated service or process), multimedia devices (e.g., streaming video with audio converted into text).
- VUI Voice User Interface
- Calls may be originated either from the voice-only client (e.g., a telephone user contacts an IM user) or from the text-only client (e.g., the interactive text messaging client user originates a call to a voice-only user).
- a single VoIM gateway converts input from one modality into another (from voice to text and from text to voice).
- Grammars and language models for the speech recognition component may be optimized for an interactive text messaging or other real-time text input interaction.
- a robust voice user interface (called herein the Personal Virtual Assistant or PVA) may be hosted on an enterprise communications server operating in real time.
- the communications server may integrate with a number of known corporate telephone systems but can process any type of voice channel.
- the PVA can provide the following services: caller identification, mixed-initiative multimodal call screening and negotiation, personalized call routing, structured message-taking, cross-modal messaging, voice dialing and voice-enabled message access and configuration.
- the PVA can communicate with the caller, classify the call and handle the disposition of the incoming call.
- the PVA application When a caller dials a callee's telephone number using a voice-only device, (and the callee is a PVA user) in some embodiments of the invention, the PVA application answers the caller's call and prompts the caller to identify himself/herself.
- An automated speech recognition component is used to process the caller's voice input. The caller may be asked the relative importance of the call, and other information such as the subject of the call and so on.
- An indication that the callee's PVA is trying to reach the callee may be sent to the callee. In some embodiments of the invention, the indication is sent in the form of a “toast”, a small notification window that pops-up on the screen, typically although not necessarily in the lower right side of the desktop.
- the toast indicates an incoming email, or that a member of a contacts list is available.
- the toast remains on the screen briefly and spontaneously disappears.
- the indication may include one or more of the following: the name of the caller, the relative importance of the call, the subject of the call or other information.
- the PVA may forward the call to a specific voice-capable destination, generate a structured voicemail, send a text alert to a suitable PVA user client, establish a dialog including a (potentially) cross-modal dialog between a voice-only caller and a text-only PVA user client, or transfer the call to another user.
- the PVA may support dialing from desktop and remote phones.
- a call may be answered, and the PVA may issue a user-configurable greeting, perhaps giving the caller an indication that the caller is interacting with an automated system (e.g., “Hi, this is Bill's Virtual Assistant, please speak clearly.”).
- the caller may be identified.
- the PVA may interact with the caller to identify the caller either as a recognized member of the callee's Contacts list or as someone who is not recognized, that is, who is not a member of the callee's Contacts list.
- An initial guess may be derived from the call's ANI (Automatic Number Identification is a feature of telephony intelligent network services which permits subscribers to display or capture the telephone numbers of calling parties), however, the ANI is often not available, not recognized (i.e., not in the Contacts list), ambiguous (associated with multiple contacts), or misleading (e.g., the caller is using a telephone belonging to someone else).
- the PVA can prompt the caller to identify himself/herself and confirm the elicited or ANI-derived identity, enabling caller-specific handling logic to be performed, instead of caller-independent handling logic, which is the norm today.
- the PVA may ask the caller to provide information which enables the PVA to characterize the call (phase three of call negotiation). For example, a caller from a specific contact group (e.g., contact group is Family, Supervisor, or Customer, etc.) may be able to provide information so that the call is handled in a particular way by the PVA.
- a specific contact group e.g., contact group is Family, Supervisor, or Customer, etc.
- the PVA may play an announcement that was specified by the PVA user (the callee, in this example) to the caller such as “Let me see if Bill is at his desk” and may use a combination of caller identity, caller characterization, callee presence and configured rules to attempt to alert the callee to the call and to obtain the callee's instructions for handling the call.
- the callee's desktop phone may ring, and a desktop alert such as a popup on the callee's computer can present the incoming call information with application call-handling choices (e.g., “Bill, Larry is calling and says it is urgent. Should I: 1) Take a message 2) Take a message and say you will call back in [N] minutes 3) Transfer call to [Cellular Telephone]”).
- application call-handling choices e.g., “Bill, Larry is calling and says it is urgent. Should I: 1) Take a message 2) Take a message and say you will call back in [N] minutes 3) Transfer call to [Cellular Telephone]”.
- the callee has the option to pick up the desktop telephone to take the call directly or can communicate with the PVA via the popup or other channels.
- the callee is not at his desk but can be contacted by alternate means, such as by an interactive text channel such as Instant Messenger (IM) or the like, a similar alert and list of options can be sent via the interactive text channel, through which the election may be made.
- IM Instant Messenger
- the callee can respond to a call-waiting ring by pressing a key which causes the PVA to play an audible message that includes the call information, so that the callee can indicate how the call should be handled before returning to the original call (i.e., the call the callee was taking when he received the call-waiting ring).
- the PVA may wait for the callee election, (with a suitable timeout period after which the PVA continues without callee election), before proceeding to the next phase of call negotiation.
- the next phase of call negotiation is caller election.
- the PVA may present a list of call handling options to the caller.
- the list may include one or more options. If there is only one available option, the PVA may inform the caller before proceeding without waiting for caller input (e.g., “Bill cannot take any calls right now, but he would like you to leave a message, and he will call you back in 30 minutes.” and proceed to the voicemail prompt, or alternatively, the PVA may merely proceed. If there are two or more choices, the PVA may prompt the caller to select one of the choices, (e.g., “Bill is away from his desk. Would you like me to forward your call to his cell, take a message and send him an alert, or set up an interactive text session with him?”).
- the PVA may include the name of the destination user if there is only one specified on the transfer list (“ . . . or be transferred to [name]”) or, if there are multiple destination users specified on the transfer list, the PVA may give a generic prompt such as “ . . . or be transferred to someone who can help you.”)
- the destination user may be a PVA user or may alternatively, be someone who is not a PVA user.
- the degree of access offered to the caller may reflect static rules in place for that particular caller and/or may depend on caller-specific instructions provided by the alerted callee to the PVA. Default rules may stipulate that leaving a message is always an option.
- the PVA may learn from interactions with callers. For example if a caller has elected to use cross modal communications for the past [N] times (may be specifiable by the PVA user) when calling a callee, this mode of communication may become the default for that caller-callee pair. Other alternatives may be removed unless the PVA user (or one of the PVA users in the case that both the caller and the callee are PVA users) elects to change the mode of communication for the caller/callee pair.
- Call handling includes call forwarding, saving a structured voicemail and sending an alert, instantiating an instant (cross-modal) voice messenger session with the callee and call transfer.
- Voice-only calls may be forwarded to voice-capable destinations such as to the callee's mobile telephone or other communications device.
- a text alert providing “envelope” information (information about when the call was received) may be delivered to one or more of the callee's available text-mode clients. The caller may be guided through the process of leaving a structured voicemail, so that voicemail-waiting alerts are able to provide more information to the callee.
- the caller may be prompted for the general topic of the call (e.g., “What is this regarding?”) and the best time and means for the callee to reply.
- the PVA may also prompt the caller to record a free-format body of the message.
- the structured information may be included in any text-mode or voicemail-waiting alert which the PVA sends to the callee, facilitating timely and efficient message retrieval.
- a cross-modal “instant voice messenger” conversation may be mediated by the PVA between a voice-only caller and a text-only callee (the callee is a PVA user).
- the PVA may prompt the caller to dictate a message, which may then be converted to text using a large vocabulary recognizer that has been tuned to an interactive text-messaging service such as Instant Messenger.
- the text of the caller's message may be sent to the callee's interactive text-messaging service (e.g., to an IM client) and the callee's reply may be played back to the caller using the PVA's text-to-speech (TTS) engine.
- TTS text-to-speech
- Call transfer provides preferential call handling for priority calls when the callee is unable to to take the call directly or does not want a caller to leave a message (perhaps because the callee will be out of touch for an extended period of time).
- the PVA may be configured to provide the caller with a list of alternative callees to whom the caller may be directed.
- the original callee may have recorded an explanatory prompt (“I will be out of the office until May 15th. For questions about release schedules, ask for Ray; for questions about features, ask for Chris . . . . How may I direct your call?”
- Voice-enabled message retrieval may include a robust interface for voice-only message retrieval including multi-modal support.
- a PVA user may call the PVA using an SIP telephone or communications device and use spoken queries to get sorted and filtered lists of message summaries delivered to the user's device.
- a user interface on the device may permit rapid viewing and selection and can play back the selected message in voice mode.
- the PVA may provide voice dialing and conferencing support, from both the desktop telephone and from remote dial-in, voice-enabling call-related functions accessible from the desktop telephone keypad or a desktop Communicator client.
- the callee may have full access to his contact list.
- the PVA may support voice commands such as “Call Larry”, “Call Larry in his office”, “Call Larry on his cellular telephone and conference in Tal” and “Keep trying Larry”. For the latter case, the PVA may call the callee back when Larry answers the call. While in a call, the PVA user may access relevant voice commands such as transfer, conferencing, etc. by pressing a key on the device.
- a user logged into an interactive text messaging client may send a text message via a network such as the Internet to the VoIM gateway.
- the VoIM gateway may convert the text message to a voice message and may select an appropriate pathway (VoIP Gateway or VoIP Telephone) depending on the recipient device. If the VoIP gateway is chosen, the VoIP gateway may forward the message to a PSTN telephone.
- FIG. 1 and the following discussion are intended to provide a brief general description of a suitable computing environment in which the invention may be implemented. It should be understood, however, that handheld, portable, and other computing devices of all kinds are contemplated for use in connection with the present invention. While a general purpose computer is described below, this is but one example, and the present invention requires only a thin client having network server interoperability and interaction. Thus, the present invention may be implemented in an environment of networked hosted services in which very little or minimal client resources are implicated, e.g., a networked environment in which the client device serves merely as a browser or interface to the World Wide Web.
- the invention can be implemented via an application programming interface (API), for use by a developer, and/or included within the network browsing software which will be described in the general context of computer-executable instructions, such as program modules, being executed by one or more computers, such as client workstations, servers, or other devices.
- program modules include routines, programs, objects, components, data structures and the like that perform particular tasks or implement particular abstract data types.
- the functionality of the program modules may be combined or distributed as desired in various embodiments.
- those skilled in the art will appreciate that the invention may be practiced with other computer system configurations.
- PCs personal computers
- automated teller machines server computers
- hand-held or laptop devices multi-processor systems
- microprocessor-based systems programmable consumer electronics
- network PCs minicomputers
- mainframe computers mainframe computers
- program modules may be located in both local and remote computer storage media including memory storage devices.
- FIG. 1 thus illustrates an example of a suitable computing system environment 100 in which the invention may be implemented, although as made clear above, the computing system environment 100 is only one example of a suitable computing environment and is not intended to suggest any limitation as to the scope of use or functionality of the invention. Neither should the computing environment 100 be interpreted as having any dependency or requirement relating to any one or combination of components illustrated in the exemplary operating environment 100 .
- an exemplary system for implementing the invention includes a general purpose computing device in the form of a computer 110 .
- Components of computer 110 may include, but are not limited to, a processing unit 120 , a system memory 130 , and a system bus 121 that couples various system components including the system memory to the processing unit 120 .
- the system bus 121 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures.
- such architectures include Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus (also known as Mezzanine bus).
- ISA Industry Standard Architecture
- MCA Micro Channel Architecture
- EISA Enhanced ISA
- VESA Video Electronics Standards Association
- PCI Peripheral Component Interconnect
- Computer 110 typically includes a variety of computer readable media.
- Computer readable media can be any available media that can be accessed by computer 110 and includes both volatile and nonvolatile media, removable and non-removable media.
- Computer readable media may comprise computer storage media and communication media.
- Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
- Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CDROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computer 110 .
- Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
- modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared, and other wireless media. Combinations of any of the above should also be included within the scope of computer readable media.
- the system memory 130 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 131 and random access memory (RAM) 132 .
- ROM read only memory
- RAM random access memory
- BIOS basic input/output system
- RAM 132 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 120 .
- FIG. 1 illustrates operating system 134 , application programs 135 , other program modules 136 , and program data 137 .
- the computer 110 may also include other removable/non-removable, volatile/nonvolatile computer storage media.
- FIG. 1 illustrates a hard disk drive 141 that reads from or writes to non-removable, nonvolatile magnetic media, a magnetic disk drive 151 that reads from or writes to a removable, nonvolatile magnetic disk 152 , and an optical disk drive 155 that reads from or writes to a removable, nonvolatile optical disk 156 , such as a CD ROM or other optical media.
- removable/non-removable, volatile/nonvolatile computer storage media that can be used in the exemplary operating environment include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, and the like.
- the hard disk drive 141 is typically connected to the system bus 121 through a non-removable memory interface such as interface 140
- magnetic disk drive 151 and optical disk drive 155 are typically connected to the system bus 121 by a removable memory interface, such as interface 150 .
- hard disk drive 141 is illustrated as storing operating system 144 , application programs 145 , other program modules 146 , and program data 147 . Note that these components can either be the same as or different from operating system 134 , application programs 135 , other program modules 136 , and program data 137 . Operating system 144 , application programs 145 , other program modules 146 , and program data 147 are given different numbers here to illustrate that, at a minimum, they are different copies.
- a user may enter commands and information into the computer 110 through input devices such as a keyboard 162 and pointing device 161 , commonly referred to as a mouse, trackball or touch pad.
- Other input devices may include a microphone, joystick, game pad, satellite dish, scanner, or the like.
- a user input interface 160 that is coupled to the system bus 121 , but may be connected by other interface and bus structures, such as a parallel port, game port or a universal serial bus (USB).
- USB universal serial bus
- a monitor 191 or other type of display device is also connected to the system bus 121 via an interface, such as a video interface 190 .
- a graphics interface 182 such as Northbridge, may also be connected to the system bus 121 .
- Northbridge is a chipset that communicates with the CPU, or host processing unit 120 , and assumes responsibility for accelerated graphics port (AGP) communications.
- One or more graphics processing units (GPUs) 184 may communicate with graphics interface 182 .
- GPUs 184 generally include on-chip memory storage, such as register storage and GPUs 184 communicate with a video memory 186 .
- GPUs 184 are but one example of a coprocessor and thus a variety of coprocessing devices may be included in computer 110 .
- a monitor 191 or other type of display device is also connected to the system bus 121 via an interface, such as a video interface 190 , which may in turn communicate with video memory 186 .
- computers may also include other peripheral output devices such as speakers 197 and printer 196 , which may be connected through an output peripheral interface 195 .
- the computer 110 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 180 .
- the remote computer 180 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 110 , although only a memory storage device 181 has been illustrated in FIG. 1 .
- the logical connections depicted in FIG. 1 include a local area network (LAN) 171 and a wide area network (WAN) 173 , but may also include other networks.
- LAN local area network
- WAN wide area network
- Such networking environments are commonplace in offices, enterprise-wide computer networks, intranets and the Internet.
- the computer 110 When used in a LAN networking environment, the computer 110 is connected to the LAN 171 through a network interface or adapter 170 .
- the computer 110 When used in a WAN networking environment, the computer 110 typically includes a modem 172 or other means for establishing communications over the WAN 173 , such as the Internet.
- the modem 172 which may be internal or external, may be connected to the system bus 121 via the user input interface 160 , or other appropriate mechanism.
- program modules depicted relative to the computer 110 may be stored in the remote memory storage device.
- FIG. 1 illustrates remote application programs 185 as residing on memory device 181 . It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers may be used.
- a computer 110 or other client device can be deployed as part of a computer network.
- the present invention pertains to any computer system having any number of memory or storage units, and any number of applications and processes occurring across any number of storage units or volumes.
- the present invention may apply to an environment with server computers and client computers deployed in a network environment, having remote or local storage.
- the present invention may also apply to a standalone computing device, having programming language functionality, interpretation and execution capabilities.
- FIG. 2 a illustrates an exemplary system 200 for cross-modal communication in accordance with embodiments of the invention.
- System 200 may reside on one or more computers such as the one illustrated above with respect to FIG. 1 .
- Personal Virtual Assistant (PVA) 202 may include one or more components including a (not shown) speech server comprising a conversational workflow engine including a language processing core and workflow-based APIs and a communication server such as Microsoft's Live Communication Server.
- a caller 204 may make a telephone call to a callee 212 (a PVA user) via any telephony communications device including a telephone, cellular telephone, PDA, smart phone, or any PC based soft-phone, and so on.
- caller 204 may represent any SIP (Session Initiation Protocol) telephony client including a communications server client such as a Microsoft Office Communicator client, a POTS telephone using VoIP gateways or any other telephone or communications device using the SIP protocol.
- Caller 204 may represent a PVA user, and thus the incoming call may be a call placed by another PVA.
- caller 204 may be a non-PVA user.
- the caller may be an interactive text messenger client as illustrated in FIG. 2 b, caller 213 making a call to a telephone user callee 205 .
- Call-handling graphical user interfaces may be provided to complement the voice interface and handle the text side of cross-modal interactions.
- Embodiments of the invention include one or more of the following areas of speech-enabled functionality: incoming call negotiation, call handling, voice mailbox access, voice dialing and PVA configuration.
- Incoming call negotiation may include one or more of the following: caller identification, call characterization, user alert and election and user election.
- PVA 202 may receive information such as contact information from a contact database 208 and rules including generalized rules 202 a and user-specific rules 202 b.
- a PVA user such as callee 212 may specify rules concerning how all or some portion of incoming calls are to be handled (e.g., “If I do not answer my telephone by the fourth ring, transfer my call to my assistant, Kathleen.” “If the caller identifies himself as “Sam”, tell him not to call me any more and hang up!”). Rules may apply to all calls received by the callee, may apply to all calls of a particular type, may apply to all calls from a particular (specified) caller, may apply to all calls from a specified group of callers, and so on. Information about the whereabouts of the callee 212 may be provided by a presence server 210 . Presence information can be obtained by information available from other applications running on the computer and may include information concerning whether the PVA user is logged onto his computer, whether or not the PVA is logged onto an interactive text messaging application, information available from a scheduling or calendar application
- caller 204 may or may not be a PVA user
- callee 212 is a PVA user.
- a dialog with PVA 202 may be initiated by PVA 202 .
- PVA 202 may greet the caller and (optionally) indicate that the PVA is an automated speech processor.
- caller Larry calls PVA user Bill's desktop telephone by dialing Bill's telephone number.
- the PVA may answer the call and greet the caller, e.g., PVA: “Hi, this is Bill's Virtual Assistant; please speak clearly.”
- PVA 202 may receive information from the IP network concerning the identity of the caller. For example, the ANI feature of the IP network may deliver the telephone number of the telephone Larry is using to PVA 202 .
- the PVA 202 may examine Contact list 208 for the number received from the network. If the received number is found in the Contact list 208 and (optionally) is unique, the PVA 202 may attempt to confirm that the caller is the person associated with the number, e.g., PVA: “Is this Larry calling?” If this is Larry calling, he may confirm his identity, e.g., Larry: “Yes”.
- a speech recognition component of PVA 202 is activated.
- the dialog that collects the information from the caller may be open-ended where the speech recognizer attempts to make a best-guess at name, topic, etc. based on global grammars or it may implement categories for selection based on predefined menus such as PVA: “What is the topic? Please say ‘work-related’, ‘non-work related’ or ‘confidential’ enabling much simpler grammars to be used.
- the dialog may be enhanced by sharing the presence of the callee with the caller, if the callee's rules allow this information to be shared with the caller. In this case, an extra step may be added that tells the caller that the callee is [wherever the callee is] before continuing on.
- PVA 202 may attempt to confirm identity. For example, suppose it is someone else, (perhaps using Larry's phone), the caller may identify himself, e.g., “No, this is Tal, on Larry's telephone.” The PVA 202 may check for the received name (e.g., “Tal”) in the Contact list 208 . If the received name is found in the Contact list 208 and is unique, the PVA 202 may determine that identity has been confirmed and go on to the negotiation phase. The PVA 202 may determine that the received name is not unique and may reprompt for identity.
- the received name e.g., “Tal”
- PVA 202 may conclude that the caller is unknown, record the non-Contact name and continue on to the negotiation phase. If no ANI information is available, PVA 202 may prompt for identity, (e.g., “May I ask who is calling?”) and then follow the rest of the procedure described above. If during the above described process, the name is not recognized with a high degree of confidence, the PVA 202 may reprompt, e.g., PVA: “I did not understand you, please repeat your name?”
- the urgency of the call may be elicited from the caller.
- the PVA 202 may ask “Is your call urgent?” and use the received information to classify the call to determine how to handle the call, as described more fully below.
- members of the Contacts list 208 can be grouped (e.g., “Florence” is a member of the Family group, “Larry” is a member of the Supervisor group, etc.) and only some groups may be prompted for urgency information, based on user-specified rules.
- PVA user the location or presence of the callee (PVA user) is determined If the PVA user is at his desk, logged into his computer and not using the telephone when the call is received, PVA 202 may cause the telephone to ring, and display a popup message on the PVA user's computer screen.
- An example of a popup message may be, PVA: “[name of caller] is calling and he says it is urgent.”
- PVA 202 may provide a list of options including but not limited to: PVA: “(1) Take Message, will reply in [N] minutes (2) Transfer to [name specified in user rules].
- the PVA user in some embodiments of the invention has one or more of the following options: [Case 1] The callee picks up the telephone to talk to the caller. [Case 2] The callee enters a number (e.g., types ‘20’) into the Take Message edit box and selects it. In response, PVA 202 may proceed to take a structured voicemail from the caller, and tells the caller that the callee will return his call within the callee-specified (in the example, within 20 minutes) reply time. [Case 3] The callee leaves the edit box empty, and selects the Take Message option.
- a number e.g., types ‘20’
- PVA 202 may proceed to take a structured voicemail from the caller, and tells the caller that the callee will return his call within the callee-specified (in the example, within 20 minutes) reply time.
- the callee leaves the edit box empty, and selects the Take Message option
- PVA 202 may proceed to take a structured voicemail, and does not specify a reply time to the caller.
- the callee selects Option 2 (Transfer call).
- PVA 202 may proceed to the Call Transfer function, with destination being the destination specified by the callee.
- the PVA 202 may ring the telephone with a special ring denoting that there is a second incoming call (call waiting ring).
- the callee (PVA user) may put his first caller on hold.
- the PVA 202 may indicate to the callee that he has a call from the confirmed identity caller and indicate (if collected) the urgency classification of the call, e.g., PVA: “[name of caller] is calling and he says it is urgent.”
- the callee may respond as described above.
- PVA 202 may display a popup on the callee's computer screen, such as “[name of caller] is calling and he says it is urgent. Should I (1) Forward his call to your mobile (2) Take a message and tell him you will call back in [N] minutes (3) Transfer the call to [person specified by PVA user]?” [Case 1] If the callee selects option 1, PVA 202 may proceed to Forward Call processing. In response to this selection, PVA 202 may forward the call to the PVA user's mobile telephone. [Case 2] The callee selects Option 2 or 3.
- PVA 202 may proceed to take a structured voicemail or perform Call Transfer processing, respectively. If the PVA user is not at his/her desk, is not logged onto his computer and receives an urgent call, the PVA 202 may discover via presence processing that the callee cannot be alerted via the PVA, and may proceed to the Caller Election phase.
- the PVA 202 may tell the caller that the callee is not at his desk, and that the caller can leave a message, e.g., PVA: “I'm sorry, but [callee name] is not at his desk, so I'll take a message for him, and he will get back to you as soon as he can.” If the caller is permitted by the user-specified rules to elect to forward a call, the PVA 202 can indicate to the caller that his call can either be forwarded to the callee's mobile communications device or he can be sent an alert, e.g., PVA: “I'm sorry, [name of caller], but [name of callee] is not at his desk.
- PVA 202 may also dispose of a call by transferring the call. PVA 202 may transfer a call to a single number or may dispose of the call by selecting from a list of numbers to which to transfer the call. Suppose for example, that the PVA user (the callee) will be out of the office for some period of time and wants the PVA to dispose of the call based on the subject matter of the incoming call. When a call comes in to the callee's number, PVA 202 may play a prompt recorded by the PVA user for this purpose, e.g., Bill: “I'm out of the office until June 15. For schedule questions, contact Ray. For beta questions, speak with Alex. For M4 questions, talk to Chris.
- PVA 202 may process the caller's reply and transfer the call based on the recognized name or key word in the caller's response; e.g., Larry: “Oh, um, put me through to Ray.”
- PVA may respond with a message such as PVA: “Please wait while I transfer your call to Ray.”
- the person to whom the message is transferred may be a PVA user, and thus the call transferred to “Ray” may itself be handled by a PVA.
- FIG. 2 b illustrates an exemplary system 201 for cross-modal communication in accordance with embodiments of the invention.
- System 201 may reside on one or more computers such as the one illustrated above with respect to FIG. 1 .
- Personal Virtual Assistant (PVA) 202 may include one or more components including a (not shown) speech server comprising a conversational workflow engine including a language processing core and workflow-based APIs and a communication server such as Microsoft's Live Communication Server.
- a caller 213 (a PVA user) may make a telephone call to a callee 205 via an interactive text messaging client connected to a VoIM gateway as described above.
- callee 205 may represent any SIP (Session Initiation Protocol) telephony client including a communications server client such as a Microsoft Office Communicator client, a POTS telephone using VoIP gateways or any other telephone or communications device using the SIP protocol.
- Callee 205 may represent a PVA user, and thus the incoming call may be a call placed by another PVA.
- callee 205 may be a non-PVA user.
- Call-handling graphical user interfaces may be provided to complement the voice interface and handle the text side of cross-modal interactions.
- Embodiments of the invention include one or more of the following areas of speech-enabled functionality: incoming call negotiation, call handling, voice mailbox access, voice dialing and PVA configuration.
- Incoming call negotiation may include one or more of the following: caller identification, call characterization, user alert and election and user election.
- PVA 202 may receive information such as contact information from a contact database 208 and rules including generalized rules 202 a and user-specific rules 202 b.
- a PVA user such as callee 205 may specify rules concerning how all or some portion of incoming calls are to be handled (e.g., “If I do not answer my telephone by the fourth ring, transfer my call to my assistant, Kathleen.” “If the caller identifies himself as “Sam”, tell him not to call me any more and hang up!”).
- Rules may apply to all calls received by the callee, may apply to all calls of a particular type, may apply to all calls from a particular (specified) caller, may apply to all calls from a specified group of callers, and so on.
- Information about the whereabouts of the callee 212 may be provided by a presence server 210 . Presence information can be obtained by information available from other applications running on the computer and may include information concerning whether the PVA user is logged onto his computer, whether or not the PVA is logged onto an interactive text messaging application, information available from a scheduling or calendar application
- callee 205 may or may not be a PVA user
- caller 213 is a PVA user.
- a dialog with PVA 202 may be initiated by PVA 202 .
- FIG. 3 is an exemplary flow diagram of a method for cross-modal communication in accordance with embodiments of the invention.
- Caller information may include information such as ANI information, DNIS (dialed number identification service) information, SIP invite address or other information.
- additional information may be elicited from the caller using automated speech recognition processing. If the information received at 302 and 304 is sufficient, this information may be used to look up the rules-defined access privileges for the caller ( 306 ).
- the caller telephone number is looked up in the contact list.
- the contact list may include caller telephone number, caller name, the group (e.g., Family, Supervisor, Customer, etc.) to which the caller belongs.
- special rules may be applied.
- additional information may be requested from the caller. For example, if the caller belongs to a Family group, the caller may be prompted for call classification information such as urgency of the call. Rules in the rules database may be based on this classification. For example, an urgent call from a Family member may receive some kind of preferential treatment, as specified in the rules for this type of call from this type of caller.
- callee presence information is obtained. Presence information can be obtained by information concerning whether the PVA user is logged onto his computer, whether or not the PVA is logged onto an interactive text messaging application, information available from a scheduling or calendar application.
- callee determines if the callee has access to user alert. If the callee does have access to user alert and the callee is on-line ( 314 ), an alert is sent ( 316 ). If callee input is received specifying the call disposition desired ( 326 ), the call is handled according to the callee input ( 328 ). If no callee input is received after a specifiable time period or the callee does not have access to user alert, at 320 , the caller is provided with a list of whatever choices are enabled by the access privileges of the caller. These may include forward call, voicemail, transfer to another number, instant voice messenger and so on as described above. The caller's choice is obtained using automated speech recognition ( 322 ). At 324 if the option chosen is to initiate an instant messaging session with the callee, cross-modal communication as described below is initiated.
- FIG. 4 is a method for cross-modal communication when the caller is a voice-device user in accordance with embodiments of the invention.
- the call is received by the PVA.
- the PVA answers the call and receives whatever information is available for the call as described above.
- the available information is processed according to PVA user-configurable rules. Assuming that enough information is available to make a final call disposition, that disposition is enacted. If there is not enough information to make a call disposition decision, the source and nature of the next information to obtain is determined
- the options available are presented to the caller.
- the PVA in response to receiving a user selection of initiating an interactive text messaging session with the callee, the PVA initiates an interactive text messaging session.
- the PVA may prompt the caller for a message for the callee.
- the PVA may convert the voice message to text at 414 , send the text message to the IM client (the callee) at 416 , receive a text answer from the callee at 418 , convert the text answer to voice at 420 , send the voice message to the caller at 422 , receive the voice message response from the caller at 424 and continue processing at 414 until the cross-modal conversation terminates.
- FIG. 5 is a method for cross-modal communication when the caller is an interactive text-messaging client in accordance with embodiments of the invention.
- the PVA receives a request to make a cross-modal call from a an interactive text-messaging client who is a PVA user.
- the PVA receives the request including whatever information is available for the call as described above.
- the available information is processed according to PVA user-configurable rules. Assuming that enough information is available to make a final call disposition, that disposition is enacted. If there is not enough information to make a call disposition decision, the source and nature of the next information to obtain is determined
- the options available are presented to the caller.
- the PVA in response to receiving a user selection of one of the options, the PVA initiates a telephony call.
- the PVA may prompt the caller for a message for the callee.
- the PVA may convert the text message to voice at 514 , send the voice message to the telephony client (the callee) at 516 , receive a voice answer from the callee at 518 , convert the voice answer to text at 520 , send the text message to the caller at 522 , receive the text message response from the caller at 524 and continue processing at 514 until the cross-modal conversation terminates.
- the various techniques described herein may be implemented in connection with hardware or software or, where appropriate, with a combination of both.
- the methods and apparatus of the present invention may take the form of program code (i.e., instructions) embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention.
- the computing device will generally include a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device.
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 13/412,413 filed on Mar. 5, 2012, which is a continuation of U.S. patent application Ser. No. 11/695,826 filed on Apr. 3, 2007, now U.S. Pat. No. 8,131,556. The entire contents of U.S. patent application Ser. No. 13/412,413 and U.S. patent application Ser. No. 11/695,826 are herein incorporated herein by reference in their entireties.
- Telephony encompasses the general use of equipment to provide voice communication over a distance. Plain old telephone service, or POTS, is the most basic form of residential and small business telephone service. POTS became available soon after the telephone system was introduced in the late 19th century and, from the standpoint of the user, has existed more or less unchanged ever since. POTS services include bi-directional or full duplex voice path with a limited frequency range, a dial tone and ringing signal, subscriber dialing, operator services such as directory assistance, and long distance and conference calling. During the 1970s and 1980s, new network services became available due to the creation of electronic telephone exchanges and computerization. New services included voice mail, caller identification, call waiting, reminder calls and other similar services.
- Advances in digital electronics have revolutionized telephony by providing alternate means of voice communication than those provided by traditional (analog) telephone systems. IP Telephony is a form of telephony which uses the TCP/IP protocol popularized by the Internet to transmit digitized voice data. The routing of voice conversations over the Internet or through other IP networks is also called VoIP (Voice over Internet Protocol). Digital telephony was introduced to improve voice services, but was subsequently found to be very useful in the creation of new network services because it can transfer data quickly over telephone lines. Computer Telephony Integration (CTI) enables a computer to control phone functions such as making and receiving voice, fax, and data calls. The Session Initiation Protocol (SIP) is a signaling protocol used for creating, modifying and terminating sessions (voice or video calls) with one or more participants. Sessions include Internet telephone calls, multimedia distribution and multimedia conferences. Development of the SIP protocol was motivated by the need for a signaling and call setup protocol for IP-based communications that could support the call processing functions and features present in the public switched telephone network (PSTN) using proxy servers and user agents.
- Interactive text messaging is a form of real-time communication between two or more people based on typed text. Interactive text messaging differs from email in that text conversations are conducted in real-time using computers connected over a network such as the Internet. A client program on a text messaging user's computer connects to an interactive text messaging service. The user types a message and the service sends the text message to the recipient. Popular interactive text messaging services on the Internet include .NET Messenger Service, AOL Instant Messenger, Excite/Pal, Gadu-Gadu, Google Talk, iChat, ICQ, Jabber, Qnext, QQ, Meetro, Skype, Trillian, Yahoo! Messenger and Rediff Bol Instant Messenger.
- All the advances in telephony and interactive text messaging, however, have not solved some very basic problems associated with actually reaching a desired party. Communication in real-time is currently impossible between people who have access only to devices of different modalities. For example, when one user only has access to an ordinary (voice-only) telephone and another only has access to a (text only) Instant Messaging client, there is no way for the two users to communicate in real-time. The below addresses these and other problems.
- Communications between users of different modalities is enabled by a single integrated platform that allows both the input of voice (from a telephone, for example) to be realized as text (as an interactive text message) and allows the input of text (from the interactive text messaging application) to be realized as voice (on the telephone). Real-time communication may be enabled between any permutation of any number of text devices (desktop, PDA, mobile telephone) and voice devices (mobile telephone, regular telephone, etc.). A call to a text device user may be initiated by a voice device user or vice versa.
-
FIG. 1 is a block diagram illustrating an exemplary computing environment in which aspects of the invention may be implemented; -
FIG. 2 a is a block diagram of a system for cross-modal communication in accordance with some embodiments of the invention; -
FIG. 2 b is another block diagram of a system for cross-modal communication in accordance with embodiments of the invention; and -
FIG. 3-5 are flow diagrams of methods for cross-modal communication in accordance with some embodiments of the invention. - Overview
- Currently, communication in real-time is impossible between people who only have access to devices of different modalities. One example is when one user only has a voice-only telephone and the other user only has access to an interactive text messaging service, (as frequently occurs, for example, when the interactive text messaging service user is in a meeting and a telephone user tries to call him). A VoIM Gateway (“Voice over IM”) is a platform which enables real-time communication between users of different modalities. The platform allows both the input of voice (from the telephone) to be realized as text (in the interactive text messaging client) and the input of text (from the text messaging client) to be realized as voice (on the telephone). While described in terms of an interactive text messaging application such as IM, other forms of text messaging utilizing SMS (e.g., to cellular telephones), HTML, email, etc. are contemplated. Any permutation of text devices (desktop, PDA, mobile telephone having text capabilities, etc.) and voice devices (mobile telephone, standard telephone, etc.) may be used. Embodiments of the invention may permit multiple-user scenarios (i.e., more than two users), different originating devices, different speech recognition grammars and aspects of Voice User Interface (VUI) (interacting with a computer through a voice/speech platform in order to initiate an automated service or process), multimedia devices (e.g., streaming video with audio converted into text). Calls may be originated either from the voice-only client (e.g., a telephone user contacts an IM user) or from the text-only client (e.g., the interactive text messaging client user originates a call to a voice-only user). A single VoIM gateway converts input from one modality into another (from voice to text and from text to voice). Grammars and language models for the speech recognition component may be optimized for an interactive text messaging or other real-time text input interaction.
- With respect to calls originated from a voice-only client, a robust voice user interface, (called herein the Personal Virtual Assistant or PVA) may be hosted on an enterprise communications server operating in real time. The communications server may integrate with a number of known corporate telephone systems but can process any type of voice channel. The PVA can provide the following services: caller identification, mixed-initiative multimodal call screening and negotiation, personalized call routing, structured message-taking, cross-modal messaging, voice dialing and voice-enabled message access and configuration. The PVA can communicate with the caller, classify the call and handle the disposition of the incoming call.
- When a caller dials a callee's telephone number using a voice-only device, (and the callee is a PVA user) in some embodiments of the invention, the PVA application answers the caller's call and prompts the caller to identify himself/herself. An automated speech recognition component is used to process the caller's voice input. The caller may be asked the relative importance of the call, and other information such as the subject of the call and so on. An indication that the callee's PVA is trying to reach the callee may be sent to the callee. In some embodiments of the invention, the indication is sent in the form of a “toast”, a small notification window that pops-up on the screen, typically although not necessarily in the lower right side of the desktop. The toast indicates an incoming email, or that a member of a contacts list is available. The toast remains on the screen briefly and spontaneously disappears. The indication may include one or more of the following: the name of the caller, the relative importance of the call, the subject of the call or other information. In response to instructions received from direct interaction of the PVA with the callee or to default instructions stored in a rules database (generic or caller-specific), the PVA may forward the call to a specific voice-capable destination, generate a structured voicemail, send a text alert to a suitable PVA user client, establish a dialog including a (potentially) cross-modal dialog between a voice-only caller and a text-only PVA user client, or transfer the call to another user. The PVA may support dialing from desktop and remote phones.
- For example, in a greeting phase of call negotiation, a call may be answered, and the PVA may issue a user-configurable greeting, perhaps giving the caller an indication that the caller is interacting with an automated system (e.g., “Hi, this is Bill's Virtual Assistant, please speak clearly.”). In the next phase of call negotiation, the caller may be identified. The PVA may interact with the caller to identify the caller either as a recognized member of the callee's Contacts list or as someone who is not recognized, that is, who is not a member of the callee's Contacts list. An initial guess may be derived from the call's ANI (Automatic Number Identification is a feature of telephony intelligent network services which permits subscribers to display or capture the telephone numbers of calling parties), however, the ANI is often not available, not recognized (i.e., not in the Contacts list), ambiguous (associated with multiple contacts), or misleading (e.g., the caller is using a telephone belonging to someone else). The PVA can prompt the caller to identify himself/herself and confirm the elicited or ANI-derived identity, enabling caller-specific handling logic to be performed, instead of caller-independent handling logic, which is the norm today.
- Before alerting the callee to the call, the PVA may ask the caller to provide information which enables the PVA to characterize the call (phase three of call negotiation). For example, a caller from a specific contact group (e.g., contact group is Family, Supervisor, or Customer, etc.) may be able to provide information so that the call is handled in a particular way by the PVA. (For example, a member of a Family or Supervisor contact group may be extended the opportunity to say the call is urgent to facilitate special call handling.) In the fourth phase of call negotiation, the PVA may play an announcement that was specified by the PVA user (the callee, in this example) to the caller such as “Let me see if Bill is at his desk” and may use a combination of caller identity, caller characterization, callee presence and configured rules to attempt to alert the callee to the call and to obtain the callee's instructions for handling the call. For example, if the callee's presence indicates that the callee is at his/her desk, the callee's desktop phone may ring, and a desktop alert such as a popup on the callee's computer can present the incoming call information with application call-handling choices (e.g., “Bill, Larry is calling and says it is urgent. Should I: 1) Take a message 2) Take a message and say you will call back in [N] minutes 3) Transfer call to [Cellular Telephone]”). The callee has the option to pick up the desktop telephone to take the call directly or can communicate with the PVA via the popup or other channels. If the callee is not at his desk but can be contacted by alternate means, such as by an interactive text channel such as Instant Messenger (IM) or the like, a similar alert and list of options can be sent via the interactive text channel, through which the election may be made. If the callee is already on the desktop telephone but is not online (i.e., not able to receive a desktop alert), the callee can respond to a call-waiting ring by pressing a key which causes the PVA to play an audible message that includes the call information, so that the callee can indicate how the call should be handled before returning to the original call (i.e., the call the callee was taking when he received the call-waiting ring). In each of these cases, the PVA may wait for the callee election, (with a suitable timeout period after which the PVA continues without callee election), before proceeding to the next phase of call negotiation. The next phase of call negotiation is caller election.
- In the caller election phase, the PVA may present a list of call handling options to the caller. The list may include one or more options. If there is only one available option, the PVA may inform the caller before proceeding without waiting for caller input (e.g., “Bill cannot take any calls right now, but he would like you to leave a message, and he will call you back in 30 minutes.” and proceed to the voicemail prompt, or alternatively, the PVA may merely proceed. If there are two or more choices, the PVA may prompt the caller to select one of the choices, (e.g., “Bill is away from his desk. Would you like me to forward your call to his cell, take a message and send him an alert, or set up an interactive text session with him?”). If one of the callee-specified ways to handle the call is to transfer the call, the PVA may include the name of the destination user if there is only one specified on the transfer list (“ . . . or be transferred to [name]”) or, if there are multiple destination users specified on the transfer list, the PVA may give a generic prompt such as “ . . . or be transferred to someone who can help you.”) The destination user may be a PVA user or may alternatively, be someone who is not a PVA user. The degree of access offered to the caller may reflect static rules in place for that particular caller and/or may depend on caller-specific instructions provided by the alerted callee to the PVA. Default rules may stipulate that leaving a message is always an option.
- In some embodiments of the invention, the PVA may learn from interactions with callers. For example if a caller has elected to use cross modal communications for the past [N] times (may be specifiable by the PVA user) when calling a callee, this mode of communication may become the default for that caller-callee pair. Other alternatives may be removed unless the PVA user (or one of the PVA users in the case that both the caller and the callee are PVA users) elects to change the mode of communication for the caller/callee pair.
- Call handling includes call forwarding, saving a structured voicemail and sending an alert, instantiating an instant (cross-modal) voice messenger session with the callee and call transfer. Voice-only calls may be forwarded to voice-capable destinations such as to the callee's mobile telephone or other communications device. A text alert providing “envelope” information (information about when the call was received) may be delivered to one or more of the callee's available text-mode clients. The caller may be guided through the process of leaving a structured voicemail, so that voicemail-waiting alerts are able to provide more information to the callee. In addition to the identity and call characterization described above, the caller may be prompted for the general topic of the call (e.g., “What is this regarding?”) and the best time and means for the callee to reply. The PVA may also prompt the caller to record a free-format body of the message. The structured information may be included in any text-mode or voicemail-waiting alert which the PVA sends to the callee, facilitating timely and efficient message retrieval.
- A cross-modal “instant voice messenger” conversation may be mediated by the PVA between a voice-only caller and a text-only callee (the callee is a PVA user). The PVA may prompt the caller to dictate a message, which may then be converted to text using a large vocabulary recognizer that has been tuned to an interactive text-messaging service such as Instant Messenger. The text of the caller's message may be sent to the callee's interactive text-messaging service (e.g., to an IM client) and the callee's reply may be played back to the caller using the PVA's text-to-speech (TTS) engine. Thus the callee is able to “chat” using an interactive text messaging service with a voice-only caller in situations where it would be disruptive or otherwise undesirable to accept a voice call and engage in an audible conversation.
- Call transfer provides preferential call handling for priority calls when the callee is unable to to take the call directly or does not want a caller to leave a message (perhaps because the callee will be out of touch for an extended period of time). The PVA may be configured to provide the caller with a list of alternative callees to whom the caller may be directed. The original callee may have recorded an explanatory prompt (“I will be out of the office until May 15th. For questions about release schedules, ask for Ray; for questions about features, ask for Chris . . . . How may I direct your call?”
- Voice-enabled message retrieval may include a robust interface for voice-only message retrieval including multi-modal support. A PVA user may call the PVA using an SIP telephone or communications device and use spoken queries to get sorted and filtered lists of message summaries delivered to the user's device. A user interface on the device may permit rapid viewing and selection and can play back the selected message in voice mode.
- The PVA may provide voice dialing and conferencing support, from both the desktop telephone and from remote dial-in, voice-enabling call-related functions accessible from the desktop telephone keypad or a desktop Communicator client. The callee may have full access to his contact list. The PVA may support voice commands such as “Call Larry”, “Call Larry in his office”, “Call Larry on his cellular telephone and conference in Tal” and “Keep trying Larry”. For the latter case, the PVA may call the callee back when Larry answers the call. While in a call, the PVA user may access relevant voice commands such as transfer, conferencing, etc. by pressing a key on the device.
- With respect to a call originated from the interactive text messaging user to a voice only user, a user logged into an interactive text messaging client may send a text message via a network such as the Internet to the VoIM gateway. The VoIM gateway may convert the text message to a voice message and may select an appropriate pathway (VoIP Gateway or VoIP Telephone) depending on the recipient device. If the VoIP gateway is chosen, the VoIP gateway may forward the message to a PSTN telephone.
- Exemplary Computing Environment
-
FIG. 1 and the following discussion are intended to provide a brief general description of a suitable computing environment in which the invention may be implemented. It should be understood, however, that handheld, portable, and other computing devices of all kinds are contemplated for use in connection with the present invention. While a general purpose computer is described below, this is but one example, and the present invention requires only a thin client having network server interoperability and interaction. Thus, the present invention may be implemented in an environment of networked hosted services in which very little or minimal client resources are implicated, e.g., a networked environment in which the client device serves merely as a browser or interface to the World Wide Web. - Although not required, the invention can be implemented via an application programming interface (API), for use by a developer, and/or included within the network browsing software which will be described in the general context of computer-executable instructions, such as program modules, being executed by one or more computers, such as client workstations, servers, or other devices. Generally, program modules include routines, programs, objects, components, data structures and the like that perform particular tasks or implement particular abstract data types. Typically, the functionality of the program modules may be combined or distributed as desired in various embodiments. Moreover, those skilled in the art will appreciate that the invention may be practiced with other computer system configurations. Other well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to, personal computers (PCs), automated teller machines, server computers, hand-held or laptop devices, multi-processor systems, microprocessor-based systems, programmable consumer electronics, network PCs, minicomputers, mainframe computers, and the like. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network or other data transmission medium. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
-
FIG. 1 thus illustrates an example of a suitablecomputing system environment 100 in which the invention may be implemented, although as made clear above, thecomputing system environment 100 is only one example of a suitable computing environment and is not intended to suggest any limitation as to the scope of use or functionality of the invention. Neither should thecomputing environment 100 be interpreted as having any dependency or requirement relating to any one or combination of components illustrated in theexemplary operating environment 100. - With reference to
FIG. 1 , an exemplary system for implementing the invention includes a general purpose computing device in the form of acomputer 110. Components ofcomputer 110 may include, but are not limited to, aprocessing unit 120, asystem memory 130, and a system bus 121 that couples various system components including the system memory to theprocessing unit 120. The system bus 121 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. By way of example, and not limitation, such architectures include Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus (also known as Mezzanine bus). -
Computer 110 typically includes a variety of computer readable media. Computer readable media can be any available media that can be accessed bycomputer 110 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer readable media may comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CDROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed bycomputer 110. Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared, and other wireless media. Combinations of any of the above should also be included within the scope of computer readable media. - The
system memory 130 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 131 and random access memory (RAM) 132. A basic input/output system 133 (BIOS), containing the basic routines that help to transfer information between elements withincomputer 110, such as during start-up, is typically stored inROM 131.RAM 132 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processingunit 120. By way of example, and not limitation,FIG. 1 illustratesoperating system 134, application programs 135,other program modules 136, andprogram data 137. - The
computer 110 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,FIG. 1 illustrates ahard disk drive 141 that reads from or writes to non-removable, nonvolatile magnetic media, amagnetic disk drive 151 that reads from or writes to a removable, nonvolatilemagnetic disk 152, and anoptical disk drive 155 that reads from or writes to a removable, nonvolatileoptical disk 156, such as a CD ROM or other optical media. Other removable/non-removable, volatile/nonvolatile computer storage media that can be used in the exemplary operating environment include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, and the like. Thehard disk drive 141 is typically connected to the system bus 121 through a non-removable memory interface such asinterface 140, andmagnetic disk drive 151 andoptical disk drive 155 are typically connected to the system bus 121 by a removable memory interface, such asinterface 150. - The drives and their associated computer storage media discussed above and illustrated in
FIG. 1 provide storage of computer readable instructions, data structures, program modules and other data for thecomputer 110. InFIG. 1 , for example,hard disk drive 141 is illustrated as storingoperating system 144,application programs 145,other program modules 146, andprogram data 147. Note that these components can either be the same as or different fromoperating system 134, application programs 135,other program modules 136, andprogram data 137.Operating system 144,application programs 145,other program modules 146, andprogram data 147 are given different numbers here to illustrate that, at a minimum, they are different copies. A user may enter commands and information into thecomputer 110 through input devices such as akeyboard 162 andpointing device 161, commonly referred to as a mouse, trackball or touch pad. Other input devices (not shown) may include a microphone, joystick, game pad, satellite dish, scanner, or the like. These and other input devices are often connected to theprocessing unit 120 through auser input interface 160 that is coupled to the system bus 121, but may be connected by other interface and bus structures, such as a parallel port, game port or a universal serial bus (USB). - A
monitor 191 or other type of display device is also connected to the system bus 121 via an interface, such as avideo interface 190. Agraphics interface 182, such as Northbridge, may also be connected to the system bus 121. Northbridge is a chipset that communicates with the CPU, orhost processing unit 120, and assumes responsibility for accelerated graphics port (AGP) communications. One or more graphics processing units (GPUs) 184 may communicate withgraphics interface 182. In this regard,GPUs 184 generally include on-chip memory storage, such as register storage andGPUs 184 communicate with avideo memory 186.GPUs 184, however, are but one example of a coprocessor and thus a variety of coprocessing devices may be included incomputer 110. Amonitor 191 or other type of display device is also connected to the system bus 121 via an interface, such as avideo interface 190, which may in turn communicate withvideo memory 186. In addition to monitor 191, computers may also include other peripheral output devices such asspeakers 197 andprinter 196, which may be connected through an output peripheral interface 195. - The
computer 110 may operate in a networked environment using logical connections to one or more remote computers, such as aremote computer 180. Theremote computer 180 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to thecomputer 110, although only amemory storage device 181 has been illustrated inFIG. 1 . The logical connections depicted inFIG. 1 include a local area network (LAN) 171 and a wide area network (WAN) 173, but may also include other networks. Such networking environments are commonplace in offices, enterprise-wide computer networks, intranets and the Internet. - When used in a LAN networking environment, the
computer 110 is connected to theLAN 171 through a network interface oradapter 170. When used in a WAN networking environment, thecomputer 110 typically includes amodem 172 or other means for establishing communications over theWAN 173, such as the Internet. Themodem 172, which may be internal or external, may be connected to the system bus 121 via theuser input interface 160, or other appropriate mechanism. In a networked environment, program modules depicted relative to thecomputer 110, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,FIG. 1 illustratesremote application programs 185 as residing onmemory device 181. It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers may be used. - One of ordinary skill in the art can appreciate that a
computer 110 or other client device can be deployed as part of a computer network. In this regard, the present invention pertains to any computer system having any number of memory or storage units, and any number of applications and processes occurring across any number of storage units or volumes. The present invention may apply to an environment with server computers and client computers deployed in a network environment, having remote or local storage. The present invention may also apply to a standalone computing device, having programming language functionality, interpretation and execution capabilities. - Cross-Modal Communication
-
FIG. 2 a illustrates anexemplary system 200 for cross-modal communication in accordance with embodiments of the invention.System 200 may reside on one or more computers such as the one illustrated above with respect toFIG. 1 . Personal Virtual Assistant (PVA) 202 may include one or more components including a (not shown) speech server comprising a conversational workflow engine including a language processing core and workflow-based APIs and a communication server such as Microsoft's Live Communication Server. Acaller 204 may make a telephone call to a callee 212 (a PVA user) via any telephony communications device including a telephone, cellular telephone, PDA, smart phone, or any PC based soft-phone, and so on. Thus,caller 204 may represent any SIP (Session Initiation Protocol) telephony client including a communications server client such as a Microsoft Office Communicator client, a POTS telephone using VoIP gateways or any other telephone or communications device using the SIP protocol.Caller 204 may represent a PVA user, and thus the incoming call may be a call placed by another PVA. Alternatively,caller 204 may be a non-PVA user. Alternatively, the caller may be an interactive text messenger client as illustrated inFIG. 2 b,caller 213 making a call to atelephone user callee 205. Call-handling graphical user interfaces may be provided to complement the voice interface and handle the text side of cross-modal interactions. Embodiments of the invention include one or more of the following areas of speech-enabled functionality: incoming call negotiation, call handling, voice mailbox access, voice dialing and PVA configuration. Incoming call negotiation may include one or more of the following: caller identification, call characterization, user alert and election and user election.PVA 202 may receive information such as contact information from acontact database 208 and rules includinggeneralized rules 202 a and user-specific rules 202 b. For example, a PVA user such ascallee 212 may specify rules concerning how all or some portion of incoming calls are to be handled (e.g., “If I do not answer my telephone by the fourth ring, transfer my call to my assistant, Kathleen.” “If the caller identifies himself as “Sam”, tell him not to call me any more and hang up!”). Rules may apply to all calls received by the callee, may apply to all calls of a particular type, may apply to all calls from a particular (specified) caller, may apply to all calls from a specified group of callers, and so on. Information about the whereabouts of the callee 212 may be provided by apresence server 210. Presence information can be obtained by information available from other applications running on the computer and may include information concerning whether the PVA user is logged onto his computer, whether or not the PVA is logged onto an interactive text messaging application, information available from a scheduling or calendar application - While
caller 204 may or may not be a PVA user, inFIG. 2 a,callee 212 is a PVA user. In response to receivingcaller 204's call toPVA user 212, a dialog withPVA 202 may be initiated byPVA 202. During phase one (the greeting phase)PVA 202 may greet the caller and (optionally) indicate that the PVA is an automated speech processor. Suppose, for example, that caller Larry calls PVA user Bill's desktop telephone by dialing Bill's telephone number. After a specified number of rings, the PVA may answer the call and greet the caller, e.g., PVA: “Hi, this is Bill's Virtual Assistant; please speak clearly.” -
PVA 202 may receive information from the IP network concerning the identity of the caller. For example, the ANI feature of the IP network may deliver the telephone number of the telephone Larry is using toPVA 202. ThePVA 202 may examineContact list 208 for the number received from the network. If the received number is found in theContact list 208 and (optionally) is unique, thePVA 202 may attempt to confirm that the caller is the person associated with the number, e.g., PVA: “Is this Larry calling?” If this is Larry calling, he may confirm his identity, e.g., Larry: “Yes”. Upon receiving voice input fromcaller 204, a speech recognition component ofPVA 202 is activated. The dialog that collects the information from the caller may be open-ended where the speech recognizer attempts to make a best-guess at name, topic, etc. based on global grammars or it may implement categories for selection based on predefined menus such as PVA: “What is the topic? Please say ‘work-related’, ‘non-work related’ or ‘confidential’ enabling much simpler grammars to be used. The dialog may be enhanced by sharing the presence of the callee with the caller, if the callee's rules allow this information to be shared with the caller. In this case, an extra step may be added that tells the caller that the callee is [wherever the callee is] before continuing on. When the identity of the caller is established, processing continues to the next phase (negotiation). If the identity of the caller has not yet been established,PVA 202 may attempt to confirm identity. For example, suppose it is someone else, (perhaps using Larry's phone), the caller may identify himself, e.g., “No, this is Tal, on Larry's telephone.” ThePVA 202 may check for the received name (e.g., “Tal”) in theContact list 208. If the received name is found in theContact list 208 and is unique, thePVA 202 may determine that identity has been confirmed and go on to the negotiation phase. ThePVA 202 may determine that the received name is not unique and may reprompt for identity. If no match can be found in theContact List 208,PVA 202 may conclude that the caller is unknown, record the non-Contact name and continue on to the negotiation phase. If no ANI information is available,PVA 202 may prompt for identity, (e.g., “May I ask who is calling?”) and then follow the rest of the procedure described above. If during the above described process, the name is not recognized with a high degree of confidence, thePVA 202 may reprompt, e.g., PVA: “I did not understand you, please repeat your name?” - During the classification portion of the negotiation phase, the urgency of the call may be elicited from the caller. For example, the
PVA 202 may ask “Is your call urgent?” and use the received information to classify the call to determine how to handle the call, as described more fully below. In some embodiments of the invention, members of the Contacts list 208 can be grouped (e.g., “Florence” is a member of the Family group, “Larry” is a member of the Supervisor group, etc.) and only some groups may be prompted for urgency information, based on user-specified rules. - During a user alert and election portion of the negotiation phase, the location or presence of the callee (PVA user) is determined If the PVA user is at his desk, logged into his computer and not using the telephone when the call is received,
PVA 202 may cause the telephone to ring, and display a popup message on the PVA user's computer screen. An example of a popup message may be, PVA: “[name of caller] is calling and he says it is urgent.”PVA 202 may provide a list of options including but not limited to: PVA: “(1) Take Message, will reply in [N] minutes (2) Transfer to [name specified in user rules]. The PVA user in some embodiments of the invention has one or more of the following options: [Case 1] The callee picks up the telephone to talk to the caller. [Case 2] The callee enters a number (e.g., types ‘20’) into the Take Message edit box and selects it. In response,PVA 202 may proceed to take a structured voicemail from the caller, and tells the caller that the callee will return his call within the callee-specified (in the example, within 20 minutes) reply time. [Case 3] The callee leaves the edit box empty, and selects the Take Message option. Inresponse PVA 202 may proceed to take a structured voicemail, and does not specify a reply time to the caller. [Case 4] The callee selects Option 2 (Transfer call). In response,PVA 202 may proceed to the Call Transfer function, with destination being the destination specified by the callee. - If the PVA user is at his/her desk, on the telephone but not logged onto his computer, the
PVA 202 may ring the telephone with a special ring denoting that there is a second incoming call (call waiting ring). The callee (PVA user) may put his first caller on hold. In response thePVA 202 may indicate to the callee that he has a call from the confirmed identity caller and indicate (if collected) the urgency classification of the call, e.g., PVA: “[name of caller] is calling and he says it is urgent.” The callee may respond as described above. If the PVA user is not at his/her desk but is logged onto his computer,PVA 202 may display a popup on the callee's computer screen, such as “[name of caller] is calling and he says it is urgent. Should I (1) Forward his call to your mobile (2) Take a message and tell him you will call back in [N] minutes (3) Transfer the call to [person specified by PVA user]?” [Case 1] If the callee selects option 1,PVA 202 may proceed to Forward Call processing. In response to this selection,PVA 202 may forward the call to the PVA user's mobile telephone. [Case 2] The callee selects Option 2 or 3. In response to these selections,PVA 202 may proceed to take a structured voicemail or perform Call Transfer processing, respectively. If the PVA user is not at his/her desk, is not logged onto his computer and receives an urgent call, thePVA 202 may discover via presence processing that the callee cannot be alerted via the PVA, and may proceed to the Caller Election phase. - If the caller is limited to leaving a structured voicemail by the user-specified rules, the
PVA 202 may tell the caller that the callee is not at his desk, and that the caller can leave a message, e.g., PVA: “I'm sorry, but [callee name] is not at his desk, so I'll take a message for him, and he will get back to you as soon as he can.” If the caller is permitted by the user-specified rules to elect to forward a call, thePVA 202 can indicate to the caller that his call can either be forwarded to the callee's mobile communications device or he can be sent an alert, e.g., PVA: “I'm sorry, [name of caller], but [name of callee] is not at his desk. Would you like me to forward your call to his mobile, or take a message here and send him an alert?” [Case 1] If the caller opts to leave a message, the caller may say so and PVA may proceed to structured voicemail handling. [Case 2] If the caller opts to forward the call, the caller may say so: e.g., Caller: “Please forward the call.” and PVA may proceed to forward the call to callee's mobile number. -
PVA 202 may also dispose of a call by transferring the call.PVA 202 may transfer a call to a single number or may dispose of the call by selecting from a list of numbers to which to transfer the call. Suppose for example, that the PVA user (the callee) will be out of the office for some period of time and wants the PVA to dispose of the call based on the subject matter of the incoming call. When a call comes in to the callee's number,PVA 202 may play a prompt recorded by the PVA user for this purpose, e.g., Bill: “I'm out of the office until June 15. For schedule questions, contact Ray. For beta questions, speak with Alex. For M4 questions, talk to Chris. For all other issues, please ask to speak with Kathleen, or leave a message which I will get when I return. How shall I direct your call?”PVA 202 may process the caller's reply and transfer the call based on the recognized name or key word in the caller's response; e.g., Larry: “Oh, um, put me through to Ray.” In response to detecting the word “Ray” PVA may respond with a message such as PVA: “Please wait while I transfer your call to Ray.” The person to whom the message is transferred may be a PVA user, and thus the call transferred to “Ray” may itself be handled by a PVA. PVA: “Larry, please wait while I establish an IVM session with Bill . . . . OK, please speak your message clearly.” -
FIG. 2 b illustrates anexemplary system 201 for cross-modal communication in accordance with embodiments of the invention.System 201 may reside on one or more computers such as the one illustrated above with respect toFIG. 1 . Personal Virtual Assistant (PVA) 202 may include one or more components including a (not shown) speech server comprising a conversational workflow engine including a language processing core and workflow-based APIs and a communication server such as Microsoft's Live Communication Server. A caller 213 (a PVA user) may make a telephone call to acallee 205 via an interactive text messaging client connected to a VoIM gateway as described above. Thus,callee 205 may represent any SIP (Session Initiation Protocol) telephony client including a communications server client such as a Microsoft Office Communicator client, a POTS telephone using VoIP gateways or any other telephone or communications device using the SIP protocol.Callee 205 may represent a PVA user, and thus the incoming call may be a call placed by another PVA. Alternatively,callee 205 may be a non-PVA user. Call-handling graphical user interfaces may be provided to complement the voice interface and handle the text side of cross-modal interactions. Embodiments of the invention include one or more of the following areas of speech-enabled functionality: incoming call negotiation, call handling, voice mailbox access, voice dialing and PVA configuration. Incoming call negotiation may include one or more of the following: caller identification, call characterization, user alert and election and user election.PVA 202 may receive information such as contact information from acontact database 208 and rules includinggeneralized rules 202 a and user-specific rules 202 b. For example, a PVA user such ascallee 205 may specify rules concerning how all or some portion of incoming calls are to be handled (e.g., “If I do not answer my telephone by the fourth ring, transfer my call to my assistant, Kathleen.” “If the caller identifies himself as “Sam”, tell him not to call me any more and hang up!”). Rules may apply to all calls received by the callee, may apply to all calls of a particular type, may apply to all calls from a particular (specified) caller, may apply to all calls from a specified group of callers, and so on. Information about the whereabouts of the callee 212 may be provided by apresence server 210. Presence information can be obtained by information available from other applications running on the computer and may include information concerning whether the PVA user is logged onto his computer, whether or not the PVA is logged onto an interactive text messaging application, information available from a scheduling or calendar application - Thus, while callee 205 may or may not be a PVA user, in
FIG. 2 b,caller 213 is a PVA user. In response to receivingcaller 213's call to callee 205, a dialog withPVA 202 may be initiated byPVA 202. -
FIG. 3 is an exemplary flow diagram of a method for cross-modal communication in accordance with embodiments of the invention. At 302 caller information is received. Caller information may include information such as ANI information, DNIS (dialed number identification service) information, SIP invite address or other information. At 304 additional information may be elicited from the caller using automated speech recognition processing. If the information received at 302 and 304 is sufficient, this information may be used to look up the rules-defined access privileges for the caller (306). In some embodiments of the invention, the caller telephone number is looked up in the contact list. The contact list may include caller telephone number, caller name, the group (e.g., Family, Supervisor, Customer, etc.) to which the caller belongs. Based on belonging to a particular group, special rules (from rules database 210) may be applied. At 308, based on the access privileges, additional information may be requested from the caller. For example, if the caller belongs to a Family group, the caller may be prompted for call classification information such as urgency of the call. Rules in the rules database may be based on this classification. For example, an urgent call from a Family member may receive some kind of preferential treatment, as specified in the rules for this type of call from this type of caller. At 310, callee presence information is obtained. Presence information can be obtained by information concerning whether the PVA user is logged onto his computer, whether or not the PVA is logged onto an interactive text messaging application, information available from a scheduling or calendar application. At 312 it is determined if the callee has access to user alert. If the callee does have access to user alert and the callee is on-line (314), an alert is sent (316). If callee input is received specifying the call disposition desired (326), the call is handled according to the callee input (328). If no callee input is received after a specifiable time period or the callee does not have access to user alert, at 320, the caller is provided with a list of whatever choices are enabled by the access privileges of the caller. These may include forward call, voicemail, transfer to another number, instant voice messenger and so on as described above. The caller's choice is obtained using automated speech recognition (322). At 324 if the option chosen is to initiate an instant messaging session with the callee, cross-modal communication as described below is initiated. -
FIG. 4 is a method for cross-modal communication when the caller is a voice-device user in accordance with embodiments of the invention. At 402 the call is received by the PVA. At 404, the PVA answers the call and receives whatever information is available for the call as described above. At 406 the available information is processed according to PVA user-configurable rules. Assuming that enough information is available to make a final call disposition, that disposition is enacted. If there is not enough information to make a call disposition decision, the source and nature of the next information to obtain is determined At 408 the options available are presented to the caller. At 410, in response to receiving a user selection of initiating an interactive text messaging session with the callee, the PVA initiates an interactive text messaging session. At 412, the PVA may prompt the caller for a message for the callee. Upon receiving the voice message, the PVA may convert the voice message to text at 414, send the text message to the IM client (the callee) at 416, receive a text answer from the callee at 418, convert the text answer to voice at 420, send the voice message to the caller at 422, receive the voice message response from the caller at 424 and continue processing at 414 until the cross-modal conversation terminates. -
FIG. 5 is a method for cross-modal communication when the caller is an interactive text-messaging client in accordance with embodiments of the invention. At 502 the PVA receives a request to make a cross-modal call from a an interactive text-messaging client who is a PVA user. At 504, the PVA receives the request including whatever information is available for the call as described above. At 506 the available information is processed according to PVA user-configurable rules. Assuming that enough information is available to make a final call disposition, that disposition is enacted. If there is not enough information to make a call disposition decision, the source and nature of the next information to obtain is determined At 508 the options available are presented to the caller. At 510, in response to receiving a user selection of one of the options, the PVA initiates a telephony call. At 512, the PVA may prompt the caller for a message for the callee. Upon receiving the text message, the PVA may convert the text message to voice at 514, send the voice message to the telephony client (the callee) at 516, receive a voice answer from the callee at 518, convert the voice answer to text at 520, send the text message to the caller at 522, receive the text message response from the caller at 524 and continue processing at 514 until the cross-modal conversation terminates. - The various techniques described herein may be implemented in connection with hardware or software or, where appropriate, with a combination of both. Thus, the methods and apparatus of the present invention, or certain aspects or portions thereof, may take the form of program code (i.e., instructions) embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention. In the case of program code execution on programmable computers, the computing device will generally include a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device. One or more programs that may utilize the creation and/or implementation of domain-specific programming models aspects of the present invention, e.g., through the use of a data processing API or the like, are preferably implemented in a high level procedural or object oriented programming language to communicate with a computer system. However, the program(s) can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language, and combined with hardware implementations.
- While the present invention has been described in connection with the preferred embodiments of the various figures, it is to be understood that other similar embodiments may be used or modifications and additions may be made to the described embodiments for performing the same function of the present invention without deviating therefrom. Therefore, the present invention should not be limited to any single embodiment, but rather should be construed in breadth and scope in accordance with the appended claims.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/456,008 US20150030144A1 (en) | 2007-04-03 | 2014-08-11 | Communications using different modalities |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/695,826 US8131556B2 (en) | 2007-04-03 | 2007-04-03 | Communications using different modalities |
US13/412,413 US8805688B2 (en) | 2007-04-03 | 2012-03-05 | Communications using different modalities |
US14/456,008 US20150030144A1 (en) | 2007-04-03 | 2014-08-11 | Communications using different modalities |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/412,413 Continuation US8805688B2 (en) | 2007-04-03 | 2012-03-05 | Communications using different modalities |
Publications (1)
Publication Number | Publication Date |
---|---|
US20150030144A1 true US20150030144A1 (en) | 2015-01-29 |
Family
ID=39827725
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/695,826 Expired - Fee Related US8131556B2 (en) | 2007-04-03 | 2007-04-03 | Communications using different modalities |
US13/412,413 Expired - Fee Related US8805688B2 (en) | 2007-04-03 | 2012-03-05 | Communications using different modalities |
US14/456,008 Abandoned US20150030144A1 (en) | 2007-04-03 | 2014-08-11 | Communications using different modalities |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/695,826 Expired - Fee Related US8131556B2 (en) | 2007-04-03 | 2007-04-03 | Communications using different modalities |
US13/412,413 Expired - Fee Related US8805688B2 (en) | 2007-04-03 | 2012-03-05 | Communications using different modalities |
Country Status (1)
Country | Link |
---|---|
US (3) | US8131556B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180338038A1 (en) * | 2017-05-16 | 2018-11-22 | Google Llc | Handling calls on a shared speech-enabled device |
Families Citing this family (154)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US9420072B2 (en) | 2003-04-25 | 2016-08-16 | Z124 | Smartphone databoost |
US8520511B2 (en) * | 2003-09-11 | 2013-08-27 | Qualcomm Incorporated | Automatic handling of incoming communications at a wireless device |
US7742581B2 (en) | 2004-11-24 | 2010-06-22 | Value-Added Communications, Inc. | Electronic messaging exchange |
US9876915B2 (en) | 2005-01-28 | 2018-01-23 | Value-Added Communications, Inc. | Message exchange |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8983051B2 (en) | 2007-04-03 | 2015-03-17 | William F. Barton | Outgoing call classification and disposition |
US8131556B2 (en) * | 2007-04-03 | 2012-03-06 | Microsoft Corporation | Communications using different modalities |
US20080273678A1 (en) * | 2007-05-01 | 2008-11-06 | Igor Balk | Systems and methods for phone call management |
US8041765B1 (en) * | 2007-05-17 | 2011-10-18 | Avaya Inc. | Method and apparatus providing sender information by way of a personal virtual assistant (PVA) |
US7986914B1 (en) * | 2007-06-01 | 2011-07-26 | At&T Mobility Ii Llc | Vehicle-based message control using cellular IP |
US10002189B2 (en) | 2007-12-20 | 2018-06-19 | Apple Inc. | Method and apparatus for searching using an active ontology |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
TWI435589B (en) * | 2008-03-18 | 2014-04-21 | Wistron Corp | Voip integrating system and method thereof |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US8881020B2 (en) * | 2008-06-24 | 2014-11-04 | Microsoft Corporation | Multi-modal communication through modal-specific interfaces |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
EP2377286B1 (en) | 2008-12-17 | 2021-03-03 | Telefonaktiebolaget LM Ericsson (publ) | A method of and a network server and mobile user equipment for providing chat/voip services in a mobile telecommunications network |
CN101465825B (en) * | 2008-12-31 | 2012-07-11 | 阿里巴巴集团控股有限公司 | Instant communication method, instant communication server, voice server and system |
US9621714B2 (en) | 2009-01-27 | 2017-04-11 | Value-Added Communications, Inc. | System and method for electronic notification in institutional communication |
US8265239B2 (en) * | 2009-02-25 | 2012-09-11 | International Business Machines Corporation | Callee centric location and presence enabled voicemail using session initiated protocol enabled signaling for IP multimedia subsystem networks |
US20100238842A1 (en) * | 2009-03-19 | 2010-09-23 | Microsoft Corporation | Phone conferencing architecture with optimized services management |
US8594296B2 (en) * | 2009-05-20 | 2013-11-26 | Microsoft Corporation | Multimodal callback tagging |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US9213776B1 (en) | 2009-07-17 | 2015-12-15 | Open Invention Network, Llc | Method and system for searching network resources to locate content |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US20110244840A1 (en) * | 2010-03-30 | 2011-10-06 | Telenav, Inc. | Communication system with caller identification mechanism and method of operation thereof |
US8751682B2 (en) | 2010-09-27 | 2014-06-10 | Z124 | Data transfer using high speed connection, high integrity connection, and descriptor |
US8788576B2 (en) | 2010-09-27 | 2014-07-22 | Z124 | High speed parallel data exchange with receiver side data handling |
WO2012040855A2 (en) * | 2010-09-29 | 2012-04-05 | Research In Motion Limited | Method to exchange application specific information for outgoing calls between two mobile devices |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US20130244631A1 (en) * | 2011-03-30 | 2013-09-19 | Devika Phutane | Systems and methods for generating a shortened notification of an incoming call |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US8499051B2 (en) * | 2011-07-21 | 2013-07-30 | Z124 | Multiple messaging communication optimization |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
US20130060587A1 (en) * | 2011-09-02 | 2013-03-07 | International Business Machines Corporation | Determining best time to reach customers in a multi-channel world ensuring right party contact and increasing interaction likelihood |
US9774721B2 (en) | 2011-09-27 | 2017-09-26 | Z124 | LTE upgrade module |
US9495012B2 (en) | 2011-09-27 | 2016-11-15 | Z124 | Secondary single screen mode activation through user interface activation |
US8924853B2 (en) * | 2011-10-07 | 2014-12-30 | Blackberry Limited | Apparatus, and associated method, for cognitively translating media to facilitate understanding |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US8861689B1 (en) | 2012-06-08 | 2014-10-14 | Amazon Technologies, Inc. | System and method to facilitate communication between users via different modalities |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
CN113470640B (en) | 2013-02-07 | 2022-04-26 | 苹果公司 | Voice trigger of digital assistant |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
KR101922663B1 (en) | 2013-06-09 | 2018-11-28 | 애플 인크. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
DE112014003653B4 (en) * | 2013-08-06 | 2024-04-18 | Apple Inc. | Automatically activate intelligent responses based on activities from remote devices |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9775009B2 (en) * | 2014-04-23 | 2017-09-26 | Cellco Partnership | Method and system for coordinating a communication response |
TWI566107B (en) | 2014-05-30 | 2017-01-11 | 蘋果公司 | Method for processing a multi-part voice command, non-transitory computer readable storage medium and electronic device |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10375129B2 (en) * | 2014-06-17 | 2019-08-06 | Microsoft Technology Licensing, Llc | Facilitating conversations with automated location mapping |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9412379B2 (en) | 2014-09-16 | 2016-08-09 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method for initiating a wireless communication link using voice recognition |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
CN104660771A (en) * | 2015-03-18 | 2015-05-27 | 上海斐讯数据通信技术有限公司 | Short message management method, system and electronic terminal |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179588B1 (en) | 2016-06-09 | 2019-02-22 | Apple Inc. | Intelligent automated assistant in a home environment |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
CN105872267A (en) * | 2016-06-27 | 2016-08-17 | 维沃移动通信有限公司 | Short message management method and mobile terminal |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US10771629B2 (en) | 2017-02-06 | 2020-09-08 | babyTel Inc. | System and method for transforming a voicemail into a communication session |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
US10749827B2 (en) | 2017-05-11 | 2020-08-18 | Global Tel*Link Corporation | System and method for inmate notification and training in a controlled environment facility |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | Low-latency intelligent automated assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10841755B2 (en) | 2017-07-01 | 2020-11-17 | Phoneic, Inc. | Call routing using call forwarding options in telephony networks |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
CN109299216B (en) * | 2018-10-29 | 2019-07-23 | 山东师范大学 | A kind of cross-module state Hash search method and system merging supervision message |
US11468890B2 (en) | 2019-06-01 | 2022-10-11 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11438452B1 (en) | 2019-08-09 | 2022-09-06 | Apple Inc. | Propagating context information in a privacy preserving manner |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11115373B1 (en) * | 2020-06-11 | 2021-09-07 | Movius Interactive Corporation | Multi-channel engagement platform converter |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US11805189B1 (en) * | 2020-09-17 | 2023-10-31 | Intrado Life & Safety, Inc. | Publish and subscribe call center architecture |
US11909918B2 (en) * | 2022-01-28 | 2024-02-20 | Zoom Video Communications, Inc. | Mapping a universal contact center service access point to a service access point specific to a determined modality |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5724410A (en) * | 1995-12-18 | 1998-03-03 | Sony Corporation | Two-way voice messaging terminal having a speech to text converter |
US6618704B2 (en) * | 2000-12-01 | 2003-09-09 | Ibm Corporation | System and method of teleconferencing with the deaf or hearing-impaired |
US20040019487A1 (en) * | 2002-03-11 | 2004-01-29 | International Business Machines Corporation | Multi-modal messaging |
US20040267527A1 (en) * | 2003-06-25 | 2004-12-30 | International Business Machines Corporation | Voice-to-text reduction for real time IM/chat/SMS |
US20050021344A1 (en) * | 2003-07-24 | 2005-01-27 | International Business Machines Corporation | Access to enhanced conferencing services using the tele-chat system |
US7058578B2 (en) * | 2002-09-24 | 2006-06-06 | Rockwell Electronic Commerce Technologies, L.L.C. | Media translator for transaction processing system |
US20080192732A1 (en) * | 2007-02-13 | 2008-08-14 | John Riley | Method and system for multi-modal communications |
US8131556B2 (en) * | 2007-04-03 | 2012-03-06 | Microsoft Corporation | Communications using different modalities |
Family Cites Families (82)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5815566A (en) | 1991-10-10 | 1998-09-29 | Executone Information Systems, Inc. | Apparatus and method for dynamic inbound/outbound call management and for scheduling appointments |
US5329578A (en) | 1992-05-26 | 1994-07-12 | Northern Telecom Limited | Personal communication service with mobility manager |
US7142846B1 (en) | 1994-01-05 | 2006-11-28 | Henderson Daniel A | Method and apparatus for improved paging receiver and system |
US20010036821A1 (en) | 1994-04-19 | 2001-11-01 | Jay L. Gainsboro | Computer-based method and apparatus for controlling, monitoring, recording and reporting wireless communications |
US6804332B1 (en) | 1994-09-30 | 2004-10-12 | Wildfire Communications, Inc. | Network based knowledgeable assistant |
US5930700A (en) | 1995-11-29 | 1999-07-27 | Bell Communications Research, Inc. | System and method for automatically screening and directing incoming calls |
US7136475B1 (en) | 1999-07-27 | 2006-11-14 | Aspect Communications Corporation | Call Management system with call control from user workstation computers |
US5946386A (en) | 1996-03-11 | 1999-08-31 | Xantel Corporation | Call management system with call control from user workstation computers |
US6377993B1 (en) | 1997-09-26 | 2002-04-23 | Mci Worldcom, Inc. | Integrated proxy interface for web based data management reports |
US6198808B1 (en) * | 1997-12-31 | 2001-03-06 | Weblink Wireless, Inc. | Controller for use with communications systems for converting a voice message to a text message |
US6219638B1 (en) * | 1998-11-03 | 2001-04-17 | International Business Machines Corporation | Telephone messaging and editing system |
US6477551B1 (en) * | 1999-02-16 | 2002-11-05 | International Business Machines Corporation | Interactive electronic messaging system |
US20030147518A1 (en) | 1999-06-30 | 2003-08-07 | Nandakishore A. Albal | Methods and apparatus to deliver caller identification information |
US7886008B2 (en) | 1999-07-28 | 2011-02-08 | Rpost International Limited | System and method for verifying delivery and integrity of electronic messages |
US7136458B1 (en) | 1999-12-23 | 2006-11-14 | Bellsouth Intellectual Property Corporation | Voice recognition for filtering and announcing message |
US6625258B1 (en) | 1999-12-27 | 2003-09-23 | Nortel Networks Ltd | System and method for providing unified communication services support |
US6415018B1 (en) | 2000-02-08 | 2002-07-02 | Lucent Technologies Inc. | Telecommunication system and method for handling special number calls having geographic sensitivity |
US7058036B1 (en) * | 2000-02-25 | 2006-06-06 | Sprint Spectrum L.P. | Method and system for wireless instant messaging |
US6757362B1 (en) | 2000-03-06 | 2004-06-29 | Avaya Technology Corp. | Personal virtual assistant |
WO2001067241A1 (en) | 2000-03-06 | 2001-09-13 | Conita Technologies, Inc. | Virtual assistant engine |
CA2304353A1 (en) | 2000-04-06 | 2001-10-06 | Stephen Murphy | Telephone call manager |
US6965920B2 (en) | 2000-07-12 | 2005-11-15 | Peter Henrik Pedersen | Profile responsive electronic message management system |
US7400712B2 (en) | 2001-01-18 | 2008-07-15 | Lucent Technologies Inc. | Network provided information using text-to-speech and speech recognition and text or speech activated network control sequences for complimentary feature access |
US6964023B2 (en) * | 2001-02-05 | 2005-11-08 | International Business Machines Corporation | System and method for multi-modal focus detection, referential ambiguity resolution and mood classification using multi-modal input |
US7289522B2 (en) | 2001-03-20 | 2007-10-30 | Verizon Business Global Llc | Shared dedicated access line (DAL) gateway routing discrimination |
US6996076B1 (en) | 2001-03-29 | 2006-02-07 | Sonus Networks, Inc. | System and method to internetwork wireless telecommunication networks |
GB0113570D0 (en) * | 2001-06-04 | 2001-07-25 | Hewlett Packard Co | Audio-form presentation of text messages |
AU2002344786A1 (en) | 2001-06-18 | 2003-01-02 | Jeffrey L. Crandell | Apparatus, systems and methods for managing incoming and outgoing communication |
US7333507B2 (en) * | 2001-08-31 | 2008-02-19 | Philip Bravin | Multi modal communications system |
US7640006B2 (en) * | 2001-10-03 | 2009-12-29 | Accenture Global Services Gmbh | Directory assistance with multi-modal messaging |
US7254384B2 (en) * | 2001-10-03 | 2007-08-07 | Accenture Global Services Gmbh | Multi-modal messaging |
US20030135569A1 (en) * | 2002-01-15 | 2003-07-17 | Khakoo Shabbir A. | Method and apparatus for delivering messages based on user presence, preference or location |
EP1495625B1 (en) * | 2002-04-02 | 2011-09-28 | Verizon Business Global LLC | Providing of presence information to a telephony services system |
US20040198327A1 (en) | 2002-04-22 | 2004-10-07 | Mavis Bates | Telecommunications call completion |
US7383570B2 (en) | 2002-04-25 | 2008-06-03 | Intertrust Technologies, Corp. | Secure authentication systems and methods |
US7065185B1 (en) * | 2002-06-28 | 2006-06-20 | Bellsouth Intellectual Property Corp. | Systems and methods for providing real-time conversation using disparate communication devices |
US7139916B2 (en) | 2002-06-28 | 2006-11-21 | Ebay, Inc. | Method and system for monitoring user interaction with a computer |
US7640293B2 (en) * | 2002-07-17 | 2009-12-29 | Research In Motion Limited | Method, system and apparatus for messaging between wireless mobile terminals and networked computers |
JP4007893B2 (en) | 2002-10-03 | 2007-11-14 | 株式会社エヌ・ティ・ティ・ドコモ | Server device, program, and recording medium |
US7715538B2 (en) | 2003-04-29 | 2010-05-11 | At&T Intellectual Property I, L.P. | Privacy screening services |
US20040267531A1 (en) * | 2003-06-30 | 2004-12-30 | Whynot Stephen R. | Method and system for providing text-to-speech instant messaging |
US7308082B2 (en) * | 2003-07-24 | 2007-12-11 | International Business Machines Corporation | Method to enable instant collaboration via use of pervasive messaging |
US7725395B2 (en) | 2003-09-19 | 2010-05-25 | Microsoft Corp. | System and method for devising a human interactive proof that determines whether a remote client is a human or a computer program |
US20050125541A1 (en) * | 2003-12-04 | 2005-06-09 | Randall Frank | Integrating multiple communication modes |
US7197646B2 (en) | 2003-12-19 | 2007-03-27 | Disney Enterprises, Inc. | System and method for preventing automated programs in a network |
US7804949B2 (en) | 2003-12-31 | 2010-09-28 | Alcatel Lucent | Client-based integration of PBX and messaging systems |
US7295833B2 (en) | 2004-04-09 | 2007-11-13 | At&T Mobility Ii Llc | Spam control for sharing content on mobile devices |
WO2005115035A1 (en) * | 2004-05-20 | 2005-12-01 | Research In Motion Limited | Handling an audio conference related to a text-based message |
GB0412888D0 (en) | 2004-06-10 | 2004-07-14 | Nortel Networks Ltd | Method of operating a contact center |
US8050272B2 (en) | 2004-06-29 | 2011-11-01 | Damaka, Inc. | System and method for concurrent sessions in a peer-to-peer hybrid communications network |
US8190686B2 (en) | 2004-08-17 | 2012-05-29 | Alcatel Lucent | Spam filtering for mobile communication devices |
US7451921B2 (en) | 2004-09-01 | 2008-11-18 | Eric Morgan Dowling | Methods, smart cards, and systems for providing portable computer, VoIP, and application services |
US7983706B2 (en) * | 2004-09-07 | 2011-07-19 | At&T Intellectual Property I, L.P. | System and method for voice and text based service interworking |
EP1643739A1 (en) * | 2004-10-04 | 2006-04-05 | Alcatel | Method for exchanging messages between end-user and textmessage server |
US7533419B2 (en) | 2004-10-29 | 2009-05-12 | Microsoft Corporation | Human interactive proof service |
US7912192B2 (en) | 2005-02-15 | 2011-03-22 | At&T Intellectual Property Ii, L.P. | Arrangement for managing voice over IP (VoIP) telephone calls, especially unsolicited or unwanted calls |
US20060182236A1 (en) * | 2005-02-17 | 2006-08-17 | Siemens Communications, Inc. | Speech conversion for text messaging |
US7974608B2 (en) | 2005-03-03 | 2011-07-05 | Alcatel-Lucent Usa Inc. | Anonymous call blocking in wireless networks |
US7551574B1 (en) | 2005-03-31 | 2009-06-23 | Trapeze Networks, Inc. | Method and apparatus for controlling wireless network access privileges based on wireless client location |
US8107495B2 (en) * | 2005-05-13 | 2012-01-31 | Yahoo! Inc. | Integrating access to audio messages and instant messaging with VOIP |
KR100834818B1 (en) | 2005-05-25 | 2008-06-05 | 삼성전자주식회사 | Method for providing location based service in mobile communication system |
US20070022481A1 (en) | 2005-07-22 | 2007-01-25 | Goldman Stuart O | Network support for restricting call terminations in a security risk area |
US20070026372A1 (en) | 2005-07-27 | 2007-02-01 | Huelsbergen Lorenz F | Method for providing machine access security by deciding whether an anonymous responder is a human or a machine using a human interactive proof |
US20070033258A1 (en) | 2005-08-04 | 2007-02-08 | Walter Vasilaky | System and method for an email firewall and use thereof |
US8605718B2 (en) | 2005-08-30 | 2013-12-10 | Babitech Ltd. | Immediate communication system |
US7920692B2 (en) | 2005-10-03 | 2011-04-05 | Verizon Data Services Llc | PBX call management |
US20080089499A1 (en) | 2005-12-09 | 2008-04-17 | American Telecom Services, Inc. | Apparatus, System, Method and Computer Program Product for Pre-Paid Long Distance Telecommunications and Charitable Fee Sharing |
US8549651B2 (en) | 2007-02-02 | 2013-10-01 | Facebook, Inc. | Determining a trust level in a social network environment |
US8077849B2 (en) | 2006-01-10 | 2011-12-13 | Utbk, Inc. | Systems and methods to block communication calls |
US7515698B2 (en) | 2006-04-17 | 2009-04-07 | At & T Intellectual Property I, L.P. | System and method for providing telephone call notification and management in a network environment |
US7552467B2 (en) | 2006-04-24 | 2009-06-23 | Jeffrey Dean Lindsay | Security systems for protecting an asset |
US20070286395A1 (en) * | 2006-05-24 | 2007-12-13 | International Business Machines Corporation | Intelligent Multimedia Dial Tone |
US8744054B2 (en) | 2006-06-29 | 2014-06-03 | Apple Inc. | Method and system for automatic call redialing |
US8601538B2 (en) | 2006-08-22 | 2013-12-03 | Fuji Xerox Co., Ltd. | Motion and interaction based CAPTCHA |
US8837704B2 (en) | 2006-08-31 | 2014-09-16 | Microsoft Corporation | Client controlled dynamic call forwarding |
US8292741B2 (en) | 2006-10-26 | 2012-10-23 | Cfph, Llc | Apparatus, processes and articles for facilitating mobile gaming |
US8000276B2 (en) | 2007-02-05 | 2011-08-16 | Wefi, Inc. | Providing easy access to radio networks |
US20080208589A1 (en) * | 2007-02-27 | 2008-08-28 | Cross Charles W | Presenting Supplemental Content For Digital Media Using A Multimodal Application |
US7827265B2 (en) | 2007-03-23 | 2010-11-02 | Facebook, Inc. | System and method for confirming an association in a web-based social network |
US20080247529A1 (en) | 2007-04-03 | 2008-10-09 | Microsoft Corporation | Incoming Call Classification And Disposition |
US8983051B2 (en) | 2007-04-03 | 2015-03-17 | William F. Barton | Outgoing call classification and disposition |
US20090083826A1 (en) | 2007-09-21 | 2009-03-26 | Microsoft Corporation | Unsolicited communication management via mobile device |
-
2007
- 2007-04-03 US US11/695,826 patent/US8131556B2/en not_active Expired - Fee Related
-
2012
- 2012-03-05 US US13/412,413 patent/US8805688B2/en not_active Expired - Fee Related
-
2014
- 2014-08-11 US US14/456,008 patent/US20150030144A1/en not_active Abandoned
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5724410A (en) * | 1995-12-18 | 1998-03-03 | Sony Corporation | Two-way voice messaging terminal having a speech to text converter |
US6618704B2 (en) * | 2000-12-01 | 2003-09-09 | Ibm Corporation | System and method of teleconferencing with the deaf or hearing-impaired |
US20040019487A1 (en) * | 2002-03-11 | 2004-01-29 | International Business Machines Corporation | Multi-modal messaging |
US7058578B2 (en) * | 2002-09-24 | 2006-06-06 | Rockwell Electronic Commerce Technologies, L.L.C. | Media translator for transaction processing system |
US20040267527A1 (en) * | 2003-06-25 | 2004-12-30 | International Business Machines Corporation | Voice-to-text reduction for real time IM/chat/SMS |
US20050021344A1 (en) * | 2003-07-24 | 2005-01-27 | International Business Machines Corporation | Access to enhanced conferencing services using the tele-chat system |
US20080192732A1 (en) * | 2007-02-13 | 2008-08-14 | John Riley | Method and system for multi-modal communications |
US8131556B2 (en) * | 2007-04-03 | 2012-03-06 | Microsoft Corporation | Communications using different modalities |
US8805688B2 (en) * | 2007-04-03 | 2014-08-12 | Microsoft Corporation | Communications using different modalities |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180338038A1 (en) * | 2017-05-16 | 2018-11-22 | Google Llc | Handling calls on a shared speech-enabled device |
US10791215B2 (en) | 2017-05-16 | 2020-09-29 | Google Llc | Handling calls on a shared speech-enabled device |
US10911594B2 (en) | 2017-05-16 | 2021-02-02 | Google Llc | Handling calls on a shared speech-enabled device |
US11057515B2 (en) * | 2017-05-16 | 2021-07-06 | Google Llc | Handling calls on a shared speech-enabled device |
US11089151B2 (en) | 2017-05-16 | 2021-08-10 | Google Llc | Handling calls on a shared speech-enabled device |
US11595514B2 (en) | 2017-05-16 | 2023-02-28 | Google Llc | Handling calls on a shared speech-enabled device |
US11622038B2 (en) | 2017-05-16 | 2023-04-04 | Google Llc | Handling calls on a shared speech-enabled device |
Also Published As
Publication number | Publication date |
---|---|
US20120170722A1 (en) | 2012-07-05 |
US8805688B2 (en) | 2014-08-12 |
US20080249778A1 (en) | 2008-10-09 |
US8131556B2 (en) | 2012-03-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8805688B2 (en) | Communications using different modalities | |
US20080247529A1 (en) | Incoming Call Classification And Disposition | |
US8983051B2 (en) | Outgoing call classification and disposition | |
US8625749B2 (en) | Content sensitive do-not-disturb (DND) option for a communication system | |
US20070047726A1 (en) | System and method for providing contextual information to a called party | |
US8504633B2 (en) | Method and system for communicating a data file | |
US7283829B2 (en) | Management of call requests in multi-modal communication environments | |
US7702792B2 (en) | Method and system for managing communication sessions between a text-based and a voice-based client | |
US7463723B2 (en) | Method to enable instant collaboration via use of pervasive messaging | |
KR100749456B1 (en) | Source Audio Identifiers for Digital Communications | |
US9673989B2 (en) | Method and systems for connection into conference calls | |
US20090086950A1 (en) | Active call processing and notifications | |
US8027447B2 (en) | Call processing based on electronic calendar information | |
US20070172042A1 (en) | Call urgency screening | |
US20120121077A1 (en) | System and method for brokering communication dependent tasks | |
EP1345397A1 (en) | Instant messaging for called party notification | |
US20070206760A1 (en) | Service-initiated voice chat | |
US8358763B2 (en) | Camping on a conference or telephony port | |
GB2578121A (en) | System and method for hands-free advanced control of real-time data stream interactions | |
CA2706392C (en) | Method and apparatus for enabling a calling party to leave a voice message for a called party in response to a command provided by the calling party | |
US9116930B2 (en) | Method and system for configuring a contact database associated with a user | |
EP2337324A1 (en) | Method and apparatus for sending call context information | |
US8897427B2 (en) | Method and apparatus for enabling a calling party to leave a voice message for a called party | |
TWI388191B (en) | Video automatic switchboard system and its operating method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BARTON, WILLIAM F.;GALANES, FRANCISCO M.;OCKENE, LAWRENCE M.;AND OTHERS;SIGNING DATES FROM 20070323 TO 20070328;REEL/FRAME:034111/0031 |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034747/0417 Effective date: 20141014 Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:039025/0454 Effective date: 20141014 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE |